The University of Southampton
University of Southampton Institutional Repository

End-to-End learning for visual navigation of forest environments

End-to-End learning for visual navigation of forest environments
End-to-End learning for visual navigation of forest environments
Off-road navigation in forest environments is a challenging problem in field robotics. Rovers are required to infer their traversability over a priori unknown and dynamically changing forest terrain using noisy onboard navigation sensors. The problem is compounded for small-sized rovers, such as that of a swarm. Their size-proportional low-viewpoint affords them a restricted view for navigation, which may be partially occluded by forest vegetation. Hand-crafted features, typically employed for terrain traversability analysis, are often brittle and may fail to discriminate obstacles in varying lighting and weather conditions. We design a low-cost navigation system tailored for small-sized forest rovers using self-learned features. The MobileNet-V1 and MobileNet-V2 models, trained following an end-to-end learning approach, are deployed to steer a mobile platform, with a human-in-the-loop, towards traversable paths while avoiding obstacles. Receiving a 128 × 96 pixel RGB image from a monocular camera as input, the algorithm running on a Raspberry Pi 4, exhibited robustness to motion blur, low lighting, shadows and high-contrast lighting conditions. It was able to successfully navigate a total of over 3 km of real-world forest terrain comprising shrubs, dense bushes, tall grass, fallen branches, fallen tree trunks, and standing trees, in over five different weather conditions and four different times of day.
end-to-end learning, low-cost sensors, low-viewpoint forest navigation, multiclass classification, off-road visual navigation, small-sized rovers, sparse swarms
1999-4907
Niu, Chaoyue
ab5e47f0-384f-411b-ae79-d29c8aea681b
Zauner, Klaus-Peter
c8b22dbd-10e6-43d8-813b-0766f985cc97
Tarapore, Danesh
fe8ec8ae-1fad-4726-abef-84b538542ee4
Niu, Chaoyue
ab5e47f0-384f-411b-ae79-d29c8aea681b
Zauner, Klaus-Peter
c8b22dbd-10e6-43d8-813b-0766f985cc97
Tarapore, Danesh
fe8ec8ae-1fad-4726-abef-84b538542ee4

Niu, Chaoyue, Zauner, Klaus-Peter and Tarapore, Danesh (2023) End-to-End learning for visual navigation of forest environments. Forests, 14 (2), [268]. (doi:10.3390/f14020268).

Record type: Article

Abstract

Off-road navigation in forest environments is a challenging problem in field robotics. Rovers are required to infer their traversability over a priori unknown and dynamically changing forest terrain using noisy onboard navigation sensors. The problem is compounded for small-sized rovers, such as that of a swarm. Their size-proportional low-viewpoint affords them a restricted view for navigation, which may be partially occluded by forest vegetation. Hand-crafted features, typically employed for terrain traversability analysis, are often brittle and may fail to discriminate obstacles in varying lighting and weather conditions. We design a low-cost navigation system tailored for small-sized forest rovers using self-learned features. The MobileNet-V1 and MobileNet-V2 models, trained following an end-to-end learning approach, are deployed to steer a mobile platform, with a human-in-the-loop, towards traversable paths while avoiding obstacles. Receiving a 128 × 96 pixel RGB image from a monocular camera as input, the algorithm running on a Raspberry Pi 4, exhibited robustness to motion blur, low lighting, shadows and high-contrast lighting conditions. It was able to successfully navigate a total of over 3 km of real-world forest terrain comprising shrubs, dense bushes, tall grass, fallen branches, fallen tree trunks, and standing trees, in over five different weather conditions and four different times of day.

Text
forests-14-00268 - Version of Record
Available under License Creative Commons Attribution.
Download (3MB)

More information

Accepted/In Press date: 27 January 2023
Published date: February 2023
Additional Information: Publisher Copyright: © 2023 by the authors.
Keywords: end-to-end learning, low-cost sensors, low-viewpoint forest navigation, multiclass classification, off-road visual navigation, small-sized rovers, sparse swarms

Identifiers

Local EPrints ID: 474919
URI: http://eprints.soton.ac.uk/id/eprint/474919
ISSN: 1999-4907
PURE UUID: f9942b48-7f0b-4c15-b2db-3daa1433602b
ORCID for Chaoyue Niu: ORCID iD orcid.org/0000-0001-7626-0317
ORCID for Danesh Tarapore: ORCID iD orcid.org/0000-0002-3226-6861

Catalogue record

Date deposited: 07 Mar 2023 17:32
Last modified: 17 Mar 2024 03:46

Export record

Altmetrics

Contributors

Author: Chaoyue Niu ORCID iD
Author: Klaus-Peter Zauner
Author: Danesh Tarapore ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×