The University of Southampton
University of Southampton Institutional Repository

An embarrassingly simple approach for visual navigation of forest environments

An embarrassingly simple approach for visual navigation of forest environments
An embarrassingly simple approach for visual navigation of forest environments
Navigation in forest environments is a challenging and open problem in the area of field robotics. Rovers in forest environments are required to infer the traversability of a priori unknown terrains, comprising a number of different types of compliant and rigid obstacles, under varying lighting and weather conditions. The challenges are further compounded for inexpensive small-sized (portable) rovers. While such rovers may be useful for collaboratively monitoring large tracts of forests as a swarm, with low environmental impact, their small-size affords them only a low viewpoint of their proximal terrain. Moreover, their limited view may frequently be partially occluded by compliant obstacles in close proximity such as shrubs and tall grass. Perhaps, consequently, most studies on off-road navigation typically use large-sized rovers equipped with expensive exteroceptive navigation sensors. We design a low-cost navigation system tailored for small-sized forest rovers. For navigation, a light-weight convolution neural network is used to predict depth images from RGB input images from a low-viewpoint monocular camera. Subsequently, a simple coarse-grained navigation algorithm aggregates the predicted depth information to steer our mobile platform towards open traversable areas in the forest while avoiding obstacles. In this study, the steering commands output from our navigation algorithm direct an operator pushing the mobile platform. Our navigation algorithm has been extensively tested in high-fidelity forest simulations and in field trials. Using no more than a 16 × 16 pixel depth prediction image from a 32 × 32 pixel RGB image, our algorithm running on a Raspberry Pi was able to successfully navigate a total of over 750 m of real-world forest terrain comprising shrubs, dense bushes, tall grass, fallen branches, fallen tree trunks, small ditches and mounds, and standing trees, under five different weather conditions and four different times of day. Furthermore, our algorithm exhibits robustness to changes in the mobile platform’s camera pitch angle, motion blur, low lighting at dusk, and high-contrast lighting conditions.
compliant obstacles, depth prediction, forest simulation, low-cost sensors, low-viewpoint forest navigation, off-road navigation, small-sized rovers, sparse swarms
Niu, Chaoyue
ab5e47f0-384f-411b-ae79-d29c8aea681b
Newlands, Callum
cde2fa7a-8b95-4baf-aaf8-77571c37016b
Zauner, Klaus-Peter
c8b22dbd-10e6-43d8-813b-0766f985cc97
Tarapore, Danesh
fe8ec8ae-1fad-4726-abef-84b538542ee4
Niu, Chaoyue
ab5e47f0-384f-411b-ae79-d29c8aea681b
Newlands, Callum
cde2fa7a-8b95-4baf-aaf8-77571c37016b
Zauner, Klaus-Peter
c8b22dbd-10e6-43d8-813b-0766f985cc97
Tarapore, Danesh
fe8ec8ae-1fad-4726-abef-84b538542ee4

Niu, Chaoyue, Newlands, Callum, Zauner, Klaus-Peter and Tarapore, Danesh (2023) An embarrassingly simple approach for visual navigation of forest environments. Frontiers in Robotics and AI, 10, [1086798]. (doi:10.3389/frobt.2023.1086798).

Record type: Article

Abstract

Navigation in forest environments is a challenging and open problem in the area of field robotics. Rovers in forest environments are required to infer the traversability of a priori unknown terrains, comprising a number of different types of compliant and rigid obstacles, under varying lighting and weather conditions. The challenges are further compounded for inexpensive small-sized (portable) rovers. While such rovers may be useful for collaboratively monitoring large tracts of forests as a swarm, with low environmental impact, their small-size affords them only a low viewpoint of their proximal terrain. Moreover, their limited view may frequently be partially occluded by compliant obstacles in close proximity such as shrubs and tall grass. Perhaps, consequently, most studies on off-road navigation typically use large-sized rovers equipped with expensive exteroceptive navigation sensors. We design a low-cost navigation system tailored for small-sized forest rovers. For navigation, a light-weight convolution neural network is used to predict depth images from RGB input images from a low-viewpoint monocular camera. Subsequently, a simple coarse-grained navigation algorithm aggregates the predicted depth information to steer our mobile platform towards open traversable areas in the forest while avoiding obstacles. In this study, the steering commands output from our navigation algorithm direct an operator pushing the mobile platform. Our navigation algorithm has been extensively tested in high-fidelity forest simulations and in field trials. Using no more than a 16 × 16 pixel depth prediction image from a 32 × 32 pixel RGB image, our algorithm running on a Raspberry Pi was able to successfully navigate a total of over 750 m of real-world forest terrain comprising shrubs, dense bushes, tall grass, fallen branches, fallen tree trunks, small ditches and mounds, and standing trees, under five different weather conditions and four different times of day. Furthermore, our algorithm exhibits robustness to changes in the mobile platform’s camera pitch angle, motion blur, low lighting at dusk, and high-contrast lighting conditions.

Text
frobt-10-1086798 - Version of Record
Available under License Creative Commons Attribution.
Download (49MB)

More information

Accepted/In Press date: 14 June 2023
Published date: 28 June 2023
Additional Information: Funding Information: The authors acknowledge the use of the IRIDIS High-Performance Computing Facility, and associated support services at the University of Southampton, in the completion of this work. Publisher Copyright: Copyright © 2023 Niu, Newlands, Zauner and Tarapore.
Keywords: compliant obstacles, depth prediction, forest simulation, low-cost sensors, low-viewpoint forest navigation, off-road navigation, small-sized rovers, sparse swarms

Identifiers

Local EPrints ID: 478778
URI: http://eprints.soton.ac.uk/id/eprint/478778
PURE UUID: e542fc8b-0419-4911-8e40-e6e4270bb3c0
ORCID for Chaoyue Niu: ORCID iD orcid.org/0000-0001-7626-0317
ORCID for Danesh Tarapore: ORCID iD orcid.org/0000-0002-3226-6861

Catalogue record

Date deposited: 10 Jul 2023 16:35
Last modified: 17 Mar 2024 03:46

Export record

Altmetrics

Contributors

Author: Chaoyue Niu ORCID iD
Author: Callum Newlands
Author: Klaus-Peter Zauner
Author: Danesh Tarapore ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×