The University of Southampton
University of Southampton Institutional Repository

Automated detection of gunshots in tropical forests using convolutional neural networks

Automated detection of gunshots in tropical forests using convolutional neural networks
Automated detection of gunshots in tropical forests using convolutional neural networks
Unsustainable hunting is one of the leading drivers of global biodiversity loss, yet very few direct measures exist due to the difficulty in monitoring this cryptic activity. Where guns are commonly used for hunting, such as in the tropical forests of the Americas and Africa, acoustic detection can potentially provide a solution to this monitoring challenge. The emergence of low cost autonomous recording units (ARUs) brings into reach the ability to monitor hunting pressure over wide spatial and temporal scales. However, ARUs produce immense amounts of data, and long term and large-scale monitoring is not possible without efficient automated sound classification techniques. We tested the effectiveness of a sequential two-stage detection pipeline for detecting gunshots from acoustic data collected in the tropical forests of Belize. The pipeline involved an on-board detection algorithm which was developed and tested in a prior study, followed by a spectrogram based convolutional neural network (CNN), which was developed in this manuscript. As gunshots are rare events, we focussed on developing a classification pipeline that maximises recall at the cost of increased false positives, with the aim of using the classifier to assist human annotation of files. We trained the CNN on annotated data collected across two study sites in Belize, comprising 597 gunshots and 28,195 background sounds. Predictions from the annotated validation dataset comprising 150 gunshots and 7044 background sounds collected from the same sites yielded a recall of 0.95 and precision of 0.85. The combined recall of the two-step pipeline was estimated at 0.80. We subsequently applied the CNN to an un-annotated dataset of over 160,000 files collected in a spatially distinct study site to test for generalisability and precision under a more realistic monitoring scenario. Our model was able to generalise to this dataset, and classified gunshots with 0.57 precision and estimated 80% recall, producing a substantially more manageable dataset for human verification. Using a classifier-guided listening approach such as ours can make wide scale monitoring of threats such as hunting a feasible option for conservation management.
1470-160X
Katsis, Lydia, Katerina Diane
a90d89d0-22f0-47fd-94a6-bb7f2d9614cf
Hill, Andrew
bfc05b70-7a90-40ab-8240-4d1f56aa3e4d
Pina Covarrubias, Evelyn
11128d21-ddb4-4f07-b9e6-cd5abf2e83bc
Prince, Peter
13940cd1-98ab-4dca-a9ce-2403b2e61daa
Rogers, Alex
60b99721-b556-4805-ab34-deb808a8666c
Doncaster, Charles
0eff2f42-fa0a-4e35-b6ac-475ad3482047
Snaddon, Jake
31a601f7-c9b0-45e2-b59b-fda9a0c5a54b
Katsis, Lydia, Katerina Diane
a90d89d0-22f0-47fd-94a6-bb7f2d9614cf
Hill, Andrew
bfc05b70-7a90-40ab-8240-4d1f56aa3e4d
Pina Covarrubias, Evelyn
11128d21-ddb4-4f07-b9e6-cd5abf2e83bc
Prince, Peter
13940cd1-98ab-4dca-a9ce-2403b2e61daa
Rogers, Alex
60b99721-b556-4805-ab34-deb808a8666c
Doncaster, Charles
0eff2f42-fa0a-4e35-b6ac-475ad3482047
Snaddon, Jake
31a601f7-c9b0-45e2-b59b-fda9a0c5a54b

Katsis, Lydia, Katerina Diane, Hill, Andrew, Pina Covarrubias, Evelyn, Prince, Peter, Rogers, Alex, Doncaster, Charles and Snaddon, Jake (2022) Automated detection of gunshots in tropical forests using convolutional neural networks. Ecological Indicators, 141, [103128]. (doi:10.1016/j.ecolind.2022.109128).

Record type: Article

Abstract

Unsustainable hunting is one of the leading drivers of global biodiversity loss, yet very few direct measures exist due to the difficulty in monitoring this cryptic activity. Where guns are commonly used for hunting, such as in the tropical forests of the Americas and Africa, acoustic detection can potentially provide a solution to this monitoring challenge. The emergence of low cost autonomous recording units (ARUs) brings into reach the ability to monitor hunting pressure over wide spatial and temporal scales. However, ARUs produce immense amounts of data, and long term and large-scale monitoring is not possible without efficient automated sound classification techniques. We tested the effectiveness of a sequential two-stage detection pipeline for detecting gunshots from acoustic data collected in the tropical forests of Belize. The pipeline involved an on-board detection algorithm which was developed and tested in a prior study, followed by a spectrogram based convolutional neural network (CNN), which was developed in this manuscript. As gunshots are rare events, we focussed on developing a classification pipeline that maximises recall at the cost of increased false positives, with the aim of using the classifier to assist human annotation of files. We trained the CNN on annotated data collected across two study sites in Belize, comprising 597 gunshots and 28,195 background sounds. Predictions from the annotated validation dataset comprising 150 gunshots and 7044 background sounds collected from the same sites yielded a recall of 0.95 and precision of 0.85. The combined recall of the two-step pipeline was estimated at 0.80. We subsequently applied the CNN to an un-annotated dataset of over 160,000 files collected in a spatially distinct study site to test for generalisability and precision under a more realistic monitoring scenario. Our model was able to generalise to this dataset, and classified gunshots with 0.57 precision and estimated 80% recall, producing a substantially more manageable dataset for human verification. Using a classifier-guided listening approach such as ours can make wide scale monitoring of threats such as hunting a feasible option for conservation management.

Text
1-s2.0-S1470160X22006008-main - Version of Record
Available under License Creative Commons Attribution.
Download (2MB)

More information

Accepted/In Press date: 28 June 2022
e-pub ahead of print date: 4 July 2022
Published date: 4 July 2022

Identifiers

Local EPrints ID: 468267
URI: http://eprints.soton.ac.uk/id/eprint/468267
ISSN: 1470-160X
PURE UUID: befa4e97-0c9f-42fe-9012-e9c00f694fb2
ORCID for Evelyn Pina Covarrubias: ORCID iD orcid.org/0000-0002-3564-7467
ORCID for Charles Doncaster: ORCID iD orcid.org/0000-0001-9406-0693
ORCID for Jake Snaddon: ORCID iD orcid.org/0000-0003-3549-5472

Catalogue record

Date deposited: 09 Aug 2022 16:37
Last modified: 12 Nov 2024 02:35

Export record

Altmetrics

Contributors

Author: Andrew Hill
Author: Evelyn Pina Covarrubias ORCID iD
Author: Peter Prince
Author: Alex Rogers
Author: Jake Snaddon ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×