Open access

Combining two user-friendly machine learning tools increases species detection from acoustic recordings

Publication: Canadian Journal of Zoology
2 January 2024

Abstract

Passive acoustic monitoring usually generates large datasets that require machine learning algorithms to scan sound files, although the complexity of developing machine learning algorithms can be a barrier. We assessed the ability and speed of two user-friendly machine learning tools, Kaleidoscope Pro and BirdNET, for detecting the American toad (Anaxyrus americanus (Holbrook, 1836)) in sound recordings. We developed a two-step approach, combining both tools to maximize species detection while minimizing the time needed for output verification. When considered separately, Kaleidoscope Pro successfully detected the American toad in 85.9% of recordings in the validation dataset, while BirdNET detected the species in 58.4% of recordings. Combining the two tools in the two-step approach increased the detection rate to 93.3%. We applied the two-step approach to a large acoustic dataset (n = 6194 recordings). We started by scanning the dataset using Kaleidoscope Pro (species detected in 417 recordings), then we used BirdNET on the remaining recordings without confirmed presence. The two-step approach reduced the scanning time, the time needed for output verification, and added 37 additional species detections in 45 min. Our findings highlight that combining machine learning tools can improve species detectability while minimizing time and effort.

1. Introduction

The technological advances in recent decades revolutionized the way we currently monitor habitats and species. Among the emerging techniques for biomonitoring, several automated and non-invasive methods have rapidly become standard tools in ecology, such as camera trapping, remote sensing, and passive acoustic monitoring (Lahoz-Monfort and Magrath 2021). These automated and non-invasive techniques offer researchers the ability to expand the spatial and temporal scales of their studies and contribute to collect large amounts of data. However, datasets obtained through automated techniques often pose issues for investigators because manual processing of automated data is time-consuming, tedious, and subject to human bias. To overcome these issues, machine learning algorithms can effectively process such large datasets (e.g., Priyadarshani et al. 2018; Stowell 2022; Xie et al. 2022).
Passive acoustic monitoring is increasingly being used to detect different groups such as anurans, bats, birds, or insects (Sugai et al. 2019; Hoefer et al. 2023). Surveys relying on passive acoustic monitoring easily generate substantial volumes of recordings, making it impossible to visually inspect or listen to all files (e.g., Pérez-Granados and Schuchmann 2020). The development of machine learning algorithms has become crucial to deal with these large numbers of files (Stowell 2022; Xie et al. 2022). Unfortunately, implementing some of the state-of-the-art machine learning models can be complex and intimidating for ecologists and managers without an engineering or computing background. Indeed, the difficulty in using sound detection tools is a limiting factor for passive acoustic monitoring surveys (Wood et al. 2023a). However, a new generation of user-friendly and ready-to-use machine learning tools have recently emerged and may further improve the effectiveness of automated audio recognition, such as BirdNET and Kaleidoscope Pro (e.g., Manzano-Rubio et al. 2022; Bota et al. 2023; Wood et al. 2023a).
BirdNET is an automated sound classifier that is free, open source, and based on a convolutional neural network architecture for automated identification of over 6000 wildlife species, including birds, anurans, and mammals (Kahl et al. 2021; Pérez-Granados 2023; Wood et al. 2023a, 2023b). For each 3 s fragment of an audio file, BirdNET provides a species identification accompanied by a confidence score, allowing researchers to filter the output according to a desired confidence level. Although BirdNET is a promising tool, its effectiveness for wildlife monitoring has yet to be extensively assessed (reviewed by Pérez-Granados 2023), with only a single case study testing its capabilities for anuran monitoring (Wood et al. 2023a). BirdNET can be run through a user-friendly interface or the command line, and requires no expertise in machine learning (Wood et al. 2023a). Another ready-to-use and user-friendly machine learning tool for audio recognition is Kaleidoscope Pro (Manzano-Rubio et al. 2022), which requires a paid subscription for an annual license. Unlike BirdNET, the Kaleidoscope Pro workflow relies on the automated detection of candidate sounds based on the input of signal parameters and their classification, through unsupervised machine learning (hidden Markov models), into clusters of species vocalizations (Pérez-Granados and Schuchmann 2020). Both BirdNET and Kaleidoscope Pro can easily be trained to develop species-specific algorithms without the need for technical expertise. Despite its potential, current knowledge on the ability of automated classification of anuran vocalization in large acoustic datasets remains limited (e.g., Huang et al. 2014; Wood et al. 2023a).
In this paper, we aim to evaluate two user-friendly machine learning tools, BirdNET and Kaleidoscope Pro, to detect the American toad (Anaxyrus americanus (Holbrook, 1836)) in recordings. Specifically, we evaluated the efficacy of each approach in detecting the species compared to a human and then evaluated the efficacy of both methods combined. Additionally, we measured the computing time required to scan the validation acoustic dataset (n = 371 3 min recordings) and the amount of human time needed to verify the output. Then, we evaluated the effectiveness and speed of a two-step approach to detect the presence of the species using a large field acoustic dataset collected in northern Canada (n = 6194 3 min recordings). By sharing our assessments and insights, we hope to provide valuable guidance to apply automated detection and machine learning approaches in wildlife monitoring. For clarity, we use the term “detections” to denote potential multiple instances of predictions made by BirdNET or Kaleidoscope Pro in a given recording, whereas we use the term presence to indicate that the species was confirmed at least once by a human in a 3 min recording.

2. Materials and methods

2.1. Study area and pond selection

We collected acoustic data in the Eeyou Istchee James Bay region of northwestern Quebec, Canada, situated between the latitudes of 49° and 53°N, and longitudes of 71° and 79°W. The study area spans approximately 400 000 km2 and lies within the traditional Cree and Abitibiwinni First Nation territory. The landscape of the study area consists of a mosaic of forests dominated by black spruce, rocky hills bordering coniferous forests, and ombrotrophic to minerotrophic peatlands. The region experiences a subpolar and subhumid climate, characterized by mean temperatures between −0.5 and −4 °C, whereas annual precipitation varies between 700 and 900 mm. Most of the snow generally falls from August to May.
We selected 50 ponds smaller than 2 ha in size, maintaining a distance of at least 800 m between ponds to achieve pond independence. These ponds represented the two main types of ponds in the study area: 12 beaver ponds and 38 peatland ponds. Numbers of each pond type reflected the proportion of ponds available for each type. For more comprehensive details regarding our methodology and pond selection process, we refer interested readers to Feldman et al. (2023).

2.2. Study species

The American toad is a widespread species in North America, occurring in a variety of breeding and foraging habitats (Dodd 2013). However, the habitat requirements of the species in the northern part of its range, including the study region, are not well documented (Fortin et al. 2012; Feldman et al. 2023). The calling activity of the American toad spans from May to July, with choruses occurring between mid-May and early June, predominantly at night between 10 pm and 2 am (Taylor 2006). The male call is characterized by a prolonged musical whistled trill at a constant pitch (Fig. 1; Hunter et al. 1999), which makes the species a good candidate to evaluate acoustic recognition algorithms. Although encountering toads in small groups is common, full choruses are rare (Taylor 2006). Hence, investigating the acoustic activity of breeding American toads, particularly in the less-explored northern borders of their distribution, can provide valuable insights into the adaptive responses of anurans to climate change and increasing anthropogenic activities.
Fig. 1.
Fig. 1. Sonogram of a typical advertisement call of the American toad (Anaxyrus americanus) and an image of the species. Photo: MJF.

2.3. Acoustic monitoring protocol

We deployed automated acoustic recorders at each of the 50 ponds (SM4 Song meter, Wildlife Acoustics Inc., Maynard, MA, USA). The SM4 recorders were placed 2–10 m from the water’s edge and 1.5 m above ground. Recorders were programmed to record 3 min segments in .wav format every hour from 19h00 to 23h00 over seven consecutive days. Sound files were encoded at a sampling rate of 44.1 kHz and a 16-bit sample resolution. Each pond was sampled twice a year in 2018 and 2019 with visits spaced 5–7 weeks apart from May to July. We collected a total of 6194 files across the 2 years, yielding a total of 309.7 h of recording. It is noteworthy that 11.5% of these files were unusable, despite our diligent maintenance and battery supplementation. The nature of these technical issues remained unknown. Data were retrieved at the end of the 7-day recording period. This study exclusively employed acoustic recording units for passive acoustic monitoring without direct interaction with live animals. As such, no formal animal ethics approval was required for this research for data collection.

2.4. Acoustic recording analyses

2.4.1. BirdNET

The acoustic dataset was analyzed using the default values for overlap (0 s) and sensitivity (1.0) and the minimum threshold for confidence score (0.01) on BirdNET-Analyzer (version 2.2.0; Kahl et al. 2021). We applied the “American toad” filter to classify sounds only for the target species (Manzano-Rubio et al. 2022; see extended description of BirdNET settings in Kahl et al. 2021; Pérez-Granados et al. 2023 and Supplementary image S1). Every 3 min recording with BirdNET predictions was listened to and inspected by a human at the timestamp of the 3 s spectrograms annotated by BirdNET to verify whether the American toad was present or absent in the recording. If a human did not confirm the presence of the American toad in the first BirdNET prediction, subsequent predictions of the species within the same 3 min recording were reviewed. If the American presence was not confirmed, the species was marked as non-detected in the file. In the latter case, the recording was considered as mislabelled by BirdNET (i.e., false positive).

2.4.2. Kaleidoscope Pro

We used Kaleidoscope Pro (version 5.4.7, Wildlife Acoustics) to analyze the same acoustic dataset described above for BirdNET. To include specific parameters for the American toad, we measured the duration and minimum and maximum frequency of 39 calls of the American toad from the study area using Raven Pro 1.6 (Cornell Lab of Ornithology 2023; see Supplementary Table S1). The signal parameter inputs included the minimum and maximum length of detection (3–30 s), minimum and maximum frequency range (1.3–2.1 kHz), and a maximum intersyllable gap of 0.5 s. Kaleidoscope Pro reported a series of candidate sounds that met these criteria. Candidate sounds were automatically grouped through unsupervised machine learning by using K-means clustering through hidden Markov models (default values, see Pérez-Granados and Schuchmann 2020; see settings in Pérez-Granados et al. 2023; and Supplementary image S2). Within a cluster, Kaleidoscope Pro sorts candidate sounds by similitude. Thus, most signals of a definite cluster belong to the same type of vocalization of a given species, and the first songs of each cluster are the most similar and representative of each cluster. We reviewed each cluster and labelled them as “American toad cluster” or “Other” based on the detection of an American toad call within the first 50 sounds (i.e., the most representative) of each cluster. Previous work showed that this procedure can identify over 99% of candidate sounds of two bird species while also reducing over 95% the time required to verify output (Pérez-Granados and Schuchmann 2020). Candidate sounds within the “American toad cluster” were acoustically and visually checked by a human until the presence of the species was confirmed. Once a detection was confirmed within a 3 min recording, we did not review other detections from the same recording. Otherwise, we reviewed all detections within the “American toad cluster”. We considered a 3 min recording as mislabelled by Kaleidoscope Pro (i.e., false positive) when candidate sounds checked by a human did not reveal the species’ presence. Candidate sounds of the cluster “Other” were not checked and therefore not considered in subsequent analyses (Pérez-Granados and Schuchmann 2020).

2.5. Automated software comparison

To assess the ability of each automated tool to detect the presence of the American toad, we created a validation dataset of referenced recordings (see Pérez-Granados et al. 2023). The validation dataset comprised 371 (3 min) recordings randomly selected from 34 ponds with known presence of the species. For each recording, we reported whether the species was detected or not after checking spectrograms in Raven Pro 1.6 (Cornell Lab of Ornithology 2023). Recordings were reviewed blinded with respect to site location, date, and hour of recording. To evaluate the effectiveness of the two machine learning approaches for detecting the American toad, we estimated the percentage of presences detected by Kaleidoscope Pro, BirdNET, and by both methods combined with respect to the total number of recordings with confirmed presence in the validation dataset.
We assessed the effectiveness and speed of a two-step approach to scan and detect the American toad in a large acoustic dataset (“field acoustic dataset”) consisting of 6194 files of 3 min recordings. Here, the two-step approach aimed to maximize the number of recordings with confirmed presence while expediting the time required for a human to verify the output. The two-step approach consisted of (1) scanning the entire dataset using Kaleidoscope Pro, which is faster than BirdNET and then (2) using BirdNET to scan files where the species was not detected by Kaleidoscope Pro (n = 5778). We estimated the percentage of total files with confirmed presence, by summing the number of files where the species had been detected by Kaleidoscope Pro and those where the species was detected by BirdNET. We evaluated the computing time required to manually inspect files for acoustic analysis, compared to processing files using each machine learning approach separately, or using the two-step approach. The time required for acoustic analyses was divided by (i) the time required by machine learning tool to scan the entire acoustic dataset and (ii) time for output verification by a human, which included the creation of the final database without misidentifications in a manageable format (Excel file in our case).

3. Results

3.1. Automated software comparison

The validation dataset comprised 149 3 min recordings with confirmed presence of the species and 222 3 min recordings where the species was not detected. Kaleidoscope Pro detected the American toad in 85.9% of 149 3 min recordings with confirmed presence, although it also reported candidate sounds for 12.7% of 3 min recordings that did not contain the species (Table 1). BirdNET detected the species in 58.4% of the 3 min recordings with confirmed presence (Table 1). Additionally, BirdNET misidentified the species presence in 9.3% of 3 min recordings with confirmed presence. The best results were obtained when applying both machine learning approaches. By using both approaches, we detected the American toad in 139 of the 149 3 min recordings with confirmed presence (93.3% of the total, Table 1).
Table 1.
Table 1. Confusion matrix of the ability of Kaleidoscope Pro, BirdNET, and both tools applied together to detect the presence of the American toad (Anaxyrus americanus) in recordings.

3.2. Field acoustic dataset

Using the two-step approach, we detected the American toad in a total of 462 3 min recordings within the field acoustic dataset. In the first step, Kaleidoscope Pro detected the species in 417 3 min recordings, which represented 90.3% of the total of 3 min recordings with confirmed presence of the species (Table 2). In the second step, we ran BirdNET on the field acoustic dataset but excluding the 417 3 min recordings with presence already confirmed by Kaleidoscope Pro. BirdNET predicted the American toad in a total of 83 additional 3 min recordings, whereas the species presence was confirmed by a human in 45 of these recordings (54.2% of the 3 min recordings with predictions by BirdNET). BirdNET misidentified the American toad in 38 3 min recordings (0.7% of the 5778 3 min recordings analysed) but increased the percentage of confirmed presences within the dataset by up to 9.7% (from 417 to 462, Table 2).
Table 2.
Table 2. Total number of recordings with annotated, verified, and not-detected presence of the American toad (Anaxyrus americanus) using the two-step approach.
The time required for visual inspection of the field acoustic dataset and data entry was estimated to be about 153.9 h, based on the estimate of 1.5 min needed to visually inspect each 3 min recording of the validation dataset. This value equals to 50% of the total amount of recording time. Kaleidoscope Pro analyses on the 6194 files took 32 mins and extracted a total of 37 629 candidate sounds grouped into 63 clusters. A total of 7534 candidate sounds (20% of the total) were included within the “American toad cluster”, which included seven of the original 63 clusters created by Kaleidoscope Pro (11.1% of the total). A total of 1675 recordings were manually labelled as belonging to the “American toad cluster”, and output verification was carried out by a human in 40 min. Therefore, the total time for scanning by Kaleidoscope Pro and output verification by a human required approximately 72 min, which represents around 0.4% of the total recording time. In contrast, time required by BirdNET to scan the 6194 files was 11.3 h, resulting in 932 BirdNET predictions in 83 sound recordings (Table 2). Human time devoted to verifying the output of BirdNET analyses was 723 min (3.9% of the total recording time).
The two-step method on the field acoustic dataset was performed in a total of 795 min (4.3% of the recording time). Most of the time (710 min) was devoted to data scanning by machine learning, which did not require human supervision. Verifying the output of the two-step approach required only 85 min for a human (Table 2). Overall, the two-step approach increased by 11% the number of recordings with detections originally identified by Kaleidoscope Pro alone (Table 2).

4. Discussion

The inherent challenges of automated acoustic recognition of target species in large acoustic datasets limited the widespread use of passive acoustic monitoring (Wood et al. 2023a). However, recent advancements of algorithms (e.g., Kahl et al. 2021; Wood et al. 2023a) together with the development of user-friendly software for automated recognition have opened up new avenues for automated wildlife recognition from sound recordings (Manzano-Rubio et al. 2022; Bota et al. 2023). Here, we demonstrated the ability of two user-friendly machine learning approaches (Kaleidoscope Pro and BirdNET) for detecting the presence of the American toad.
Kaleidoscope Pro was faster and detected the target species in a larger number of recordings than BirdNET. Our findings align with a recent study that compared both Kaleidoscope Pro and BirdNET in detecting the presence of a threatened bird species (Botaurus stellaris (Linnaeus, 1758), see Manzano-Rubio et al. 2022). The target species considered by Manzano-Rubio et al. (2022) and the American toad have simple vocalizations (Fig. 1), which may partly explain the high ability of Kaleidoscope Pro for detecting both species. However, convolutional neural networks, such as BirdNET, may outperform Kaleidoscope Pro at detecting more complex vocalizations. Further research should investigate into the ability of Kaleidoscope Pro and BirdNET for detecting a wider range of species. In our dataset, Kaleidoscope Pro produced a higher number of false positives than BirdNET, but the computing time and the time required to verify the output of Kaleidoscope Pro by a human was considerably reduced compared to BirdNET. Such speed can be attributed to the ease in verifying the output directly within Kaleidoscope Pro and to the workflow we employed. Users can easily navigate through Kaleidoscope Pro output while checking the sonograms, confirm the classification by pressing a key, and move on to the next candidate sound. In contrast, verifying the BirdNET output required to individually open the file associated with each recording and locate the 3 s spectrogram annotated. However, the most recent version of BirdNET (version 2.4, released in June 2023) now allows users to extract all the segments (in .wav file) where the target species were predicted. Although our study was conducted before this new addition to BirdNET, we expect that this new feature will substantially accelerate output verification. It is worth of highlight that we used an Intel(R) Core(TM) i7 (8th Gen, CPU 1.80 GHz, 1.99 GH, 8 GB RAM) with the acoustic recordings stored and analysed from an external hard drive. Therefore, the speed of data transfer from the hard drive to the laptop was the main limiting factor of computing time (i.e., computing time would be lower if processed from the internal memory of the laptop). However, it is important to consider that many monitoring programmes typically store their acoustic data in hard drives. Therefore, the timing values provided in our study can be useful for managers and researchers as an estimate of the time required for automated detection using Kaleidoscope Pro or BirdNET.
The workflow we applied in our study greatly expedited the review process. Initially, we leveraged the unsupervised clustering performed by Kaleidoscope Pro, focusing our attention on candidate sounds from clusters with a high probability of containing the target species. This approach reduced the number of candidate sounds to be verified by up to 80% from the original output (see also Pérez-Granados and Schuchmann 2020). We acknowledge the possibility of a few missed American toad vocalizations using this approach. Nonetheless, the number of missed presences (false negatives) at the recording level is expected to be minimal. To expedite the reviewing process, we focused solely on confirming the presence of the species in each recording. Once the American toad was confirmed within a recording, we did not check subsequent candidate sounds within that recording.
BirdNET was originally developed for automated bird song recognition (>6000 bird species included), but it has expanded to include tens of anurans and a few mammals (e.g., Wood et al. 2023a, 2023b). Our assessment, to the best of our knowledge, is the second study evaluating the performance of BirdNET for monitoring anurans (see Wood et al. 2023a). We hope that our results will encourage investigators seeking automated anuran detection to consider BirdNET. The ability of BirdNET to detect the American toad was low compared to Kaleidoscope Pro. Nonetheless, BirdNET was still able to detect the species using the default values in over half of the recordings, despite their short duration of 3 min. We encourage researchers wishing to use BirdNET to further assess the influence of the sensitivity, overlap, and confidence score parameters on wildlife detection. Moreover, further research assessing BirdNET’s effectiveness in detecting a wider range of anurans would be particularly valuable, as our current knowledge is limited to three species from North America (Wood et al. 2023a and our study).
Surprisingly, BirdNET detected a large number of presences undetected by Kaleidoscope Pro (8% in the validation dataset and 10% in the field acoustic dataset). Upon visual inspection and received sound level, it became evident that most of BirdNET’s detections corresponded to American toad vocalizations emitted far from the recorder. This pattern suggests that BirdNET is more sensitive than Kaleidoscope Pro in detecting the species when it vocalized at larger distances. However, the reason why BirdNET did not detect the species more frequently than Kaleidoscope Pro remains unclear. Differences in the training dataset used for creating the American toad algorithm in BirdNET, such as the inclusion of weak vocalizations, may have contributed to these discrepancies. Unfortunately, because the contents of the BirdNET dataset are not publicly released, it is not possible to determine the number, type and quality of the vocalizations used for training the algorithm. Further research should investigate the underlying reasons for the differences observed between the two machine learning approaches we compared, although their distinct structures (e.g., convolutional neural network architecture on BirdNET and hidden Markov models in Kaleidoscope Pro) may complicate direct comparisons. Moreover, the unsupervised nature of Kaleidoscope Pro machine learning makes the classification process more obscure than with BirdNET.
Our two-step approach consisted in using the speed of Kaleidoscope Pro in scanning and verifying the whole field acoustic dataset and then running BirdNET exclusively on the recordings where Kaleidoscope Pro had not detected the species. This sequential workflow significantly reduced the overall time required for data scanning and output verification in BirdNET. The two-step approach improved the detection probability of the American toad by approximately 10%, with only 45 additional minutes dedicated to BirdNET output verification. Importantly, this approach enabled us to remove all false positives from the final dataset, amounting to a mere 0.5% of the total recording time. Removing false positives is needed for efficient use of acoustic algorithms in large acoustic datasets and to provide reliable results for further analyses highly sensitive to false positives such as occupancy analyses (Guillera‐Arroita et al. 2017; Wood et al. 2023a).
Our acoustic monitoring protocol was set to record from 7 pm to 11 pm, aiming to monitor the assemblage of pond-breeding anuran species in the study area (see Feldman et al. 2023). However, the maximum vocal activity of the American toad occurs between 10 pm and 2 am (Taylor 2006). Therefore, our protocol does not align with the peak of calling activity of the American toad and may have resulted in a potential underestimation of species occurrence. Although it does not have an impact on our conclusions, further research aiming to acoustically monitor the American toad should extend the recording schedule to cover the hours of maximum calling activity of the species.

5. Conclusions

In this study, we demonstrated the ability of two user-friendly machine learning approaches as efficient tools for automated acoustic recognition of an anuran at the northern portion of its distribution. Our assessment opens a new door for managers and researchers with minimal technical expertise to leverage the power of automated recognition software. Moreover, we developed a streamlined two-step approach that increases the probability of detecting the species while reducing the time of acoustic processing and manipulation by humans, showing that both machine learning approaches can be easily combined. The manual verification approach that we used allowed us to remove every false positive from the database while expediting the process (i.e., once the presence was verified, no more detections for the same 3 min recording were checked). We hope that our assessments and recommendations will encourage managers and researchers to embrace passive acoustic monitoring as a viable technique for wildlife monitoring. Our results indicate that automated detection may no longer be a limiting factor for passive acoustic monitoring surveys.

Acknowledgements

We thank M. Brigham and two anonymous reviewers whose comments improved the manuscript. CP acknowledges the support of the Ministerio of Educación y Formación Profesional through the Beatriz Galindo Fellowship (Beatriz Galindo—Convocatoria 2020). MJF gratefully acknowledges the field assistance of R. Chevallier in setting up the acoustic equipment, and L. Imbeau and N. Fenton for their contributions to the conception and design of the study. The Natural Sciences and Engineering Research Council of Canada—UQAT Industrial Research Chair on Nordic Biodiversity in a Mining Context funded the field work.

References

Bota G., Manzano-Rubio R., Catalán L., Gómez-Catasús J., Pérez-Granados C. 2023. Hearing to the unseen: AudioMoth and BirdNET as a cheap and easy method for monitoring cryptic bird species. Sensors, 23: 7176.
Cornell Lab of Ornithology. 2023. Raven Pro: Interactive Sound Analysis Software (Version 1.6.4) [computer software]. The Cornell Lab of Ornithology, Ithaca, NY.
Dodd C.K. 2013. Frogs of the United States and Canada. John Hopkins University Press.
Feldman M.J., Mazerolle M.J., Imbeau L., Fenton N.J. 2023. Occupancy and abundance of pond-breeding anurans in boreal landscapes. J. Herpetol. 57: 159–171.
Fortin C., Galois P., Dutil B., Ponge L., Ouellet M. 2012. Inventaire de l'herpétofaune dans la région des monts Otish. Nat. Can. 136: 22–31.
Guillera-Arroita G., Lahoz-Monfort J.J., van Rooyen A.R., Weeks A.R., Tingley R. 2017. Dealing with false-positive and false-negative errors about species occurrence at multiple levels. Methods Ecol. Evol. 8: 1081–1091.
Hoefer S., McKnight D.T., Allen-Ankins S., Nordberg E.J., Schwarzkopf L. 2023. Passive acoustic monitoring in terrestrial vertebrates: a review. Bioacoustics, 1–26.
Huang C.J., Chen Y.J., Chen H.M., Jian J.J., Tseng S.C., Yang Y.J., Hsu P.A. 2014. Intelligent feature extraction and classification of anuran vocalizations. Appl. Soft Comput. 19: 1–7.
Hunter M.L. Jr., Calhoun A.J., McCollough M. 1999. Maine amphibians and reptiles. University of Maine Press, Orono, ME.
Kahl S., Wood C.M., Eibl M., Klinck H. 2021. BirdNET: a deep learning solution for avian diversity monitoring. Ecol. Inform. 61: 101236.
Lahoz-Monfort J.J., Magrath M.J. 2021. A comprehensive overview of technologies for species and habitat monitoring and conservation. BioScience, 71: 1038–1062.
Manzano-Rubio R., Bota G., Brotons L., Soto-Largo E., Pérez-Granados C. 2022. Low-cost open-source recorders and ready-to-use machine learning approaches provide effective monitoring of threatened species. Ecol. Inform. 72: 101910.
Pérez-Granados C. 2023. BirdNET: applications, performance, pitfalls and future opportunities. IBIS, 165: 1068–1075.
Pérez-Granados C., Schuchmann K.L. 2020. Monitoring the annual vocal activity of two enigmatic nocturnal neotropical birds: the Common Potoo (Nyctibius griseus) and the Great Potoo (Nyctibius grandis). J. Ornithol. 161: 1129–1141.
Pérez-Granados C., Feldman M.J., Mazerolle M.J. 2023. Automated acoustic recognition for dummies: combination of two user-friendly machine learning approaches optimize species detection: data repository. Mendeley Data, V3. Availablefrom https://data.mendeley.com/datasets/ydzwvgtwcp/3.
Priyadarshani N., Marsland S., Castro I. 2018. Automated birdsong recognition in complex acoustic environments: a review. J. Avian Biol. 49: jav–01447.
Stowell D. 2022. Computational bioacoustics with deep learning: a review and roadmap. PeerJ, 10: e13152.
Sugai L.S.M., Silva T.S.F., Ribeiro J.W., Jr., Llusia D. 2019. Terrestrial passive acoustic monitoring: review and perspectives. BioScience, 69: 15–25.
Taylor P. 2006. Calling periods for frogs and toads near Pinawa, Manitoba, with an update on Mink Frog and Green Frog distribution. Blue Jay, 64: 44–55.
Wood C.M., Barceinas Cruz A., Kahl S. 2023b. Pairing a user-friendly machine-learning animal sound detector with passive acoustic surveys for occupancy modeling of an endangered primate. Am. J. Primatol. 85: e23507.
Wood C.M., Kahl S., Barnes S., Van Horne R., Brown C. 2023a. Passive acoustic surveys and the BirdNET algorithm reveal detailed spatiotemporal variation in the vocal activity of two anurans. Bioacoustics, 32: 532–543.
Xie J., Zhong Y., Zhang J., Liu S., Ding C., Triantafyllopoulos A. 2022. A review of automatic recognition technology for bird vocalizations in the deep learning era. Ecol. Inform. 7: 101927.

Supplementary material

Supplementary Material 1 (DOCX / 296 KB).

Information & Authors

Information

Published In

cover image Canadian Journal of Zoology
Canadian Journal of Zoology
Volume 102Number 4April 2024
Pages: 403 - 409

History

Received: 24 August 2023
Accepted: 26 October 2023
Accepted manuscript online: 6 November 2023
Version of record online: 2 January 2024

Data Availability Statement

The dataset including the raw databases generated and used during the study, settings introduced in Kaleidoscope Pro and BirdNET, the audio recordings of the validation dataset, and the metadata information are openly available at Mendeley Data at https://data.mendeley.com/datasets/ydzwvgtwcp/3 and with doi:10.17632/ydzwvgtwcp.3.

Key Words

  1. American toad
  2. BirdNET
  3. convolutional neural network
  4. Kaleidoscope Pro
  5. Anaxyrus americanus (Holbrook
  6. 1836)
  7. passive acoustic monitoring

Authors

Affiliations

Ecology Department, Alicante University, Alicante, Spain
Conservation Biology Group. Landscape Dynamics and Biodiversity programme. Forest Science and Technology Center of Catalonia (CTFC), Solsona, Catalonia, Spain
Author Contributions: Data curation, Formal analysis, Investigation, Methodology, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, and Writing – review & editing.
Conservation Biology Group. Landscape Dynamics and Biodiversity programme. Forest Science and Technology Center of Catalonia (CTFC), Solsona, Catalonia, Spain
Institut de Recherche sur les Forêts (IRF), Chaire industrielle CRSNG-UQAT sur la biodiversité en contexte minier, Centre d’étude de la forêt, Université du Québec en Abitibi Témiscamingue (UQAT), Rouyn-Noranda, Québec, Canada
Author Contributions: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, and Writing – review & editing.
Centre d’étude de la forêt, Département des sciences du bois et de la forêt, Université Laval, Québec, Canada
Author Contributions: Conceptualization, Data curation, Funding acquisition, Investigation, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, and Writing – review & editing.

Author Contributions

Conceptualization: MJF, MJM
Data curation: CP, MJF, MJM
Formal analysis: CP, MJF
Funding acquisition: MJM
Investigation: CP, MJF, MJM
Methodology: CP, MJF
Project administration: MJM
Resources: CP, MJF, MJM
Software: CP, MJF, MJM
Supervision: CP, MJF, MJM
Validation: CP, MJF, MJM
Visualization: CP, MJF, MJM
Writing – original draft: CP, MJF, MJM
Writing – review & editing: CP, MJF, MJM

Competing Interests

The authors declare there are no competing interests.

Funding Information

Natural Sciences and Engineering Research Council of Canada

Metrics & Citations

Metrics

Other Metrics

Citations

Cite As

Export Citations

If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.

Cited by

1. An Easily Customizable Approach for Automated Species-Specific Detection of Anuran Calls Using the European Green Toad as an Example
2. Passive acoustic monitoring and automated detection of the American bullfrog

View Options

View options

PDF

View PDF

Login options

Check if you access through your login credentials or your institution to get full access on this article.

Subscribe

Click on the button below to subscribe to Canadian Journal of Zoology

Purchase options

Purchase this article to get full access to it.

Restore your content access

Enter your email address to restore your content access:

Note: This functionality works only for purchases done as a guest. If you already have an account, log in to access the content to which you are entitled.

Media

Tables

Media

Share Options

Share

Share the article link

Share on social media

Cookies Notification

We use cookies to improve your website experience. To learn about our use of cookies and how you can manage your cookie settings, please see our Cookie Policy.
×