I. Introduction
Avian vocalization studies are essential for ecological research, conservation planning, species distribution modelling, and mitigating risks like bird-plane collisions. Traditionally, these studies required labor-intensive manual surveys by trained observers. However, the advent of machine learning and digital signal processing technologies has shifted the focus towards automated bird population surveys, offering a more efficient and cost-effective approach. Bird Audio Detection (BAD), a key aspect of these automated surveys, aims to identify bird calls within audio recordings. Despite its potential, BAD faces challenges due to the use of weakly labelled data (WLD) and the complexity of forest soundscapes. This paper proposes a novel approach using Convolutional Neural Networks (CNNs) for acoustic bird audio classification. The method involves creating spectrograms from the audio data for the CNN model, which is trained to recognize specific bird species' vocal features within forest recordings. The model's results are compiled into a CSV file, quantifying the bird's vocal presence, and contributing significantly to ecological research. The paper will further discuss previous methods used in the audio field, delve into the methodology, and suggest future research directions. The traditional approach to collecting bird species data has involved manual surveys, which are labour-intensive and require observers trained in bird recognition. With the advent of machine learning and digital signal processing technologies, there has been a significant shift towards automated bird population surveys, which have the potential to provide vast amounts of valuable data with less effort and expense than human surveys.