<?xml version="1.0" encoding="UTF-8" standalone="yes"?>
<OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:dim="http://www.dspace.org/xmlns/dspace/dim" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd">
    <responseDate>2026-05-04T18:33:46.322Z</responseDate>
    <request verb="GetRecord" identifier="ezaposleni.singidunum.ac.rs/rest/sciNaucniRezultati/oai:2:11776" metadataPrefix="dim">http://ezaposleni.singidunum.ac.rs/rest/sciNaucniRezultati/oai</request>
    <GetRecord>
        <record>
            <header>
                <identifier>ezaposleni.singidunum.ac.rs/rest/sciNaucniRezultati/oai:2:11776</identifier>
                <datestamp>2026-01-04T12:51:07Z</datestamp>
                <setSpec>2</setSpec>
            </header>
            <metadata>
                <dim:dim>
                    <dim:field mdschema="dc" element="title" lang="en">Classifying birds of South America via audio analysis using convolutional networks and boosting models optimized by metaheuristics</dim:field>
                    <dim:field mdschema="dc" element="date" qualifier="issued">2026</dim:field>
                    <dim:field mdschema="dc" element="identifier" qualifier="uri">http://ezaposleni.singidunum.ac.rs/rest/sciNaucniRezultati/oai/record/2/11776</dim:field>
                    <dim:field mdschema="dc" element="identifier" qualifier="uri">https://link.springer.com/article/10.1007/s42044-025-00363-4</dim:field>
                    <dim:field mdschema="dc" element="contributor" qualifier="author" authority="id:54523" confidence="-1">T. Dogandzic</dim:field>
                    <dim:field mdschema="dc" element="contributor" qualifier="author" authority="orcid::0000-0002-4351-068X" confidence="-1">M. Zivkovic</dim:field>
                    <dim:field mdschema="dc" element="contributor" qualifier="author" authority="orcid::0000-0001-9402-7391" confidence="-1">L. Jovanovic</dim:field>
                    <dim:field mdschema="dc" element="contributor" qualifier="author" authority="orcid::0000-0001-7412-7870" confidence="-1">J. Perisic</dim:field>
                    <dim:field mdschema="dc" element="contributor" qualifier="author" authority="etfid:1178" confidence="-1">M. Milovanovic</dim:field>
                    <dim:field mdschema="dc" element="contributor" qualifier="author" authority="orcid::0000-0002-5511-2531" confidence="-1">M. Antonijevic</dim:field>
                    <dim:field mdschema="dc" element="contributor" qualifier="author" authority="id:54529" confidence="-1">M. Hammad</dim:field>
                    <dim:field mdschema="dc" element="contributor" qualifier="author" authority="orcid::0000-0002-2062-924X" confidence="-1">N. Bacanin</dim:field>
                    <dim:field mdschema="dc" element="description" qualifier="abstract">Accurate categorization of bird calls is essential for effective biodiversity surveillance and for assessing ecosystem conditions. It also enables monitoring of avian species distributions and detecting population changes. This is particularly important for species that are difficult to observe visually, but whose presence can still be identified through their distinctive calls. Moreover, in the long term, collecting audio samples can help capture changes in populations that may indicate shifts in ecosystem health. This research investigates the use of a two-level framework composed of a convolutional neural network (CNN), Categorical Boosting (CatBoost), and the Light Gradient Boosting Machine (LightGBM) for accurate sound classification of six South American bird species. Additionally, this study proposes a modified version of the variable neighborhood search (VNS) algorithm, employed to tune both layers of the framework for this specific task. A thorough comparative analysis was conducted against other powerful optimizers, accompanied by rigorous statistical evaluation of the results and SHapley Additive exPlanations (SHAP) analysis of the best models. The proposed approach achieved a superior classification accuracy of nearly 76.5% with a Matthews correlation coefficient of 0.722354, significantly outperforming the other evaluated methods. These results demonstrate both strong predictive performance and the suitability of the framework for classifying calls from different bird species.</dim:field>
                    <dim:field mdschema="dc" element="type">article</dim:field>
                    <dim:field mdschema="dc" element="identifier" qualifier="doi">10.1007/s42044-025-00363-4</dim:field>
                    <dim:field mdschema="dc" element="citation" qualifier="volume">9</dim:field>
                    <dim:field mdschema="dc" element="citation" qualifier="issue">13</dim:field>
                    <dim:field mdschema="dc" element="identifier" qualifier="issn">2520-8446</dim:field>
                    <dim:field mdschema="dc" element="source">Iran Journal of Computer Science</dim:field>
                </dim:dim>
            </metadata>
        </record>
    </GetRecord>
</OAI-PMH>
