Confidence Based Acoustic Event Detection

Xianjun Xia, Roberto Togneri, Ferdous Sohel, David Huang

    Research output: Chapter in Book/Conference paperConference paperpeer-review

    11 Citations (Scopus)

    Abstract

    Acoustic event detection, the determination of the acoustic event type and the localisation of the event, has been widely applied in many real-world applications. Many works adopt the multi-label classification technique to perform the polyphonic acoustic event detection with a global threshold to detect the active acoustic events. However, the manually labeled boundaries are error-prone and cannot always be accurate, especially when the frame length is too short to be accurately labeled by human annotators. To deal with this, a confidence is assigned to each frame and acoustic event detection is performed using a multi-variable regression approach in this paper. Experimental results on the latest TUT sound event 2017 database of polyphonic events demonstrate the superior performance of the proposed approach compared to the multi-label classification based AED method.

    Original languageEnglish
    Title of host publication2018 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2018 - Proceedings
    Place of PublicationUSA
    PublisherIEEE, Institute of Electrical and Electronics Engineers
    Pages306-310
    Number of pages5
    Volume2018-April
    ISBN (Print)9781538646588
    DOIs
    Publication statusPublished - 10 Sept 2018
    Event2018 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2018 - Calgary, Canada
    Duration: 15 Apr 201820 Apr 2018

    Conference

    Conference2018 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2018
    Abbreviated titleICASSP 2018
    Country/TerritoryCanada
    CityCalgary
    Period15/04/1820/04/18

    Fingerprint

    Dive into the research topics of 'Confidence Based Acoustic Event Detection'. Together they form a unique fingerprint.

    Cite this