Visual active learning for labeling: A case for soundscape ecology data

Research output: Contribution to journalArticlepeer-review

Abstract

Labeling of samples is a recurrent and time-consuming task in data analysis and machine learning and yet generally overlooked in terms of visual analytics approaches to improve the process. As the number of tailored applications of learning models increases, it is crucial that more effective approaches to labeling are developed. In this paper, we report the development of a methodology and a framework to support labeling, with an application case as background. The methodology performs visual active learning and label propagation with 2D embeddings as layouts to achieve faster and interactive labeling of samples. The framework is realized through SoundscapeX, a tool to support labeling in soundscape ecology data. We have applied the framework to a set of audio recordings collected for a Long Term Ecological Research Project in the Cantareira-Mantiqueira Corridor (LTER CCM), localized in the transition between northeastern São Paulo state and southern Minas Gerais state in Brazil. We employed a pre-label data set of groups of animals to test the efficacy of the approach. The results showed the best accuracy at 94.58% in the prediction of labeling for birds and insects; and 91.09% for the prediction of the sound event as frogs and insects.

Original languageEnglish
Article number265
JournalInformation (Switzerland)
Volume12
Issue number7
DOIs
Publication statusPublished - Jul 2021

Keywords

  • Active learning
  • Clustering
  • Labeling
  • Sampling
  • Soundscape ecology
  • Visualization

Fingerprint

Dive into the research topics of 'Visual active learning for labeling: A case for soundscape ecology data'. Together they form a unique fingerprint.

Cite this