Paper ID | AUD-31.1 | ||
Paper Title | Impact of Sound Duration and Inactive Frames on Sound Event Detection Performance | ||
Authors | Keisuke Imoto, Doshisha University, Japan; Sakiko Mishima, Yumi Arai, Reishi Kondo, NEC Corporation, Japan | ||
Session | AUD-31: Detection and Classification of Acoustic Scenes and Events 6: Events | ||
Location | Gather.Town | ||
Session Time: | Friday, 11 June, 13:00 - 13:45 | ||
Presentation Time: | Friday, 11 June, 13:00 - 13:45 | ||
Presentation | Poster | ||
Topic | Audio and Acoustic Signal Processing: [AUD-CLAS] Detection and Classification of Acoustic Scenes and Events | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | In many methods of sound event detection (SED), a segmented time frame is regarded as one data sample to model training. The durations of sound events greatly depend on the sound event class, e.g., the sound event ``fan'' has a long duration, whereas the sound event ``mouse clicking'' is instantaneous. Thus, the difference in the duration between sound event classes results in a serious data imbalance in SED. Moreover, most sound events tend to occur occasionally; therefore, there are many more inactive time frames of sound events than active frames. This also causes a severe data imbalance between active and inactive frames. In this paper, we investigate the impact of sound duration and inactive frames on SED performance by introducing four loss functions, such as simple reweighting loss, inverse frequency loss, asymmetric focal loss, and focal batch Tversky loss. Then, we provide insights into how we tackle this imbalance problem. |