IEEE ICASSP 2022

2022 IEEE International Conference on Acoustics, Speech and Signal Processing

7-13 May 2022
  • Virtual (all paper presentations)
22-27 May 2022
  • Main Venue: Marina Bay Sands Expo & Convention Center, Singapore
27-28 October 2022
  • Satellite Venue: Crowne Plaza Shenzhen Longgang City Centre, Shenzhen, China

ICASSP 2022
AUD-15.2

CAA-NET: CONDITIONAL ATROUS CNNS WITH ATTENTION FOR EXPLAINABLE DEVICE-ROBUST ACOUSTIC SCENE CLASSIFICATION

Zhao Ren, University of Augsburg & Leibniz University Hannover, Germany; Qiuqiang Kong, ByteDance AI Laboratory, China; Jing Han, University of Cambridge, United Kingdom of Great Britain and Northern Ireland; Mark Plumbley, University of Surrey, United Kingdom of Great Britain and Northern Ireland; Björn Schuller, Imperial College London & University of Augsburg, United Kingdom of Great Britain and Northern Ireland

Session:
Detection and Classification of Acoustic Scenes and Events IV: On-device Considerations

Track:
Audio and Acoustic Signal Processing

Location:
Gather Area K

Presentation Time:
Tue, 10 May, 21:00 - 21:45 China Time (UTC +8)
Tue, 10 May, 13:00 - 13:45 UTC

Session Chair:
Qiuqiang Kong, Bytedance
Presentation
Discussion
Resources
No resources available.
Session AUD-15
AUD-15.1: SELF-SUPERVISED REPRESENTATION LEARNING FOR UNSUPERVISED ANOMALOUS SOUND DETECTION UNDER DOMAIN SHIFT
Han Chen, Yan Song, Li-Rong Dai, Ian McLoughlin, University of Science and Technology of China, China; Lin Liu, iFLYTEK Research, iFLYTEK CO., LTD, China
AUD-15.2: CAA-NET: CONDITIONAL ATROUS CNNS WITH ATTENTION FOR EXPLAINABLE DEVICE-ROBUST ACOUSTIC SCENE CLASSIFICATION
Zhao Ren, University of Augsburg & Leibniz University Hannover, Germany; Qiuqiang Kong, ByteDance AI Laboratory, China; Jing Han, University of Cambridge, United Kingdom of Great Britain and Northern Ireland; Mark Plumbley, University of Surrey, United Kingdom of Great Britain and Northern Ireland; Björn Schuller, Imperial College London & University of Augsburg, United Kingdom of Great Britain and Northern Ireland
AUD-15.3: FEDERATED SELF-TRAINING FOR DATA-EFFICIENT AUDIO RECOGNITION
Vasileios Tsouvalas, Tanir Ozcelebi, Eindhoven University of Technology, Netherlands; Aaqib Saeed, Philips Research, Netherlands
AUD-15.4: FEDERATED SELF-SUPERVISED LEARNING FOR ACOUSTIC EVENT CLASSIFICATION
Meng Feng, Massachusetts Institute of Technology, United States of America; Chieh-Chi Kao, Qingming Tang, Ming Sun, Viktor Rozgic, Spyros Matsoukas, Chao Wang, Amazon, United States of America
AUD-15.5: TEMPORAL KNOWLEDGE DISTILLATION FOR ON-DEVICE AUDIO CLASSIFICATION
Kwanghee Choi, Martin Kersner, Jacob Morton, Buru Chang, Hyperconnect, Korea, Republic of
AUD-15.6: STREAMING ON-DEVICE DETECTION OF DEVICE DIRECTED SPEECH FROM VOICE AND TOUCH-BASED INVOCATION
Ognjen (Oggi) Rudovic, Akanksha Bindal, Vineet Garg, Pramod Simha, Pranay Dighe, Sachin Kajarekar, Apple Inc., United States of America