Title | ||
---|---|---|
Editorial Editorial of Special Issue on Self-Supervised Learning for Speech and Audio Processing |
Abstract | ||
---|---|---|
The papers in this special section focus on self-supervised learning for speech and audio processing. A current trend in the machine learning community is the adoption of self-supervised approaches to pretrain deep networks. Self-supervised learning utilizes proxy-supervised learning tasks (or pretext tasks)—for example, distinguishing parts of the input signal from distractors or reconstructing masked input segments conditioned on unmasked segments—to obtain training data from unlabeled corpora. These approaches make it possible to use the tremendous amount of unlabeled data available on the web to train large neural models. Recent self-supervised approaches for speech and audio processing are also gaining attention. |
Year | DOI | Venue |
---|---|---|
2022 | 10.1109/JSTSP.2022.3205434 | IEEE Journal of Selected Topics in Signal Processing |
DocType | Volume | Issue |
Journal | 16 | 6 |
ISSN | Citations | PageRank |
1932-4553 | 0 | 0.34 |
References | Authors | |
3 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Hung-Yi Lee | 1 | 217 | 45.30 |
Shinji Watanabe | 2 | 1158 | 139.38 |
Karen Livescu | 3 | 1254 | 71.43 |
Abdel-rahman Mohamed | 4 | 3772 | 266.13 |
Tara N. Sainath | 5 | 3497 | 232.43 |