摘要
In previous works, only parameter weights of ASR models are optimized under fixed-topology architecture. However, the design of successful model architecture has always relied on human experience and intuition. Besides, many hyperparameters related to model architecture need to be manually tuned. Therefore in this paper, we propose an ASR approach with efficient gradient-based architecture search, DARTS-ASR. In order to examine the generalizability of DARTS-ASR, we apply our approach not only on many languages to perform monolingual ASR, but also on a multilingual ASR setting. Following previous works, we conducted experiments on a multilingual dataset, IARPA BABEL. The experiment results show that our approach outperformed the baseline fixed-topology architecture by 10.2% and 10.0% relative reduction on character error rates under monolingual and multilingual ASR settings respectively. Furthermore, we perform some analysis on the searched architectures by DARTS-ASR.
原文 | 英語 |
---|---|
主出版物標題 | Interspeech 2020 |
發行者 | International Speech Communication Association |
頁面 | 1803-1807 |
頁數 | 5 |
ISBN(列印) | 9781713820697 |
DOIs | |
出版狀態 | 已出版 - 2020 |
事件 | 21st Annual Conference of the International Speech Communication Association, INTERSPEECH 2020 - Shanghai, 中國 持續時間: 25 10 2020 → 29 10 2020 |
出版系列
名字 | Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH |
---|---|
卷 | 2020-October |
ISSN(列印) | 2308-457X |
ISSN(電子) | 1990-9772 |
Conference
Conference | 21st Annual Conference of the International Speech Communication Association, INTERSPEECH 2020 |
---|---|
國家/地區 | 中國 |
城市 | Shanghai |
期間 | 25/10/20 → 29/10/20 |
文獻附註
Publisher Copyright:Copyright © 2020 ISCA