每年專案
摘要
When recording conversations, there may be multiple people talking at once. While our human ears can filter out unwanted sounds, this can be challenging for automatic speech recognition (ASR) systems, leading to reduced accuracy. To address this issue, preprocessing mechanisms such as speech separation and targeted speaker extraction are necessary to separate each person’s speech. With the development of deep learning, the quality of separated speech has improved significantly. Our objective is to focus on speaker extraction, which entails implementing a primary system for speech extraction and a secondary subsystem for delivering target information. To accomplish this, we have chosen a temporal convolutional network (TCN) architecture as the foundation of our speech extraction model. A TCN enables convolutional neural networks (CNNs) to manage time series modeling, and it can be constructed in various model lengths. Furthermore, we have integrated attention enhancement into the secondary subsystem to provide the speech extraction model with comprehensive and effective target information, which helps to improve the model’s ability to estimate masks. As a result, the quality of the target speaker extraction will be greatly enhanced with a more precise mask.
原文 | ???core.languages.en_GB??? |
---|---|
文章編號 | 307 |
期刊 | Electronics (Switzerland) |
卷 | 13 |
發行號 | 2 |
DOIs | |
出版狀態 | 已出版 - 1月 2024 |
指紋
深入研究「Target Speaker Extraction Using Attention-Enhanced Temporal Convolutional Network」主題。共同形成了獨特的指紋。專案
- 1 已完成