MEConformer: Highly representative embedding extractor for speaker verification via incorporating selective convolution into deep speaker encoder

计算机科学 卷积神经网络 判别式 变压器 编码器 嵌入 说话人识别 语音识别 特征提取 字错误率 模式识别(心理学) 人工智能 操作系统 物理 量子力学 电压
作者
Qiuyu Zheng,Zengzhao Chen,Zhifeng Wang,Hai Liu,Mingxing Lin
出处
期刊:Expert Systems With Applications [Elsevier]
卷期号:244: 123004-123004
标识
DOI:10.1016/j.eswa.2023.123004
摘要

Transformer models have demonstrated superior performance across various domains, including computer vision, natural language processing, and speech recognition. The success of these models can be attributed to their robust parallel capacity and high computation speed, primarily reliant on the attention layer. In the domain of speaker recognition, state-of-the-art results have been achieved using convolutional neural network (CNN) architectures, particularly with speaker embeddings represented by x-vectors and r-vectors. However, existing CNN-based methods tend to focus on local features while overlooking the global dependence of voiceprint features, resulting in the loss of crucial information. Moreover, the presence of noise in audio data is an influential factor that cannot be disregarded, as it significantly impacts the extraction of discriminative speaker embeddings. To address these challenges, we propose a novel model called the Multi-Scale Expand Convolution Transformer (MEConformer). This model aims to convert variable-length audio into a fixed low-dimensional representation. The MEConformer leverages a CNN framework with expanded receptive fields to capture frame-level features effectively. Additionally, we introduce a transformer encoder that incorporates contextual dependencies, enabling the extraction of both frame-level and discourse-level feature representations. Furthermore, we present a multi-scale residual aggregation strategy, which facilitates the efficient transmission of voiceprint information across the model. By combining these innovative components, the MEConformer achieves a state-of-the-art Equal Error Rate (EER) of 3.72% on the VoxCeleb1 test set. Furthermore, it demonstrates EERs of 5.94% and 3.72% on the VoxCeleb1-H and VoxCeleb-E datasets, respectively. The code for the proposed MEConformer model will be made publicly available at https://codeocean.com/capsule/4563012/tree.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
更新
大幅提高文件上传限制,最高150M (2024-4-1)

科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
刚刚
gstaihn发布了新的文献求助10
3秒前
Gavin完成签到,获得积分10
3秒前
小岳同学发布了新的文献求助30
7秒前
漂亮夏兰完成签到 ,获得积分10
8秒前
8秒前
8秒前
spirit完成签到 ,获得积分10
11秒前
GQ完成签到,获得积分10
11秒前
大个应助Crimson采纳,获得10
12秒前
13秒前
liuuuuu发布了新的文献求助10
13秒前
达西西发布了新的文献求助10
14秒前
wulixin完成签到,获得积分10
15秒前
华仔应助ning采纳,获得10
24秒前
liuuuuu完成签到,获得积分10
24秒前
liu bo完成签到,获得积分10
26秒前
27秒前
铁妹儿完成签到 ,获得积分10
28秒前
orixero应助ning采纳,获得10
29秒前
29秒前
NexusExplorer应助wen采纳,获得10
30秒前
安济应助科研通管家采纳,获得10
33秒前
科研通AI2S应助科研通管家采纳,获得10
33秒前
33秒前
33秒前
爆米花应助科研通管家采纳,获得10
33秒前
秋雪瑶应助科研通管家采纳,获得10
33秒前
SciGPT应助ning采纳,获得10
33秒前
加油完成签到 ,获得积分10
33秒前
38秒前
今后应助ning采纳,获得10
38秒前
达西西完成签到 ,获得积分10
42秒前
小金刀发布了新的文献求助10
45秒前
桐桐应助乐观的雅青采纳,获得10
45秒前
47秒前
47秒前
向路路发布了新的文献求助10
48秒前
CodeCraft应助如约而至采纳,获得10
48秒前
54秒前
高分求助中
Sustainable Land Management: Strategies to Cope with the Marginalisation of Agriculture 1000
Corrosion and Oxygen Control 600
Python Programming for Linguistics and Digital Humanities: Applications for Text-Focused Fields 500
Heterocyclic Stilbene and Bibenzyl Derivatives in Liverworts: Distribution, Structures, Total Synthesis and Biological Activity 500
重庆市新能源汽车产业大数据招商指南(两链两图两池两库两平台两清单两报告) 400
Division and square root. Digit-recurrence algorithms and implementations 400
行動データの計算論モデリング 強化学習モデルを例として 400
热门求助领域 (近24小时)
化学 材料科学 医学 生物 有机化学 工程类 生物化学 纳米技术 物理 内科学 计算机科学 化学工程 复合材料 遗传学 基因 物理化学 催化作用 电极 光电子学 量子力学
热门帖子
关注 科研通微信公众号,转发送积分 2547452
求助须知:如何正确求助?哪些是违规求助? 2176252
关于积分的说明 5603165
捐赠科研通 1897045
什么是DOI,文献DOI怎么找? 946545
版权声明 565383
科研通“疑难数据库(出版商)”最低求助积分说明 503793