个人信息Personal Information
教师英文名称:Wei-Qiang Zhang
教师拼音名称:Zhang Wei Qiang
电子邮箱:
办公地点:电子工程馆5-111
联系方式:010-62781847
学位:博士学位
毕业院校:清华大学
学科:信号与信息处理
-
2023年-2026年:NSFC面上项目“基于自监督预训练模型的异常声音检测”,项目主持。
2021年-2023年:工信部人工智能产业创新任务揭榜挂帅项目“人工智能训练资源库”,课题主持。
2019年-2022年:NSFC联合重点项目“复杂环境下语音数据的说话人识别及关键词检索”,项目主持。
2019年-2023年:国家重点研发计划重点专项课题“基于语音信息的分析”,课题主持。
2019年-2021年:教育部项目“人工智能安全应用及人工智能安全防护关键技术研究”,课题主持。
2014年-2017年:NSFC面上项目“噪声和短语音条件下的说话人识别”,项目主持。
2011年-2013年:NSFC青年项目“面向海量语音信息处理的垃圾过滤和数据选择方法研究”,项目主持。
2010年-2013年:NSFC重大研究计划重点支持项目“多人多方对话中的语音分离、内容分析与理解”,项目参与。
2009年-2011年:国家863重点项目“陪护机器人”,项目参与。
-
H. Wang and W.-Q. Zhang, “Unstructured pruning and low rank factorisation of self-supervised pre-trained speech models,” IEEE Journal of Selected Topics in Signal Processing, 2024. doi: 10.1109/JSTSP.2024.3433616.
Y.-F. Shao, P. Jiang, Y. Dong, W. Li, and W.-Q. Zhang, “AE-IRMamba: Low complexity inverted residual Mamba for identification of piezoelectric ceramic and optical fiber acoustic emission sensors signals,” IEEE Sensors Journal, 2024. doi: 10.1109/JSEN.2024.3457913.
H. Zhang, N. Si, Y. Chen, X. Yang, D. Qu, and W.-Q. Zhang, “Improving speech translation by cross-modal multi-grained contrastive learning,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 31, pp. 1075–1086, 2023. doi: 10.1109/TASLP.2023.3244521.
J. Yao, X. Chen, X.-L. Zhang, W.-Q. Zhang, and K. Yang, “Symmetric saliency-based adversarial attack to speaker identification,” IEEE Signal Processing Letters, vol. 30, 2023. doi: 10.1109/LSP.2023.3236509.
Y. Qin, L. Sun, H. Chen, W. Yang, W.-Q. Zhang, J. Fei, and G. Wang, “MVKT-ECG: Efficient single-lead ECG classification for multi-label arrhythmia by multi-view knowledge transferring,” Computers in Biology and Medicine, vol. 166, Art. no. 107503, Sept. 2023. doi: 10.1016/j.compbiomed.2023.107503.
C. Wu, F. Wu, T. Qi, W.-Q. Zhang, X. Xie, and Y. Huang, “Removing AI’s sentiment manipulation of personalized news delivery,” Humanities & Social Sciences Communications, vol. 9, no. 459, 2022. doi: 10.1057/s41599-022-01473-1.
J. Zhao and W.-Q. Zhang, “Improving automatic speech recognition performance for low-resource languages with self-supervised models,” IEEE Journal of Selected Topics in Signal Processing, vol. 16, no. 6, pp. 1227–1241, Oct. 2022. doi: 10.1109/JSTSP.2022.3184480.
Z. Zhao and W.-Q. Zhang, “End-to-end keyword search system based on attention mechanism and energy scorer for low resource languages,” Neural Networks, vol. 139, pp. 326-334, Jul. 2021. doi: 10.1016/j.neunet.2021.04.002.
C. Lu, Y. Liu, W.-Q. Zhang and S. Zhang, “Tightness of a new and enhanced semidefinite relaxation for MIMO detection,” SIAM Journal on Optimization, vol. 29, no. 1, pp. 719-742, Jan. 2019. doi: 10.1137/17M115075X.
J. Kang, W.-Q. Zhang, W.-W. Liu, J. Liu, and M. T. Johnson, “Lattice based transcription loss for end-to-end speech recognition,” Journal of Signal Processing Systems, vol. 90, no. 7, pp. 1013-1023, Sept. 2018. doi: 10.1007/s11265-017-1292-0.
C. Lu, Z. Deng, W.-Q. Zhang, and S.-C. Fang, “Argument division based branch-and-bound algorithm for unit-modulus constrained complex quadratic programming,” Journal of Global Optimization, vol. 70, no. 1, pp. 171-187, Jan. 2018. doi: 10.1007/s10898-017-0551-8.
X.-K. Yang, L. He, D. Qu, and W.-Q. Zhang, “Semi-supervised minimum redundancy maximum relevance feature selection for audio classification,” Multimedia Tools and Applications, vol. 77, no. 1, pp. 713-739, Jan. 2018. doi: 10.1007/s11042-016-4287-0.
X. Yang, D. Qu, W.-L. Zhang, and W.-Q. Zhang, “An adapted data selection for deep learning-based audio segmentation in multi-genre broadcast channel,” Digital Signal Processing, vol. 81, pp. 8-15, Oct. 2018. doi: 10.1016/j.dsp.2018.03.004.
W.-W. Liu, M. Cai, W.-Q. Zhang, J. Liu, and M. T. Johnson, “Discriminative boosting algorithm for diversified front-end phonotactic language recognition,” Journal of Signal Processing Systems, vol. 82, no. 2, pp. 229-239, Feb. 2016. doi: 10.1007/s11265-015-1017-1.
W.-Q. Zhang, “Fast Doppler rate estimation based on fourth-order moment spectrum,” Electronics Letters, vol. 51, no. 23, pp. 1926–1928, Nov. 2015. doi: 10.1049/el.2015.2182.
Z.-Y. Li, W.-Q. Zhang, and J. Liu, “Multi-resolution time frequency feature and complementary combination for short utterance speaker recognition,” Multimedia Tools and Applications, vol. 74, no. 74, pp. 937-953, Feb. 2015. doi: 10.1007/s11042-013-1705-4.
W.-Q. Zhang, W.-W. Liu, Z.-Y. Li, Y.-Z. Shi, and J. Liu, “Spoken language recognition based on gap-weighted subsequence kernels,” Speech Communication, vol. 60, pp. 1-12, May 2014. doi: 10.1016/j.specom.2014.01.005.
Y.-Z. Shi, W.-Q. Zhang, J. Liu, and M. Johnson, “Efficient one-pass decoding with NNLM for speech recognition,” IEEE Signal Processing Letters, vol. 21, no. 4, pp. 377-381, Apr. 2014. doi: 10.1109/LSP.2014.2303136.
W.-L. Zhang, D. Qu, W.-Q. Zhang, and B.-C. Li. “Rapid speaker adaptation using compressive sensing,” Speech Communication, vol. 55, no. 10, pp. 950-963, Nov.-Dec. 2013. doi: 10.1016/j.specom.2013.06.012.
W.-L. Zhang, W.-Q. Zhang, B.-C. Li, D. Qu, and M. T. Johnson, “Bayesian speaker adaptation based on a new hierarchical probabilistic model,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 20, no. 7, pp. 2002-2015, Sept. 2012. doi: 10.1109/TASL.2012.2193390.
W.-Q. Zhang, L. He, Y. Deng, J. Liu, and M. Johnson, “Time-frequency cepstral feature and constrained heteroscedastic linear discriminant analysis for language recognition,” IEEE Transactions on Audio, Speech and Language Processing, vol. 19, no. 2, pp. 266-272, Feb. 2011. doi: 10.1109/TASL.2010.2047680.
W.-Q. Zhang, T. Hou, and J. Liu, “Discriminative score fusion for language identification,” Chinese Journal of Electronics, vol. 19, no. 1, pp. 124–128, Jan. 2010. doi: 10.23919/CJE.2010.10159256.
W.-Q. Zhang and J. Liu, “An equalized heteroscedastic linear discriminant analysis algorithm,” IEEE Signal Processing Letters, vol. 15, pp. 585-588, 2008. doi: 10.1109/LSP.2008.2001561.
R. Tao, W.-Q. Zhang, and E.-Q. Chen, “Two-stage method for joint time delay and Doppler shift estimation,” IET Radar, Sonar & Navigation, vol. 2, no. 1, pp. 71-77, Feb. 2008. doi: 10.1049/iet-rsn:20060014.
R. Tao, B. Deng, W.-Q. Zhang, and Y. Wang, “Sampling and sampling rate conversion of band limited signals in the fractional Fourier transform domain,” IEEE Transactions on Signal Processing, vol. 56, no. 1, pp. 158-171, Jan. 2008. doi: 10.1109/TSP.2007.901666.
-
B. Han, Z. Lv, A. Jiang, W. Huang, Z. Chen, Y. Deng, J. Ding, C. Lu, W.-Q. Zhang, P. Fan, J. Liu, and Y. Qian, “Exploring large scale pre-trained models for robust machine anomalous sound detection,” in Proc. ICASSP, 2024, pp. 1327–1330. doi: 10.1109/ICASSP48485.2024.10447183.
J. Li and W.-Q. Zhang, “Whisper-based transfer learning for Alzheimer disease classification: Leveraging speech segments with full transcripts as prompts,” in Proc. ICASSP, 2024, pp. 11211–11215. doi: 10.1109/ICASSP48485.2024.10448004.
H. Wang, G. Hu, G. Lin, W.-Q. Zhang, and J. Li, “Simul-Whisper: Attention-guided streaming Whisper with truncation detection,” in Proc. Interspeech, 2024, pp. 4483–4487. doi: 10.21437/Interspeech.2024-1814.
J. Li, Y. Pu, Q. Sun, and W.-Q. Zhang, “Improving Whisper’s recognition performance for under-represented language Kazakh leveraging unpaired speech and text,” in Proc. Interspeech, 2024, pp. 2514–2518. doi: 10.21437/Interspeech.2024-1790.
A. Jiang, B. Han, Z. Lv, Y. Deng, W.-Q. Zhang, X. Chen, Y. Qian, J. Liu, and P. Fan, “AnoPatch: Towards better consistency in machine anomalous sound detection,” in Proc. Interspeech, 2024, pp. 107–111. doi: 10.21437/Interspeech.2024-1761.
X. Zheng, A. Jiang, B. Han, Y. Qian, P. Fan, J. Liu, and W.-Q. Zhang, “Improving anomalous sound detection via low-rank adaptation fine-tuning of pre-trained audio models,” in Proc. SLT, 2024.
X. Chen, Y. Pu, J. Li, and W.-Q. Zhang, “Cross-lingual Alzheimer’s disease detection based on paralinguistic and pre-trained features,” in Proc. ICASSP, 2023. doi: 10.1109/ICASSP49357.2023.10095522.
A. Jiang, W.-Q. Zhang, Y. Deng, P. Fan, and J. Liu, “Unsupervised anomaly detection and localization of machine audio: A GAN-based approach,” in Proc. ICASSP, 2023. doi: 10.1109/ICASSP49357.2023.10096813.
H. Wang, S. Wang, W.-Q. Zhang, and J. Bai, “DistilXLSR: A light weight cross-lingual speech representation model,” in Proc. Interspeech, 2023, pp. 2273–2277. doi: 10.21437/Interspeech.2023-1444.
H. Wang, S. Wang, W.-Q. Zhang, H. Suo, and Y. Wan, “Task-agnostic structured pruning of speech representation models,” in Proc. Interspeech, 2023, pp. 231–235. doi: 10.21437/Interspeech.2023-1442.
Z. Cui, W. Wu, C. Zhang, W.-Q. Zhang, and J. Wu, “Transferring speech-generic and depression-specific knowledge for Alzheimer’s disease detection,” in Proc. ASRU, 2023. doi: 10.1109/ASRU57964.2023.10389785.
Y. Wang, C. Tang, Z. Ma, Z. Zheng, X. Chen, and W.-Q. Zhang, “Exploring effective distillation of self-supervised speech models for automatic speech recognition,” in Proc. ASRU, 2023. doi: 10.1109/ASRU57964.2023.10389746.
Q. Hou, A. Jiang, W.-Q. Zhang, P. Fan, and J. Liu, “Decoupling detectors for scalable anomaly detection in AIoT systems with multiple machines,” in Proc. GLOBECOM, 2023, pp. 5943–5948. doi: 10.1109/GLOBECOM54140.2023.10436800.
J. Zhao, H. Wang, J. Li, S. Chai, G. Wang, G. Chen, and W.-Q. Zhang, “The THUEE system description for the IARPA OpenASR21 challenge,” in Proc. Interspeech, 2022. doi: 10.21437/Interspeech.2022-269.
J. Zhao, G. Shi, G.-B. Wang, and W.-Q. Zhang, “Automatic speech recognition for low-resource languages: The THUEE systems for the IARPA OpenASR20 evaluation,” in Proc. ASRU, 2021, pp. 335–341. doi: 10.1109/ASRU51503.2021.9688260.
L. Xue, K. Song, D. Wu, X. Tan, N. L. Zhang, T. Qin, W.-Q. Zhang, and T.-Y. Liu, “DeepRapper: Neural rap generation with rhyme and rhythm modeling,” in Proc. ACL, 2021, pp. 69-81. doi: 10.18653/v1/2021.acl-long.6.
G. Chen, S. Chai, G. Wang, J. Du, W.-Q. Zhang, C. Weng, D. Su, D. Povey, J. Trmal, J. Zhang, M. Jin, S. Khudanpur, S. Watanabe, S. Zhao, W. Zou, X. Li, X. Yao, Y. Wang, Y. Wang, Z. You, and Z. Yan, “GigaSpeech: An evolving, multi-domain ASR corpus with 10,000 hours of transcribed audio,” in Proc. Interspeech, 2021, pp. 3670-3674. doi: 10.21437/Interspeech.2021-1965.
J. Zhao, Z. Lv, A. Han, G. Wang, G. Shi, J. Kang, J. Yan, P. Hu, S. Huang, and W.-Q. Zhang, “The TNT team system descriptions of Cantonese and Mongolian for IARPA OpenASR20,” in Proc. Interspeech, 2021, pp. 4344-4348. doi: 10.21437/Interspeech.2021-1063.
H. Yu, J. Zhao, S. Yang, Z. Wu, Y. Nie, and W.-Q. Zhang, “Language recognition based on unsupervised pretrained models,” in Proc. Interspeech, 2021, pp. 3271-3275. doi: 10.21437/Interspeech.2021-807.
Y. Yan, X. Tan, B. Li, G. Zhang, T. Qin, S. Zhao, Y. Shen, W.-Q. Zhang, and T.-Y. Liu, “Adaptive text to speech for spontaneous style,” in Proc. Interspeech, 2021, pp. 4668-4672. doi: 10.21437/Interspeech.2021-584.
K. He, Y. Shen, W.-Q. Zhang, and J. Liu, “Staged training strategy and multi-activation for audio tagging with noisy and sparse multi-label data,” in Proc. ICASSP, 2020, pp. 631-635. doi: 10.1109/ICASSP40776.2020.9053776.
J. Xie, R. Yan, S. Xiao, L. Peng, M. T. Johnson, and W.-Q. Zhang, “Dynamic temporal residual learning for speech recognition,” in Proc. ICASSP, 2020, pp. 7709-7713. doi: 10.1109/ICASSP40776.2020.9054653.
Z. Zhao and W.-Q. Zhang, “End-to-end keyword search based on attention and energy scorer for low resource languages,” in Proc. Interspeech, 2020, pp. 2587-2591. doi: 10.21437/Interspeech.2020-2613.
R. Li, T. Liang, D. Song, Y. Liu, Y. Wu, C. Xu, P. Ouyang, X. Zhang, X. Chen, W.-Q. Zhang, S. Yin, and L. He, “THUEE system for NIST SRE19 CTS challenge,” in Proc. Interspeech, 2020, pp. 2232-2236. doi: 10.21437/Interspeech.2020-1245.
Z. Li, L. He, J. Li, L. Wang, and W.-Q. Zhang, “Towards discriminative representations and unbiased predictions: Class-specific angular softmax for speech emotion recognition,” in Proc. Interspeech, 2019, pp. 1696-1700. doi: 10.21437/Interspeech.2019-1683.
K. He, Y. Shen, and W.-Q. Zhang, “Hierarchical pooling structure for weakly labeled sound event detection,” in Proc. Interspeech, 2019, pp. 3624-3628. doi: 10.21437/Interspeech.2019-2049.
H. Yang and W.-Q. Zhang, “Music genre classification using duplicated convolutional layers in neural networks,” in Proc. Interspeech, 2019, pp. 3382-3386. doi: 10.21437/Interspeech.2019-1298.
Y. Shen, K. He, and W.-Q. Zhang, “Learning how to listen: A temporal-frequential attention model for sound event detection,” in Proc. Interspeech, 2019, pp. 2563-2567. doi: 10.21437/Interspeech.2019-2045.
J. Kang, W.-Q. Zhang, and J. Liu, “Gated convolutional networks based hybrid acoustic models for low resource speech recognition,” in Proc. ASRU, 2017, pp. 157-164. doi: 10.1109/ASRU.2017.8268930.
Z.-Q. Lv, J. Kang, W.-Q. Zhang, and J. Liu, “An LSTM-CTC based verification system for proxy-word based OOV keyword search,” in Proc. ICASSP, 2017, pp. 5655-5659. doi: 10.1109/ICASSP.2017.7953239.
Y. Tian, L. He, M. Cai, W.-Q. Zhang, and J. Liu, “Deep neural networks based speaker modeling at different levels of phonetic granularity,” in Proc. ICASSP, 2017, pp. 5440-5444. doi: 10.1109/ICASSP.2017.7953196.
X.-K. Yang, D. Qu, W.-L. Zhang, and W.-Q. Zhang, “The NDSC transcription system for the 2016 multi-genre broadcast challenge,” in Proc. SLT, 2016, pp. 273-278. doi: 10.1109/SLT.2016.7846276.
Z.-Q. Lv, M. Cai, W.-Q. Zhang, and J. Liu, “A novel discriminative score calibration method for keyword search,” in Proc. Interspeech, 2016, pp. 745-749. doi: 10.21437/Interspeech.2016-606.
Y. Tian, M. Cai, H. Liang, W.-Q. Zhang, and J. Liu, “Improving deep neural networks based speaker verification using unlabeled data,” in Proc. Interspeech, 2016, pp. 1863-1867. doi: 10.21437/Interspeech.2016-614.
Z.-Q. Lv, M. Cai, C. Lu, J. Kang, L.-K. Hui, W.-Q. Zhang, and J. Liu, “Improved system fusion for keyword search,” in Proc. ASRU, 2015, pp. 231-236. doi: 10.1109/ASRU.2015.7404799.
M. Cai, Z.-Q. Lv, B.-L. Song, Y.-Z. Shi, W.-L. Wu, C. Lu, W.-Q. Zhang, and J. Liu, “The THUEE system for the OpenKWS14 keyword search evaluation,” in Proc. ICASSP, 2015, pp. 4734-4738. doi: 10.1109/ICASSP.2015.7178869.
J. Kang, C. Lu, M. Cai, W.-Q. Zhang, and J. Liu, “Neuron sparseness versus connection sparseness in deep neural network for large vocabulary speech recognition,” in Proc. ICASSP, 2015, pp. 4954-4958. doi: 10.1109/ICASSP.2015.7178913.
Y.-Z. Shi, W.-Q. Zhang, M. Cai, and J. Liu, “Variance regularization of RNNLM for speech recognition,” in Proc. ICASSP, 2014, pp. 4931-4935. doi: 10.1109/ICASSP.2014.6854532.
W.-W. Liu, W.-Q. Zhang, Y.-Z. Shi, A. Ji, J. Xu, and J. Liu, “Improved phonotactic language recognition based on RNN feature reconstruction,” in Proc. ICASSP, 2014, pp. 5359-5363. doi: 10.1109/ICASSP.2014.6854619.
W.-W. Liu, W.-Q. Zhang, and J. Liu, “Phonotactic language identification based on time-gap-weighted lattice kernels,” in Proc. Interspeech, 2014, pp. 3022-3026. doi: 10.21437/Interspeech.2014-606.
W.-L. Zhang, D. Qu, W.-Q. Zhang, and B.-C. Li, “Speaker adaptation based on sparse and low-rank eigenphone matrix estimation,” in Proc. Interspeech, 2014, pp. 2792-2796. doi: 10.21437/Interspeech.2014-496.
Z.-Y. Li, W.-Q. Zhang, W.-W. Liu, Y. Tian, and J. Liu, “Text-independent speaker verification via state alignment,” in Proc. Odyssey, 2014, pp. 68–72. Available: doi: 10.21437/Odyssey.2014-10.
Y.-Z. Shi, W.-Q. Zhang, M. Cai, and J. Liu, “Temporal kernel neural network language model,” in Proc. ICASSP, 2013, pp. 8247-8251. doi: 10.1109/ICASSP.2013.6639273.
W.-Q. Zhang, Z.-Y. Li, W. Liu, and J. Liu, “THU-EE system fusion for the NIST 2012 speaker recognition evaluation,” in Proc. Interspeech, 2013, pp. 2474-2478. doi: 10.21437/Interspeech.2013-413.
W. Liu, W.-Q. Zhang, Zhang, Z.-Y. Li, and J. Liu. “Parallel absolute-relative feature based phonotactic language recognition,” in Proc. Interspeech, 2013, pp. 59-63. doi: 10.21437/Interspeech.2013-38.
W.-L. Zhang, W.-Q. Zhang, and B.-C. Li, “Compact acoustic modeling based on acoustic manifold using a mixture of factor analyzers,” in Proc. ASRU, 2013, pp. 37-42. doi: 10.1109/ASRU.2013.6707702.
Z.-Y. Li, W.-Q. Zhang, L. He, and J. Liu, “Complementary combination in i-vector level for language recognition,” in Proc. Odyssey, 2012, pp. 334-337. Available: https://www.isca-archive.org/odyssey_2012/li12_odyssey.pdf
Y.-Z. Shi, W.-Q. Zhang, and J. Liu, “Robust audio fingerprinting based on local spectral luminance maxima scheme,” in Proc. Interspeech, 2011, pp. 2485-2488. doi: 10.21437/Interspeech.2011-636.
W.-L. Zhang, W.-Q. Zhang, and B.-C. Li, “Speaker adaptation based on speaker-dependent eigenphone estimation,” in Proc. ASRU, 2011, pp. 48-52. doi: 10.1109/ASRU.2011.6163904.
W.-Q. Zhang, Y. Deng, L. He, and J. Liu, “Variant time-frequency cepstral features for speaker recognition,” in Proc. Interspeech, 2010, pp. 2122-2125. doi: 10.21437/Interspeech.2010-160.
S. Meng, W.-Q. Zhang, and J. Liu, “Combining Chinese spoken term detection systems via side-information conditioned linear logistic regression,” in Proc. Interspeech, 2010, pp. 685-688. doi: 10.21437/Interspeech.2010-260.
J. Yang, J. Liu, and W.-Q. Zhang, “A fast query by humming system based on notes,” in Proc. Interspeech, 2010, pp. 2898-2901. doi: 10.21437/Interspeech.2010-753.
W.-Q. Zhang, Y. Shan, and J. Liu, “Multiple background models for speaker verification,” in Proc. Odyssey, 2010, pp. 47-51. Available: https://www.isca-archive.org/odyssey_2010/zhang10_odyssey.pdf
W.-Q. Zhang and J. Liu, “Two-stage method for specific audio retrieval,” in Proc. ICASSP, 2007, pp. IV-85-88. doi: 10.1109/ICASSP.2007.367169.