Chao Weng

2.5k total citations · 2 hit papers
71 papers, 1.4k citations indexed

About

Chao Weng is a scholar working on Artificial Intelligence, Signal Processing and Computer Vision and Pattern Recognition. According to data from OpenAlex, Chao Weng has authored 71 papers receiving a total of 1.4k indexed citations (citations by other indexed papers that have themselves been cited), including 61 papers in Artificial Intelligence, 58 papers in Signal Processing and 5 papers in Computer Vision and Pattern Recognition. Recurrent topics in Chao Weng's work include Speech Recognition and Synthesis (60 papers), Speech and Audio Processing (54 papers) and Music and Audio Processing (45 papers). Chao Weng is often cited by papers focused on Speech Recognition and Synthesis (60 papers), Speech and Audio Processing (54 papers) and Music and Audio Processing (45 papers). Chao Weng collaborates with scholars based in China, United States and Hong Kong. Chao Weng's co-authors include Dong Yu, Dan Su, Shinji Watanabe, Chengzhu Yu, Helen Meng, Jianwei Yu, Meng Yu, Biing-Hwang Juang, Yuexian Zou and Dongchao Yang and has published in prestigious journals such as IEEE Signal Processing Letters, Computer Speech & Language and IEEE/ACM Transactions on Audio Speech and Language Processing.

In The Last Decade

Chao Weng

67 papers receiving 1.3k citations

Hit Papers

Diffsound: Discrete Diffusion Model for Text-to-Sound Gen... 2023 2026 2024 2025 2023 2024 25 50 75 100

Peers — A (Enhanced Table)

Peers by citation overlap · career bar shows stage (early→late) cites · hero ref

Name h Career Trend Papers Cites
Chao Weng China 22 1.0k 992 209 88 43 71 1.4k
Ehsan Variani United States 11 939 0.9× 884 0.9× 83 0.4× 63 0.7× 38 0.9× 22 1.1k
Xiao-Lei Zhang China 16 816 0.8× 838 0.8× 176 0.8× 173 2.0× 92 2.1× 83 1.2k
Khe Chai Sim Singapore 21 1.6k 1.5× 1.2k 1.2× 157 0.8× 31 0.4× 41 1.0× 111 1.8k
Tian Tan China 14 578 0.6× 516 0.5× 133 0.6× 58 0.7× 46 1.1× 25 825
Liang Lu United States 19 932 0.9× 851 0.9× 72 0.3× 84 1.0× 38 0.9× 64 1.1k
Shigeki Karita Japan 13 1.5k 1.5× 1.1k 1.1× 116 0.6× 53 0.6× 36 0.8× 26 1.8k
Dimitrios Dimitriadis United States 18 569 0.5× 473 0.5× 92 0.4× 54 0.6× 28 0.7× 57 813
Tom Ko China 13 1.6k 1.5× 1.3k 1.3× 167 0.8× 31 0.4× 34 0.8× 44 1.8k
Jahn Heymann Germany 17 1.4k 1.3× 1.5k 1.5× 101 0.5× 372 4.2× 102 2.4× 27 1.8k
Atsunori Ogawa Japan 20 1.1k 1.0× 1.0k 1.0× 81 0.4× 141 1.6× 63 1.5× 136 1.5k

Countries citing papers authored by Chao Weng

Since Specialization
Citations

This map shows the geographic impact of Chao Weng's research. It shows the number of citations coming from papers published by authors working in each country. You can also color the map by specialization and compare the number of citations received by Chao Weng with the expected number of citations based on a country's size and research output (numbers larger than one mean the country cites Chao Weng more than expected).

Fields of papers citing papers by Chao Weng

Since Specialization
Physical SciencesHealth SciencesLife SciencesSocial Sciences

This network shows the impact of papers produced by Chao Weng. Nodes represent research fields, and links connect fields that are likely to share authors. Colored nodes show fields that tend to cite the papers produced by Chao Weng. The network helps show where Chao Weng may publish in the future.

Co-authorship network of co-authors of Chao Weng

This figure shows the co-authorship network connecting the top 25 collaborators of Chao Weng. A scholar is included among the top collaborators of Chao Weng based on the total number of citations received by their joint publications. Widths of edges represent the number of papers authors have co-authored together. Node borders signify the number of papers an author published with Chao Weng. Chao Weng is excluded from the visualization to improve readability, since they are connected to all nodes in the network.

All Works

20 of 20 papers shown
1.
Yang, Dongchao, Songxiang Liu, Rongjie Huang, Chao Weng, & Helen Meng. (2024). InstructTTS: Modelling Expressive TTS in Discrete Latent Space With Natural Language Style Prompt. IEEE/ACM Transactions on Audio Speech and Language Processing. 32. 2913–2925. 21 indexed citations
3.
Chen, Haoxin, Yong Zhang, Menghan Xia, et al.. (2024). VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models. 7310–7320. 46 indexed citations breakdown →
4.
Cui, Jianwei, Yu Gu, Chao Weng, et al.. (2024). Sifisinger: A High-Fidelity End-to-End Singing Voice Synthesizer Based on Source-Filter Model. 33. 11126–11130. 1 indexed citations
5.
Yang, Dongchao, Jianwei Yu, Helin Wang, et al.. (2023). Diffsound: Discrete Diffusion Model for Text-to-Sound Generation. IEEE/ACM Transactions on Audio Speech and Language Processing. 31. 1720–1733. 108 indexed citations breakdown →
6.
Yang, Dongchao, et al.. (2023). NoreSpeech: Knowledge Distillation based Conditional Diffusion Model for Noise-robust Expressive TTS. 4798–4802. 9 indexed citations
7.
Yu, Jianwei, et al.. (2023). High Fidelity Speech Enhancement with Band-split RNN. 2483–2487. 21 indexed citations
8.
Yu, Jianwei, et al.. (2022). Integrating Lattice-Free MMI Into End-to-End Speech Recognition. IEEE/ACM Transactions on Audio Speech and Language Processing. 31. 25–38. 9 indexed citations
9.
Subramanian, Aswin Shanmugam, Chao Weng, Shinji Watanabe, Meng Yu, & Dong Yu. (2022). Deep learning based multi-source localization with source splitting and its effectiveness in multi-talker speech recognition. Computer Speech & Language. 75. 101360–101360. 59 indexed citations
10.
Yu, Jianwei, et al.. (2022). Improving Mandarin End-to-End Speech Recognition With Word N-Gram Language Model. IEEE Signal Processing Letters. 29. 812–816. 8 indexed citations
11.
Li, Chenyi, et al.. (2022). Enhancing Speaking Styles in Conversational Text-to-Speech Synthesis with Graph-Based Multi-Modal Context Modeling. ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 7917–7921. 13 indexed citations
12.
Chen, Guoguo, Wei-Qiang Zhang, Chao Weng, et al.. (2021). GigaSpeech: An Evolving, Multi-Domain ASR Corpus with 10,000 Hours of Transcribed Audio. 3670–3674. 117 indexed citations
13.
Li, Xu, Na Li, Chao Weng, et al.. (2021). Replay and Synthetic Speech Detection with Res2Net Architecture. 6354–6358. 113 indexed citations
14.
Lam, Max W. Y., Jun Wang, Chao Weng, Dan Su, & Dong Yu. (2021). Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition. 316–320. 1 indexed citations
15.
Zhang, Deyuan, et al.. (2021). Video Human Action Recognition with Channel Attention on ST-GCN. Journal of Physics Conference Series. 2010(1). 12131–12131. 1 indexed citations
16.
You, Zhao, Dan Su, Jie Chen, Chao Weng, & Dong Yu. (2020). Dfsmn-San with Persistent Memory Model for Automatic Speech Recognition. abs 1907 1470. 7704–7708. 6 indexed citations
17.
Weng, Chao, Jia Cui, Guangsen Wang, et al.. (2018). Improving Attention Based Sequence-to-Sequence Models for End-to-End English Conversational Speech Recognition. 761–765. 38 indexed citations
18.
Huang, Zhen, et al.. (2014). Deep learning vector quantization for acoustic information retrieval. 10 indexed citations
19.
Weng, Chao & Biing‐Hwang Juang. (2013). Latent semantic rational kernels for topic spotting on spontaneous conversational speech. 8302–8306. 2 indexed citations
20.
Weng, Chao, Biing‐Hwang Juang, & Daniel Povey. (2012). Discriminative training using non-uniform criteria for keyword spotting on spontaneous speech. 559–562. 3 indexed citations

Rankless uses publication and citation data sourced from OpenAlex, an open and comprehensive bibliographic database. While OpenAlex provides broad and valuable coverage of the global research landscape, it—like all bibliographic datasets—has inherent limitations. These include incomplete records, variations in author disambiguation, differences in journal indexing, and delays in data updates. As a result, some metrics and network relationships displayed in Rankless may not fully capture the entirety of a scholar's output or impact.

Explore authors with similar magnitude of impact

Rankless by CCL
2026