Wei-Ning Hsu

4.0k total citations
52 papers, 1.4k citations indexed

About

Wei-Ning Hsu is a scholar working on Artificial Intelligence, Signal Processing and Information Systems. According to data from OpenAlex, Wei-Ning Hsu has authored 52 papers receiving a total of 1.4k indexed citations (citations by other indexed papers that have themselves been cited), including 45 papers in Artificial Intelligence, 26 papers in Signal Processing and 2 papers in Information Systems. Recurrent topics in Wei-Ning Hsu's work include Speech Recognition and Synthesis (37 papers), Music and Audio Processing (23 papers) and Speech and Audio Processing (21 papers). Wei-Ning Hsu is often cited by papers focused on Speech Recognition and Synthesis (37 papers), Music and Audio Processing (23 papers) and Speech and Audio Processing (21 papers). Wei-Ning Hsu collaborates with scholars based in United States, Israel and France. Wei-Ning Hsu's co-authors include James Glass, Yu-An Chung, Hao Tang, Yossi Adi, Hsuan-Tien Lin, Abdelrahman Mohamed, Abdelrahman Mohamed, Adam Polyak, Jade Copet and Emmanuel Dupoux and has published in prestigious journals such as Transactions of the Association for Computational Linguistics, Natural Product Communications and arXiv (Cornell University).

In The Last Decade

Wei-Ning Hsu

49 papers receiving 1.3k citations

Peers — A (Enhanced Table)

Peers by citation overlap · career bar shows stage (early→late) cites · hero ref

Name h Career Trend Papers Cites
Wei-Ning Hsu United States 20 1.2k 691 145 75 35 52 1.4k
Thomas Fang Zheng China 16 813 0.7× 608 0.9× 97 0.7× 63 0.8× 37 1.1× 102 972
Khe Chai Sim Singapore 21 1.6k 1.3× 1.2k 1.7× 157 1.1× 50 0.7× 50 1.4× 111 1.8k
Andros Tjandra Japan 16 712 0.6× 387 0.6× 132 0.9× 40 0.5× 22 0.6× 39 923
Pietro Laface Italy 21 1.4k 1.1× 1.1k 1.6× 131 0.9× 154 2.1× 18 0.5× 104 1.6k
Zoltán Tüske Germany 18 740 0.6× 524 0.8× 83 0.6× 26 0.3× 18 0.5× 47 871
Erik McDermott Japan 20 1.9k 1.6× 1.5k 2.2× 266 1.8× 162 2.2× 38 1.1× 68 2.2k
Shuang Xu China 9 593 0.5× 385 0.6× 232 1.6× 44 0.6× 13 0.4× 16 893
Xu Tan China 23 914 0.8× 398 0.6× 400 2.8× 34 0.5× 24 0.7× 61 1.1k
Björn Hoffmeister Germany 18 715 0.6× 489 0.7× 82 0.6× 37 0.5× 21 0.6× 38 850

Countries citing papers authored by Wei-Ning Hsu

Since Specialization
Citations

This map shows the geographic impact of Wei-Ning Hsu's research. It shows the number of citations coming from papers published by authors working in each country. You can also color the map by specialization and compare the number of citations received by Wei-Ning Hsu with the expected number of citations based on a country's size and research output (numbers larger than one mean the country cites Wei-Ning Hsu more than expected).

Fields of papers citing papers by Wei-Ning Hsu

Since Specialization
Physical SciencesHealth SciencesLife SciencesSocial Sciences

This network shows the impact of papers produced by Wei-Ning Hsu. Nodes represent research fields, and links connect fields that are likely to share authors. Colored nodes show fields that tend to cite the papers produced by Wei-Ning Hsu. The network helps show where Wei-Ning Hsu may publish in the future.

Co-authorship network of co-authors of Wei-Ning Hsu

This figure shows the co-authorship network connecting the top 25 collaborators of Wei-Ning Hsu. A scholar is included among the top collaborators of Wei-Ning Hsu based on the total number of citations received by their joint publications. Widths of edges represent the number of papers authors have co-authored together. Node borders signify the number of papers an author published with Wei-Ning Hsu. Wei-Ning Hsu is excluded from the visualization to improve readability, since they are connected to all nodes in the network.

All Works

20 of 20 papers shown
1.
2.
Majumder, Navonil, et al.. (2024). Tango 2: Aligning Diffusion-based Text-to-Audio Generations through Direct Preference Optimization. 564–572. 12 indexed citations
3.
Wang, Changhan, Hirofumi Inaguma, Peng‐Jen Chen, et al.. (2023). Simple and Effective Unsupervised Speech Translation. 10771–10784. 9 indexed citations
4.
Kharitonov, Eugene, Jade Copet, Yossi Adi, et al.. (2023). Generative Spoken Dialogue Language Modeling. Transactions of the Association for Computational Linguistics. 11. 250–266. 20 indexed citations
5.
Chen, Peng‐Jen, Kevin Tran, Yilin Yang, et al.. (2023). Speech-to-Speech Translation for a Real-world Unwritten Language. 4969–4983. 12 indexed citations
7.
Hsu, Wei-Ning, Bowen Shi, Itai Gat, et al.. (2023). Expresso: A Benchmark and Analysis of Discrete Expressive Speech Resynthesis. SPIRE - Sciences Po Institutional REpository. 4823–4827. 10 indexed citations
8.
Chou, Ju-Chieh, Wei-Ning Hsu, Karen Livescu, et al.. (2023). Toward Joint Language Modeling for Speech Units and Text. 6582–6593. 3 indexed citations
10.
Kreuk, Felix, Adam Polyak, Jade Copet, et al.. (2022). Textless Speech Emotion Conversion using Discrete & Decomposed Representations. 11200–11214. 14 indexed citations
11.
Lee, Ann, Adam Polyak, Yossi Adi, et al.. (2022). Text-Free Prosody-Aware Generative Spoken Language Modeling. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 8666–8681. 26 indexed citations
12.
Chen, Peng‐Jen, Changhan Wang, Juan Pino, et al.. (2022). Enhanced Direct Speech-to-Speech Translation Using Self-supervised Pre-training and Data Augmentation. Interspeech 2022. 5195–5199. 24 indexed citations
13.
Lee, Ann, Peng‐Jen Chen, Changhan Wang, et al.. (2022). Direct Speech-to-Speech Translation With Discrete Units. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 3327–3339. 63 indexed citations
14.
Shi, Bowen, Wei-Ning Hsu, & Abdelrahman Mohamed. (2022). Robust Self-Supervised Audio-Visual Speech Recognition. Interspeech 2022. 2118–2122. 64 indexed citations
15.
Polyak, Adam, Yossi Adi, Jade Copet, et al.. (2021). Speech Resynthesis from Discrete Disentangled Self-Supervised Representations. arXiv (Cornell University). 3615–3619. 144 indexed citations
16.
Hsu, Wei-Ning, Yao-Hung Hubert Tsai, Benjamin Bolte, Ruslan Salakhutdinov, & Abdelrahman Mohamed. (2021). Hubert: How Much Can a Bad Teacher Benefit ASR Pre-Training?. 6533–6537. 83 indexed citations
17.
Lakhotia, Kushal, Wei-Ning Hsu, Yossi Adi, et al.. (2021). On Generative Spoken Language Modeling from Raw Audio. Transactions of the Association for Computational Linguistics. 9. 1336–1354. 10 indexed citations
18.
Chung, Yu-An, Wei-Ning Hsu, Hao Tang, & James Glass. (2019). An Unsupervised Autoregressive Model for Speech Representation Learning. 146–150. 210 indexed citations
19.
Hsu, Wei-Ning, Yu Zhang, & James Glass. (2017). Unsupervised Learning of Disentangled and Interpretable Representations from Sequential Data. Neural Information Processing Systems. 30. 1878–1889. 62 indexed citations
20.
Romeo, Salvatore, Giovanni Da San Martino, Alberto Barrón‐Cedeño, et al.. (2016). Neural Attention for Learning to Rank Questions in Community Question Answering. International Conference on Computational Linguistics. 1734–1745. 22 indexed citations

Rankless uses publication and citation data sourced from OpenAlex, an open and comprehensive bibliographic database. While OpenAlex provides broad and valuable coverage of the global research landscape, it—like all bibliographic datasets—has inherent limitations. These include incomplete records, variations in author disambiguation, differences in journal indexing, and delays in data updates. As a result, some metrics and network relationships displayed in Rankless may not fully capture the entirety of a scholar's output or impact.

Explore authors with similar magnitude of impact

Rankless by CCL
2026