


default search action
Yu Shi 0001
Person information
- affiliation: Microsoft, Redmond, WA, USA
Other persons with the same name
- Yu Shi — disambiguation page
- Yu Shi 0002 — Facebook, Menlo Park, CA, USA (and 1 more)
- Yu Shi 0003
— Chinese Academy of Sciences (CAS), Chongqing Institute of Green and Intelligent Technology (CIGIT), Chongqing, China
- Yu Shi 0004
— Wuhan Institute of Technology, Hubei Key Laboratory of Optical Information and Pattern Recognition, Wuhan, China (and 2 more)
- Yu Shi 0005
— Jilin University, International Center of Future Science, Changchun, China
Refine list

refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [c32]Ziyi Yang, Mahmoud Khademi, Yichong Xu, Reid Pryzant, Yuwei Fang, Chenguang Zhu, Dongdong Chen, Yao Qian, Xuemei Gao, Yi-Ling Chen, Robert Gmyr, Naoyuki Kanda, Noel Codella, Bin Xiao, Yu Shi, Lu Yuan, Takuya Yoshioka, Michael Zeng, Xuedong Huang:
i-Code V2: An Autoregressive Generation Framework over Vision, Language, and Speech Data. NAACL-HLT (Findings) 2024: 1615-1627 - 2023
- [j3]Junwei Liao
, Sefik Emre Eskimez
, Liyang Lu
, Yu Shi
, Ming Gong
, Linjun Shou
, Hong Qu
, Michael Zeng
:
Improving Readability for Automatic Speech Recognition Transcription. ACM Trans. Asian Low Resour. Lang. Inf. Process. 22(5): 142:1-142:23 (2023) - [c31]Ziyi Yang, Yuwei Fang, Chenguang Zhu, Reid Pryzant, Dongdong Chen, Yu Shi, Yichong Xu, Yao Qian, Mei Gao, Yi-Ling Chen, Liyang Lu, Yujia Xie, Robert Gmyr, Noel Codella, Naoyuki Kanda, Bin Xiao, Lu Yuan, Takuya Yoshioka, Michael Zeng, Xuedong Huang:
i-Code: An Integrative and Composable Multimodal Learning Framework. AAAI 2023: 10880-10890 - [c30]Pengcheng He, Baolin Peng, Song Wang, Yang Liu, Ruochen Xu, Hany Hassan, Yu Shi, Chenguang Zhu, Wayne Xiong, Michael Zeng, Jianfeng Gao, Xuedong Huang:
Z-Code++: A Pre-trained Language Model Optimized for Abstractive Summarization. ACL (1) 2023: 5095-5112 - [c29]Haibin Yu, Yuxuan Hu, Yao Qian, Ma Jin, Linquan Liu, Shujie Liu, Yu Shi, Yanmin Qian, Edward Lin, Michael Zeng:
Code-Switching Text Generation and Injection in Mandarin-English ASR. ICASSP 2023: 1-5 - [i16]Haibin Yu, Yuxuan Hu, Yao Qian, Ma Jin, Linquan Liu, Shujie Liu, Yu Shi, Yanmin Qian, Edward Lin, Michael Zeng:
Code-Switching Text Generation and Injection in Mandarin-English ASR. CoRR abs/2303.10949 (2023) - [i15]Ziyi Yang, Mahmoud Khademi, Yichong Xu, Reid Pryzant, Yuwei Fang, Chenguang Zhu, Dongdong Chen, Yao Qian, Mei Gao, Yi-Ling Chen, Robert Gmyr, Naoyuki Kanda, Noel Codella, Bin Xiao, Yu Shi, Lu Yuan, Takuya Yoshioka, Michael Zeng, Xuedong Huang:
i-Code V2: An Autoregressive Generation Framework over Vision, Language, and Speech Data. CoRR abs/2305.12311 (2023) - 2022
- [j2]Junwei Liao
, Yu Shi, Yong Xu:
Automatic Speech Recognition Post-Processing for Readability: Task, Dataset and a Two-Stage Pre-Trained Approach. IEEE Access 10: 117053-117066 (2022) - [c28]Wei Wang, Shuo Ren, Yao Qian, Shujie Liu, Yu Shi, Yanmin Qian, Michael Zeng:
Optimizing Alignment of Speech and Language Latent Spaces for End-To-End Speech Recognition and Understanding. ICASSP 2022: 7802-7806 - [i14]Ziyi Yang, Yuwei Fang, Chenguang Zhu, Reid Pryzant, Dongdong Chen, Yu Shi, Yichong Xu, Yao Qian, Mei Gao, Yi-Ling Chen, Liyang Lu, Yujia Xie, Robert Gmyr, Noel Codella, Naoyuki Kanda, Bin Xiao, Lu Yuan, Takuya Yoshioka, Michael Zeng, Xuedong Huang:
i-Code: An Integrative and Composable Multimodal Learning Framework. CoRR abs/2205.01818 (2022) - [i13]Pengcheng He, Baolin Peng, Liyang Lu, Song Wang, Jie Mei, Yang Liu, Ruochen Xu, Hany Hassan Awadalla, Yu Shi, Chenguang Zhu, Wayne Xiong, Michael Zeng, Jianfeng Gao, Xuedong Huang:
Z-Code++: A Pre-trained Language Model Optimized for Abstractive Summarization. CoRR abs/2208.09770 (2022) - [i12]Xiaofei Wang, Zhuo Chen, Yu Shi, Jian Wu, Naoyuki Kanda, Takuya Yoshioka:
Breaking trade-offs in speech separation with sparsely-gated mixture of experts. CoRR abs/2211.06493 (2022) - 2021
- [c27]Yao Qian, Ximo Bian, Yu Shi, Naoyuki Kanda, Leo Shen, Zhen Xiao, Michael Zeng:
Speech-Language Pre-Training for End-to-End Spoken Language Understanding. ICASSP 2021: 7458-7462 - [c26]Junwei Liao, Yu Shi, Ming Gong, Linjun Shou, Sefik Emre Eskimez, Liyang Lu, Hong Qu, Michael Zeng:
Generating Human Readable Transcript for Automatic Speech Recognition with Pre-Trained Language Model. ICASSP 2021: 7578-7582 - [c25]Junwei Liao, Yu Shi, Ming Gong, Linjun Shou, Hong Qu, Michael Zeng:
Improving Zero-shot Neural Machine Translation on Language-specific Encoders- Decoders. IJCNN 2021: 1-8 - [c24]Shahram Ghorbani, Yashesh Gaur, Yu Shi, Jinyu Li
:
Listen, Look and Deliberate: Visual Context-Aware Speech Recognition Using Pre-Trained Text-Video Representations. SLT 2021: 621-628 - [i11]Yao Qian, Ximo Bian, Yu Shi, Naoyuki Kanda, Leo Shen, Zhen Xiao, Michael Zeng:
Speech-language Pre-training for End-to-end Spoken Language Understanding. CoRR abs/2102.06283 (2021) - [i10]Junwei Liao, Yu Shi, Ming Gong, Linjun Shou, Hong Qu, Michael Zeng:
Improving Zero-shot Neural Machine Translation on Language-specific Encoders-Decoders. CoRR abs/2102.06578 (2021) - [i9]Junwei Liao, Yu Shi, Ming Gong, Linjun Shou, Sefik Emre Eskimez, Liyang Lu, Hong Qu, Michael Zeng:
Generating Human Readable Transcript for Automatic Speech Recognition with Pre-trained Language Model. CoRR abs/2102.11114 (2021) - [i8]Linhao Zhang, Yu Shi, Linjun Shou, Ming Gong, Houfeng Wang, Michael Zeng:
A Joint and Domain-Adaptive Approach to Spoken Language Understanding. CoRR abs/2107.11768 (2021) - [i7]Wei Wang, Shuo Ren, Yao Qian, Shujie Liu, Yu Shi, Yanmin Qian, Michael Zeng:
Optimizing Alignment of Speech and Language Latent Spaces for End-to-End Speech Recognition and Understanding. CoRR abs/2110.12138 (2021) - [i6]Lu Yuan, Dongdong Chen, Yi-Ling Chen, Noel Codella, Xiyang Dai, Jianfeng Gao, Houdong Hu, Xuedong Huang, Boxin Li, Chunyuan Li, Ce Liu, Mengchen Liu, Zicheng Liu, Yumao Lu, Yu Shi, Lijuan Wang, Jianfeng Wang, Bin Xiao, Zhen Xiao, Jianwei Yang, Michael Zeng, Luowei Zhou, Pengchuan Zhang:
Florence: A New Foundation Model for Computer Vision. CoRR abs/2111.11432 (2021) - [i5]Ken'ichi Kumatani, Robert Gmyr, Felipe Cruz Salinas, Linquan Liu, Wei Zuo, Devang Patel, Eric Sun, Yu Shi:
Building a great multi-lingual teacher with sparsely-gated mixture of experts for speech recognition. CoRR abs/2112.05820 (2021) - 2020
- [c23]Ruochen Xu, Chenguang Zhu, Yu Shi, Michael Zeng, Xuedong Huang:
Mixed-Lingual Pre-training for Cross-lingual Summarization. AACL/IJCNLP 2020: 536-541 - [c22]Huaishao Luo, Yu Shi, Ming Gong, Linjun Shou, Tianrui Li:
MaP: A Matrix-based Prediction Approach to Improve Span Extraction in Machine Reading Comprehension. AACL/IJCNLP 2020: 687-695 - [c21]Yao Qian, Yu Shi, Michael Zeng:
Discriminative Transfer Learning for Optimizing ASR and Semantic Labeling in Task-Oriented Spoken Dialog. INTERSPEECH 2020: 3915-3919 - [i4]Junwei Liao, Sefik Emre Eskimez, Liyang Lu, Yu Shi, Ming Gong, Linjun Shou, Hong Qu, Michael Zeng:
Improving Readability for Automatic Speech Recognition Transcription. CoRR abs/2004.04438 (2020) - [i3]Huaishao Luo, Yu Shi, Ming Gong, Linjun Shou, Tianrui Li:
MaP: A Matrix-based Prediction Approach to Improve Span Extraction in Machine Reading Comprehension. CoRR abs/2009.14348 (2020) - [i2]Ruochen Xu, Chenguang Zhu, Yu Shi, Michael Zeng, Xuedong Huang:
Mixed-Lingual Pre-training for Cross-lingual Summarization. CoRR abs/2010.08892 (2020) - [i1]Shahram Ghorbani, Yashesh Gaur, Yu Shi, Jinyu Li:
Listen, Look and Deliberate: Visual context-aware speech recognition using pre-trained text-video representations. CoRR abs/2011.04084 (2020)
2010 – 2019
- 2010
- [c20]Yongqiang Wang, Qiang Huo, Yu Shi:
A Study of Discriminative Training for HMM-Based Online Handwritten Chinese/Japanese Character Recognition. ICFHR 2010: 518-523 - [c19]Guangchuan Shi, Yu Shi, Qiang Huo:
A study of irrelevant variability normalization based training and unsupervised online adaptation for LVCSR. INTERSPEECH 2010: 1357-1360
2000 – 2009
- 2009
- [c18]Lei Ma, Qiang Huo, Yu Shi:
A Study of Feature Design for Online Handwritten Chinese Character Recognition Based on Continuous-Density Hidden Markov Models. ICDAR 2009: 526-530 - 2008
- [c17]Zhen Xuan Luo, Yu Shi, Frank K. Soong:
Symbol graph based discriminative training and rescoring for improved math symbol recognition. ICASSP 2008: 1953-1956 - [c16]Peng Yu, Yu Shi, Frank Seide:
Approximateword-lattice indexing with text indexers: Time-Anchored Lattice Expansion. ICASSP 2008: 5248-5251 - [c15]Yu Shi, Frank K. Soong:
A symbol graph based handwritten math expression recognition. ICPR 2008: 1-4 - [c14]Yu Shi, Frank Seide, Frank K. Soong:
GPU-accelerated Gaussian clustering for fMPE discriminative training. INTERSPEECH 2008: 944-947 - 2007
- [c13]Frank Seide, Peng Yu, Yu Shi:
Towards spoken-document retrieval for the enterprise: Approximate word-lattice indexing with text indexers. ASRU 2007: 629-634 - [c12]Yanlu Xie, Yu Shi, Frank K. Soong, Beiqian Dai:
A Segmentation Posterior Based Endpointing Algorithm. ICASSP (4) 2007: 813-816 - [c11]Yu Shi, HaiYang Li, Frank K. Soong:
A Unified Framework for Symbol Segmentation and Recognition of Handwritten Mathematical Expressions. ICDAR 2007: 854-858 - 2006
- [c10]Yu Shi, Frank K. Soong, Jian-Lai Zhou:
Auto-Segmentation Based Partitioning and Clustering Approach to Robust Endpointing. ICASSP (1) 2006: 793-796 - [c9]Yu Shi, Frank K. Soong, Jian-Lai Zhou:
Auto-segmentation based VAD for robust ASR. INTERSPEECH 2006 - [c8]Yu Shi, Frank K. Soong, Jian-Lai Zhou:
Integrating Hypotheses of Multiple Recognizers for Improving Mandarin LVCSR Performance. ISCSLP 2006 - [c7]Dongwen Ying, Yu Shi, Frank K. Soong, Jianwu Dang, Xugang Lu:
A Robust Voice Activity Detection Based on Noise Eigenspace Projection. ISCSLP (Selected Papers) 2006: 76-86 - 2004
- [c6]Yu Shi, Eric Chang:
Studies in massively speaker-specific speech recognition. ICASSP (1) 2004: 825-828 - [c5]Chao Huang, Yu Shi, Jianlai Zhou, Min Chu, Terry Wang, Eric Chang:
Segmental tonal modeling for phone set design in Mandarin LVCSR. ICASSP (1) 2004: 901-904 - [c4]Jian-Lai Zhou, Ye Tian, Yu Shi, Chao Huang, Eric Chang:
Tone articulation modeling for Mandarin spontaneous speech recognition. ICASSP (1) 2004: 997-1000 - 2003
- [c3]Yu Shi, Eric Chang:
Spectrogram-based formant tracking via particle filters. ICASSP (1) 2003: 168-171 - 2002
- [j1]Eric Chang, Frank Seide, Helen M. Meng, Zhuoran Chen, Yu Shi, Yuk-Chi Li:
A system for spoken query information retrieval on mobile devices. IEEE Trans. Speech Audio Process. 10(8): 531-541 (2002) - [c2]Yu Shi, Eric Chang, Hu Peng, Min Chu:
Power spectral density based channel equalization of large speech database for concatenative TTS system. INTERSPEECH 2002: 2369-2372 - 2001
- [c1]Eric Chang, Yu Shi, Jian-Lai Zhou, Chao Huang:
Speech lab in a box: a Mandarin speech toolbox to jumpstart speech related research. INTERSPEECH 2001: 2799-2802
Coauthor Index

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from ,
, and
to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and
to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2025-02-21 20:32 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint