default search action
Junichi Yamagishi
Person information
- affiliation: National Institute of Informatics, Tokyo, Japan
- affiliation (former): University of Edinburgh, Scotland, UK
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2025
- [j80]Aidan Pine, Erica Cooper, David Guzmán, Eric Joanis, Anna Kazantseva, Ross Krekoski, Roland Kuhn, Samuel Larkin, Patrick Littell, Delaney Lothian, Akwiratékha' Martin, Korin Richmond, Marc Tessier, Cassia Valentini-Botinhao, Dan Wells, Junichi Yamagishi:
Speech Generation for Indigenous Language Education. Comput. Speech Lang. 90: 101723 (2025) - 2024
- [j79]Khanh-Duy Nguyen, Huy H. Nguyen, Trung-Nghia Le, Junichi Yamagishi, Isao Echizen:
Analysis of Fine-Grained Counting Methods for Masked Face Counting: A Comparative Study. IEEE Access 12: 27426-27443 (2024) - [j78]Hichem Felouat, Huy H. Nguyen, Trung-Nghia Le, Junichi Yamagishi, Isao Echizen:
eKYC-DF: A Large-Scale Deepfake Dataset for Developing and Evaluating eKYC Systems. IEEE Access 12: 30876-30892 (2024) - [j77]Chang Zeng, Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi:
Joint speaker encoder and neural back-end model for fully end-to-end automatic speaker verification with multiple enrollment utterances. Comput. Speech Lang. 86: 101619 (2024) - [j76]Michele Panariello, Natalia A. Tomashenko, Xin Wang, Xiaoxiao Miao, Pierre Champion, Hubert Nourtel, Massimiliano Todisco, Nicholas W. D. Evans, Emmanuel Vincent, Junichi Yamagishi:
The VoicePrivacy 2022 Challenge: Progress and Perspectives in Voice Anonymisation. IEEE ACM Trans. Audio Speech Lang. Process. 32: 3477-3491 (2024) - [j75]Cheng Gong, Xin Wang, Erica Cooper, Dan Wells, Longbiao Wang, Jianwu Dang, Korin Richmond, Junichi Yamagishi:
ZMM-TTS: Zero-Shot Multilingual and Multispeaker Speech Synthesis Conditioned on Self-Supervised Discrete Speech Representations. IEEE ACM Trans. Audio Speech Lang. Process. 32: 4036-4051 (2024) - [c273]Shirin Dabbaghi Varnosfaderani, Canasai Kruengkrai, Ramin Yahyapour, Junichi Yamagishi:
Bridging Textual and Tabular Worlds for Fact Verification: A Lightweight, Attention-Based Model. LREC/COLING 2024: 2515-2519 - [c272]Aditya Ravuri, Erica Cooper, Junichi Yamagishi:
Uncertainty as a Predictor: Leveraging Self-Supervised Learning for Zero-Shot MOS Prediction. ICASSP Workshops 2024: 580-584 - [c271]Xin Wang, Junichi Yamagishi:
Can Large-Scale Vocoded Spoofed Data Improve Speech Spoofing Countermeasure with a Self-Supervised Front End? ICASSP 2024: 10311-10315 - [c270]Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi, Nicholas W. D. Evans, Massimiliano Todisco, Jean-François Bonastre, Mickael Rouvier:
Synvox2: Towards A Privacy-Friendly Voxceleb2 Dataset. ICASSP 2024: 11421-11425 - [c269]Wanying Ge, Xin Wang, Junichi Yamagishi, Massimiliano Todisco, Nicholas W. D. Evans:
Spoofing Attack Augmentation: Can Differently-Trained Attack Models Improve Generalisation? ICASSP 2024: 12531-12535 - [c268]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Exploring Self-Supervised Vision Transformers for Deepfake Detection: A Comparative Analysis. IJCB 2024: 1-10 - [i151]Shirin Dabbaghi Varnosfaderani, Canasai Kruengkrai, Ramin Yahyapour, Junichi Yamagishi:
Bridging Textual and Tabular Worlds for Fact Verification: A Lightweight, Attention-Based Model. CoRR abs/2403.17361 (2024) - [i150]Natalia A. Tomashenko, Xiaoxiao Miao, Pierre Champion, Sarina Meyer, Xin Wang, Emmanuel Vincent, Michele Panariello, Nicholas W. D. Evans, Junichi Yamagishi, Massimiliano Todisco:
The VoicePrivacy 2024 Challenge Evaluation Plan. CoRR abs/2404.02677 (2024) - [i149]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Exploring Self-Supervised Vision Transformers for Deepfake Detection: A Comparative Analysis. CoRR abs/2405.00355 (2024) - [i148]Jee-weon Jung, Xin Wang, Nicholas W. D. Evans, Shinji Watanabe, Hye-jin Shim, Hemlata Tak, Sidhhant Arora, Junichi Yamagishi, Joon Son Chung:
To what extent can ASV systems naturally defend against spoofing attacks? CoRR abs/2406.05339 (2024) - [i147]Lin Zhang, Xin Wang, Erica Cooper, Mireia Díez, Federico Landini, Nicholas W. D. Evans, Junichi Yamagishi:
Spoof Diarization: "What Spoofed When" in Partially Spoofed Audio. CoRR abs/2406.07816 (2024) - [i146]Yun Liu, Xuechen Liu, Xiaoxiao Miao, Junichi Yamagishi:
Target Speaker Extraction with Curriculum Learning. CoRR abs/2406.07845 (2024) - [i145]Zhengyang Chen, Xuechen Liu, Erica Cooper, Junichi Yamagishi, Yanmin Qian:
Generating Speakers by Prompting Listener Impressions for Pre-trained Multi-Speaker Text-to-Speech Systems. CoRR abs/2406.08812 (2024) - [i144]Cheng Gong, Erica Cooper, Xin Wang, Chunyu Qiang, Mengzhe Geng, Dan Wells, Longbiao Wang, Jianwu Dang, Marc Tessier, Aidan Pine, Korin Richmond, Junichi Yamagishi:
An Initial Investigation of Language Adaptation for TTS Systems under Low-resource Scenarios. CoRR abs/2406.08911 (2024) - [i143]Xin Wang, Tomi Kinnunen, Kong Aik Lee, Paul-Gauthier Noé, Junichi Yamagishi:
Revisiting and Improving Scoring Fusion for Spoofing-aware Speaker Verification Using Compositional Data Analysis. CoRR abs/2406.10836 (2024) - [i142]Xin Wang, Héctor Delgado, Hemlata Tak, Jee-weon Jung, Hye-jin Shim, Massimiliano Todisco, Ivan Kukanov, Xuechen Liu, Md. Sahidullah, Tomi Kinnunen, Nicholas W. D. Evans, Kong Aik Lee, Junichi Yamagishi:
ASVspoof 5: Crowdsourced Speech Data, Deepfakes, and Adversarial Attacks at Scale. CoRR abs/2408.08739 (2024) - [i141]Xuechen Liu, Xin Wang, Junichi Yamagishi:
A Preliminary Case Study on Long-Form In-the-Wild Audio Spoofing Detection. CoRR abs/2408.14066 (2024) - [i140]Zhengyang Chen, Shuai Wang, Mingyang Zhang, Xuechen Liu, Junichi Yamagishi, Yanmin Qian:
Disentangling the Prosody and Semantic Information with Pre-trained Model for In-Context Learning based Zero-Shot Voice Conversion. CoRR abs/2409.05004 (2024) - [i139]Chang Zeng, Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi:
Spoofing-Aware Speaker Verification Robust Against Domain and Channel Mismatches. CoRR abs/2409.06327 (2024) - [i138]Wen-Chin Huang, Szu-Wei Fu, Erica Cooper, Ryandhimas E. Zezario, Tomoki Toda, Hsin-Min Wang, Junichi Yamagishi, Yu Tsao:
The VoiceMOS Challenge 2024: Beyond Speech Quality Prediction. CoRR abs/2409.07001 (2024) - [i137]Jesujoba O. Alabi, Xuechen Liu, Dietrich Klakow, Junichi Yamagishi:
AfriHuBERT: A self-supervised speech representation model for African languages. CoRR abs/2409.20201 (2024) - [i136]Yun Liu, Xuechen Liu, Junichi Yamagishi:
Improving curriculum learning for target speaker extraction with synthetic speakers. CoRR abs/2410.00811 (2024) - [i135]Natalia A. Tomashenko, Xiaoxiao Miao, Emmanuel Vincent, Junichi Yamagishi:
The First VoicePrivacy Attacker Challenge Evaluation Plan. CoRR abs/2410.07428 (2024) - 2023
- [j74]Ching-Chun Chang, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Cyber Vaccine for Deepfake Immunity. IEEE Access 11: 105027-105039 (2023) - [j73]Lin Zhang, Xin Wang, Erica Cooper, Nicholas W. D. Evans, Junichi Yamagishi:
The PartialSpoof Database and Countermeasures for the Detection of Short Fake Speech Segments Embedded in an Utterance. IEEE ACM Trans. Audio Speech Lang. Process. 31: 813-825 (2023) - [j72]Xuechen Liu, Xin Wang, Md. Sahidullah, Jose Patino, Héctor Delgado, Tomi Kinnunen, Massimiliano Todisco, Junichi Yamagishi, Nicholas W. D. Evans, Andreas Nautsch, Kong Aik Lee:
ASVspoof 2021: Towards Spoofed and Deepfake Speech Detection in the Wild. IEEE ACM Trans. Audio Speech Lang. Process. 31: 2507-2522 (2023) - [j71]Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi, Natalia A. Tomashenko:
Speaker Anonymization Using Orthogonal Householder Neural Network. IEEE ACM Trans. Audio Speech Lang. Process. 31: 3681-3695 (2023) - [j70]Kunkun Pang, Dafei Qin, Yingruo Fan, Julian Habekost, Takaaki Shiratori, Junichi Yamagishi, Taku Komura:
BodyFormer: Semantics-guided 3D Body Gesture Synthesis with Transformer. ACM Trans. Graph. 42(4): 43:1-43:12 (2023) - [c267]Canasai Kruengkrai, Junichi Yamagishi:
Revisiting Pathologies of Neural Models under Input Reduction. ACL (Findings) 2023: 11504-11517 - [c266]Lifan Zhong, Erica Cooper, Junichi Yamagishi, Nobuaki Minematsu:
Exploring Isolated Musical Notes as Pre-training Data for Predominant Instrument Recognition in Polyphonic Music. APSIPA ASC 2023: 2312-2319 - [c265]Erica Cooper, Wen-Chin Huang, Yu Tsao, Hsin-Min Wang, Tomoki Toda, Junichi Yamagishi:
The Voicemos Challenge 2023: Zero-Shot Subjective Speech Quality Prediction for Multiple Domains. ASRU 2023: 1-7 - [c264]Hemant Yadav, Erica Cooper, Junichi Yamagishi, Sunayana Sitaram, Rajiv Ratn Shah:
Partial Rank Similarity Minimization Method for Quality MOS Prediction of Unseen Speech Synthesis Systems in Zero-Shot and Semi-Supervised Setting. ASRU 2023: 1-7 - [c263]Haoyu Li, Yun Liu, Junichi Yamagishi:
Joint Noise Reduction and Listening Enhancement for Full-End Speech Enhancement. ICASSP 2023: 1-5 - [c262]Paul-Gauthier Noé, Xiaoxiao Miao, Xin Wang, Junichi Yamagishi, Jean-François Bonastre, Driss Matrouf:
Hiding Speaker's Sex in Speech Using Zero-Evidence Speaker Representation in an Analysis/Synthesis Pipeline. ICASSP 2023: 1-5 - [c261]Xuan Shi, Erica Cooper, Xin Wang, Junichi Yamagishi, Shrikanth Narayanan:
Can Knowledge of End-to-End Text-to-Speech Models Improve Neural Midi-to-Audio Synthesis Systems? ICASSP 2023: 1-5 - [c260]Xin Wang, Junichi Yamagishi:
Spoofed Training Data for Speech Spoofing Countermeasure Can Be Efficiently Created Using Neural Vocoders. ICASSP 2023: 1-5 - [c259]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
How Close Are Other Computer Vision Tasks to Deepfake Detection? IJCB 2023: 1-10 - [c258]Erica Cooper, Junichi Yamagishi:
Investigating Range-Equalizing Bias in Mean Opinion Score Ratings of Synthesized Speech. INTERSPEECH 2023: 1104-1108 - [c257]Chang Zeng, Xin Wang, Xiaoxiao Miao, Erica Cooper, Junichi Yamagishi:
Improving Generalization Ability of Countermeasures for New Mismatch Scenario by Combining Multiple Advanced Regularization Terms. INTERSPEECH 2023: 1998-2002 - [c256]Lin Zhang, Xin Wang, Erica Cooper, Nicholas W. D. Evans, Junichi Yamagishi:
Range-Based Equal Error Rate for Spoof Localization. INTERSPEECH 2023: 3212-3216 - [c255]Sung Hwan Mun, Hye-jin Shim, Hemlata Tak, Xin Wang, Xuechen Liu, Md. Sahidullah, Myeonghun Jeong, Min Hyun Han, Massimiliano Todisco, Kong Aik Lee, Junichi Yamagishi, Nicholas W. D. Evans, Tomi Kinnunen, Nam Soo Kim, Jee-weon Jung:
Towards Single Integrated Spoofing-aware Speaker Verification Embeddings. INTERSPEECH 2023: 3989-3993 - [c254]Hieu-Thi Luong, Junichi Yamagishi:
Controlling Multi-Class Human Vocalization Generation via a Simple Segment-based Labeling Scheme. INTERSPEECH 2023: 4379-4383 - [c253]Huy H. Nguyen, Trung-Nghia Le, Junichi Yamagishi, Isao Echizen:
Analysis of Master Vein Attacks on Finger Vein Recognition Systems. WACV 2023: 1900-1908 - [d2]Junichi Yamagishi:
Model checkpoints for "XFEVER: Exploring Fact Verification across Languages". Zenodo, 2023 - [i134]Ching-Chun Chang, Huy Hong Nguyen, Junichi Yamagishi, Isao Echizen:
Cyber Vaccine for Deepfake Immunity. CoRR abs/2303.02659 (2023) - [i133]Lin Zhang, Xin Wang, Erica Cooper, Nicholas W. D. Evans, Junichi Yamagishi:
Range-Based Equal Error Rate for Spoof Localization. CoRR abs/2305.17739 (2023) - [i132]Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi, Natalia A. Tomashenko:
Language-independent speaker anonymization using orthogonal Householder neural network. CoRR abs/2305.18823 (2023) - [i131]Sung Hwan Mun, Hye-jin Shim, Hemlata Tak, Xin Wang, Xuechen Liu, Md. Sahidullah, Myeonghun Jeong, Min Hyun Han, Massimiliano Todisco, Kong Aik Lee, Junichi Yamagishi, Nicholas W. D. Evans, Tomi Kinnunen, Nam Soo Kim, Jee-weon Jung:
Towards single integrated spoofing-aware speaker verification embeddings. CoRR abs/2305.19051 (2023) - [i130]Lifan Zhong, Erica Cooper, Junichi Yamagishi, Nobuaki Minematsu:
Exploring Isolated Musical Notes as Pre-training Data for Predominant Instrument Recognition in Polyphonic Music. CoRR abs/2306.08850 (2023) - [i129]Xin Wang, Junichi Yamagishi:
Can large-scale vocoded spoofed data improve speech spoofing countermeasure with a self-supervised front end? CoRR abs/2309.06014 (2023) - [i128]Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi, Nicholas W. D. Evans, Massimiliano Todisco, Jean-François Bonastre, Mickael Rouvier:
SynVox2: Towards a privacy-friendly VoxCeleb2 dataset. CoRR abs/2309.06141 (2023) - [i127]Nicolas Jonason, Xin Wang, Erica Cooper, Lauri Juvela, Bob L. T. Sturm, Junichi Yamagishi:
DDSP-based Neural Waveform Synthesis of Polyphonic Guitar Performance from String-wise MIDI Input. CoRR abs/2309.07658 (2023) - [i126]Wanying Ge, Xin Wang, Junichi Yamagishi, Massimiliano Todisco, Nicholas W. D. Evans:
Spoofing attack augmentation: can differently-trained attack models improve generalisation? CoRR abs/2309.09586 (2023) - [i125]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
How Close are Other Computer Vision Tasks to Deepfake Detection? CoRR abs/2310.00922 (2023) - [i124]Hemant Yadav, Erica Cooper, Junichi Yamagishi, Sunayana Sitaram, Rajiv Ratn Shah:
Partial Rank Similarity Minimization Method for Quality MOS Prediction of Unseen Speech Synthesis Systems in Zero-Shot and Semi-supervised setting. CoRR abs/2310.05078 (2023) - [i123]Kunkun Pang, Dafei Qin, Yingruo Fan, Julian Habekost, Takaaki Shiratori, Junichi Yamagishi, Taku Komura:
BodyFormer: Semantics-guided 3D Body Gesture Synthesis with Transformer. CoRR abs/2310.06851 (2023) - [i122]Yi-Chen Chang, Canasai Kruengkrai, Junichi Yamagishi:
XFEVER: Exploring Fact Verification across Languages. CoRR abs/2310.16278 (2023) - [i121]Xuechen Liu, Xin Wang, Erica Cooper, Xiaoxiao Miao, Junichi Yamagishi:
Speaker-Text Retrieval via Contrastive Learning. CoRR abs/2312.06055 (2023) - [i120]Cheng Gong, Xin Wang, Erica Cooper, Dan Wells, Longbiao Wang, Jianwu Dang, Korin Richmond, Junichi Yamagishi:
ZMM-TTS: Zero-shot Multilingual and Multispeaker Speech Synthesis Conditioned on Self-supervised Discrete Speech Representations. CoRR abs/2312.14398 (2023) - [i119]Aditya Ravuri, Erica Cooper, Junichi Yamagishi:
Uncertainty as a Predictor: Leveraging Self-Supervised Learning for Zero-Shot MOS Prediction. CoRR abs/2312.15616 (2023) - 2022
- [j69]Natalia A. Tomashenko, Xin Wang, Emmanuel Vincent, Jose Patino, Brij Mohan Lal Srivastava, Paul-Gauthier Noé, Andreas Nautsch, Nicholas W. D. Evans, Junichi Yamagishi, Benjamin O'Brien, Anaïs Chanclu, Jean-François Bonastre, Massimiliano Todisco, Mohamed Maouche:
The VoicePrivacy 2020 Challenge: Results and findings. Comput. Speech Lang. 74: 101362 (2022) - [j68]Huy H. Nguyen, Minoru Kuribayashi, Junichi Yamagishi, Isao Echizen:
Effects of Image Processing Operations on Adversarial Noise and Their Use in Detecting and Correcting Adversarial Images. IEICE Trans. Inf. Syst. 105-D(1): 65-77 (2022) - [j67]Cheng-Hung Hu, Yu-Huai Peng, Junichi Yamagishi, Yu Tsao, Hsin-Min Wang:
SVSNet: An End-to-End Speaker Voice Similarity Assessment Model. IEEE Signal Process. Lett. 29: 767-771 (2022) - [j66]Xuan Shi, Erica Cooper, Junichi Yamagishi:
Use of Speaker Recognition Approaches for Learning and Evaluating Embedding Representations of Musical Instrument Sounds. IEEE ACM Trans. Audio Speech Lang. Process. 30: 367-377 (2022) - [j65]Anssi Kanervisto, Ville Hautamäki, Tomi Kinnunen, Junichi Yamagishi:
Optimizing Tandem Speaker Verification and Anti-Spoofing Systems. IEEE ACM Trans. Audio Speech Lang. Process. 30: 477-488 (2022) - [j64]Brij Mohan Lal Srivastava, Mohamed Maouche, Md. Sahidullah, Emmanuel Vincent, Aurélien Bellet, Marc Tommasi, Natalia A. Tomashenko, Xin Wang, Junichi Yamagishi:
Privacy and Utility of X-Vector Based Speaker Anonymization. IEEE ACM Trans. Audio Speech Lang. Process. 30: 2383-2395 (2022) - [j63]Huy H. Nguyen, Sébastien Marcel, Junichi Yamagishi, Isao Echizen:
Master Face Attacks on Face Recognition Systems. IEEE Trans. Biom. Behav. Identity Sci. 4(3): 398-411 (2022) - [c252]Canasai Kruengkrai, Junichi Yamagishi:
Mitigating the Diminishing Effect of Elastic Weight Consolidation. COLING 2022: 4568-4574 - [c251]Wen-Chin Huang, Erica Cooper, Junichi Yamagishi, Tomoki Toda:
LDNet: Unified Listener Dependent Modeling in MOS Prediction for Synthetic Speech. ICASSP 2022: 896-900 - [c250]Xin Wang, Junichi Yamagishi:
Estimating the Confidence of Speech Spoofing Countermeasure. ICASSP 2022: 6372-6376 - [c249]Chang Zeng, Xin Wang, Erica Cooper, Xiaoxiao Miao, Junichi Yamagishi:
Attention Back-End for Automatic Speaker Verification with Multiple Enrollment Utterances. ICASSP 2022: 6717-6721 - [c248]Erica Cooper, Wen-Chin Huang, Tomoki Toda, Junichi Yamagishi:
Generalization Ability of MOS Prediction Networks. ICASSP 2022: 8442-8446 - [c247]Cheng-I Jeff Lai, Erica Cooper, Yang Zhang, Shiyu Chang, Kaizhi Qian, Yi-Lun Liao, Yung-Sung Chuang, Alexander H. Liu, Junichi Yamagishi, David D. Cox, James R. Glass:
On the Interplay between Sparsity, Naturalness, Intelligibility, and Prosody in Speech Synthesis. ICASSP 2022: 8447-8451 - [c246]Li-Kuang Chen, Canasai Kruengkrai, Junichi Yamagishi:
Outlier-Aware Training for Improving Group Accuracy Disparities. AACL/IJCNLP 2022 (Student Research Workshop) 2022: 54-60 - [c245]Chang Zeng, Lin Zhang, Meng Liu, Junichi Yamagishi:
Spoofing-Aware Attention based ASV Back-end with Multiple Enrollment Utterances and a Sampling Strategy for the SASV Challenge 2022. INTERSPEECH 2022: 2883-2887 - [c244]Haoyu Li, Junichi Yamagishi:
DDS: A new device-degraded speech dataset for speech enhancement. INTERSPEECH 2022: 2913-2917 - [c243]Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi, Natalia A. Tomashenko:
Analyzing Language-Independent Speaker Anonymization Framework under Unseen Conditions. INTERSPEECH 2022: 4426-4430 - [c242]Wen-Chin Huang, Erica Cooper, Yu Tsao, Hsin-Min Wang, Tomoki Toda, Junichi Yamagishi:
The VoiceMOS Challenge 2022. INTERSPEECH 2022: 4536-4540 - [c241]Junichi Yamagishi:
Lessons Learned from ASVSpoof and Remaining Challenges. DDAM@MM 2022: 1-2 - [c240]Xin Wang, Junichi Yamagishi:
Investigating Self-Supervised Front Ends for Speech Spoofing Countermeasures. Odyssey 2022: 100-106 - [c239]Hemlata Tak, Massimiliano Todisco, Xin Wang, Jee-weon Jung, Junichi Yamagishi, Nicholas W. D. Evans:
Automatic Speaker Verification Spoofing and Deepfake Detection Using Wav2vec 2.0 and Data Augmentation. Odyssey 2022: 112-119 - [c238]Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi, Natalia A. Tomashenko:
Language-Independent Speaker Anonymization Approach Using Self-Supervised Pre-Trained Models. Odyssey 2022: 279-286 - [c237]Xin Wang, Junichi Yamagishi:
Investigating Active-Learning-Based Training Data Selection for Speech Spoofing Countermeasure. SLT 2022: 585-592 - [p6]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Capsule-Forensics Networks for Deepfake Detection. Handbook of Digital Face Manipulation and Detection 2022: 275-301 - [p5]Ruben Tolosana, Christian Rathgeb, Rubén Vera-Rodríguez, Christoph Busch, Luisa Verdoliva, Siwei Lyu, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen, Peter Rot, Klemen Grm, Vitomir Struc, Antitza Dantcheva, Zahid Akhtar, Sergio Romero-Tapiador, Julian Fiérrez, Aythami Morales, Javier Ortega-Garcia, Els Kindt, Catherine Jasserand, Tarmo Kalvet, Marek Tiits:
Future Trends in Digital Face Manipulation and Detection. Handbook of Digital Face Manipulation and Detection 2022: 463-482 - [d1]Junichi Yamagishi, Seyyed Saeed Sarfjoo:
Device Recorded VCTK (DR-VCTK). Zenodo, 2022 - [i118]Xin Wang, Junichi Yamagishi:
A Practical Guide to Logical Access Voice Presentation Attack Detection. CoRR abs/2201.03321 (2022) - [i117]Anssi Kanervisto, Ville Hautamäki, Tomi Kinnunen, Junichi Yamagishi:
Optimizing Tandem Speaker Verification and Anti-Spoofing Systems. CoRR abs/2201.09709 (2022) - [i116]Trung-Nghia Le, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Robust Deepfake On Unrestricted Media: Generation And Detection. CoRR abs/2202.06228 (2022) - [i115]Hemlata Tak, Massimiliano Todisco, Xin Wang, Jee-weon Jung, Junichi Yamagishi, Nicholas W. D. Evans:
Automatic speaker verification spoofing and deepfake detection using wav2vec 2.0 and data augmentation. CoRR abs/2202.12233 (2022) - [i114]Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi, Natalia A. Tomashenko:
Language-Independent Speaker Anonymization Approach using Self-Supervised Pre-Trained Models. CoRR abs/2202.13097 (2022) - [i113]Wen-Chin Huang, Erica Cooper, Yu Tsao, Hsin-Min Wang, Tomoki Toda, Junichi Yamagishi:
The VoiceMOS Challenge 2022. CoRR abs/2203.11389 (2022) - [i112]Haoyu Li, Yun Liu, Junichi Yamagishi:
Joint Noise Reduction and Listening Enhancement for Full-End Speech Enhancement. CoRR abs/2203.11500 (2022) - [i111]Natalia A. Tomashenko, Xin Wang, Xiaoxiao Miao, Hubert Nourtel, Pierre Champion, Massimiliano Todisco, Emmanuel Vincent, Nicholas W. D. Evans, Junichi Yamagishi, Jean-François Bonastre:
The VoicePrivacy 2022 Challenge Evaluation Plan. CoRR abs/2203.12468 (2022) - [i110]Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi, Natalia A. Tomashenko:
Analyzing Language-Independent Speaker Anonymization Framework under Unseen Conditions. CoRR abs/2203.14834 (2022) - [i109]Lin Zhang, Xin Wang, Erica Cooper, Nicholas W. D. Evans, Junichi Yamagishi:
The PartialSpoof Database and Countermeasures for the Detection of Short Generated Audio Segments Embedded in a Speech Utterance. CoRR abs/2204.05177 (2022) - [i108]Natalia A. Tomashenko, Brij Mohan Lal Srivastava, Xin Wang, Emmanuel Vincent, Andreas Nautsch, Junichi Yamagishi, Nicholas W. D. Evans, Jose Patino, Jean-François Bonastre, Paul-Gauthier Noé, Massimiliano Todisco:
The VoicePrivacy 2020 Challenge Evaluation Plan. CoRR abs/2205.07123 (2022) - [i107]Chang Zeng, Xiaoxiao Miao, Xin Wang, Erica Cooper, Junichi Yamagishi:
Joint Speaker Encoder and Neural Back-end Model for Fully End-to-End Automatic Speaker Verification with Multiple Enrollment Utterances. CoRR abs/2209.00485 (2022) - [i106]Xuechen Liu, Xin Wang, Md. Sahidullah, Jose Patino, Héctor Delgado, Tomi Kinnunen, Massimiliano Todisco, Junichi Yamagishi, Nicholas W. D. Evans, Andreas Nautsch, Kong Aik Lee:
ASVspoof 2021: Towards Spoofed and Deepfake Speech Detection in the Wild. CoRR abs/2210.02437 (2022) - [i105]Xin Wang, Junichi Yamagishi:
Spoofed training data for speech spoofing countermeasure can be efficiently created using neural vocoders. CoRR abs/2210.10570 (2022) - [i104]Huy H. Nguyen, Trung-Nghia Le, Junichi Yamagishi, Isao Echizen:
Analysis of Master Vein Attacks on Finger Vein Recognition Systems. CoRR abs/2210.10667 (2022) - [i103]Li-Kuang Chen, Canasai Kruengkrai, Junichi Yamagishi:
Outlier-Aware Training for Improving Group Accuracy Disparities. CoRR abs/2210.15183 (2022) - [i102]Xuan Shi, Erica Cooper, Xin Wang, Junichi Yamagishi, Shrikanth Narayanan:
Can Knowledge of End-to-End Text-to-Speech Models Improve Neural MIDI-to-Audio Synthesis Systems? CoRR abs/2211.13868 (2022) - [i101]Paul-Gauthier Noé, Xiaoxiao Miao, Xin Wang, Junichi Yamagishi, Jean-François Bonastre, Driss Matrouf:
Hiding speaker's sex in speech using zero-evidence speaker representation in an analysis/synthesis pipeline. CoRR abs/2211.16065 (2022) - 2021
- [j62]Yusuke Yasuda, Xin Wang, Junichi Yamagishi:
Investigation of learning abilities on linguistic features in sequence-to-sequence text-to-speech synthesis. Comput. Speech Lang. 67: 101183 (2021) - [j61]Noboru Babaguchi, Isao Echizen, Junichi Yamagishi, Naoko Nitta, Yuta Nakashima, Kazuaki Nakamura, Kazuhiro Kono, Fuming Fang, Seiko Myojin, Zhenzhong Kuang, Huy H. Nguyen, Ngoc-Dung T. Tieu:
Preventing Fake Information Generation Against Media Clone Attacks. IEICE Trans. Inf. Syst. 104-D(1): 2-11 (2021) - [j60]Isao Echizen, Noboru Babaguchi, Junichi Yamagishi, Naoko Nitta, Yuta Nakashima, Kazuaki Nakamura, Kazuhiro Kono, Fuming Fang, Seiko Myojin, Zhenzhong Kuang, Huy H. Nguyen, Ngoc-Dung T. Tieu:
Generation and Detection of Media Clones. IEICE Trans. Inf. Syst. 104-D(1): 12-23 (2021) - [j59]Berrak Sisman, Junichi Yamagishi, Simon King, Haizhou Li:
An Overview of Voice Conversion and Its Challenges: From Statistical Modeling to Deep Learning. IEEE ACM Trans. Audio Speech Lang. Process. 29: 132-157 (2021) - [j58]Haoyu Li, Junichi Yamagishi:
Multi-Metric Optimization Using Generative Adversarial Networks for Near-End Speech Intelligibility Enhancement. IEEE ACM Trans. Audio Speech Lang. Process. 29: 3000-3011 (2021) - [j57]Andreas Nautsch, Xin Wang, Nicholas W. D. Evans, Tomi H. Kinnunen, Ville Vestman, Massimiliano Todisco, Héctor Delgado, Md. Sahidullah, Junichi Yamagishi, Kong Aik Lee:
ASVspoof 2019: Spoofing Countermeasures for the Detection of Synthesized, Converted and Replayed Speech. IEEE Trans. Biom. Behav. Identity Sci. 3(2): 252-265 (2021) - [c236]Canasai Kruengkrai, Junichi Yamagishi, Xin Wang:
A Multi-Level Attention Model for Evidence-Based Fact Checking. ACL/IJCNLP (Findings) 2021: 2447-2460 - [c235]Marc Treu, Trung-Nghia Le, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Fashion-Guided Adversarial Attack on Person Segmentation. CVPR Workshops 2021: 943-952 - [c234]Khanh-Duy Nguyen, Hai-Dang Nguyen, Trung-Nghia Le, Junichi Yamagishi, Isao Echizen:
Effectiveness of Detection-based and Regression-based Approaches for Estimating Mask-Wearing Ratio. FG 2021: 1-8 - [c233]Yusuke Yasuda, Xin Wang, Junichi Yamagishi:
End-to-End Text-to-Speech Using Latent Duration Based on VQ-VAE. ICASSP 2021: 5694-5698 - [c232]Shuhei Kato, Yusuke Yasuda, Xin Wang, Erica Cooper, Junichi Yamagishi:
How Similar or Different is Rakugo Speech Synthesizer to Professional Performers? ICASSP 2021: 6488-6492 - [c231]Jennifer Williams, Yi Zhao, Erica Cooper, Junichi Yamagishi:
Learning Disentangled Phone and Speaker Representations in a Semi-Supervised VQ-VAE Paradigm. ICASSP 2021: 7053-7057 - [c230]Trung-Nghia Le, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
OpenForensics: Large-Scale Challenging Dataset For Multi-Face Forgery Detection And Segmentation In-The-Wild. ICCV 2021: 10097-10107 - [c229]Xin Wang, Junichi Yamagishi:
A Comparative Study on Recent Neural Spoofing Countermeasures for Synthetic Speech Detection. Interspeech 2021: 4259-4263 - [c228]Lin Zhang, Xin Wang, Erica Cooper, Junichi Yamagishi, Jose Patino, Nicholas W. D. Evans:
An Initial Investigation for Detecting Partially Spoofed Audio. Interspeech 2021: 4264-4268 - [c227]Tomi Kinnunen, Andreas Nautsch, Md. Sahidullah, Nicholas W. D. Evans, Xin Wang, Massimiliano Todisco, Héctor Delgado, Junichi Yamagishi, Kong Aik Lee:
Visualizing Classifier Adjacency Relations: A Case Study in Speaker Verification and Voice Anti-Spoofing. Interspeech 2021: 4299-4303 - [c226]Yang Ai, Haoyu Li, Xin Wang, Junichi Yamagishi, Zhen-Hua Ling:
Denoising-and-Dereverberation Hierarchical Neural Vocoder for Robust Waveform Generation. SLT 2021: 477-484 - [c225]Haoyu Li, Yang Ai, Junichi Yamagishi:
Enhancing Low-Quality Voice Recordings Using Disentangled Channel Factor and Neural Waveform Model. SLT 2021: 734-741 - [c224]Jennifer Williams, Jason Fong, Erica Cooper, Junichi Yamagishi:
Exploring Disentanglement with Multilingual and Monolingual VQ-VAE. SSW 2021: 124-129 - [c223]Erica Cooper, Xin Wang, Junichi Yamagishi:
Text-to-Speech Synthesis Techniques for MIDI-to-Audio Synthesis. SSW 2021: 130-135 - [c222]Hieu-Thi Luong, Junichi Yamagishi:
Preliminary study on using vector quantization latent spaces for TTS/VC systems with consistent performance. SSW 2021: 136-141 - [c221]Erica Cooper, Junichi Yamagishi:
How do Voices from Past Speech Synthesis Challenges Compare Today? SSW 2021: 183-188 - [i100]Andreas Nautsch, Xin Wang, Nicholas W. D. Evans, Tomi Kinnunen, Ville Vestman, Massimiliano Todisco, Héctor Delgado, Md. Sahidullah, Junichi Yamagishi, Kong Aik Lee:
ASVspoof 2019: spoofing countermeasures for the detection of synthesized, converted and replayed speech. CoRR abs/2102.05889 (2021) - [i99]Chang Zeng, Xin Wang, Erica Cooper, Junichi Yamagishi:
Attention Back-end for Automatic Speaker Verification with Multiple Enrollment Utterances. CoRR abs/2104.01541 (2021) - [i98]Lin Zhang, Xin Wang, Erica Cooper, Junichi Yamagishi, Jose Patino, Nicholas W. D. Evans:
An Initial Investigation for Detecting Partially Spoofed Audio. CoRR abs/2104.02518 (2021) - [i97]Marc Treu, Trung-Nghia Le, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Fashion-Guided Adversarial Attack on Person Segmentation. CoRR abs/2104.08422 (2021) - [i96]Haoyu Li, Junichi Yamagishi:
Multi-Metric Optimization using Generative Adversarial Networks for Near-End Speech Intelligibility Enhancement. CoRR abs/2104.08499 (2021) - [i95]Erica Cooper, Xin Wang, Junichi Yamagishi:
Text-to-Speech Synthesis Techniques for MIDI-to-Audio Synthesis. CoRR abs/2104.12292 (2021) - [i94]Jennifer Williams, Jason Fong, Erica Cooper, Junichi Yamagishi:
Exploring Disentanglement with Multilingual and Monolingual VQ-VAE. CoRR abs/2105.01573 (2021) - [i93]Erica Cooper, Junichi Yamagishi:
How do Voices from Past Speech Synthesis Challenges Compare Today? CoRR abs/2105.02373 (2021) - [i92]Canasai Kruengkrai, Junichi Yamagishi, Xin Wang:
A Multi-Level Attention Model for Evidence-Based Fact Checking. CoRR abs/2106.00950 (2021) - [i91]Tomi Kinnunen, Andreas Nautsch, Md. Sahidullah, Nicholas W. D. Evans, Xin Wang, Massimiliano Todisco, Héctor Delgado, Junichi Yamagishi, Kong Aik Lee:
Visualizing Classifier Adjacency Relations: A Case Study in Speaker Verification and Voice Anti-Spoofing. CoRR abs/2106.06362 (2021) - [i90]Hieu-Thi Luong, Junichi Yamagishi:
Preliminary study on using vector quantization latent spaces for TTS/VC systems with consistent performance. CoRR abs/2106.13479 (2021) - [i89]Cheng-Hung Hu, Yu-Huai Peng, Junichi Yamagishi, Yu Tsao, Hsin-Min Wang:
SVSNet: An End-to-end Speaker Voice Similarity Assessment Model. CoRR abs/2107.09392 (2021) - [i88]Xuan Shi, Erica Cooper, Junichi Yamagishi:
Use of speaker recognition approaches for learning timbre representations of musical instrument sounds from raw waveforms. CoRR abs/2107.11506 (2021) - [i87]Lin Zhang, Xin Wang, Erica Cooper, Junichi Yamagishi:
Multi-Task Learning in Utterance-Level and Segmental-Level Spoof Detection. CoRR abs/2107.14132 (2021) - [i86]Trung-Nghia Le, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
OpenForensics: Large-Scale Challenging Dataset For Multi-Face Forgery Detection And Segmentation In-The-Wild. CoRR abs/2107.14480 (2021) - [i85]Jean-François Bonastre, Héctor Delgado, Nicholas W. D. Evans, Tomi Kinnunen, Kong Aik Lee, Xuechen Liu, Andreas Nautsch, Paul-Gauthier Noé, Jose Patino, Md. Sahidullah, Brij Mohan Lal Srivastava, Massimiliano Todisco, Natalia A. Tomashenko, Emmanuel Vincent, Xin Wang, Junichi Yamagishi:
Benchmarking and challenges in security and privacy for voice biometrics. CoRR abs/2109.00281 (2021) - [i84]Héctor Delgado, Nicholas W. D. Evans, Tomi Kinnunen, Kong Aik Lee, Xuechen Liu, Andreas Nautsch, Jose Patino, Md. Sahidullah, Massimiliano Todisco, Xin Wang, Junichi Yamagishi:
ASVspoof 2021: Automatic Speaker Verification Spoofing and Countermeasures Challenge Evaluation Plan. CoRR abs/2109.00535 (2021) - [i83]Junichi Yamagishi, Xin Wang, Massimiliano Todisco, Md. Sahidullah, Jose Patino, Andreas Nautsch, Xuechen Liu, Kong Aik Lee, Tomi Kinnunen, Nicholas W. D. Evans, Héctor Delgado:
ASVspoof 2021: accelerating progress in spoofed and deepfake speech detection. CoRR abs/2109.00537 (2021) - [i82]Natalia A. Tomashenko, Xin Wang, Emmanuel Vincent, Jose Patino, Brij Mohan Lal Srivastava, Paul-Gauthier Noé, Andreas Nautsch, Nicholas W. D. Evans, Junichi Yamagishi, Benjamin O'Brien, Anaïs Chanclu, Jean-François Bonastre, Massimiliano Todisco, Mohamed Maouche:
The VoicePrivacy 2020 Challenge: Results and findings. CoRR abs/2109.00648 (2021) - [i81]Huy H. Nguyen, Sébastien Marcel, Junichi Yamagishi, Isao Echizen:
Master Face Attacks on Face Recognition Systems. CoRR abs/2109.03398 (2021) - [i80]Haoyu Li, Junichi Yamagishi:
DDS: A new device-degraded speech dataset for speech enhancement. CoRR abs/2109.07931 (2021) - [i79]Cheng-I Jeff Lai, Erica Cooper, Yang Zhang, Shiyu Chang, Kaizhi Qian, Yi-Lun Liao, Yung-Sung Chuang, Alexander H. Liu, Junichi Yamagishi, David D. Cox, James R. Glass:
On the Interplay Between Sparsity, Naturalness, Intelligibility, and Prosody in Speech Synthesis. CoRR abs/2110.01147 (2021) - [i78]Xin Wang, Junichi Yamagishi:
Estimating the confidence of speech spoofing countermeasure. CoRR abs/2110.04775 (2021) - [i77]Hieu-Thi Luong, Junichi Yamagishi:
LaughNet: synthesizing laughter utterances from waveform silhouettes and a single laughter example. CoRR abs/2110.04946 (2021) - [i76]Wen-Chin Huang, Erica Cooper, Junichi Yamagishi, Tomoki Toda:
LDNet: Unified Listener Dependent Modeling in MOS Prediction for Synthetic Speech. CoRR abs/2110.09103 (2021) - [i75]Xin Wang, Junichi Yamagishi:
Investigating self-supervised front ends for speech spoofing countermeasures. CoRR abs/2111.07725 (2021) - [i74]Khanh-Duy Nguyen, Huy H. Nguyen, Trung-Nghia Le, Junichi Yamagishi, Isao Echizen:
Effectiveness of Detection-based and Regression-based Approaches for Estimating Mask-Wearing Ratio. CoRR abs/2111.12888 (2021) - 2020
- [j56]Shuhei Kato, Yusuke Yasuda, Xin Wang, Erica Cooper, Shinji Takaki, Junichi Yamagishi:
Modeling of Rakugo Speech and Its Limitations: Toward Speech Synthesis That Entertains Audiences. IEEE Access 8: 138149-138161 (2020) - [j55]Jean-François Bonastre, Tomi Kinnunen, Anthony Larcher, Junichi Yamagishi:
Introduction to the special issue "Speaker and language characterization and recognition: Voice modeling, conversion, synthesis and ethical aspects". Comput. Speech Lang. 60 (2020) - [j54]Xin Wang, Junichi Yamagishi, Massimiliano Todisco, Héctor Delgado, Andreas Nautsch, Nicholas W. D. Evans, Md. Sahidullah, Ville Vestman, Tomi Kinnunen, Kong Aik Lee, Lauri Juvela, Paavo Alku, Yu-Huai Peng, Hsin-Te Hwang, Yu Tsao, Hsin-Min Wang, Sébastien Le Maguer, Markus Becker, Zhen-Hua Ling:
ASVspoof 2019: A large-scale public database of synthesized, converted and replayed speech. Comput. Speech Lang. 64: 101114 (2020) - [j53]Xin Wang, Shinji Takaki, Junichi Yamagishi, Simon King, Keiichi Tokuda:
A Vector Quantized Variational Autoencoder (VQ-VAE) Autoregressive Neural F0 Model for Statistical Parametric Speech Synthesis. IEEE ACM Trans. Audio Speech Lang. Process. 28: 157-170 (2020) - [j52]Xin Wang, Shinji Takaki, Junichi Yamagishi:
Neural Source-Filter Waveform Models for Statistical Parametric Speech Synthesis. IEEE ACM Trans. Audio Speech Lang. Process. 28: 402-415 (2020) - [j51]Tomi Kinnunen, Héctor Delgado, Nicholas W. D. Evans, Kong Aik Lee, Ville Vestman, Andreas Nautsch, Massimiliano Todisco, Xin Wang, Md. Sahidullah, Junichi Yamagishi, Douglas A. Reynolds:
Tandem Assessment of Spoofing Countermeasures and Automatic Speaker Verification: Fundamentals. IEEE ACM Trans. Audio Speech Lang. Process. 28: 2195-2210 (2020) - [j50]Hieu-Thi Luong, Junichi Yamagishi:
NAUTILUS: A Versatile Voice Cloning System. IEEE ACM Trans. Audio Speech Lang. Process. 28: 2967-2981 (2020) - [c220]David Ifeoluwa Adelani, Haotian Mai, Fuming Fang, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Generating Sentiment-Preserving Fake Online Reviews Using Neural Language Models and Their Human- and Machine-Based Detection. AINA 2020: 1341-1354 - [c219]Rong Huang, Fuming Fang, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
A Method for Identifying Origin of Digital Images Using a Convolutional Neural Network. APSIPA 2020: 1293-1299 - [c218]Ngoc-Dung T. Tieu, Junichi Yamagishi, Isao Echizen:
Color Transfer to Anonymized Gait Images While Maintaining Anonymization. APSIPA 2020: 1406-1413 - [c217]Yi Zhao, Wen-Chin Huang, Xiaohai Tian, Junichi Yamagishi, Rohan Kumar Das, Tomi Kinnunen, Zhen-Hua Ling, Tomoki Toda:
Voice Conversion Challenge 2020 -- Intra-lingual semi-parallel and cross-lingual voice conversion --. Blizzard Challenge / Voice Conversion Challenge 2020 - [c216]Rohan Kumar Das, Tomi Kinnunen, Wen-Chin Huang, Zhen-Hua Ling, Junichi Yamagishi, Yi Zhao, Xiaohai Tian, Tomoki Toda:
Predictions of Subjective Ratings and Spoofing Assessments of Voice Conversion Challenge 2020 Submissions. Blizzard Challenge / Voice Conversion Challenge 2020 - [c215]Hieu-Thi Luong, Junichi Yamagishi:
Latent linguistic embedding for cross-lingual text-to-speech and voice conversion. Blizzard Challenge / Voice Conversion Challenge 2020 - [c214]Erica Cooper, Cheng-I Lai, Yusuke Yasuda, Fuming Fang, Xin Wang, Nanxin Chen, Junichi Yamagishi:
Zero-Shot Multi-Speaker Text-To-Speech with State-Of-The-Art Neural Speaker Embeddings. ICASSP 2020: 6184-6188 - [c213]Yi Zhao, Xin Wang, Lauri Juvela, Junichi Yamagishi:
Transferring Neural Speech Waveform Synthesizers to Musical Instrument Sounds Generation. ICASSP 2020: 6269-6273 - [c212]Yusuke Yasuda, Xin Wang, Junichi Yamagishi:
Effect of Choice of Probability Distribution, Randomness, and Search Methods for Alignment Modeling in Sequence-to-Sequence Text-to-Speech Synthesis Using Hard Alignment. ICASSP 2020: 6724-6728 - [c211]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen, Sébastien Marcel:
Generating Master Faces for Use in Performing Wolf Attacks on Face Recognition Systems. IJCB 2020: 1-10 - [c210]Rong Huang, Fuming Fang, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Security of Facial Forensics Models Against Adversarial Attacks. ICIP 2020: 2236-2240 - [c209]Haoyu Li, Szu-Wei Fu, Yu Tsao, Junichi Yamagishi:
iMetricGAN: Intelligibility Enhancement for Speech-in-Noise Using Generative Adversarial Network-Based Metric Learning. INTERSPEECH 2020: 1336-1340 - [c208]Natalia A. Tomashenko, Brij Mohan Lal Srivastava, Xin Wang, Emmanuel Vincent, Andreas Nautsch, Junichi Yamagishi, Nicholas W. D. Evans, Jose Patino, Jean-François Bonastre, Paul-Gauthier Noé, Massimiliano Todisco:
Introducing the VoicePrivacy Initiative. INTERSPEECH 2020: 1693-1697 - [c207]Andreas Nautsch, Jose Patino, Natalia A. Tomashenko, Junichi Yamagishi, Paul-Gauthier Noé, Jean-François Bonastre, Massimiliano Todisco, Nicholas W. D. Evans:
The Privacy ZEBRA: Zero Evidence Biometric Recognition Assessment. INTERSPEECH 2020: 1698-1702 - [c206]Brij Mohan Lal Srivastava, Natalia A. Tomashenko, Xin Wang, Emmanuel Vincent, Junichi Yamagishi, Mohamed Maouche, Aurélien Bellet, Marc Tommasi:
Design Choices for X-Vector Based Speaker Anonymization. INTERSPEECH 2020: 1713-1717 - [c205]Xin Wang, Junichi Yamagishi:
Using Cyclic Noise as the Source Signal for Neural Source-Filter-Based Speech Waveform Model. INTERSPEECH 2020: 1992-1996 - [c204]Haoyu Li, Junichi Yamagishi:
Noise Tokens: Learning Neural Noise Templates for Environment-Aware Speech Enhancement. INTERSPEECH 2020: 2452-2456 - [c203]Yang Ai, Xin Wang, Junichi Yamagishi, Zhen-Hua Ling:
Reverberation Modeling for Source-Filter-Based Neural Vocoder. INTERSPEECH 2020: 3560-3564 - [c202]Erica Cooper, Cheng-I Lai, Yusuke Yasuda, Junichi Yamagishi:
Can Speaker Augmentation Improve Multi-Speaker End-to-End TTS? INTERSPEECH 2020: 3979-3983 - [c201]Yi Zhao, Haoyu Li, Cheng-I Lai, Jennifer Williams, Erica Cooper, Junichi Yamagishi:
Improved Prosody from Learned F0 Codebook Representations for VQ-VAE Speech Waveform Reconstruction. INTERSPEECH 2020: 4417-4421 - [c200]Anssi Kanervisto, Ville Hautamäki, Tomi Kinnunen, Junichi Yamagishi:
An Initial Investigation on Optimizing Tandem Speaker Verification and Countermeasure Systems Using Reinforcement Learning. Odyssey 2020: 151-158 - [e1]Junichi Yamagishi, Zhenhua Ling, Rohan Kumar Das, Simon King, Tomi Kinnunen, Tomoki Toda, Wen-Chin Huang, Xiao Zhou, Xiaohai Tian, Yi Zhao:
Joint Workshop for the Blizzard Challenge and Voice Conversion Challenge 2020, Shanghai, China, October 30, 2020. ISCA 2020 [contents] - [i73]Anssi Kanervisto, Ville Hautamäki, Tomi Kinnunen, Junichi Yamagishi:
An initial investigation on optimizing tandem speaker verification and countermeasure systems using reinforcement learning. CoRR abs/2002.03801 (2020) - [i72]Haoyu Li, Szu-Wei Fu, Yu Tsao, Junichi Yamagishi:
iMetricGAN: Intelligibility Enhancement for Speech-in-Noise using Generative Adversarial Network-based Metric Learning. CoRR abs/2004.00932 (2020) - [i71]Natalia A. Tomashenko, Brij Mohan Lal Srivastava, Xin Wang, Emmanuel Vincent, Andreas Nautsch, Junichi Yamagishi, Nicholas W. D. Evans, Jose Patino, Jean-François Bonastre, Paul-Gauthier Noé, Massimiliano Todisco:
Introducing the VoicePrivacy Initiative. CoRR abs/2005.01387 (2020) - [i70]Yang Ai, Xin Wang, Junichi Yamagishi, Zhen-Hua Ling:
Reverberation Modeling for Source-Filter-based Neural Vocoder. CoRR abs/2005.07379 (2020) - [i69]Yi Zhao, Haoyu Li, Cheng-I Lai, Jennifer Williams, Erica Cooper, Junichi Yamagishi:
Improved Prosody from Learned F0 Codebook Representations for VQ-VAE Speech Waveform Reconstruction. CoRR abs/2005.07884 (2020) - [i68]Brij Mohan Lal Srivastava, Natalia A. Tomashenko, Xin Wang, Emmanuel Vincent, Junichi Yamagishi, Mohamed Maouche, Aurélien Bellet, Marc Tommasi:
Design Choices for X-vector Based Speaker Anonymization. CoRR abs/2005.08601 (2020) - [i67]Andreas Nautsch, Jose Patino, Natalia A. Tomashenko, Junichi Yamagishi, Paul-Gauthier Noé, Jean-François Bonastre, Massimiliano Todisco, Nicholas W. D. Evans:
The Privacy ZEBRA: Zero Evidence Biometric Recognition Assessment. CoRR abs/2005.09413 (2020) - [i66]Yusuke Yasuda, Xin Wang, Junichi Yamagishi:
Investigation of learning abilities on linguistic features in sequence-to-sequence text-to-speech synthesis. CoRR abs/2005.10390 (2020) - [i65]Hieu-Thi Luong, Junichi Yamagishi:
NAUTILUS: a Versatile Voice Cloning System. CoRR abs/2005.11004 (2020) - [i64]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen, Sébastien Marcel:
Generating Master Faces for Use in Performing Wolf Attacks on Face Recognition Systems. CoRR abs/2006.08376 (2020) - [i63]Tomi Kinnunen, Héctor Delgado, Nicholas W. D. Evans, Kong Aik Lee, Ville Vestman, Andreas Nautsch, Massimiliano Todisco, Xin Wang, Md. Sahidullah, Junichi Yamagishi, Douglas A. Reynolds:
Tandem Assessment of Spoofing Countermeasures and Automatic Speaker Verification: Fundamentals. CoRR abs/2007.05979 (2020) - [i62]Berrak Sisman, Junichi Yamagishi, Simon King, Haizhou Li:
An Overview of Voice Conversion and its Challenges: From Statistical Modeling to Deep Learning. CoRR abs/2008.03648 (2020) - [i61]Yi Zhao, Wen-Chin Huang, Xiaohai Tian, Junichi Yamagishi, Rohan Kumar Das, Tomi Kinnunen, Zhen-Hua Ling, Tomoki Toda:
Voice Conversion Challenge 2020: Intra-lingual semi-parallel and cross-lingual voice conversion. CoRR abs/2008.12527 (2020) - [i60]Rohan Kumar Das, Tomi Kinnunen, Wen-Chin Huang, Zhen-Hua Ling, Junichi Yamagishi, Yi Zhao, Xiaohai Tian, Tomoki Toda:
Predictions of Subjective Ratings and Spoofing Assessments of Voice Conversion Challenge 2020 Submissions. CoRR abs/2009.03554 (2020) - [i59]Saurabh Gupta, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Viable Threat on News Reading: Generating Biased News Using Natural Language Models. CoRR abs/2010.02150 (2020) - [i58]Hieu-Thi Luong, Junichi Yamagishi:
Latent linguistic embedding for cross-lingual text-to-speech and voice conversion. CoRR abs/2010.03717 (2020) - [i57]Yusuke Yasuda, Xin Wang, Junichi Yamagishi:
End-to-End Text-to-Speech using Latent Duration based on VQ-VAE. CoRR abs/2010.09602 (2020) - [i56]Antoine Perquin, Erica Cooper, Junichi Yamagishi:
Grapheme or phoneme? An Analysis of Tacotron's Embedded Representations. CoRR abs/2010.10694 (2020) - [i55]Jennifer Williams, Yi Zhao, Erica Cooper, Junichi Yamagishi:
Learning Disentangled Phone and Speaker Representations in a Semi-Supervised VQ-VAE Paradigm. CoRR abs/2010.10727 (2020) - [i54]Shuhei Kato, Yusuke Yasuda, Xin Wang, Erica Cooper, Junichi Yamagishi:
How Similar or Different Is Rakugo Speech Synthesizer to Professional Performers? CoRR abs/2010.11549 (2020) - [i53]Yang Ai, Haoyu Li, Xin Wang, Junichi Yamagishi, Zhen-Hua Ling:
Denoising-and-Dereverberation Hierarchical Neural Vocoder for Robust Waveform Generation. CoRR abs/2011.03955 (2020) - [i52]Erica Cooper, Xin Wang, Yi Zhao, Yusuke Yasuda, Junichi Yamagishi:
Pretraining Strategies, Waveform Model Choice, and Acoustic Configurations for Multi-Speaker End-to-End Speech Synthesis. CoRR abs/2011.04839 (2020) - [i51]Haoyu Li, Yang Ai, Junichi Yamagishi:
Enhancing Low-Quality Voice Recordings Using Disentangled Channel Factor and Neural Waveform Model. CoRR abs/2011.05038 (2020)
2010 – 2019
- 2019
- [j49]Ngoc-Dung T. Tieu, Huy H. Nguyen, Hoang-Quoc Nguyen-Son, Junichi Yamagishi, Isao Echizen:
Spatio-temporal generative adversarial network for gait anonymization. J. Inf. Secur. Appl. 46: 307-319 (2019) - [j48]Toru Nakashika, Shinji Takaki, Junichi Yamagishi:
Complex-Valued Restricted Boltzmann Machine for Speaker-Dependent Speech Parameterization From Complex Spectra. IEEE ACM Trans. Audio Speech Lang. Process. 27(2): 244-254 (2019) - [c199]Ngoc-Dung T. Tieu, Huy H. Nguyen, Fuming Fang, Junichi Yamagishi, Isao Echizen:
An RGB Gait Anonymization Model for Low-Quality Silhouettes. APSIPA 2019: 1686-1693 - [c198]Hieu-Thi Luong, Junichi Yamagishi:
Bootstrapping Non-Parallel Voice Conversion from Speaker-Adaptive Text-to-Speech. ASRU 2019: 200-207 - [c197]Huy H. Nguyen, Fuming Fang, Junichi Yamagishi, Isao Echizen:
Multi-task Learning for Detecting and Segmenting Manipulated Facial Images and Videos. BTAS 2019: 1-8 - [c196]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Capsule-forensics: Using Capsule Networks to Detect Forged Images and Videos. ICASSP 2019: 2307-2311 - [c195]Xin Wang, Shinji Takaki, Junichi Yamagishi:
Neural Source-filter-based Waveform Model for Statistical Parametric Speech Synthesis. ICASSP 2019: 5916-5920 - [c194]Cheng-I Lai, Alberto Abad, Korin Richmond, Junichi Yamagishi, Najim Dehak, Simon King:
Attentive Filtering Networks for Audio Replay Attack Detection. ICASSP 2019: 6316-6320 - [c193]Fuming Fang, Xin Wang, Junichi Yamagishi, Isao Echizen:
Audiovisual Speaker Conversion: Jointly and Simultaneously Transforming Facial Expression and Acoustic Characteristics. ICASSP 2019: 6795-6799 - [c192]Shreyas Seshadri, Lauri Juvela, Junichi Yamagishi, Okko Räsänen, Paavo Alku:
Cycle-consistent Adversarial Networks for Non-parallel Vocal Effort Based Speaking Style Conversion. ICASSP 2019: 6835-6839 - [c191]Yusuke Yasuda, Xin Wang, Shinji Takaki, Junichi Yamagishi:
Investigation of Enhanced Tacotron Text-to-speech Synthesis Systems with Self-attention for Pitch Accent Language. ICASSP 2019: 6905-6909 - [c190]Lauri Juvela, Bajibabu Bollepalli, Junichi Yamagishi, Paavo Alku:
Waveform Generation for Text-to-speech Synthesis Using Pitch-synchronous Multi-scale Generative Adversarial Networks. ICASSP 2019: 6915-6919 - [c189]Shinji Takaki, Toru Nakashika, Xin Wang, Junichi Yamagishi:
STFT Spectral Loss for Training a Neural Speech Waveform Model. ICASSP 2019: 7065-7069 - [c188]Lauri Juvela, Bajibabu Bollepalli, Junichi Yamagishi, Paavo Alku:
GELP: GAN-Excited Linear Prediction for Speech Synthesis from Mel-Spectrogram. INTERSPEECH 2019: 694-698 - [c187]Massimiliano Todisco, Xin Wang, Ville Vestman, Md. Sahidullah, Héctor Delgado, Andreas Nautsch, Junichi Yamagishi, Nicholas W. D. Evans, Tomi H. Kinnunen, Kong Aik Lee:
ASVspoof 2019: Future Horizons in Spoofed and Fake Audio Detection. INTERSPEECH 2019: 1008-1012 - [c186]Mingyang Zhang, Xin Wang, Fuming Fang, Haizhou Li, Junichi Yamagishi:
Joint Training Framework for Text-to-Speech and Voice Conversion Using Multi-Source Tacotron and WaveNet. INTERSPEECH 2019: 1298-1302 - [c185]Hieu-Thi Luong, Xin Wang, Junichi Yamagishi, Nobuyuki Nishizawa:
Training Multi-Speaker Neural Text-to-Speech Systems Using Speaker-Imbalanced Speech Corpora. INTERSPEECH 2019: 1303-1307 - [c184]Chen-Chou Lo, Szu-Wei Fu, Wen-Chin Huang, Xin Wang, Junichi Yamagishi, Yu Tsao, Hsin-Min Wang:
MOSNet: Deep Learning-Based Objective Assessment for Voice Conversion. INTERSPEECH 2019: 1541-1545 - [c183]Yi Zhao, Atsushi Ando, Shinji Takaki, Junichi Yamagishi, Satoshi Kobashikawa:
Does the Lombard Effect Improve Emotional Communication in Noise? - Analysis of Emotional Speech Acted in Noise. INTERSPEECH 2019: 3292-3296 - [c182]Xin Wang, Junichi Yamagishi:
Neural Harmonic-plus-Noise Waveform Model with Trainable Maximum Voice Frequency for Text-to-Speech Synthesis. SSW 2019: 1-6 - [c181]Shuhei Kato, Yusuke Yasuda, Xin Wang, Erica Cooper, Shinji Takaki, Junichi Yamagishi:
Rakugo speech synthesis using segment-to-segment neural transduction and style tokens - toward speech synthesis for entertaining audiences. SSW 2019: 111-116 - [c180]Fuming Fang, Xin Wang, Junichi Yamagishi, Isao Echizen, Massimiliano Todisco, Nicholas W. D. Evans, Jean-François Bonastre:
Speaker Anonymization Using X-vector and Neural Waveform Models. SSW 2019: 155-160 - [c179]Yusuke Yasuda, Xin Wang, Junichi Yamagishi:
Initial investigation of encoder-decoder end-to-end TTS using marginalization of monotonic hard alignments. SSW 2019: 211-216 - [p4]Md. Sahidullah, Héctor Delgado, Massimiliano Todisco, Tomi Kinnunen, Nicholas W. D. Evans, Junichi Yamagishi, Kong-Aik Lee:
Introduction to Voice Presentation Attack Detection and Recent Advances. Handbook of Biometric Anti-Spoofing, 2nd Ed. 2019: 321-361 - [i50]Md. Sahidullah, Héctor Delgado, Massimiliano Todisco, Tomi Kinnunen, Nicholas W. D. Evans, Junichi Yamagishi, Kong-Aik Lee:
Introduction to Voice Presentation Attack Detection and Recent Advances. CoRR abs/1901.01085 (2019) - [i49]Yi Zhao, Atsushi Ando, Shinji Takaki, Junichi Yamagishi, Satoshi Kobashikawa:
Does the Lombard Effect Improve Emotional Communication in Noise? - Analysis of Emotional Speech Acted in Noise -. CoRR abs/1903.12316 (2019) - [i48]Mingyang Zhang, Xin Wang, Fuming Fang, Haizhou Li, Junichi Yamagishi:
Joint training framework for text-to-speech and voice conversion using multi-source Tacotron and WaveNet. CoRR abs/1903.12389 (2019) - [i47]Shinji Takaki, Hirokazu Kameoka, Junichi Yamagishi:
Training a Neural Speech Waveform Model using Spectral Losses of Short-Time Fourier Transform and Continuous Wavelet Transform. CoRR abs/1903.12392 (2019) - [i46]Hieu-Thi Luong, Xin Wang, Junichi Yamagishi, Nobuyuki Nishizawa:
Training Multi-Speaker Neural Text-to-Speech Systems using Speaker-Imbalanced Speech Corpora. CoRR abs/1904.00771 (2019) - [i45]Lauri Juvela, Bajibabu Bollepalli, Junichi Yamagishi, Paavo Alku:
GELP: GAN-Excited Linear Prediction for Speech Synthesis from Mel-spectrogram. CoRR abs/1904.03976 (2019) - [i44]Massimiliano Todisco, Xin Wang, Ville Vestman, Md. Sahidullah, Héctor Delgado, Andreas Nautsch, Junichi Yamagishi, Nicholas W. D. Evans, Tomi Kinnunen, Kong Aik Lee:
ASVspoof 2019: Future Horizons in Spoofed and Fake Audio Detection. CoRR abs/1904.05441 (2019) - [i43]Chen-Chou Lo, Szu-Wei Fu, Wen-Chin Huang, Xin Wang, Junichi Yamagishi, Yu Tsao, Hsin-Min Wang:
MOSNet: Deep Learning based Objective Assessment for Voice Conversion. CoRR abs/1904.08352 (2019) - [i42]Xin Wang, Shinji Takaki, Junichi Yamagishi:
Neural source-filter waveform models for statistical parametric speech synthesis. CoRR abs/1904.12088 (2019) - [i41]Fuming Fang, Xin Wang, Junichi Yamagishi, Isao Echizen, Massimiliano Todisco, Nicholas W. D. Evans, Jean-François Bonastre:
Speaker Anonymization Using X-vector and Neural Waveform Models. CoRR abs/1905.13561 (2019) - [i40]Huy H. Nguyen, Fuming Fang, Junichi Yamagishi, Isao Echizen:
Multi-task Learning For Detecting and Segmenting Manipulated Facial Images and Videos. CoRR abs/1906.06876 (2019) - [i39]Hieu-Thi Luong, Junichi Yamagishi:
A Unified Speaker Adaptation Method for Speech Synthesis using Transcribed and Untranscribed Speech with Backpropagation. CoRR abs/1906.07414 (2019) - [i38]David Ifeoluwa Adelani, Haotian Mai, Fuming Fang, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Generating Sentiment-Preserving Fake Online Reviews Using Neural Language Models and Their Human- and Machine-based Detection. CoRR abs/1907.09177 (2019) - [i37]Xin Wang, Junichi Yamagishi:
Neural Harmonic-plus-Noise Waveform Model with Trainable Maximum Voice Frequency for Text-to-Speech Synthesis. CoRR abs/1908.10256 (2019) - [i36]Yusuke Yasuda, Xin Wang, Junichi Yamagishi:
Initial investigation of an encoder-decoder end-to-end TTS framework using marginalization of monotonic hard latent alignments. CoRR abs/1908.11535 (2019) - [i35]Hieu-Thi Luong, Junichi Yamagishi:
Bootstrapping non-parallel voice conversion from speaker-adaptive text-to-speech. CoRR abs/1909.06532 (2019) - [i34]Yi Zhao, Xin Wang, Lauri Juvela, Junichi Yamagishi:
Transferring neural speech waveform synthesizers to musical instrument sounds generation. CoRR abs/1910.12381 (2019) - [i33]Yusuke Yasuda, Xin Wang, Junichi Yamagishi:
Effect of choice of probability distribution, randomness, and search methods for alignment modeling in sequence-to-sequence text-to-speech synthesis using hard alignment. CoRR abs/1910.12383 (2019) - [i32]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Use of a Capsule Network to Detect Fake Images and Videos. CoRR abs/1910.12467 (2019) - [i31]Rong Huang, Fuming Fang, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
A Method for Identifying Origin of Digital Images Using a Convolution Neural Network. CoRR abs/1911.00655 (2019) - [i30]Rong Huang, Fuming Fang, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Security of Facial Forensics Models Against Adversarial Attacks. CoRR abs/1911.00660 (2019) - [i29]Xin Wang, Junichi Yamagishi, Massimiliano Todisco, Héctor Delgado, Andreas Nautsch, Nicholas W. D. Evans, Md. Sahidullah, Ville Vestman, Tomi Kinnunen, Kong Aik Lee, Lauri Juvela, Paavo Alku, Yu-Huai Peng, Hsin-Te Hwang, Yu Tsao, Hsin-Min Wang, Sébastien Le Maguer, Markus Becker, Fergus Henderson, Rob Clark, Yu Zhang, Quan Wang, Ye Jia, Kai Onuma, Koji Mushika, Takashi Kaneda, Yuan Jiang, Li-Juan Liu, Yi-Chiao Wu, Wen-Chin Huang, Tomoki Toda, Kou Tanaka, Hirokazu Kameoka, Ingmar Steiner, Driss Matrouf, Jean-François Bonastre, Avashna Govender, Srikanth Ronanki, Jing-Xuan Zhang, Zhen-Hua Ling:
The ASVspoof 2019 database. CoRR abs/1911.01601 (2019) - [i28]Seyyed Saeed Sarfjoo, Xin Wang, Gustav Eje Henter, Jaime Lorenzo-Trueba, Shinji Takaki, Junichi Yamagishi:
Transformation of low-quality device-recorded speech to high-quality speech using improved SEGAN model. CoRR abs/1911.03952 (2019) - [i27]Huy H. Nguyen, Minoru Kuribayashi, Junichi Yamagishi, Isao Echizen:
Detecting and Correcting Adversarial Images Using Image Processing Operations. CoRR abs/1912.05391 (2019) - 2018
- [j47]Yi Zhao, Shinji Takaki, Hieu-Thi Luong, Junichi Yamagishi, Daisuke Saito, Nobuaki Minematsu:
Wasserstein GAN and Waveform Loss-Based Acoustic Model Training for Multi-Speaker Text-to-Speech Synthesis Systems Using a WaveNet Vocoder. IEEE Access 6: 60478-60488 (2018) - [j46]Xin Wang, Shinji Takaki, Junichi Yamagishi:
Investigating very deep highway networks for parametric speech synthesis. Speech Commun. 96: 1-9 (2018) - [j45]Jaime Lorenzo-Trueba, Gustav Eje Henter, Shinji Takaki, Junichi Yamagishi, Yosuke Morino, Yuta Ochiai:
Investigating different representations for modeling and controlling multiple emotions in DNN-based speech synthesis. Speech Commun. 99: 135-143 (2018) - [j44]Xin Wang, Shinji Takaki, Junichi Yamagishi:
Autoregressive Neural F0 Model for Statistical Parametric Speech Synthesis. IEEE ACM Trans. Audio Speech Lang. Process. 26(8): 1406-1419 (2018) - [j43]Cassia Valentini-Botinhao, Junichi Yamagishi:
Speech Enhancement of Noisy and Reverberant Speech for Text-to-Speech. IEEE ACM Trans. Audio Speech Lang. Process. 26(8): 1420-1433 (2018) - [j42]Manu Airaksinen, Lauri Juvela, Bajibabu Bollepalli, Junichi Yamagishi, Paavo Alku:
A Comparison Between STRAIGHT, Glottal, and Sinusoidal Vocoding in Statistical Parametric Speech Synthesis. IEEE ACM Trans. Audio Speech Lang. Process. 26(9): 1658-1670 (2018) - [c178]Huy H. Nguyen, Ngoc-Dung T. Tieu, Hoang-Quoc Nguyen-Son, Vincent Nozick, Junichi Yamagishi, Isao Echizen:
Modular Convolutional Neural Network for Discriminating between Computer-Generated Images and Photographic Images. ARES 2018: 1:1-1:10 - [c177]Shinji Takaki, Yoshikazu Nishimura, Junichi Yamagishi:
Unsupervised Speaker Adaptation for DNN-based Speech Synthesis using Input Codes. APSIPA 2018: 649-658 - [c176]Gustav Eje Henter, Jaime Lorenzo-Trueba, Xin Wang, Mariko Kondo, Junichi Yamagishi:
Cyborg Speech: Deep Multilingual Speech Synthesis for Generating Segmental Foreign Accent with Natural Prosody. ICASSP 2018: 4799-4803 - [c175]Xin Wang, Jaime Lorenzo-Trueba, Shinji Takaki, Lauri Juvela, Junichi Yamagishi:
A Comparison of Recent Waveform Generation and Acoustic Modeling Methods for Neural-Network-Based Speech Synthesis. ICASSP 2018: 4804-4808 - [c174]Fuming Fang, Junichi Yamagishi, Isao Echizen, Jaime Lorenzo-Trueba:
High-Quality Nonparallel Voice Conversion Based on Cycle-Consistent Adversarial Network. ICASSP 2018: 5279-5283 - [c173]Lauri Juvela, Bajibabu Bollepalli, Xin Wang, Hirokazu Kameoka, Manu Airaksinen, Junichi Yamagishi, Paavo Alku:
Speech Waveform Synthesis from MFCC Sequences with Generative Adversarial Networks. ICASSP 2018: 5679-5683 - [c172]Huy H. Nguyen, Ngoc-Dung T. Tieu, Hoang-Quoc Nguyen-Son, Junichi Yamagishi, Isao Echizen:
Transformation on Computer-Generated Facial Image to Avoid Detection by Spoofing Detector. ICME 2018: 1-6 - [c171]Hieu-Thi Luong, Xin Wang, Junichi Yamagishi, Nobuyuki Nishizawa:
Investigating Accuracy of Pitch-accent Annotations in Neural Network-based Speech Synthesis and Denoising Effects. INTERSPEECH 2018: 37-41 - [c170]Massimiliano Todisco, Héctor Delgado, Kong-Aik Lee, Md. Sahidullah, Nicholas W. D. Evans, Tomi Kinnunen, Junichi Yamagishi:
Integrated Presentation Attack Detection and Automatic Speaker Verification: Common Features and Gaussian Back-end Fusion. INTERSPEECH 2018: 77-81 - [c169]Lauri Juvela, Vassilis Tsiaras, Bajibabu Bollepalli, Manu Airaksinen, Junichi Yamagishi, Paavo Alku:
Speaker-independent Raw Waveform Model for Glottal Excitation. INTERSPEECH 2018: 2012-2016 - [c168]Hieu-Thi Luong, Junichi Yamagishi:
Multimodal Speech Synthesis Architecture for Unsupervised Speaker Adaptation. INTERSPEECH 2018: 2494-2498 - [c167]Igor Jauk, Jaime Lorenzo-Trueba, Junichi Yamagishi, Antonio Bonafonte:
Expressive Speech Synthesis Using Sentiment Embeddings. INTERSPEECH 2018: 3062-3066 - [c166]Tomi Kinnunen, Jaime Lorenzo-Trueba, Junichi Yamagishi, Tomoki Toda, Daisuke Saito, Fernando Villavicencio, Zhen-Hua Ling:
A Spoofing Benchmark for the 2018 Voice Conversion Challenge: Leveraging from Spoofing Countermeasures for Speech Artifact Assessment. Odyssey 2018: 187-194 - [c165]Jaime Lorenzo-Trueba, Junichi Yamagishi, Tomoki Toda, Daisuke Saito, Fernando Villavicencio, Tomi Kinnunen, Zhen-Hua Ling:
The Voice Conversion Challenge 2018: Promoting Development of Parallel and Nonparallel Methods. Odyssey 2018: 195-202 - [c164]Jaime Lorenzo-Trueba, Fuming Fang, Xin Wang, Isao Echizen, Junichi Yamagishi, Tomi Kinnunen:
Can we steal your vocal identity from the Internet?: Initial investigation of cloning Obama's voice using GAN, WaveNet and low-quality found data. Odyssey 2018: 240-247 - [c163]Héctor Delgado, Massimiliano Todisco, Md. Sahidullah, Nicholas W. D. Evans, Tomi Kinnunen, Kong-Aik Lee, Junichi Yamagishi:
ASVspoof 2017 Version 2.0: meta-data analysis and baseline enhancements. Odyssey 2018: 296-303 - [c162]Tomi Kinnunen, Kong-Aik Lee, Héctor Delgado, Nicholas W. D. Evans, Massimiliano Todisco, Md. Sahidullah, Junichi Yamagishi, Douglas A. Reynolds:
t-DCF: a Detection Cost Function for the Tandem Assessment of Spoofing Countermeasures and Automatic Speaker Verification. Odyssey 2018: 312-319 - [c161]Hoang-Quoc Nguyen-Son, Huy H. Nguyen, Ngoc-Dung T. Tieu, Junichi Yamagishi, Isao Echizen:
Identifying Computer-Translated Paragraphs using Coherence Features. PACLIC 2018 - [c160]Hieu-Thi Luong, Junichi Yamagishi:
Scaling and Bias Codes for Modeling Speaker-Adaptive DNN-Based Speech Synthesis Systems. SLT 2018: 610-617 - [c159]Darius Afchar, Vincent Nozick, Junichi Yamagishi, Isao Echizen:
MesoNet: a Compact Facial Video Forgery Detection Network. WIFS 2018: 1-7 - [c158]Fuming Fang, Junichi Yamagishi, Isao Echizen, Md. Sahidullah, Tomi Kinnunen:
Transforming acoustic characteristics to deceive playback spoofing countermeasures of speaker verification systems. WIFS 2018: 1-9 - [i26]Jaime Lorenzo-Trueba, Fuming Fang, Xin Wang, Isao Echizen, Junichi Yamagishi, Tomi Kinnunen:
Can we steal your vocal identity from the Internet?: Initial investigation of cloning Obama's voice using GAN, WaveNet and low-quality found data. CoRR abs/1803.00860 (2018) - [i25]Toru Nakashika, Shinji Takaki, Junichi Yamagishi:
Complex-Valued Restricted Boltzmann Machine for Direct Speech Parameterization from Complex Spectra. CoRR abs/1803.09946 (2018) - [i24]Fuming Fang, Junichi Yamagishi, Isao Echizen, Jaime Lorenzo-Trueba:
High-quality nonparallel voice conversion based on cycle-consistent adversarial network. CoRR abs/1804.00425 (2018) - [i23]Lauri Juvela, Bajibabu Bollepalli, Xin Wang, Hirokazu Kameoka, Manu Airaksinen, Junichi Yamagishi, Paavo Alku:
Speech waveform synthesis from MFCC sequences with generative adversarial networks. CoRR abs/1804.00920 (2018) - [i22]Xin Wang, Jaime Lorenzo-Trueba, Shinji Takaki, Lauri Juvela, Junichi Yamagishi:
A comparison of recent waveform generation and acoustic modeling methods for neural-network-based speech synthesis. CoRR abs/1804.02549 (2018) - [i21]Jaime Lorenzo-Trueba, Junichi Yamagishi, Tomoki Toda, Daisuke Saito, Fernando Villavicencio, Tomi Kinnunen, Zhen-Hua Ling:
The Voice Conversion Challenge 2018: Promoting Development of Parallel and Nonparallel Methods. CoRR abs/1804.04262 (2018) - [i20]Huy H. Nguyen, Ngoc-Dung T. Tieu, Hoang-Quoc Nguyen-Son, Junichi Yamagishi, Isao Echizen:
Transformation on Computer-Generated Facial Image to Avoid Detection by Spoofing Detector. CoRR abs/1804.04418 (2018) - [i19]Tomi Kinnunen, Jaime Lorenzo-Trueba, Junichi Yamagishi, Tomoki Toda, Daisuke Saito, Fernando Villavicencio, Zhen-Hua Ling:
A Spoofing Benchmark for the 2018 Voice Conversion Challenge: Leveraging from Spoofing Countermeasures for Speech Artifact Assessment. CoRR abs/1804.08438 (2018) - [i18]Lauri Juvela, Vassilis Tsiaras, Bajibabu Bollepalli, Manu Airaksinen, Junichi Yamagishi, Paavo Alku:
Speaker-independent raw waveform model for glottal excitation. CoRR abs/1804.09593 (2018) - [i17]Tomi Kinnunen, Kong-Aik Lee, Héctor Delgado, Nicholas W. D. Evans, Massimiliano Todisco, Md. Sahidullah, Junichi Yamagishi, Douglas A. Reynolds:
t-DCF: a Detection Cost Function for the Tandem Assessment of Spoofing Countermeasures and Automatic Speaker Verification. CoRR abs/1804.09618 (2018) - [i16]Gustav Eje Henter, Xin Wang, Junichi Yamagishi:
Deep Encoder-Decoder Models for Unsupervised Learning of Controllable Speech Synthesis. CoRR abs/1807.11470 (2018) - [i15]Hieu-Thi Luong, Junichi Yamagishi:
Scaling and bias codes for modeling speaker-adaptive DNN-based speech synthesis systems. CoRR abs/1807.11632 (2018) - [i14]Yi Zhao, Shinji Takaki, Hieu-Thi Luong, Junichi Yamagishi, Daisuke Saito, Nobuaki Minematsu:
Wasserstein GAN and Waveform Loss-based Acoustic Model Training for Multi-speaker Text-to-Speech Synthesis Systems Using a WaveNet Vocoder. CoRR abs/1807.11679 (2018) - [i13]Hieu-Thi Luong, Xin Wang, Junichi Yamagishi, Nobuyuki Nishizawa:
Investigating accuracy of pitch-accent annotations in neural network-based speech synthesis and denoising effects. CoRR abs/1808.00665 (2018) - [i12]Hieu-Thi Luong, Junichi Yamagishi:
Multimodal speech synthesis architecture for unsupervised speaker adaptation. CoRR abs/1808.06288 (2018) - [i11]Darius Afchar, Vincent Nozick, Junichi Yamagishi, Isao Echizen:
MesoNet: a Compact Facial Video Forgery Detection Network. CoRR abs/1809.00888 (2018) - [i10]Fuming Fang, Junichi Yamagishi, Isao Echizen, Md. Sahidullah, Tomi Kinnunen:
Transforming acoustic characteristics to deceive playback spoofing countermeasures of speaker verification systems. CoRR abs/1809.04274 (2018) - [i9]Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Capsule-Forensics: Using Capsule Networks to Detect Forged Images and Videos. CoRR abs/1810.11215 (2018) - [i8]Shinji Takaki, Toru Nakashika, Xin Wang, Junichi Yamagishi:
STFT spectral loss for training a neural speech waveform model. CoRR abs/1810.11945 (2018) - [i7]Xin Wang, Shinji Takaki, Junichi Yamagishi:
Neural source-filter-based waveform model for statistical parametric speech synthesis. CoRR abs/1810.11946 (2018) - [i6]Yusuke Yasuda, Xin Wang, Shinji Takaki, Junichi Yamagishi:
Investigation of enhanced Tacotron text-to-speech synthesis systems with self-attention for pitch accent language. CoRR abs/1810.11960 (2018) - [i5]Lauri Juvela, Bajibabu Bollepalli, Junichi Yamagishi, Paavo Alku:
Waveform generation for text-to-speech synthesis using pitch-synchronous multi-scale generative adversarial networks. CoRR abs/1810.12598 (2018) - [i4]Fuming Fang, Xin Wang, Junichi Yamagishi, Isao Echizen:
Audiovisual speaker conversion: jointly and simultaneously transforming facial expression and acoustic characteristics. CoRR abs/1810.12730 (2018) - [i3]Cheng-I Lai, Alberto Abad, Korin Richmond, Junichi Yamagishi, Najim Dehak, Simon King:
Attentive Filtering Networks for Audio Replay Attack Detection. CoRR abs/1810.13048 (2018) - [i2]Hoang-Quoc Nguyen-Son, Ngoc-Dung T. Tieu, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Identifying Computer-Translated Paragraphs using Coherence Features. CoRR abs/1812.10896 (2018) - 2017
- [j41]Michael Pucher, Bettina Zillinger, Markus Toman, Dietmar Schabus, Cassia Valentini-Botinhao, Junichi Yamagishi, Erich Schmid, Thomas Woltron:
Influence of speaker familiarity on blind and visually impaired children's and young adults' perception of synthetic voices. Comput. Speech Lang. 46: 179-195 (2017) - [j40]Junichi Yamagishi, Tomi Kinnunen, Nicholas W. D. Evans, Phillip L. De Leon, Isabel Trancoso:
Introduction to the Issue on Spoofing and Countermeasures for Automatic Speaker Verification. IEEE J. Sel. Top. Signal Process. 11(4): 585-587 (2017) - [j39]Zhizheng Wu, Junichi Yamagishi, Tomi Kinnunen, Cemal Hanilçi, Md. Sahidullah, Aleksandr Sizov, Nicholas W. D. Evans, Massimiliano Todisco:
ASVspoof: The Automatic Speaker Verification Spoofing and Countermeasures Challenge. IEEE J. Sel. Top. Signal Process. 11(4): 588-604 (2017) - [c157]Hoang-Quoc Nguyen-Son, Ngoc-Dung T. Tieu, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen:
Identifying computer-generated text using statistical analysis. APSIPA 2017: 1504-1511 - [c156]Xin Wang, Shinji Takaki, Junichi Yamagishi:
An autoregressive recurrent mixture density network for parametric speech synthesis. ICASSP 2017: 4895-4899 - [c155]Hieu-Thi Luong, Shinji Takaki, Gustav Eje Henter, Junichi Yamagishi:
Adapting and controlling DNN-based speech synthesis using input codes. ICASSP 2017: 4905-4909 - [c154]Tomi Kinnunen, Lauri Juvela, Paavo Alku, Junichi Yamagishi:
Non-parallel voice conversion using i-vector PLDA: towards unifying speaker verification and transformation. ICASSP 2017: 5535-5539 - [c153]Tomi Kinnunen, Md. Sahidullah, Héctor Delgado, Massimiliano Todisco, Nicholas W. D. Evans, Junichi Yamagishi, Kong-Aik Lee:
The ASVspoof 2017 Challenge: Assessing the Limits of Replay Spoofing Attack Detection. INTERSPEECH 2017: 2-6 - [c152]Jaime Lorenzo-Trueba, Cassia Valentini-Botinhao, Gustav Eje Henter, Junichi Yamagishi:
Misperceptions of the Emotional Content of Natural and Vocoded Speech in a Car. INTERSPEECH 2017: 606-610 - [c151]Manuel Sam Ribeiro, Oliver Watts, Junichi Yamagishi:
Learning Word Vector Representations Based on Acoustic Counts. INTERSPEECH 2017: 799-803 - [c150]Xin Wang, Shinji Takaki, Junichi Yamagishi:
An RNN-Based Quantized F0 Model with Multi-Tier Feedback Links for Text-to-Speech Synthesis. INTERSPEECH 2017: 1059-1063 - [c149]Shinji Takaki, Hirokazu Kameoka, Junichi Yamagishi:
Direct Modeling of Frequency Spectra and Waveform Generation Based on Phase Recovery for DNN-Based Speech Synthesis. INTERSPEECH 2017: 1128-1132 - [c148]Lauri Juvela, Bajibabu Bollepalli, Junichi Yamagishi, Paavo Alku:
Reducing Mismatch in Training of DNN-Based Glottal Excitation Models in a Statistical Parametric Text-to-Speech System. INTERSPEECH 2017: 1368-1372 - [c147]Cassia Valentini-Botinhao, Junichi Yamagishi:
Speech Intelligibility in Cars: The Effect of Speaking Style, Noise and Listener Age. INTERSPEECH 2017: 2944-2948 - [c146]Takuhiro Kaneko, Shinji Takaki, Hirokazu Kameoka, Junichi Yamagishi:
Generative Adversarial Network-Based Postfilter for STFT Spectrograms. INTERSPEECH 2017: 3389-3393 - [c145]Gustav Eje Henter, Jaime Lorenzo-Trueba, Xin Wang, Junichi Yamagishi:
Principles for Learning Controllable TTS from Annotated and Latent Variation. INTERSPEECH 2017: 3956-3960 - [c144]Toru Nakashika, Shinji Takaki, Junichi Yamagishi:
Complex-Valued Restricted Boltzmann Machine for Direct Learning of Frequency Spectra. INTERSPEECH 2017: 4021-4025 - [c143]Nicolas Rahmouni, Vincent Nozick, Junichi Yamagishi, Isao Echizen:
Distinguishing computer graphics from natural images using convolution neural networks. WIFS 2017: 1-6 - [c142]Ngoc-Dung T. Tieu, Huy H. Nguyen, Hoang-Quoc Nguyen-Son, Junichi Yamagishi, Isao Echizen:
An approach for gait anonymization using deep learning. WIFS 2017: 1-6 - [p3]Keiichi Tokuda, Akinobu Lee, Yoshihiko Nankaku, Keiichiro Oura, Kei Hashimoto, Daisuke Yamamoto, Ichi Takumi, Takahiro Uchiya, Shuhei Tsutsumi, Steve Renals, Junichi Yamagishi:
User Generated Dialogue Systems: uDialogue. Human-Harmonized Information Technology (2) 2017: 77-114 - 2016
- [j38]Adriana Stan, Yoshitaka Mamiya, Junichi Yamagishi, Peter Bell, Oliver Watts, Robert A. J. Clark, Simon King:
ALISA: An automatic lightly supervised speech segmentation and alignment tool. Comput. Speech Lang. 35: 116-133 (2016) - [j37]Xin Wang, Shinji Takaki, Junichi Yamagishi:
Investigation of Using Continuous Representation of Various Linguistic Units in Neural Network Based Text-to-Speech Synthesis. IEICE Trans. Inf. Syst. 99-D(10): 2471-2480 (2016) - [j36]Zhizheng Wu, Phillip L. De Leon, Cenk Demiroglu, Ali Khodabakhsh, Simon King, Zhen-Hua Ling, Daisuke Saito, Bryan Stewart, Tomoki Toda, Mirjam Wester, Junichi Yamagishi:
Anti-Spoofing for Text-Independent Speaker Verification: An Initial Database, Comparison of Countermeasures, and Human Performance. IEEE ACM Trans. Audio Speech Lang. Process. 24(4): 768-783 (2016) - [c141]Lauri Juvela, Xin Wang, Shinji Takaki, Sangjin Kim, Manu Airaksinen, Junichi Yamagishi:
The NII speech synthesis entry for Blizzard Challenge 2016. Blizzard Challenge 2016 - [c140]Jaime Lorenzo-Trueba, Roberto Barra-Chicote, Ascensión Gallardo-Antolín, Junichi Yamagishi, Juan Manuel Montero:
Continuous Expressive Speaking Styles Synthesis based on CVSM and MR-HMM. COLING 2016: 369-376 - [c139]Thomas Merritt, Robert A. J. Clark, Zhizheng Wu, Junichi Yamagishi, Simon King:
Deep neural network-guided unit selection synthesis. ICASSP 2016: 5145-5149 - [c138]Rasmus Dall, Sandrine Brognaux, Korin Richmond, Cassia Valentini-Botinhao, Gustav Eje Henter, Julia Hirschberg, Junichi Yamagishi, Simon King:
Testing the consistency assumption: Pronunciation variant forced alignment in read and spontaneous speech synthesis. ICASSP 2016: 5155-5159 - [c137]Kei Hashimoto, Junichi Yamagishi, Isao Echizen:
Privacy-preserving sound to degrade automatic speaker verification performance. ICASSP 2016: 5500-5504 - [c136]Manuel Sam Ribeiro, Oliver Watts, Junichi Yamagishi, Robert A. J. Clark:
Wavelet-based decomposition of F0 as a secondary task for DNN-based speech synthesis with multi-task learning. ICASSP 2016: 5525-5529 - [c135]Shinji Takaki, Junichi Yamagishi:
A deep auto-encoder based low-dimensional feature extraction from FFT spectral envelopes for statistical parametric speech synthesis. ICASSP 2016: 5535-5539 - [c134]Qiong Hu, Junichi Yamagishi, Korin Richmond, Kartick Subramanian, Yannis Stylianou:
Initial investigation of speech synthesis based on complex-valued neural networks. ICASSP 2016: 5630-5634 - [c133]Takenori Yoshimura, Gustav Eje Henter, Oliver Watts, Mirjam Wester, Junichi Yamagishi, Keiichi Tokuda:
A Hierarchical Predictor of Synthetic Speech Naturalness Using Neural Networks. INTERSPEECH 2016: 342-346 - [c132]Cassia Valentini-Botinhao, Xin Wang, Shinji Takaki, Junichi Yamagishi:
Speech Enhancement for a Noise-Robust Text-to-Speech Synthesis System Using Deep Recurrent Neural Networks. INTERSPEECH 2016: 352-356 - [c131]Lauri Juvela, Hirokazu Kameoka, Manu Airaksinen, Junichi Yamagishi, Paavo Alku:
Majorisation-Minimisation Based Optimisation of the Composite Autoregressive System with Application to Glottal Inverse Filtering. INTERSPEECH 2016: 968-972 - [c130]Jean-Philippe Goldman, Pierre-Edouard Honnet, Robert A. J. Clark, Philip N. Garner, Maria Ivanova, Alexandros Lazaridis, Hui Liang, Tiago Macedo, Beat Pfister, Manuel Sam Ribeiro, Eric Wehrli, Junichi Yamagishi:
The SIWIS Database: A Multilingual Speech Database with Acted Emphasis. INTERSPEECH 2016: 1532-1535 - [c129]Tomoki Toda, Ling-Hui Chen, Daisuke Saito, Fernando Villavicencio, Mirjam Wester, Zhizheng Wu, Junichi Yamagishi:
The Voice Conversion Challenge 2016. INTERSPEECH 2016: 1632-1636 - [c128]Mirjam Wester, Zhizheng Wu, Junichi Yamagishi:
Analysis of the Voice Conversion Challenge 2016 Evaluation Results. INTERSPEECH 2016: 1637-1641 - [c127]Fernando Villavicencio, Junichi Yamagishi, Jordi Bonada, Felipe Espic:
Applying Spectral Normalisation and Efficient Envelope Estimation and Statistical Transformation for the Voice Conversion Challenge 2016. INTERSPEECH 2016: 1657-1661 - [c126]Lauri Juvela, Xin Wang, Shinji Takaki, Manu Airaksinen, Junichi Yamagishi, Paavo Alku:
Using Text and Acoustic Features in Predicting Glottal Excitation Waveforms for Parametric Speech Synthesis with Recurrent Neural Networks. INTERSPEECH 2016: 2283-2287 - [c125]Xin Wang, Shinji Takaki, Junichi Yamagishi:
Enhance the Word Vector with Prosodic Information for the Recurrent Neural Network Based TTS System. INTERSPEECH 2016: 2856-2860 - [c124]Manuel Sam Ribeiro, Oliver Watts, Junichi Yamagishi:
Syllable-Level Representations of Suprasegmental Features for DNN-Based Text-to-Speech Synthesis. INTERSPEECH 2016: 3186-3190 - [c123]Sayaka Shiota, Fernando Villavicencio, Junichi Yamagishi, Nobutaka Ono, Isao Echizen, Tomoko Matsui:
Voice Liveness Detection for Speaker Verification based on a Tandem Single/Double-channel Pop Noise Detector. Odyssey 2016: 259-263 - [c122]Mirjam Wester, Zhizheng Wu, Junichi Yamagishi:
Multidimensional scaling of systems in the Voice Conversion Challenge 2016. SSW 2016: 38-43 - [c121]Michael Pucher, Fernando Villavicencio, Junichi Yamagishi:
Development of a statistical parametric synthesis system for operatic singing in German. SSW 2016: 64-69 - [c120]Manuel Sam Ribeiro, Oliver Watts, Junichi Yamagishi:
Parallel and cascaded deep neural networks for text-to-speech synthesis. SSW 2016: 100-105 - [c119]Xin Wang, Shinji Takaki, Junichi Yamagishi:
A Comparative Study of the Performance of HMM, DNN, and RNN based Speech Synthesis Systems Trained on Very Large Speaker-Dependent Corpora. SSW 2016: 118-121 - [c118]Cassia Valentini-Botinhao, Xin Wang, Shinji Takaki, Junichi Yamagishi:
Investigating RNN-based speech enhancement methods for noise-robust Text-to-Speech. SSW 2016: 146-152 - [c117]Shinji Takaki, Sangjin Kim, Junichi Yamagishi:
Speaker Adaptation of Various Components in Deep Neural Network based Speech Synthesis. SSW 2016: 153-159 - [c116]Xin Wang, Shinji Takaki, Junichi Yamagishi:
Investigating Very Deep Highway Networks for Parametric Speech Synthesis. SSW 2016: 166-171 - [p2]Shinji Takaki, Junichi Yamagishi:
Constructing a Deep Neural Network Based Spectral Model for Statistical Speech Synthesis. Recent Advances in Nonlinear Speech Processing 2016: 117-125 - 2015
- [j35]Jaime Lorenzo-Trueba, Roberto Barra-Chicote, Rubén San Segundo, Javier Ferreiros, Junichi Yamagishi, Juan Manuel Montero:
Emotion transplantation through adaptation in HMM-based speech synthesis. Comput. Speech Lang. 34(1): 292-307 (2015) - [j34]Zhizheng Wu, Nicholas W. D. Evans, Tomi Kinnunen, Junichi Yamagishi, Federico Alegre, Haizhou Li:
Spoofing and countermeasures for speaker verification: A survey. Speech Commun. 66: 130-153 (2015) - [j33]Cassia Valentini-Botinhao, Markus Toman, Michael Pucher, Dietmar Schabus, Junichi Yamagishi:
Intelligibility of time-compressed synthetic speech: Compression method and speaking style. Speech Commun. 74: 52-64 (2015) - [j32]Ling-Hui Chen, Tuomo Raitio, Cassia Valentini-Botinhao, Zhen-Hua Ling, Junichi Yamagishi:
A Deep Generative Architecture for Postfiltering in Statistical Parametric Speech Synthesis. IEEE ACM Trans. Audio Speech Lang. Process. 23(11): 2003-2014 (2015) - [c115]Zhizheng Wu, Ali Khodabakhsh, Cenk Demiroglu, Junichi Yamagishi, Daisuke Saito, Tomoki Toda, Simon King:
SAS: A speaker verification spoofing database containing diverse attacks. ICASSP 2015: 4440-4444 - [c114]Qiong Hu, Yannis Stylianou, Ranniery Maia, Korin Richmond, Junichi Yamagishi:
Methods for applying dynamic sinusoidal models to statistical parametric speech synthesis. ICASSP 2015: 4889-4893 - [c113]Sayaka Shiota, Fernando Villavicencio, Junichi Yamagishi, Nobutaka Ono, Isao Echizen, Tomoko Matsui:
Voice liveness detection algorithms based on pop noise caused by human breath for automatic speaker verification. INTERSPEECH 2015: 239-243 - [c112]Qiong Hu, Zhizheng Wu, Korin Richmond, Junichi Yamagishi, Yannis Stylianou, Ranniery Maia:
Fusion of multiple parameterisations for DNN-based sinusoidal speech synthesis with multi-task learning. INTERSPEECH 2015: 854-858 - [c111]Manuel Sam Ribeiro, Junichi Yamagishi, Robert A. J. Clark:
A perceptual investigation of wavelet-based decomposition of f0 for text-to-speech synthesis. INTERSPEECH 2015: 1586-1590 - [c110]Michael Pucher, Markus Toman, Dietmar Schabus, Cassia Valentini-Botinhao, Junichi Yamagishi, Bettina Zillinger, Erich Schmid:
Influence of speaker familiarity on blind and visually impaired children's perception of synthetic voices in audio games. INTERSPEECH 2015: 1625-1629 - [c109]Zhizheng Wu, Tomi Kinnunen, Nicholas W. D. Evans, Junichi Yamagishi, Cemal Hanilçi, Md. Sahidullah, Aleksandr Sizov:
ASVspoof 2015: the first automatic speaker verification spoofing and countermeasures challenge. INTERSPEECH 2015: 2037-2041 - [c108]Mirjam Wester, Zhizheng Wu, Junichi Yamagishi:
Human vs machine spoofing detection on wideband and narrowband data. INTERSPEECH 2015: 2047-2051 - [c107]Thomas Merritt, Junichi Yamagishi, Zhizheng Wu, Oliver Watts, Simon King:
Deep neural network context embeddings for model selection in rich-context HMM synthesis. INTERSPEECH 2015: 2207-2211 - [c106]Pierre Lanchantin, Christophe Veaux, Mark J. F. Gales, Simon King, Junichi Yamagishi:
Reconstructing voices within the multiple-average-voice-model framework. INTERSPEECH 2015: 2232-2236 - [c105]Shinji Takaki, Sangjin Kim, Junichi Yamagishi, JongJin Kim:
Multiple feed-forward deep neural networks for statistical parametric speech synthesis. INTERSPEECH 2015: 2242-2246 - [c104]Junichi Yamagishi, Nicholas W. D. Evans:
Automatic speaker verification spoofing and countermeasures (ASVspoof 2015): open discussion and future plans. INTERSPEECH 2015 - [c103]Christophe Veaux, Junichi Yamagishi, Simon King:
A Comparison of Manual and Automatic Voice Repair for Individual with Vocal Disabilities. SLPAT@Interspeech 2015: 130-133 - [r1]Nicholas W. D. Evans, Federico Alegre, Tomi Kinnunen, Zhizheng Wu, Junichi Yamagishi:
Anti-spoofing, Voice Databases. Encyclopedia of Biometrics 2015: 123-128 - [i1]Zhenzhou Wu, Shinji Takaki, Junichi Yamagishi:
Deep Denoising Auto-encoder for Statistical Speech Synthesis. CoRR abs/1506.05268 (2015) - 2014
- [j31]Cassia Valentini-Botinhao, Junichi Yamagishi, Simon King, Ranniery Maia:
Intelligibility enhancement of HMM-generated speech in additive noise by modifying Mel cepstral coefficients to increase the glimpse proportion. Comput. Speech Lang. 28(2): 665-686 (2014) - [j30]João P. Cabral, Korin Richmond, Junichi Yamagishi, Steve Renals:
Glottal Spectral Separation for Speech Synthesis. IEEE J. Sel. Top. Signal Process. 8(2): 195-208 (2014) - [j29]Lakshmi Babu Saheer, Junichi Yamagishi, Philip N. Garner, John Dines:
Combining Vocal Tract Length Normalization With Hierarchical Linear Transformations. IEEE J. Sel. Top. Signal Process. 8(2): 262-272 (2014) - [j28]Moses Ekpenyong, Eno-Abasi Urua, Oliver Watts, Simon King, Junichi Yamagishi:
Statistical parametric speech synthesis for Ibibio. Speech Commun. 56: 243-251 (2014) - [c102]Jaime Lorenzo-Trueba, Roberto Barra-Chicote, Junichi Yamagishi, Juan Manuel Montero:
Towards Cross-Lingual Emotion Transplantation. IberSPEECH 2014: 199-208 - [c101]Pierre Lanchantin, Mark J. F. Gales, Simon King, Junichi Yamagishi:
Multiple-average-voice-based speech synthesis. ICASSP 2014: 285-289 - [c100]Oliver Watts, Siva Reddy Gangireddy, Junichi Yamagishi, Simon King, Steve Renals, Adriana Stan, Mircea Giurgiu:
Neural net word representations for phrase-break prediction without a part of speech tagger. ICASSP 2014: 2599-2603 - [c99]Qiong Hu, Yannis Stylianou, Korin Richmond, Ranniery Maia, Junichi Yamagishi, Javier Latorre:
A fixed dimension and perceptually based dynamic sinusoidal model of speech. ICASSP 2014: 6270-6274 - [c98]Jaime Lorenzo-Trueba, Julián D. Echeverry-Correa, Roberto Barra-Chicote, Rubén San-Segundo-Hernández, Javier Ferreiros, Ascensión Gallardo-Antolín, Junichi Yamagishi, Simon King, Juan Manuel Montero-Martínez:
Development of a genre-dependent TTS system with cross-speaker speaking-style transplantation. SLAM@INTERSPEECH 2014: 39-42 - [c97]Qiong Hu, Yannis Stylianou, Ranniery Maia, Korin Richmond, Junichi Yamagishi, Javier Latorre:
An investigation of the application of dynamic sinusoidal models to statistical parametric speech synthesis. INTERSPEECH 2014: 780-784 - [c96]María Luisa García Lecumberri, Roberto Barra-Chicote, Rubén Pérez Ramón, Junichi Yamagishi, Martin Cooke:
Generating segmental foreign accent. INTERSPEECH 2014: 1302-1306 - [c95]Ling-Hui Chen, Tuomo Raitio, Cassia Valentini-Botinhao, Junichi Yamagishi, Zhen-Hua Ling:
DNN-based stochastic postfilter for HMM-based speech synthesis. INTERSPEECH 2014: 1954-1958 - [c94]Cassia Valentini-Botinhao, Markus Toman, Michael Pucher, Dietmar Schabus, Junichi Yamagishi:
Intelligibility analysis of fast synthesized speech. INTERSPEECH 2014: 2922-2926 - [p1]Nicholas W. D. Evans, Tomi Kinnunen, Junichi Yamagishi, Zhizheng Wu, Federico Alegre, Phillip L. De Leon:
Speaker Recognition Anti-spoofing. Handbook of Biometric Anti-Spoofing 2014: 125-146 - 2013
- [j27]John Dines, Hui Liang, Lakshmi Babu Saheer, Matthew Gibson, William Byrne, Keiichiro Oura, Keiichi Tokuda, Junichi Yamagishi, Simon King, Mirjam Wester, Teemu Hirsimäki, Reima Karhila, Mikko Kurimo:
Personalising speech-to-speech translation: Unsupervised cross-lingual speaker adaptation for HMM-based speech synthesis. Comput. Speech Lang. 27(2): 420-437 (2013) - [j26]Sarah M. Creer, Stuart P. Cunningham, Phil D. Green, Junichi Yamagishi:
Building personalised synthetic voices for individuals with severe speech impairment. Comput. Speech Lang. 27(6): 1178-1193 (2013) - [j25]Keiichi Tokuda, Yoshihiko Nankaku, Tomoki Toda, Heiga Zen, Junichi Yamagishi, Keiichiro Oura:
Speech Synthesis Based on Hidden Markov Models. Proc. IEEE 101(5): 1234-1252 (2013) - [j24]Zhen-Hua Ling, Korin Richmond, Junichi Yamagishi:
Articulatory Control of HMM-Based Parametric Speech Synthesis Using Feature-Space-Switched Multiple Regression. IEEE Trans. Speech Audio Process. 21(1): 205-217 (2013) - [c93]Cassia Valentini-Botinhao, Elizabeth Godoy, Yannis Stylianou, Bastian Sauert, Simon King, Junichi Yamagishi:
Improving intelligibility in noise of HMM-generated speech via noise-dependent and -independent methods. ICASSP 2013: 7854-7858 - [c92]Yoshitaka Mamiya, Junichi Yamagishi, Oliver Watts, Robert A. J. Clark, Simon King, Adriana Stan:
Lightly supervised GMM VAD to use audiobook for speech synthesiser. ICASSP 2013: 7987-7991 - [c91]Nicholas W. D. Evans, Tomi Kinnunen, Junichi Yamagishi:
Spoofing and countermeasures for automatic speaker verification. INTERSPEECH 2013: 925-929 - [c90]Korin Richmond, Zhen-Hua Ling, Junichi Yamagishi, Benigno Uria:
On the evaluation of inversion mapping performance in the acoustic domain. INTERSPEECH 2013: 1012-1016 - [c89]Adriana Stan, Peter Bell, Junichi Yamagishi, Simon King:
Lightly supervised discriminative training of grapheme models for improved sentence-level alignment of speech and text data. INTERSPEECH 2013: 1525-1529 - [c88]Maria Astrinaki, Junichi Yamagishi, Simon King, Nicolas D'Alessandro, Thierry Dutoit:
Reactive accent interpolation through an interactive map application. INTERSPEECH 2013: 1877-1878 - [c87]Adriana Stan, Oliver Watts, Yoshitaka Mamiya, Mircea Giurgiu, Robert A. J. Clark, Junichi Yamagishi, Simon King:
TUNDRA: a multilingual corpus of found data for TTS research created with light supervision. INTERSPEECH 2013: 2331-2335 - [c86]Cassia Valentini-Botinhao, Junichi Yamagishi, Simon King, Yannis Stylianou:
Combining perceptually-motivated spectral shaping with loudness and duration modification for intelligibility enhancement of HMM-based synthetic speech in noise. INTERSPEECH 2013: 3567-3571 - [c85]Christophe Veaux, Junichi Yamagishi, Simon King:
The voice bank corpus: Design, collection and data analysis of a large regional accent speech database. O-COCOSDA/CASLRE 2013: 1-4 - [c84]Christophe Veaux, Junichi Yamagishi, Simon King:
Towards Personalised Synthesised Voices for Individuals with Vocal Disabilities: Voice Banking and Reconstruction. SLPAT 2013: 107-111 - [c83]Yoshitaka Mamiya, Adriana Stan, Junichi Yamagishi, Peter Bell, Oliver Watts, Robert A. J. Clark, Simon King:
Using adaptation to improve speech transcription alignment in noisy and reverberant environments. SSW 2013: 41-46 - [c82]Oliver Watts, Adriana Stan, Robert A. J. Clark, Yoshitaka Mamiya, Mircea Giurgiu, Junichi Yamagishi, Simon King:
Unsupervised and lightly-supervised learning for rapid construction of TTS systems in multiple languages from 'found' data: evaluation and analysis. SSW 2013: 101-106 - [c81]Cassia Valentini-Botinhao, Mirjam Wester, Junichi Yamagishi, Simon King:
Using neighbourhood density and selective SNR boosting to increase the intelligibility of synthetic speech in noise. SSW 2013: 113-118 - [c80]Qiong Hu, Korin Richmond, Junichi Yamagishi, Javier Latorre:
An experimental comparison of multiple vocoder types. SSW 2013: 135-140 - [c79]Jaime Lorenzo-Trueba, Roberto Barra-Chicote, Junichi Yamagishi, Oliver Watts, Juan Manuel Montero:
Towards speaking style transplantation in speech synthesis. SSW 2013: 159-163 - [c78]Maria Astrinaki, Alexis Moinet, Junichi Yamagishi, Korin Richmond, Zhen-Hua Ling, Simon King, Thierry Dutoit:
Mage - reactive articulatory feature control of HMM-based parametric speech synthesis. SSW 2013: 207-211 - [c77]Maria Astrinaki, Alexis Moinet, Junichi Yamagishi, Korin Richmond, Zhen-Hua Ling, Simon King, Thierry Dutoit:
Mage - HMM-based speech synthesis reactively controlled by the articulators. SSW 2013: 243 - [c76]Maria Astrinaki, Junichi Yamagishi, Simon King, Nicolas D'Alessandro, Thierry Dutoit:
Reactive accent interpolation through an interactive map application. SSW 2013: 245 - [c75]Christophe Veaux, Maria Astrinaki, Keiichiro Oura, Robert A. J. Clark, Junichi Yamagishi:
Real-time control of expressive speech synthesis using kinect body tracking. SSW 2013: 247-248 - 2012
- [j23]J. Sebastian Andersson, Junichi Yamagishi, Robert A. J. Clark:
Synthesis and evaluation of conversational characteristics in HMM-based speech synthesis. Speech Commun. 54(2): 175-188 (2012) - [j22]Keiichiro Oura, Junichi Yamagishi, Mirjam Wester, Simon King, Keiichi Tokuda:
Analysis of unsupervised cross-lingual speaker adaptation for HMM-based speech synthesis using KLD-based transform mapping. Speech Commun. 54(6): 703-714 (2012) - [j21]Kei Hashimoto, Junichi Yamagishi, William Byrne, Simon King, Keiichi Tokuda:
Impacts of machine translation and speech synthesis on speech-to-speech translation. Speech Commun. 54(7): 857-866 (2012) - [j20]Phillip L. De Leon, Michael Pucher, Junichi Yamagishi, Inma Hernáez, Ibon Saratxaga:
Evaluation of Speaker Verification Security and Detection of HMM-Based Synthetic Speech. IEEE Trans. Speech Audio Process. 20(8): 2280-2290 (2012) - [c74]Cassia Valentini-Botinhao, Ranniery Maia, Junichi Yamagishi, Simon King, Heiga Zen:
Cepstral analysis based on the glimpse proportion measure for improving the intelligibility of HMM-based synthetic speech in noise. ICASSP 2012: 3997-4000 - [c73]Lakshmi Babu Saheer, Junichi Yamagishi, Philip N. Garner, John Dines:
Combining vocal tract length normalization with hierarchial linear transformations. ICASSP 2012: 4493-4496 - [c72]Cassia Valentini-Botinhao, Junichi Yamagishi, Simon King:
Evaluating speech intelligibility enhancement for HMM-based synthetic speech in noise. SAPA@INTERSPEECH 2012: 22-27 - [c71]Phillip L. De Leon, Bryan Stewart, Junichi Yamagishi:
Synthetic Speech Discrimination using Pitch Pattern Statistics Derived from Image Analysis. INTERSPEECH 2012: 370-373 - [c70]Cassia Valentini-Botinhao, Junichi Yamagishi, Simon King:
Mel cepstral coefficient modification based on the Glimpse Proportion measure for improving the intelligibility of HMM-generated synthetic speech in noise. INTERSPEECH 2012: 631-634 - [c69]Christophe Veaux, Junichi Yamagishi, Simon King:
Using HMM-based Speech Synthesis to Reconstruct the Voice of Individuals with Degenerative Speech Disorders. INTERSPEECH 2012: 967-970 - [c68]Zhen-Hua Ling, Korin Richmond, Junichi Yamagishi:
Vowel Creation by Articulatory Control in HMM-based Parametric Speech Synthesis. INTERSPEECH 2012: 991-994 - [c67]Rasmus Dall, Christophe Veaux, Junichi Yamagishi, Simon King:
Analysis of speaker clustering strategies for HMM-based speech synthesis. INTERSPEECH 2012: 995-998 - [c66]Jaime Lorenzo-Trueba, Roberto Barra-Chicote, Tuomo Raitio, Nicolas Obin, Paavo Alku, Junichi Yamagishi, Juan Manuel Montero:
Towards Glottal Source Controllability in Expressive Speech Synthesis. INTERSPEECH 2012: 1620-1623 - [c65]Jaime Lorenzo-Trueba, Beatriz Martínez-González, Roberto Barra-Chicote, Verónica López-Ludeña, Javier Ferreiros, Junichi Yamagishi, Juan Manuel Montero:
Towards an Unsupervised Speaking Style Voice Building Framework: Multi-Style Speaker Diarization. INTERSPEECH 2012: 2278-2281 - [c64]Chen-Yu Yang, Georgina Brown, Liang Lu, Junichi Yamagishi, Simon King:
Noise-robust whispered speech recognition using a non-audible-murmur microphone with VTS compensation. ISCSLP 2012: 220-223 - 2011
- [j19]Adriana Stan, Junichi Yamagishi, Simon King, Matthew P. Aylett:
The Romanian speech synthesis (RSS) corpus: Building a high quality HMM-based speech synthesis system using a high sampling rate. Speech Commun. 53(3): 442-450 (2011) - [j18]Tuomo Raitio, Antti Suni, Junichi Yamagishi, Hannu Pulakka, Jani Nurminen, Martti Vainio, Paavo Alku:
HMM-Based Speech Synthesis Utilizing Glottal Inverse Filtering. IEEE Trans. Speech Audio Process. 19(1): 153-165 (2011) - [c63]Christophe Veaux, Junichi Yamagishi, Simon King:
Voice banking and voice reconstruction for MND patients. ASSETS 2011: 305-306 - [c62]João P. Cabral, Steve Renals, Junichi Yamagishi, Korin Richmond:
HMM-based speech synthesiser using the LF-model of the glottal source. ICASSP 2011: 4704-4707 - [c61]Phillip L. De Leon, Inma Hernáez, Ibon Saratxaga, Michael Pucher, Junichi Yamagishi:
Detection of synthetic speech for the problem of imposture. ICASSP 2011: 4844-4847 - [c60]Kei Hashimoto, Junichi Yamagishi, William J. Byrne, Simon King, Keiichi Tokuda:
An analysis of machine translation and speech synthesis in speech-to-speech translation system. ICASSP 2011: 5108-5111 - [c59]Cassia Valentini-Botinhao, Junichi Yamagishi, Simon King:
Evaluation of objective measures for intelligibility prediction of HMM-based synthetic speech in noise. ICASSP 2011: 5112-5115 - [c58]Sandra Andraszewicz, Junichi Yamagishi, Simon King:
Vocal attractiveness of statistical speech synthesisers. ICASSP 2011: 5368-5371 - [c57]Zhen-Hua Ling, Korin Richmond, Junichi Yamagishi:
Feature-Space Transform Tying in Unified Acoustic-Articulatory Modelling for Articulatory Control of HMM-Based Speech Synthesis. INTERSPEECH 2011: 117-120 - [c56]Cassia Valentini-Botinhao, Junichi Yamagishi, Simon King:
Can Objective Measures Predict the Intelligibility of Modified HMM-Based Synthetic Speech in Noise? INTERSPEECH 2011: 1837-1840 - [c55]Oliver Watts, Junichi Yamagishi, Simon King:
Unsupervised Continuous-Valued Word Features for Phrase-Break Prediction without a Part-of-Speech Tagger. INTERSPEECH 2011: 2157-2160 - [c54]Ming Lei, Junichi Yamagishi, Korin Richmond, Zhen-Hua Ling, Simon King, Li-Rong Dai:
Formant-Controlled HMM-Based Speech Synthesis. INTERSPEECH 2011: 2777-2780 - 2010
- [j17]John Dines, Junichi Yamagishi, Simon King:
Measuring the Gap Between HMM-Based ASR and TTS. IEEE J. Sel. Top. Signal Process. 4(6): 1046-1058 (2010) - [j16]Michael Pucher, Dietmar Schabus, Junichi Yamagishi, Friedrich Neubarth, Volker Strom:
Modeling and interpolation of Austrian German and Viennese dialect in HMM-based speech synthesis. Speech Commun. 52(2): 164-179 (2010) - [j15]Roberto Barra-Chicote, Junichi Yamagishi, Simon King, Juan Manuel Montero, Javier Macías Guarasa:
Analysis of statistical parametric and unit selection speech synthesis systems applied to emotional speech. Speech Commun. 52(5): 394-404 (2010) - [j14]Zhen-Hua Ling, Korin Richmond, Junichi Yamagishi:
An Analysis of HMM-based prediction of articulatory movements. Speech Commun. 52(10): 834-846 (2010) - [j13]Junichi Yamagishi, Bela Usabaev, Simon King, Oliver Watts, John Dines, Jilei Tian, Yong Guan, Rile Hu, Keiichiro Oura, Yi-Jian Wu, Keiichi Tokuda, Reima Karhila, Mikko Kurimo:
Thousands of Voices for HMM-Based Speech Synthesis-Analysis and Application of TTS Systems Built on Various ASR Corpora. IEEE Trans. Speech Audio Process. 18(5): 984-1004 (2010) - [j12]Oliver Watts, Junichi Yamagishi, Simon King, Kay Berkling:
Synthesis of Child Speech With HMM Adaptation and Voice Conversion. IEEE Trans. Speech Audio Process. 18(5): 1005-1016 (2010) - [c53]Junichi Yamagishi, Oliver Watts:
The CSTR/EMIME HTS System for Blizzard Challenge. Blizzard Challenge 2010 - [c52]Phillip L. De Leon, Vijendra Raj Apsingekar, Michael Pucher, Junichi Yamagishi:
Revisiting the security of speaker verification systems against imposture using synthetic speech. ICASSP 2010: 1798-1801 - [c51]Keiichiro Oura, Keiichi Tokuda, Junichi Yamagishi, Simon King, Mirjam Wester:
Unsupervised cross-lingual speaker adaptation for HMM-based speech synthesis. ICASSP 2010: 4594-4597 - [c50]Junichi Yamagishi, Simon King:
Simple methods for improving speaker-similarity of HMM-based speech synthesis. ICASSP 2010: 4610-4613 - [c49]Junichi Yamagishi, Oliver Watts, Simon King, Bela Usabaev:
Roles of the average voice in speaker-adaptive HMM-based speech synthesis. INTERSPEECH 2010: 418-421 - [c48]Oliver Watts, Junichi Yamagishi, Simon King:
The role of higher-level linguistic features in HMM-based speech synthesis. INTERSPEECH 2010: 841-844 - [c47]Michael Pucher, Dietmar Schabus, Junichi Yamagishi:
Synthesis of fast speech with interpolation of adapted HSMMs and its evaluation by blind and sighted listeners. INTERSPEECH 2010: 2186-2189 - [c46]Zhen-Hua Ling, Korin Richmond, Junichi Yamagishi:
HMM-based text-to-articulatory-movement prediction and analysis of critical articulators. INTERSPEECH 2010: 2194-2197 - [c45]Phillip L. De Leon, Michael Pucher, Junichi Yamagishi:
Evaluation of the Vulnerability of Speaker Verification to Synthetic Speech. Odyssey 2010: 28 - [c44]J. Sebastian Andersson, Junichi Yamagishi, Robert A. J. Clark:
Utilising spontaneous conversational speech in HMM-based speech synthesis. SSW 2010: 173-178 - [c43]Mirjam Wester, John Dines, Matthew Gibson, Hui Liang, Yi-Jian Wu, Lakshmi Babu Saheer, Simon King, Keiichiro Oura, Philip N. Garner, William Byrne, Yong Guan, Teemu Hirsimäki, Reima Karhila, Mikko Kurimo, Matt Shannon, Sayaka Shiota, Jilei Tian, Keiichi Tokuda, Junichi Yamagishi:
Speaker adaptation and the evaluation of speaker similarity in the EMIME speech-to-speech translation project. SSW 2010: 192-197 - [c42]Yong Guan, Jilei Tian, Yi-Jian Wu, Junichi Yamagishi, Jani Nurminen:
An unified and automatic approach of Mandarin HTS system. SSW 2010: 236-239 - [c41]Oliver Watts, Junichi Yamagishi, Simon King:
Letter-based speech synthesis. SSW 2010: 317-322 - [c40]João P. Cabral, Steve Renals, Korin Richmond, Junichi Yamagishi:
An HMM-based speech synthesiser using glottal post-filtering. SSW 2010: 365-370
2000 – 2009
- 2009
- [j11]Junichi Yamagishi, Takao Kobayashi, Yuji Nakano, Katsumi Ogata, Juri Isogai:
Analysis of Speaker Adaptation Algorithms for HMM-Based Speech Synthesis and a Constrained SMAPLR Adaptation Algorithm. IEEE Trans. Speech Audio Process. 17(1): 66-83 (2009) - [j10]Zhen-Hua Ling, Korin Richmond, Junichi Yamagishi, Ren-Hua Wang:
Integrating Articulatory Features Into HMM-Based Parametric Speech Synthesis. IEEE Trans. Speech Audio Process. 17(6): 1171-1185 (2009) - [j9]Junichi Yamagishi, Takashi Nose, Heiga Zen, Zhen-Hua Ling, Tomoki Toda, Keiichi Tokuda, Simon King, Steve Renals:
Robust Speaker-Adaptive HMM-Based Text-to-Speech Synthesis. IEEE Trans. Speech Audio Process. 17(6): 1208-1230 (2009) - [c39]J. Sebastian Andersson, João P. Cabral, Leonardo Badino, Junichi Yamagishi, Robert A. J. Clark:
Glottal Source and Prosodic Prominence Modelling in HMM-based Speech Synthesis for the Blizzard Challenge 2009. Blizzard Challenge 2009 - [c38]Junichi Yamagishi, Mike Lincoln, Simon King, John Dines, Matthew Gibson, Jilei Tian, Yong Guan:
Analysis of Unsupervised and Noise-Robust Speaker-Adaptive HMM-Based Speech Synthesis Systems toward a Unified ASR and TTS Framework. Blizzard Challenge 2009 - [c37]Junichi Yamagishi, Bela Usabaev, Simon King, Oliver Watts, John Dines, Jilei Tian, Rile Hu, Yong Guan, Keiichiro Oura, Keiichi Tokuda, Reima Karhila, Mikko Kurimo:
Thousands of voices for HMM-based speech synthesis. INTERSPEECH 2009: 420-423 - [c36]Leonardo Badino, J. Sebastian Andersson, Junichi Yamagishi, Robert A. J. Clark:
Identification of contrast and its emphatic realization in HMM based speech synthesis. INTERSPEECH 2009: 520-523 - [c35]John Dines, Junichi Yamagishi, Simon King:
Measuring the gap between HMM-based ASR and TTS. INTERSPEECH 2009: 1391-1394 - [c34]Matthew P. Aylett, Simon King, Junichi Yamagishi:
Speech synthesis without a phone inventory. INTERSPEECH 2009: 2087-2090 - [c33]Oliver Watts, Junichi Yamagishi, Simon King, Kay Berkling:
HMM adaptation and voice conversion for the synthesis of child speech: a comparison. INTERSPEECH 2009: 2627-2630 - 2008
- [j8]Junichi Yamagishi, Hisashi Kawai, Takao Kobayashi:
Phone duration modeling using gradient tree boosting. Speech Commun. 50(5): 405-415 (2008) - [c32]Junichi Yamagishi, Heiga Zen, Yi-Jian Wu, Tomoki Toda, Keiichi Tokuda:
The HTS-2008 System: Yet Another Evaluation of the Speaker-Adaptive HMM-based Speech Synthesis System in The 2008 Blizzard Challenge. Blizzard Challenge 2008 - [c31]Junichi Yamagishi, Takashi Nose, Heiga Zen, Tomoki Toda, Keiichi Tokuda:
Performance evaluation of the speaker-independent HMM-based speech synthesis system "HTS 2007" for the Blizzard Challenge 2007. ICASSP 2008: 3957-3960 - [c30]Zhen-Hua Ling, Korin Richmond, Junichi Yamagishi, Ren-Hua Wang:
Articulatory control of HMM-based parametric speech synthesis driven by phonetic knowledge. INTERSPEECH 2008: 573-576 - [c29]Junichi Yamagishi, Zhen-Hua Ling, Simon King:
Robustness of HMM-based speech synthesis. INTERSPEECH 2008: 581-584 - [c28]João P. Cabral, Steve Renals, Korin Richmond, Junichi Yamagishi:
Glottal spectral separation for parametric speech synthesis. INTERSPEECH 2008: 1829-1832 - [c27]Simon King, Keiichi Tokuda, Heiga Zen, Junichi Yamagishi:
Unsupervised adaptation for HMM-based speech synthesis. INTERSPEECH 2008: 1869-1872 - [c26]Gregor Hofer, Junichi Yamagishi, Hiroshi Shimodaira:
Speech-driven lip motion generation with a trajectory HMM. INTERSPEECH 2008: 2314-2317 - [c25]Oliver Watts, Junichi Yamagishi, Kay Berkling, Simon King:
HMM-based synthesis of child speech. WOCCI 2008: 19 - 2007
- [j7]Junichi Yamagishi, Takao Kobayashi:
Average-Voice-Based Speech Synthesis Using HSMM-Based Speaker Adaptation and Adaptive Training. IEICE Trans. Inf. Syst. 90-D(2): 533-543 (2007) - [j6]Takashi Nose, Junichi Yamagishi, Takashi Masuko, Takao Kobayashi:
A Style Control Technique for HMM-Based Expressive Speech Synthesis. IEICE Trans. Inf. Syst. 90-D(9): 1406-1413 (2007) - [c24]Korin Richmond, Volker Strom, Robert A. J. Clark, Junichi Yamagishi, Susan Fitt:
Festival multisyn voices for the 2007 Blizzard Challenge. Blizzard Challenge 2007 - [c23]Junichi Yamagishi, Heiga Zen, Tomoki Toda, Keiichi Tokuda:
Speaker-independent HMM-based speech synthesis system - HTS-2007 system for the Blizzard Challenge 2007. Blizzard Challenge 2007 - [c22]Junichi Yamagishi, Takao Kobayashi, Makoto Tachibana, Katsumi Ogata, Yuji Nakano:
Model Adaptation Approach to Speech Synthesis with Diverse Voices and Styles. ICASSP (4) 2007: 1233-1236 - [c21]Makoto Tachibana, Keigo Kawashima, Junichi Yamagishi, Takao Kobayashi:
Performance evaluation of HMM-based style classification with a small amount of training data. INTERSPEECH 2007: 2261-2264 - [c20]Gregor Hofer, Hiroshi Shimodaira, Junichi Yamagishi:
Speech driven head motion synthesis based on a trajectory model. SIGGRAPH Posters 2007: 86 - [c19]Toshio Hirai, Junichi Yamagishi, Seiichi Tenpaku:
Utilization of an HMM-based feature generation module in 5 ms segment concatenative speech synthesis. SSW 2007: 81-84 - [c18]João P. Cabral, Steve Renals, Korin Richmond, Junichi Yamagishi:
Towards an improved modeling of the glottal source in statistical parametric speech synthesis. SSW 2007: 113-118 - [c17]Junichi Yamagishi, Takao Kobayashi, Steve Renals, Simon King, Heiga Zen, Tomoki Toda, Keiichi Tokuda:
Improved average-voice-based speech synthesis using gender-mixed modeling and a parameter generation algorithm considering GV. SSW 2007: 125-130 - [c16]Heiga Zen, Takashi Nose, Junichi Yamagishi, Shinji Sako, Takashi Masuko, Alan W. Black, Keiichi Tokuda:
The HMM-based speech synthesis system (HTS) version 2.0. SSW 2007: 294-299 - 2006
- [j5]Makoto Tachibana, Junichi Yamagishi, Takashi Masuko, Takao Kobayashi:
A Style Adaptation Technique for Speech Synthesis Using HSMM and Suprasegmental Features. IEICE Trans. Inf. Syst. 89-D(3): 1092-1099 (2006) - [c15]Tomoki Toda, Hisashi Kawai, Toshio Hirai, Jinfu Ni, Nobuyuki Nishizawa, Junichi Yamagishi, Minoru Tsuzaki, Keiichi Tokuda, Satoshi Nakamura:
Developing a Test Bed of English Text-to-Speech System XIMERA for the Blizzard Challenge 2006. Blizzard Challenge 2006 - [c14]Junichi Yamagishi, Katsumi Ogata, Yuji Nakano, Juri Isogai, Takao Kobayashi:
HSMM-Based Model Adaptation Algorithms for Average-Voice-Based Speech Synthesis. ICASSP (1) 2006: 77-80 - [c13]Yuji Nakano, Makoto Tachibana, Junichi Yamagishi, Takao Kobayashi:
Constrained structural maximum a posteriori linear regression for average-voice-based speech synthesis. INTERSPEECH 2006 - [c12]Takashi Nose, Junichi Yamagishi, Takao Kobayashi:
A style control technique for speech synthesis using multiple regression HSMM. INTERSPEECH 2006 - [c11]Katsumi Ogata, Makoto Tachibana, Junichi Yamagishi, Takao Kobayashi:
Acoustic model training based on linear transformation and MAP modification for HSMM-based speech synthesis. INTERSPEECH 2006 - [c10]Makoto Tachibana, Takashi Nose, Junichi Yamagishi, Takao Kobayashi:
A technique for controlling voice quality of synthetic speech using multiple regression HSMM. INTERSPEECH 2006 - 2005
- [j4]Junichi Yamagishi, Koji Onishi, Takashi Masuko, Takao Kobayashi:
Acoustic Modeling of Speaking Styles and Emotional Expressions in HMM-Based Speech Synthesis. IEICE Trans. Inf. Syst. 88-D(3): 502-509 (2005) - [j3]Makoto Tachibana, Junichi Yamagishi, Takashi Masuko, Takao Kobayashi:
Speech Synthesis with Various Emotional Expressions and Speaking Styles by Style Interpolation and Morphing. IEICE Trans. Inf. Syst. 88-D(11): 2484-2491 (2005) - [j2]Naotake Niwase, Junichi Yamagishi, Takao Kobayashi:
Human Walking Motion Synthesis with Desired Pace and Stride Length Based on HSMM. IEICE Trans. Inf. Syst. 88-D(11): 2492-2499 (2005) - [c9]Takashi Yamazaki, Naotake Niwase, Junichi Yamagishi, Takao Kobayashi:
HumanWalking Motion Synthesis Based on Multiple Regression Hidden Semi-Markov Model. CW 2005: 445-452 - [c8]Junichi Yamagishi, Takao Kobayashi:
Adaptive Training for Hidden Semi-Markov Model. ICASSP (1) 2005: 365-368 - [c7]Juri Isogai, Junichi Yamagishi, Takao Kobayashi:
Model adaptation and adaptive training using ESAT algorithm for HMM-based speech synthesis. INTERSPEECH 2005: 2597-2600 - [c6]Makoto Tachibana, Junichi Yamagishi, Takashi Masuko, Takao Kobayashi:
Performance evaluation of style adaptation for hidden semi-Markov model based speech synthesis. INTERSPEECH 2005: 2805-2808 - 2004
- [c5]Junichi Yamagishi, Makoto Tachibana, Takashi Masuko, Takao Kobayashi:
Speaking style adaptation using context clustering decision tree for HMM-based speech synthesis. ICASSP (1) 2004: 5-8 - [c4]Junichi Yamagishi, Takashi Masuko, Takao Kobayashi:
MLLR adaptation for hidden semi-Markov model based speech synthesis. INTERSPEECH 2004: 1213-1216 - 2003
- [j1]Junichi Yamagishi, Masatsune Tamura, Takashi Masuko, Keiichi Tokuda, Takao Kobayashi:
A Training Method of Average Voice Model for HMM-Based Speech Synthesis. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 86-A(8): 1956-1963 (2003) - [c3]Junichi Yamagishi, Takashi Masuko, Keiichi Tokuda, Takao Kobayashi:
A training method for average voice model based on shared decision tree context clustering and speaker adaptive training. ICASSP (1) 2003: 716-719 - [c2]Junichi Yamagishi, Koji Onishi, Takashi Masuko, Takao Kobayashi:
Modeling of various speaking styles and emotions for HMM-based speech synthesis. INTERSPEECH 2003: 2461-2464 - 2002
- [c1]Junichi Yamagishi, Masatsune Tamura, Takashi Masuko, Keiichi Tokuda, Takao Kobayashi:
A context clustering technique for average voice model in HMM-based speech synthesis. INTERSPEECH 2002: 133-136
Coauthor Index
aka: Tomi H. Kinnunen
aka: Kong Aik Lee
aka: Zhenhua Ling
aka: Juan Manuel Montero
aka: Huy Hong Nguyen
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-12-12 21:59 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint