default search action
Shinnosuke Takamichi
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [j24]Detai Xin, Junfeng Jiang, Shinnosuke Takamichi, Yuki Saito, Akiko Aizawa, Hiroshi Saruwatari:
JVNV: A Corpus of Japanese Emotional Speech With Verbal Content and Nonverbal Expressions. IEEE Access 12: 19752-19764 (2024) - [j23]Detai Xin, Shinnosuke Takamichi, Hiroshi Saruwatari:
JNV corpus: A corpus of Japanese nonverbal vocalizations with diverse phrases and emotions. Speech Commun. 156: 103004 (2024) - [j22]Takaaki Saeki, Soumi Maiti, Xinjian Li, Shinji Watanabe, Shinnosuke Takamichi, Hiroshi Saruwatari:
Text-Inductive Graphone-Based Language Adaptation for Low-Resource Speech Synthesis. IEEE ACM Trans. Audio Speech Lang. Process. 32: 1829-1844 (2024) - [c86]Yuki Okamoto, Keisuke Imoto, Shinnosuke Takamichi, Ryotaro Nagase, Takahiro Fukumori, Yoichi Yamashita:
Environmental Sound Synthesis from Vocal Imitations and Sound Event Labels. ICASSP 2024: 411-415 - [c85]Kentaro Seki, Shinnosuke Takamichi, Takaaki Saeki, Hiroshi Saruwatari:
Diversity-Based Core-Set Selection for Text-to-Speech with Linguistic and Acoustic Features. ICASSP 2024: 12351-12355 - [c84]Shinnosuke Takamichi, Hiroki Maeda, Joonyong Park, Daisuke Saito, Hiroshi Saruwatari:
Do Learned Speech Symbols Follow Zipf's Law? ICASSP 2024: 12526-12530 - [i68]Takaaki Saeki, Soumi Maiti, Shinnosuke Takamichi, Shinji Watanabe, Hiroshi Saruwatari:
SpeechBERTScore: Reference-Aware Automatic Evaluation of Speech Generation Leveraging NLP Evaluation Metrics. CoRR abs/2401.16812 (2024) - [i67]Aya Watanabe, Shinnosuke Takamichi, Yuki Saito, Wataru Nakata, Detai Xin, Hiroshi Saruwatari:
Building speech corpus with diverse voice characteristics for its prompt-based representation. CoRR abs/2403.13353 (2024) - [i66]Detai Xin, Xu Tan, Kai Shen, Zeqian Ju, Dongchao Yang, Yuancheng Wang, Shinnosuke Takamichi, Hiroshi Saruwatari, Shujie Liu, Jinyu Li, Sheng Zhao:
RALL-E: Robust Codec Language Modeling with Chain-of-Thought Prompting for Text-to-Speech Synthesis. CoRR abs/2404.03204 (2024) - [i65]Xinjian Li, Shinnosuke Takamichi, Takaaki Saeki, William Chen, Sayaka Shiota, Shinji Watanabe:
YODAS: Youtube-Oriented Dataset for Audio and Speech. CoRR abs/2406.00899 (2024) - [i64]Yuki Saito, Takuto Igarashi, Kentaro Seki, Shinnosuke Takamichi, Ryuichi Yamamoto, Kentaro Tachibana, Hiroshi Saruwatari:
SRC4VC: Smartphone-Recorded Corpus for Voice Conversion Benchmark. CoRR abs/2406.07254 (2024) - [i63]Takuto Igarashi, Yuki Saito, Kentaro Seki, Shinnosuke Takamichi, Ryuichi Yamamoto, Kentaro Tachibana, Hiroshi Saruwatari:
Noise-Robust Voice Conversion by Conditional Denoising Training Using Latent Variables of Recording Quality and Environment. CoRR abs/2406.07280 (2024) - [i62]Kentaro Seki, Shinnosuke Takamichi, Norihiro Takamune, Yuki Saito, Kanami Imamura, Hiroshi Saruwatari:
Spatial Voice Conversion: Voice Conversion Preserving Spatial Information and Non-target Signals. CoRR abs/2406.17722 (2024) - [i61]Hitoshi Suda, Aya Watanabe, Shinnosuke Takamichi:
Who Finds This Voice Attractive? A Large-Scale Experiment Using In-the-Wild Data. CoRR abs/2407.04270 (2024) - [i60]Shunsuke Kando, Yusuke Miyao, Jason Naradowsky, Shinnosuke Takamichi:
Textless Dependency Parsing by Labeled Sequence Prediction. CoRR abs/2407.10118 (2024) - [i59]Wataru Nakata, Kentaro Seki, Hitomi Yanaka, Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
J-CHAT: Japanese Large-scale Spoken Dialogue Corpus for Spoken Dialogue Language Modeling. CoRR abs/2407.15828 (2024) - [i58]Detai Xin, Xu Tan, Shinnosuke Takamichi, Hiroshi Saruwatari:
BigCodec: Pushing the Limits of Low-Bitrate Neural Speech Codec. CoRR abs/2409.05377 (2024) - [i57]Jee-weon Jung, Wangyou Zhang, Soumi Maiti, Yihan Wu, Xin Wang, Ji-Hoon Kim, Yuta Matsunaga, Seyun Um, Jinchuan Tian, Hye-jin Shim, Nicholas W. D. Evans, Joon Son Chung, Shinnosuke Takamichi, Shinji Watanabe:
Text-To-Speech Synthesis In The Wild. CoRR abs/2409.08711 (2024) - [i56]Hiroaki Hyodo, Shinnosuke Takamichi, Tomohiko Nakamura, Junya Koguchi, Hiroshi Saruwatari:
DNN-based ensemble singing voice synthesis with interactions between singers. CoRR abs/2409.09988 (2024) - [i55]Jee-weon Jung, Yihan Wu, Xin Wang, Ji-Hoon Kim, Soumi Maiti, Yuta Matsunaga, Hye-jin Shim, Jinchuan Tian, Nicholas W. D. Evans, Joon Son Chung, Wangyou Zhang, Seyun Um, Shinnosuke Takamichi, Shinji Watanabe:
SpoofCeleb: Speech Deepfake Detection and SASV In The Wild. CoRR abs/2409.17285 (2024) - [i54]Bin Wu, Sakriani Sakti, Shinnosuke Takamichi, Satoshi Nakamura:
A Neural Transformer Framework for Simultaneous Tasks of Segmentation, Classification, and Caller Identification of Marmoset Vocalization. CoRR abs/2410.23279 (2024) - 2023
- [j21]Takaaki Saeki, Shinnosuke Takamichi, Tomohiko Nakamura, Naoko Tanji, Hiroshi Saruwatari:
SelfRemaster: Self-Supervised Speech Restoration for Historical Audio Resources. IEEE Access 11: 144831-144843 (2023) - [c83]Xinjian Li, Shinnosuke Takamichi, Takaaki Saeki, William Chen, Sayaka Shiota, Shinji Watanabe:
Yodas: Youtube-Oriented Dataset for Audio and Speech. ASRU 2023: 1-8 - [c82]Aya Watanabe, Shinnosuke Takamichi, Yuki Saito, Wataru Nakata, Detai Xin, Hiroshi Saruwatari:
COCO-NUT: Corpus of Japanese Utterance and Voice Characteristics Description for Prompt-Based Control. ASRU 2023: 1-8 - [c81]Tomohiko Nakamura, Shinnosuke Takamichi, Naoko Tanji, Satoru Fukayama, Hiroshi Saruwatari:
jaCappella Corpus: A Japanese a Cappella Vocal Ensemble Corpus. ICASSP 2023: 1-5 - [c80]Hien Ohnaka, Shinnosuke Takamichi, Keisuke Imoto, Yuki Okamoto, Kazuki Fujii, Hiroshi Saruwatari:
Visual Onoma-to-Wave: Environmental Sound Synthesis from Visual Onomatopoeias and Sound-Source Images. ICASSP 2023: 1-5 - [c79]Aya Watanabe, Shinnosuke Takamichi, Yuki Saito, Detai Xin, Hiroshi Saruwatari:
MID-Attribute Speaker Generation Using Optimal-Transport-Based Interpolation of Gaussian Mixture Models. ICASSP 2023: 1-5 - [c78]Detai Xin, Sharath Adavanne, Federico Ang, Ashish Kulkarni, Shinnosuke Takamichi, Hiroshi Saruwatari:
Improving Speech Prosody of Audiobook Text-To-Speech Synthesis with Acoustic and Textual Contexts. ICASSP 2023: 1-5 - [c77]Takaaki Saeki, Soumi Maiti, Xinjian Li, Shinji Watanabe, Shinnosuke Takamichi, Hiroshi Saruwatari:
Learning to Speak from Text: Zero-Shot Multilingual Text-to-Speech with Unsupervised Text Pretraining. IJCAI 2023: 5179-5187 - [c76]Detai Xin, Shinnosuke Takamichi, Ai Morimatsu, Hiroshi Saruwatari:
Laughter Synthesis using Pseudo Phonetic Tokens with a Large-scale In-the-wild Laughter Corpus. INTERSPEECH 2023: 17-21 - [c75]Joonyong Park, Shinnosuke Takamichi, Tomohiko Nakamura, Kentaro Seki, Detai Xin, Hiroshi Saruwatari:
How Generative Spoken Language Modeling Encodes Noisy Speech: Investigation from Phonetics to Syntactics. INTERSPEECH 2023: 1085-1089 - [c74]Yuki Saito, Shinnosuke Takamichi, Eiji Iimori, Kentaro Tachibana, Hiroshi Saruwatari:
ChatGPT-EDSS: Empathetic Dialogue Speech Synthesis Trained from ChatGPT-derived Context Word Embeddings. INTERSPEECH 2023: 3048-3052 - [c73]Yota Ueda, Shinnosuke Takamichi, Yuki Saito, Norihiro Takamune, Hiroshi Saruwatari:
HumanDiffusion: diffusion model using perceptual gradients. INTERSPEECH 2023: 4264-4268 - [c72]Yuki Saito, Eiji Iimori, Shinnosuke Takamichi, Kentaro Tachibana, Hiroshi Saruwatari:
CALLS: Japanese Empathetic Dialogue Speech Corpus of Complaint Handling and Attentive Listening in Customer Center. INTERSPEECH 2023: 5561-5565 - [c71]Kota Arai, Yutaro Hirao, Takuji Narumi, Tomohiko Nakamura, Shinnosuke Takamichi, Shigeo Yoshida:
TimToShape: Supporting Practice of Musical Instruments by Visualizing Timbre with 2D Shapes based on Crossmodal Correspondences. IUI 2023: 850-865 - [c70]Yuta Matsunaga, Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari:
Improving robustness of spontaneous speech synthesis with linguistic speech regularization and pseudo-filled-pause insertion. SSW 2023: 62-68 - [d1]Detai Xin, Junfeng Jiang, Shinnosuke Takamichi, Yuki Saito, Akiko Aizawa, Hiroshi Saruwatari:
JVNV: A Corpus of Japanese Emotional Speech with Verbal Content and Nonverbal Expressions. IEEE DataPort, 2023 - [i53]Takaaki Saeki, Soumi Maiti, Xinjian Li, Shinji Watanabe, Shinnosuke Takamichi, Hiroshi Saruwatari:
Learning to Speak from Text: Zero-Shot Multilingual Text-to-Speech with Unsupervised Text Pretraining. CoRR abs/2301.12596 (2023) - [i52]Keunwoo Choi, Jaekwon Im, Laurie M. Heller, Brian McFee, Keisuke Imoto, Yuki Okamoto, Mathieu Lagrange, Shinnosuke Takamichi:
Foley Sound Synthesis at the DCASE 2023 Challenge. CoRR abs/2304.12521 (2023) - [i51]Yuki Okamoto, Keisuke Imoto, Shinnosuke Takamichi, Ryotaro Nagase, Takahiro Fukumori, Yoichi Yamashita:
Environmental sound conversion from vocal imitations and sound event labels. CoRR abs/2305.00302 (2023) - [i50]Detai Xin, Shinnosuke Takamichi, Ai Morimatsu, Hiroshi Saruwatari:
Laughter Synthesis using Pseudo Phonetic Tokens with a Large-scale In-the-wild Laughter Corpus. CoRR abs/2305.12442 (2023) - [i49]Detai Xin, Shinnosuke Takamichi, Hiroshi Saruwatari:
JNV Corpus: A Corpus of Japanese Nonverbal Vocalizations with Diverse Phrases and Emotions. CoRR abs/2305.12445 (2023) - [i48]Yuki Saito, Eiji Iimori, Shinnosuke Takamichi, Kentaro Tachibana, Hiroshi Saruwatari:
CALLS: Japanese Empathetic Dialogue Speech Corpus of Complaint Handling and Attentive Listening in Customer Center. CoRR abs/2305.13713 (2023) - [i47]Yuki Saito, Shinnosuke Takamichi, Eiji Iimori, Kentaro Tachibana, Hiroshi Saruwatari:
ChatGPT-EDSS: Empathetic Dialogue Speech Synthesis Trained from ChatGPT-derived Context Word Embeddings. CoRR abs/2305.13724 (2023) - [i46]Joonyong Park, Shinnosuke Takamichi, Tomohiko Nakamura, Kentaro Seki, Detai Xin, Hiroshi Saruwatari:
How Generative Spoken Language Modeling Encodes Noisy Speech: Investigation from Phonetics to Syntactics. CoRR abs/2306.00697 (2023) - [i45]Yota Ueda, Shinnosuke Takamichi, Yuki Saito, Norihiro Takamune, Hiroshi Saruwatari:
HumanDiffusion: diffusion model using perceptual gradients. CoRR abs/2306.12169 (2023) - [i44]Kentaro Seki, Shinnosuke Takamichi, Takaaki Saeki, Hiroshi Saruwatari:
Diversity-based core-set selection for text-to-speech with linguistic and acoustic features. CoRR abs/2309.08127 (2023) - [i43]Shinnosuke Takamichi, Hiroki Maeda, Joonyong Park, Daisuke Saito, Hiroshi Saruwatari:
Do learned speech symbols follow Zipf's law? CoRR abs/2309.09690 (2023) - [i42]Aya Watanabe, Shinnosuke Takamichi, Yuki Saito, Wataru Nakata, Detai Xin, Hiroshi Saruwatari:
Coco-Nut: Corpus of Japanese Utterance and Voice Characteristics Description for Prompt-based Control. CoRR abs/2309.13509 (2023) - [i41]Detai Xin, Junfeng Jiang, Shinnosuke Takamichi, Yuki Saito, Akiko Aizawa, Hiroshi Saruwatari:
JVNV: A Corpus of Japanese Emotional Speech with Verbal Content and Nonverbal Expressions. CoRR abs/2310.06072 (2023) - 2022
- [j20]Hiroto Kai, Shinnosuke Takamichi, Sayaka Shiota, Hitoshi Kiya:
Lightweight and irreversible speech pseudonymization based on data-driven optimization of cascaded voice modification modules. Comput. Speech Lang. 72: 101315 (2022) - [c69]Shinnosuke Takamichi, Wataru Nakata, Naoko Tanji, Hiroshi Saruwatari:
J-MAC: Japanese multi-speaker audiobook corpus for speech synthesis. INTERSPEECH 2022: 2358-2362 - [c68]Yuto Nishimura, Yuki Saito, Shinnosuke Takamichi, Kentaro Tachibana, Hiroshi Saruwatari:
Acoustic Modeling for End-to-End Empathetic Dialogue Speech Synthesis Using Linguistic and Prosodic Contexts of Dialogue History. INTERSPEECH 2022: 3373-3377 - [c67]Takaaki Saeki, Shinnosuke Takamichi, Tomohiko Nakamura, Naoko Tanji, Hiroshi Saruwatari:
SelfRemaster: Self-Supervised Speech Restoration with Analysis-by-Synthesis Approach Using Channel Modeling. INTERSPEECH 2022: 4406-4410 - [c66]Takaaki Saeki, Detai Xin, Wataru Nakata, Tomoki Koriyama, Shinnosuke Takamichi, Hiroshi Saruwatari:
UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022. INTERSPEECH 2022: 4521-4525 - [c65]Wataru Nakata, Tomoki Koriyama, Shinnosuke Takamichi, Yuki Saito, Yusuke Ijima, Ryo Masumura, Hiroshi Saruwatari:
Predicting VQVAE-based Character Acting Style from Quotation-Annotated Text for Audiobook Speech Synthesis. INTERSPEECH 2022: 4551-4555 - [c64]Yuki Saito, Yuto Nishimura, Shinnosuke Takamichi, Kentaro Tachibana, Hiroshi Saruwatari:
STUDIES: Corpus of Japanese Empathetic Dialogue Speech Towards Friendly Voice Agent. INTERSPEECH 2022: 5155-5159 - [c63]Yuta Matsunaga, Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari:
Personalized Filled-pause Generation with Group-wise Prediction Models. LREC 2022: 385-392 - [c62]Hiroto Kai, Shinnosuke Takamichi, Sayaka Shiota, Hitoshi Kiya:
Robustness of Signal Processing-Based Pseudonymization Method Against Decryption Attack. Odyssey 2022: 287-293 - [c61]Yoshifumi Nakano, Takaaki Saeki, Shinnosuke Takamichi, Katsuhito Sudoh, Hiroshi Saruwatari:
VTTS: Visual-Text To Speech. SLT 2022: 936-942 - [i40]Shinnosuke Takamichi, Wataru Nakata, Naoko Tanji, Hiroshi Saruwatari:
J-MAC: Japanese multi-speaker audiobook corpus for speech synthesis. CoRR abs/2201.10896 (2022) - [i39]Yuta Matsunaga, Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari:
Personalized filled-pause generation with group-wise prediction models. CoRR abs/2203.09961 (2022) - [i38]Takaaki Saeki, Shinnosuke Takamichi, Tomohiko Nakamura, Naoko Tanji, Hiroshi Saruwatari:
SelfRemaster: Self-Supervised Speech Restoration with Analysis-by-Synthesis Approach Using Channel Modeling. CoRR abs/2203.12937 (2022) - [i37]Yoshifumi Nakano, Takaaki Saeki, Shinnosuke Takamichi, Katsuhito Sudoh, Hiroshi Saruwatari:
vTTS: visual-text to speech. CoRR abs/2203.14725 (2022) - [i36]Yuki Saito, Yuto Nishimura, Shinnosuke Takamichi, Kentaro Tachibana, Hiroshi Saruwatari:
STUDIES: Corpus of Japanese Empathetic Dialogue Speech Towards Friendly Voice Agent. CoRR abs/2203.14757 (2022) - [i35]Takaaki Saeki, Detai Xin, Wataru Nakata, Tomoki Koriyama, Shinnosuke Takamichi, Hiroshi Saruwatari:
UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022. CoRR abs/2204.02152 (2022) - [i34]Detai Xin, Shinnosuke Takamichi, Takuma Okamoto, Hisashi Kawai, Hiroshi Saruwatari:
Speaking-Rate-Controllable HiFi-GAN Using Feature Interpolation. CoRR abs/2204.10561 (2022) - [i33]Yuto Nishimura, Yuki Saito, Shinnosuke Takamichi, Kentaro Tachibana, Hiroshi Saruwatari:
Acoustic Modeling for End-to-End Empathetic Dialogue Speech Synthesis Using Linguistic and Prosodic Contexts of Dialogue History. CoRR abs/2206.08039 (2022) - [i32]Detai Xin, Shinnosuke Takamichi, Hiroshi Saruwatari:
Exploring the Effectiveness of Self-supervised Learning and Classifier Chains in Emotion Recognition of Nonverbal Vocalizations. CoRR abs/2206.10695 (2022) - [i31]Yuki Okamoto, Keisuke Imoto, Shinnosuke Takamichi, Takahiro Fukumori, Yoichi Yamashita:
How Should We Evaluate Synthesized Environmental Sounds. CoRR abs/2208.07679 (2022) - [i30]Yuta Matsunaga, Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari:
Empirical Study Incorporating Linguistic Knowledge on Filled Pauses for Personalized Spontaneous Speech Synthesis. CoRR abs/2210.07559 (2022) - [i29]Hien Ohnaka, Shinnosuke Takamichi, Keisuke Imoto, Yuki Okamoto, Kazuki Fujii, Hiroshi Saruwatari:
Visual onoma-to-wave: environmental sound synthesis from visual onomatopoeias and sound-source images. CoRR abs/2210.09173 (2022) - [i28]Yuta Matsunaga, Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari:
Spontaneous speech synthesis with linguistic-speech consistency training using pseudo-filled pauses. CoRR abs/2210.09815 (2022) - [i27]Aya Watanabe, Shinnosuke Takamichi, Yuki Saito, Detai Xin, Hiroshi Saruwatari:
Mid-attribute speaker generation using optimal-transport-based interpolation of Gaussian mixture models. CoRR abs/2210.09916 (2022) - [i26]Kentaro Seki, Shinnosuke Takamichi, Takaaki Saeki, Hiroshi Saruwatari:
Text-to-speech synthesis from dark data with evaluation-in-the-loop data selection. CoRR abs/2210.14850 (2022) - [i25]Detai Xin, Sharath Adavanne, Federico Ang, Ashish Kulkarni, Shinnosuke Takamichi, Hiroshi Saruwatari:
Improving Speech Prosody of Audiobook Text-to-Speech Synthesis with Acoustic and Textual Contexts. CoRR abs/2211.02336 (2022) - [i24]Tomohiko Nakamura, Shinnosuke Takamichi, Naoko Tanji, Satoru Fukayama, Hiroshi Saruwatari:
jaCappella Corpus: A Japanese a Cappella Vocal Ensemble Corpus. CoRR abs/2211.16028 (2022) - 2021
- [j19]Akihito Aiba, Minoru Yoshida, Daichi Kitamura, Shinnosuke Takamichi, Hiroshi Saruwatari:
Noise Robust Acoustic Anomaly Detection System with Nonnegative Matrix Factorization Based on Generalized Gaussian Distribution. IEICE Trans. Inf. Syst. 104-D(3): 441-449 (2021) - [j18]Takaaki Saeki, Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Real-Time Full-Band Voice Conversion with Sub-Band Modeling and Data-Driven Phase Estimation of Spectral Differentials. IEICE Trans. Inf. Syst. 104-D(7): 1002-1016 (2021) - [j17]Satoshi Mizoguchi, Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
DNN-Based Low-Musical-Noise Single-Channel Speech Enhancement Based on Higher-Order-Moments Matching. IEICE Trans. Inf. Syst. 104-D(11): 1971-1980 (2021) - [j16]Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari:
Incremental Text-to-Speech Synthesis Using Pseudo Lookahead With Large Pretrained Language Model. IEEE Signal Process. Lett. 28: 857-861 (2021) - [j15]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Perceptual-Similarity-Aware Deep Speaker Representation Learning for Multi-Speaker Generative Modeling. IEEE ACM Trans. Audio Speech Lang. Process. 29: 1033-1048 (2021) - [c60]Xuan Luo, Shinnosuke Takamichi, Tomoki Koriyama, Yuki Saito, Hiroshi Saruwatari:
Emotion-Controllable Speech Synthesis Using Emotion Soft Labels and Fine-Grained Prosody Factors. APSIPA ASC 2021: 794-799 - [c59]Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari:
Low-Latency Incremental Text-to-Speech Synthesis with Distilled Context Prediction Network. ASRU 2021: 749-756 - [c58]Yasushige Ishikawa, Shinnosuke Takamichi, Takatoyo Umemoto, Masao Aikawa, Kishio Sakamoto, Kikuko Yui, Shigeo Fujiwara, Ayako Suto, Koichi Nishiyama:
Japanese EFL Learners' Speaking Practice Utilizing Text-to-Speech Technology Within a Team-Based Flipped Learning Framework. HCI (23) 2021: 283-291 - [c57]Yota Ueda, Kazuki Fujii, Yuki Saito, Shinnosuke Takamichi, Yukino Baba, Hiroshi Saruwatari:
Humanacgan: Conditional Generative Adversarial Network with Human-Based Auxiliary Classifier and its Evaluation in Phoneme Perception. ICASSP 2021: 6468-6472 - [c56]Detai Xin, Tatsuya Komatsu, Shinnosuke Takamichi, Hiroshi Saruwatari:
Disentangled Speaker and Language Representations Using Mutual Information Minimization and Domain Adaptation for Cross-Lingual TTS. ICASSP 2021: 6608-6612 - [c55]Riku Arakawa, Zendai Kashino, Shinnosuke Takamichi, Adrien Verhulst, Masahiko Inami:
Digital Speech Makeup: Voice Conversion Based Altered Auditory Feedback for Transforming Self-Representation. ICMI 2021: 159-167 - [c54]Detai Xin, Yuki Saito, Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari:
Cross-Lingual Speaker Adaptation Using Domain Adaptation and Speaker Consistency Loss for Text-To-Speech Synthesis. Interspeech 2021: 1614-1618 - [c53]Hiroto Kai, Shinnosuke Takamichi, Sayaka Shiota, Hitoshi Kiya:
Lightweight Voice Anonymization Based on Data-Driven Optimization of Cascaded Voice Modification Modules. SLT 2021: 560-566 - [c52]Kazuya Yufune, Tomoki Koriyama, Shinnosuke Takamichi, Hiroshi Saruwatari:
Accent Modeling of Low-Resourced Dialect in Pitch Accent Language Using Variational Autoencoder. SSW 2021: 189-194 - [c51]Wataru Nakata, Tomoki Koriyama, Shinnosuke Takamichi, Naoko Tanji, Yusuke Ijima, Ryo Masumura, Hiroshi Saruwatari:
Audiobook Speech Synthesis Conditioned by Cross-Sentence Context-Aware Word Embeddings. SSW 2021: 211-215 - [i23]Yota Ueda, Kazuki Fujii, Yuki Saito, Shinnosuke Takamichi, Yukino Baba, Hiroshi Saruwatari:
HumanACGAN: conditional generative adversarial network with human-based auxiliary classifier and its evaluation in phoneme perception. CoRR abs/2102.04051 (2021) - [i22]Yuki Okamoto, Keisuke Imoto, Shinnosuke Takamichi, Ryosuke Yamanishi, Takahiro Fukumori, Yoichi Yamashita:
Onoma-to-wave: Environmental sound synthesis from onomatopoeic words. CoRR abs/2102.05872 (2021) - [i21]Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari:
Low-Latency Incremental Text-to-Speech Synthesis with Distilled Context Prediction Network. CoRR abs/2109.10724 (2021) - [i20]Tomoki Hayashi, Ryuichi Yamamoto, Takenori Yoshimura, Peter Wu, Jiatong Shi, Takaaki Saeki, Yooncheol Ju, Yusuke Yasuda, Shinnosuke Takamichi, Shinji Watanabe:
ESPnet2-TTS: Extending the Edge of TTS Research. CoRR abs/2110.07840 (2021) - [i19]Shinnosuke Takamichi, Ludwig Kürzinger, Takaaki Saeki, Sayaka Shiota, Shinji Watanabe:
JTubeSpeech: corpus of Japanese speech collected from YouTube for speech recognition and speaker verification. CoRR abs/2112.09323 (2021) - 2020
- [j14]Hiroki Tamaru, Yuki Saito, Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari:
Generative Moment Matching Network-Based Neural Double-Tracking for Synthesized and Natural Singing Voices. IEICE Trans. Inf. Syst. 103-D(3): 639-647 (2020) - [j13]Junya Koguchi, Shinnosuke Takamichi, Masanori Morise, Hiroshi Saruwatari, Shigeki Sagayama:
DNN-Based Full-Band Speech Synthesis Using GMM Approximation of Spectral Envelope. IEICE Trans. Inf. Syst. 103-D(12): 2673-2681 (2020) - [j12]Shinnosuke Takamichi, Yuki Saito, Norihiro Takamune, Daichi Kitamura, Hiroshi Saruwatari:
Phase reconstruction from amplitude spectrograms based on directional-statistics deep neural networks. Signal Process. 169: 107368 (2020) - [j11]Masashi Aso, Shinnosuke Takamichi, Norihiro Takamune, Hiroshi Saruwatari:
Acoustic model-based subword tokenization and prosodic-context extraction without language knowledge for text-to-speech synthesis. Speech Commun. 125: 53-60 (2020) - [c50]Junya Koguchi, Shinnosuke Takamichi, Masanori Morise:
PJS: phoneme-balanced Japanese singing-voice corpus. APSIPA 2020: 487-491 - [c49]Yuki Okamoto, Keisuke Imoto, Shinnosuke Takamichi, Ryosuke Yamanishi, Takahiro Fukumori, Yoichi Yamashita:
RWCP-SSD-Onomatopoeia: Onomatopoeic Word Dataset for Environmental Sound Synthesis. DCASE 2020: 125-129 - [c48]Kazuki Fujii, Yuki Saito, Shinnosuke Takamichi, Yukino Baba, Hiroshi Saruwatari:
Humangan: Generative Adversarial Network With Human-Based Discriminator And Its Evaluation In Speech Perception Modeling. ICASSP 2020: 6239-6243 - [c47]Takaaki Saeki, Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Lifter Training and Sub-Band Modeling for Computationally Efficient and High-Quality Voice Conversion Using Spectral Differentials. ICASSP 2020: 7784-7788 - [c46]Takaaki Saeki, Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Real-Time, Full-Band, Online DNN-Based Voice Conversion System Using a Single CPU. INTERSPEECH 2020: 1021-1022 - [c45]Detai Xin, Yuki Saito, Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari:
Cross-Lingual Text-To-Speech Synthesis via Domain Adaptation and Perceptual Similarity Regression in Speaker Space. INTERSPEECH 2020: 2947-2951 - [c44]Yuki Yamashita, Tomoki Koriyama, Yuki Saito, Shinnosuke Takamichi, Yusuke Ijima, Ryo Masumura, Hiroshi Saruwatari:
Investigating Effective Additional Contextual Factors in DNN-Based Spontaneous Speech Synthesis. INTERSPEECH 2020: 3201-3205 - [c43]Masashi Aso, Shinnosuke Takamichi, Hiroshi Saruwatari:
End-to-End Text-to-Speech Synthesis with Unaligned Multiple Language Units Based on Attention. INTERSPEECH 2020: 4009-4013 - [c42]Yuki Yamashita, Tomoki Koriyama, Yuki Saito, Shinnosuke Takamichi, Yusuke Ijima, Ryo Masumura, Hiroshi Saruwatari:
DNN-based Speech Synthesis Using Abundant Tags of Spontaneous Speech Corpus. LREC 2020: 6438-6443 - [c41]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
SMASH Corpus: A Spontaneous Speech Corpus Recording Third-person Audio Commentaries on Gameplay. LREC 2020: 6571-6577 - [i18]Hiroki Tamaru, Shinnosuke Takamichi, Naoko Tanji, Hiroshi Saruwatari:
JVS-MuSiC: Japanese multispeaker singing-voice corpus. CoRR abs/2001.07044 (2020) - [i17]Takaaki Saeki, Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Lifter Training and Sub-band Modeling for Computationally Efficient and High-Quality Voice Conversion Using Spectral Differentials. CoRR abs/2002.06778 (2020) - [i16]Junya Koguchi, Shinnosuke Takamichi:
PJS: phoneme-balanced Japanese singing voice corpus. CoRR abs/2006.02959 (2020) - [i15]Yuki Okamoto, Keisuke Imoto, Shinnosuke Takamichi, Ryosuke Yamanishi, Takahiro Fukumori, Yoichi Yamashita:
RWCP-SSD-Onomatopoeia: Onomatopoeic Word Dataset for Environmental Sound Synthesis. CoRR abs/2007.04719 (2020) - [i14]Shinnosuke Takamichi, Mamoru Komachi, Naoko Tanji, Hiroshi Saruwatari:
JSSS: free Japanese speech corpus for summarization and simplification. CoRR abs/2010.01793 (2020) - [i13]Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari:
Incremental Text-to-Speech Synthesis Using Pseudo Lookahead with Large Pretrained Language Model. CoRR abs/2012.12612 (2020)
2010 – 2019
- 2019
- [j10]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Vocoder-free text-to-speech synthesis incorporating generative adversarial networks using low-/multi-frequency STFT amplitude spectra. Comput. Speech Lang. 58: 347-363 (2019) - [j9]Daiki Sekizawa, Shinnosuke Takamichi, Hiroshi Saruwatari:
Prosody Correction Preserving Speaker Individuality for Chinese-Accented Japanese HMM-Based Text-to-Speech Synthesis. IEICE Trans. Inf. Syst. 102-D(6): 1218-1221 (2019) - [j8]Naoki Makishima, Shinichi Mogami, Norihiro Takamune, Daichi Kitamura, Hayato Sumino, Shinnosuke Takamichi, Hiroshi Saruwatari, Nobutaka Ono:
Independent Deeply Learned Matrix Analysis for Determined Audio Source Separation. IEEE ACM Trans. Audio Speech Lang. Process. 27(10): 1601-1615 (2019) - [c40]Kana Naruse, Shinnosuke Takamichi, Tomohiro Tanikawa, Shigeo Yoshida, Takuji Narumi, Michitaka Hirose:
Estimating confidence in voices using crowdsourcing for alleviating tension with altered auditory feedback. AsianCHI@CHI 2019: 15-22 - [c39]Hiroki Tamaru, Yuki Saito, Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari:
Generative Moment Matching Network-based Random Modulation Post-filter for DNN-based Singing Voice Synthesis and Neural Double-tracking. ICASSP 2019: 7070-7074 - [c38]Ivan Halim Parmonangan, Hiroki Tanaka, Sakriani Sakti, Shinnosuke Takamichi, Satoshi Nakamura:
Speech Quality Evaluation of Synthesized Japanese Speech Using EEG. INTERSPEECH 2019: 1228-1232 - [c37]Koichiro Yoshino, Yukitoshi Murase, Nurul Lubis, Kyoshiro Sugiyama, Hiroki Tanaka, Sakriani Sakti, Shinnosuke Takamichi, Satoshi Nakamura:
Spoken Dialogue Robot for Watching Daily Life of Elderly People. IWSDS 2019: 141-146 - [c36]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
DNN-based Speaker Embedding Using Subjective Inter-speaker Similarity for Multi-speaker Modeling in Speech Synthesis. SSW 2019: 51-56 - [c35]Riku Arakawa, Shinnosuke Takamichi, Hiroshi Saruwatari:
Implementation of DNN-based real-time voice conversion and its improvements by audio data augmentation and mask-shaped device. SSW 2019: 93-98 - [c34]Tomoki Koriyama, Shinnosuke Takamichi, Takao Kobayashi:
Sparse Approximation of Gram Matrices for GMMN-based Speech Synthesis. SSW 2019: 149-154 - [c33]Taiki Nakamura, Yuki Saito, Shinnosuke Takamichi, Yusuke Ijima, Hiroshi Saruwatari:
V2S attack: building DNN-based voice conversion from automatic speaker verification. SSW 2019: 161-165 - [c32]Masashi Aso, Shinnosuke Takamichi, Norihiro Takamune, Hiroshi Saruwatari:
Subword tokenization based on DNN-based acoustic model for end-to-end prosody generation. SSW 2019: 234-238 - [c31]Riku Arakawa, Shinnosuke Takamichi, Hiroshi Saruwatari:
TransVoice: Real-Time Voice Conversion for Augmenting Near-Field Speech Communication. UIST (Adjunct Volume) 2019: 33-35 - [i12]Hiroki Tamaru, Yuki Saito, Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari:
Generative Moment Matching Network-based Random Modulation Post-filter for DNN-based Singing Voice Synthesis and Neural Double-tracking. CoRR abs/1902.03389 (2019) - [i11]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
DNN-based Speaker Embedding Using Subjective Inter-speaker Similarity for Multi-speaker Modeling in Speech Synthesis. CoRR abs/1907.08294 (2019) - [i10]Taiki Nakamura, Yuki Saito, Shinnosuke Takamichi, Yusuke Ijima, Hiroshi Saruwatari:
V2S attack: building DNN-based voice conversion from automatic speaker verification. CoRR abs/1908.01454 (2019) - [i9]Shinnosuke Takamichi, Kentaro Mitsui, Yuki Saito, Tomoki Koriyama, Naoko Tanji, Hiroshi Saruwatari:
JVS corpus: free Japanese multi-speaker voice corpus. CoRR abs/1908.06248 (2019) - [i8]Yuki Okamoto, Keisuke Imoto, Tatsuya Komatsu, Shinnosuke Takamichi, Takumi Yagyu, Ryosuke Yamanishi, Yoichi Yamashita:
Overview of Tasks and Investigation of Subjective Evaluation Methods in Environmental Sound Synthesis and Conversion. CoRR abs/1908.10055 (2019) - [i7]Kazuki Fujii, Yuki Saito, Shinnosuke Takamichi, Yukino Baba, Hiroshi Saruwatari:
HumanGAN: generative adversarial network with human-based discriminator and its evaluation in speech perception modeling. CoRR abs/1909.11391 (2019) - 2018
- [j7]Takatomo Kano, Shinnosuke Takamichi, Sakriani Sakti, Graham Neubig, Tomoki Toda, Satoshi Nakamura:
An end-to-end model for cross-lingual transformation of paralinguistic information. Mach. Transl. 32(4): 353-368 (2018) - [j6]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Statistical Parametric Speech Synthesis Incorporating Generative Adversarial Networks. IEEE ACM Trans. Audio Speech Lang. Process. 26(1): 84-96 (2018) - [c30]Masakazu Une, Yuki Saito, Shinnosuke Takamichi, Daichi Kitamura, Ryoichi Miyazaki, Hiroshi Saruwatari:
Generative approach using the noise generation models for DNN-based speech synthesis trained from noisy speech. APSIPA 2018: 340-344 - [c29]Sayaka Shiota, Shinnosuke Takamichi, Tomoko Matsui:
Data augmentation with moment-matching networks for i-vector based speaker verification. APSIPA 2018: 345-349 - [c28]Takanori Akiyama, Shinnosuke Takamichi, Hiroshi Saruwatari:
Prosody-aware subword embedding considering Japanese intonation systems and its application to DNN-based multi-dialect speech synthesis. APSIPA 2018: 659-664 - [c27]Hitoshi Suda, Gaku Kotani, Shinnosuke Takamichi, Daisuke Saito:
A Revisit to Feature Handling for High-quality Voice Conversion Based on Gaussian Mixture Model. APSIPA 2018: 816-822 - [c26]Shinichi Mogami, Hayato Sumino, Daichi Kitamura, Norihiro Takamune, Shinnosuke Takamichi, Hiroshi Saruwatari, Nobutaka Ono:
Independent Deeply Learned Matrix Analysis for Multichannel Audio Source Separation. EUSIPCO 2018: 1557-1561 - [c25]Yuki Saito, Yusuke Ijima, Kyosuke Nishida, Shinnosuke Takamichi:
Non-Parallel Voice Conversion Using Variational Autoencoders Conditioned by Phonetic Posteriorgrams and D-Vectors. ICASSP 2018: 5274-5278 - [c24]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Text-to-Speech Synthesis Using STFT Spectra Based on Low-/Multi-Resolution Generative Adversarial Networks. ICASSP 2018: 5299-5303 - [c23]Shinnosuke Takamichi, Yuki Saito, Norihiro Takamune, Daichi Kitamura, Hiroshi Saruwatari:
Phase Reconstruction from Amplitude Spectrograms Based on Von-Mises-Distribution Deep Neural Network. IWAENC 2018: 286-290 - [c22]Shinnosuke Takamichi, Hiroshi Saruwatari:
CPJD Corpus: Crowdsourced Parallel Speech Corpus of Japanese Dialects. LREC 2018 - [i6]Shinichi Mogami, Hayato Sumino, Daichi Kitamura, Norihiro Takamune, Shinnosuke Takamichi, Hiroshi Saruwatari, Nobutaka Ono:
Independent Deeply Learned Matrix Analysis for Multichannel Audio Source Separation. CoRR abs/1806.10307 (2018) - [i5]Shinnosuke Takamichi, Yuki Saito, Norihiro Takamune, Daichi Kitamura, Hiroshi Saruwatari:
Phase reconstruction from amplitude spectrograms based on von-Mises-distribution deep neural network. CoRR abs/1807.03474 (2018) - 2017
- [j5]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Voice Conversion Using Input-to-Output Highway Networks. IEICE Trans. Inf. Syst. 100-D(8): 1925-1928 (2017) - [c21]Shinnosuke Takamichi:
Modulation spectrum-based speech parameter trajectory smoothing for DNN-based speech synthesis using FFT spectra. APSIPA 2017: 1308-1311 - [c20]Shinnosuke Takamichi, Daisuke Saito, Hiroshi Saruwatari, Nobuaki Minematsu:
The UTokyo speech synthesis system for Blizzard Challenge 2017. Blizzard Challenge 2017 - [c19]Yoshiki Mitsui, Daichi Kitamura, Shinnosuke Takamichi, Nobutaka Ono, Hiroshi Saruwatari:
Blind source separation based on independent low-rank matrix analysis with sparse regularization for time-series activity. ICASSP 2017: 21-25 - [c18]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Training algorithm to deceive Anti-Spoofing Verification for DNN-based speech synthesis. ICASSP 2017: 4900-4904 - [c17]Hiroyuki Miyoshi, Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Voice Conversion Using Sequence-to-Sequence Learning of Context Posterior Probabilities. INTERSPEECH 2017: 1268-1272 - [c16]Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari:
Sampling-Based Speech Parameter Generation Using Moment-Matching Networks. INTERSPEECH 2017: 3961-3965 - [i4]Hiroyuki Miyoshi, Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Voice Conversion Using Sequence-to-Sequence Learning of Context Posterior Probabilities. CoRR abs/1704.02360 (2017) - [i3]Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari:
Sampling-based speech parameter generation using moment-matching networks. CoRR abs/1704.03626 (2017) - [i2]Yuki Saito, Shinnosuke Takamichi, Hiroshi Saruwatari:
Statistical Parametric Speech Synthesis Incorporating Generative Adversarial Networks. CoRR abs/1709.08041 (2017) - [i1]Ryosuke Sonobe, Shinnosuke Takamichi, Hiroshi Saruwatari:
JSUT corpus: free large-scale Japanese speech corpus for end-to-end speech synthesis. CoRR abs/1711.00354 (2017) - 2016
- [j4]Shinnosuke Takamichi, Tomoki Toda, Graham Neubig, Sakriani Sakti, Satoshi Nakamura:
A Statistical Sample-Based Approach to GMM-Based Voice Conversion Using Tied-Covariance Acoustic Models. IEICE Trans. Inf. Syst. 99-D(10): 2490-2498 (2016) - [j3]Yuji Oshima, Shinnosuke Takamichi, Tomoki Toda, Graham Neubig, Sakriani Sakti, Satoshi Nakamura:
Non-Native Text-to-Speech Preserving Speaker Individuality Based on Partial Correction of Prosodic and Phonetic Characteristics. IEICE Trans. Inf. Syst. 99-D(12): 3132-3139 (2016) - [j2]Shinnosuke Takamichi, Tomoki Toda, Alan W. Black, Graham Neubig, Sakriani Sakti, Satoshi Nakamura:
Postfilters to Modify the Modulation Spectrum for Statistical Parametric Speech Synthesis. IEEE ACM Trans. Audio Speech Lang. Process. 24(4): 755-767 (2016) - [c15]Kazuhiro Kobayashi, Shinnosuke Takamichi, Satoshi Nakamura, Tomoki Toda:
The NU-NAIST Voice Conversion System for the Voice Conversion Challenge 2016. INTERSPEECH 2016: 1667-1671 - 2015
- [c14]Shinnosuke Takamichi, Kazuhiro Kobayashi, Kou Tanaka, Tomoki Toda, Satoshi Nakamura:
The NAIST Text-to-Speech System for the Blizzard Challenge 2015. Blizzard Challenge 2015 - [c13]Shinnosuke Takamichi, Tomoki Toda, Alan W. Black, Satoshi Nakamura:
Parameter generation algorithm considering Modulation Spectrum for HMM-based speech synthesis. ICASSP 2015: 4210-4214 - [c12]Shinnosuke Takamichi, Tomoki Toda, Alan W. Black, Satoshi Nakamura:
Modulation spectrum-constrained trajectory training algorithm for GMM-based Voice Conversion. ICASSP 2015: 4859-4863 - [c11]Yuji Oshima, Shinnosuke Takamichi, Tomoki Toda, Graham Neubig, Sakriani Sakti, Satoshi Nakamura:
Non-native speech synthesis preserving speaker individuality based on partial correction of prosodic and phonetic characteristics. INTERSPEECH 2015: 299-303 - [c10]Shinnosuke Takamichi, Tomoki Toda, Alan W. Black, Satoshi Nakamura:
Modulation spectrum-constrained trajectory training algorithm for HMM-based speech synthesis. INTERSPEECH 2015: 1206-1210 - [c9]Quoc Truong Do, Shinnosuke Takamichi, Sakriani Sakti, Graham Neubig, Tomoki Toda, Satoshi Nakamura:
Preserving word-level emphasis in speech-to-speech translation using linear regression HSMMs. INTERSPEECH 2015: 3665-3669 - 2014
- [j1]Shinnosuke Takamichi, Tomoki Toda, Yoshinori Shiga, Sakriani Sakti, Graham Neubig, Satoshi Nakamura:
Parameter Generation Methods With Rich Context Models for High-Quality and Flexible Text-To-Speech Synthesis. IEEE J. Sel. Top. Signal Process. 8(2): 239-250 (2014) - [c8]Shinnosuke Takamichi, Tomoki Toda, Alan W. Black, Satoshi Nakamura:
Modulation spectrum-based post-filter for GMM-based Voice Conversion. APSIPA 2014: 1-4 - [c7]Shinnosuke Takamichi, Tomoki Toda, Alan W. Black, Satoshi Nakamura:
Modified post-filter to recover modulation spectrum for HMM-based speech synthesis. GlobalSIP 2014: 547-551 - [c6]Shinnosuke Takamichi, Tomoki Toda, Graham Neubig, Sakriani Sakti, Satoshi Nakamura:
A postfilter to modify the modulation spectrum in HMM-based speech synthesis. ICASSP 2014: 290-294 - [c5]Nozomi Jinbo, Shinnosuke Takamichi, Tomoki Toda, Graham Neubig, Sakriani Sakti, Satoshi Nakamura:
A hearing impairment simulation method using audiogram-based approximation of auditory charatecteristics. INTERSPEECH 2014: 490-494 - 2013
- [c4]Shinnosuke Takamichi, Tomoki Toda, Yoshinori Shiga, Sakriani Sakti, Graham Neubig, Satoshi Nakamura:
Improvements to HMM-based speech synthesis based on parameter generation with rich context models. INTERSPEECH 2013: 364-368 - [c3]Takatomo Kano, Shinnosuke Takamichi, Sakriani Sakti, Graham Neubig, Tomoki Toda, Satoshi Nakamura:
Generalizing continuous-space translation of paralinguistic information. INTERSPEECH 2013: 2614-2618 - 2012
- [c2]Shinnosuke Takamichi, Tomoki Toda, Yoshinori Shiga, Hisashi Kawai, Sakriani Sakti, Satoshi Nakamura:
An Evaluation of Parameter Generation Methods with Rich Context Models in HMM-Based Speech Synthesis. INTERSPEECH 2012: 1139-1142 - [c1]Takatomo Kano, Sakriani Sakti, Shinnosuke Takamichi, Graham Neubig, Tomoki Toda, Satoshi Nakamura:
A method for translation of paralinguistic information. IWSLT 2012: 158-163
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-12-10 21:44 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint