default search action
Katsutoshi Itoyama
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [j24]Jiang Wang, Yuanzheng He, Daobilige Su, Katsutoshi Itoyama, Kazuhiro Nakadai, Junfeng Wu, Shoudong Huang, Youfu Li, He Kong:
SLAM-Based Joint Calibration of Multiple Asynchronous Microphone Arrays and Sound Source Localization. IEEE Trans. Robotics 40: 4024-4044 (2024) - [c95]Zirui Lin, Katsutoshi Itoyama, Kazuhiro Nakadai, Hideharu Amano:
FPGA-based Low Power Acceleration of HARK Sound Source Localization. COOL CHIPS 2024: 1-6 - [c94]Haruto Yokota, Mert Bozkurtlar, Benjamin Yen, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
A Video Vision Transformer for Sound Source Localization. EUSIPCO 2024: 106-110 - [c93]Ragib Amin Nihal, Benjamin Yen, Katsutoshi Itoyama, Kazuhiro Nakadai:
UAV-Enhanced Combination to Application: Comprehensive Analysis and Benchmarking of a Human Detection Dataset for Disaster Scenarios. ICPR (14) 2024: 145-162 - [c92]Takahiro Osaki, Yui Sudo, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Improving Noise Robustness of Automatic Speech Recognition Based on a Parallel Adapter Model with Near-Identity Initialization. IEA/AIE 2024: 454-466 - [c91]Shuhei Asaka, Katsutoshi Itoyama, Kazuhiro Nakadai:
Improving Impressions of Response Delay in AI-based Spoken Dialogue Systems. RO-MAN 2024: 1416-1421 - [c90]Mert Bozkurtlar, Benjamin Yen, Katsutoshi Itoyama, Kazuhiro Nakadai:
Real Time Sound Source Localization Using von-Mises ResNet. SII 2024: 466-471 - [i10]Ragib Amin Nihal, Benjamin Yen, Katsutoshi Itoyama, Kazuhiro Nakadai:
From Blurry to Brilliant Detection: YOLOv5-Based Aerial Object Detection with Super Resolution. CoRR abs/2401.14661 (2024) - [i9]Jiang Wang, Yuanzheng He, Daobilige Su, Katsutoshi Itoyama, Kazuhiro Nakadai, Junfeng Wu, Shoudong Huang, Youfu Li, He Kong:
SLAM-based Joint Calibration of Multiple Asynchronous Microphone Arrays and Sound Source Localization. CoRR abs/2405.19813 (2024) - [i8]Atsuo Hiroe, Katsutoshi Itoyama, Kazuhiro Nakadai:
Can all variations within the unified mask-based beamformer framework achieve identical peak extraction performance? CoRR abs/2407.15310 (2024) - [i7]Ragib Amin Nihal, Benjamin Yen, Katsutoshi Itoyama, Kazuhiro Nakadai:
UAV-Enhanced Combination to Application: Comprehensive Analysis and Benchmarking of a Human Detection Dataset for Disaster Scenarios. CoRR abs/2408.04922 (2024) - 2023
- [c89]Atsuo Hiroe, Katsutoshi Itoyama, Kazuhiro Nakadai:
Is the Ideal Ratio Mask Really the Best? - Exploring the Best Extraction Performance and Optimal Mask of Mask-based Beamformers. APSIPA ASC 2023: 1843-1850 - [c88]Haris Gulzar, Monikka Roslianna Busto, Takeharu Eda, Katsutoshi Itoyama, Kazuhiro Nakadai:
miniStreamer: Enhancing Small Conformer with Chunked-Context Masking for Streaming ASR Applications on the Edge. INTERSPEECH 2023: 3277-3281 - [c87]Takahiro Aizawa, Yoshiaki Bando, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai, Masaki Onishi:
Unsupervised Domain Adaptation of Universal Source Separation Based on Neural Full-Rank Spatial Covariance Analysis. MLSP 2023: 1-6 - [c86]Tan Sihan, Khan Nabeela Khanum, Katsutoshi Itoyama, Kazuhiro Nakadai:
Improving Sign Language Understanding Introducing Label Smoothing. RO-MAN 2023: 113-118 - [c85]Masahiko Fujita, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
An Ensemble Method for Multiple Speech Enhancement Using Deep Learning. SII 2023: 1-6 - [c84]Haris Gulzar, Muhammad Shakeel, Katsutoshi Itoyama, Kazuhiro Nakadai, Kenji Nishida, Hideharu Amano, Takeharu Eda:
FPGA based Power-Efficient Edge Server to Accelerate Speech Interface for Socially Assistive Robotics. SII 2023: 1-6 - [c83]Hidehiko Kishinami, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Reconstruction of Depth Scenes Based on Echolocation. SII 2023: 1-6 - [c82]Muhammad Shakeel, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Metric-Based Multimodal Meta-Learning for Human Movement Identification Via Footstep Recognition. SII 2023: 1-8 - [c81]Chishio Sugiyama, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Assessment of Simultaneous Calibration for Positions, Orientations, and Time Offsets in Multiple Microphone Arrays Systems. SII 2023: 1-6 - [c80]Kei Suzuki, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Audio-Visual Class Association Based on Two-stage Self-supervised Contrastive Learning towards Robust Scene Analysis. SII 2023: 1-6 - [i6]Atsuo Hiroe, Katsutoshi Itoyama, Kazuhiro Nakadai:
Is the Ideal Ratio Mask Really the Best? - Exploring the Best Extraction Performance and Optimal Mask of Mask-based Beamformers. CoRR abs/2309.12065 (2023) - 2022
- [c79]Yoshiaki Bando, Takahiro Aizawa, Katsutoshi Itoyama, Kazuhiro Nakadai:
Weakly-Supervised Neural Full-Rank Spatial Covariance Analysis for a Front-End System of Distant Speech Recognition. INTERSPEECH 2022: 3824-3828 - [c78]Yasuhiro Kagimoto, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Spotforming by NMF Using Multiple Microphone Arrays. IROS 2022: 9253-9258 - [c77]Taiki Yamada, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Outdoor evaluation of sound source localization for drone groups using microphone arrays. IROS 2022: 9296-9301 - 2021
- [j23]Yui Sudo, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Multichannel environmental sound segmentation. Appl. Intell. 51(11): 8245-8259 (2021) - [j22]Muhammad Shakeel, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Detecting earthquakes: a novel deep learning-based approach for effective disaster response. Appl. Intell. 51(11): 8305-8315 (2021) - [c76]Katsutoshi Itoyama, Yoshiya Morimoto, Shungo Masaki, Ryosuke Kojima, Kenji Nishida, Kazuhiro Nakadai:
Assessment of von Mises-Bernoulli Deep Neural Network in Sound Source Localization. Interspeech 2021: 2152-2156 - [c75]Zhi Zhong, Muhammad Shakeel, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Assessment of a Beamforming Implementation Developed for Surface Sound Source Separation. SII 2021: 369-374 - [c74]Yui Sudo, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Multi-channel Environmental Sound Segmentation utilizing Sound Source Localization and Separation U-Net. SII 2021: 382-387 - [c73]Muhammad Shakeel, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
EMC: Earthquake Magnitudes Classification on Seismic Signals via Convolutional Recurrent Networks. SII 2021: 388-393 - [c72]Taiki Yamada, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Sound Source Tracking Using Integrated Direction Likelihood for Drones with Microphone Arrays. SII 2021: 394-399 - [i5]Muhammad Shakeel, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Metric-based multimodal meta-learning for human movement identification via footstep recognition. CoRR abs/2111.07979 (2021) - 2020
- [j21]Yui Sudo, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Sound event aware environmental sound segmentation with Mask U-Net. Adv. Robotics 34(20): 1280-1290 (2020) - [j20]Ryo Nishikimi, Eita Nakamura, Masataka Goto, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Bayesian Singing Transcription Based on a Hierarchical Generative Model of Keys, Musical Notes, and F0 Trajectories. IEEE ACM Trans. Audio Speech Lang. Process. 28: 1678-1691 (2020) - [c71]Naoki Yamamoto, Kenji Nishida, Katsutoshi Itoyama, Kazuhiro Nakadai:
Detection of Ball Spin Direction using Hitting Sound in Tennis. icSPORTS 2020: 30-37 - [c70]Katsuhiro Dan, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Calibration of a Microphone Array Based on a Probabilistic Model of Microphone Positions. IEA/AIE 2020: 614-625 - [c69]Katsutoshi Itoyama, Kazuhiro Nakadai:
Synchronization of Microphones Based on Rank Minimization of Warped Spectrum for Asynchronous Distributed Recording. IROS 2020: 4842-4847 - [c68]Kazuhiro Nakadai, Shungo Masaki, Ryosuke Kojima, Osamu Sugiyama, Katsutoshi Itoyama, Kenji Nishida:
Sound Source Localization Based on von-Mises-Bernoulli Deep Neural Network. SII 2020: 658-663 - [c67]Taiki Yamada, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Sound Source Tracking by Drones with Microphone Arrays. SII 2020: 796-801 - [c66]Takashi Konno, Kenji Nishida, Katsutoshi Itoyama, Kazuhiro Nakadai:
Audio-Visual 3D Reconstruction Framework for Dynamic Scenes. SII 2020: 802-807 - [c65]Zhi Zhong, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Design and Assessment of a Scan-and-sum Beamformer for Surface Sound Source Separation. SII 2020: 808-813 - [c64]Yui Sudo, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Multi-channel Environmental sound segmentation. SII 2020: 820-825
2010 – 2019
- 2019
- [j19]Daniel Gabriel, Ryosuke Kojima, Kotaro Hoshiba, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
2D sound source position estimation using microphone arrays and its application to a VR-based bird song analysis system. Adv. Robotics 33(7-8): 403-414 (2019) - [j18]Kazuki Shimada, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Unsupervised Speech Enhancement Based on Multichannel NMF-Informed Beamforming for Noise-Robust Automatic Speech Recognition. IEEE ACM Trans. Audio Speech Lang. Process. 27(5): 960-971 (2019) - [c63]Yui Sudo, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Improvement of DOA Estimation by using Quaternion Output in Sound Event Localization and Detection. DCASE 2019: 244-247 - [c62]Kentaro Shibata, Ryo Nishikimi, Satoru Fukayama, Masataka Goto, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Joint Transcription of Lead, Bass, and Rhythm Guitars Based on a Factorial Hidden Semi-Markov Model. ICASSP 2019: 236-240 - [c61]Yui Sudo, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Environmental sound segmentation utilizing Mask U-Net. IROS 2019: 5340-5345 - [c60]Daniel Gabriel, Ryosuke Kojima, Kotaro Hoshiba, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Design and assessment of multiple-sound source localization using microphone arrays. SII 2019: 199-204 - [p2]Masashi Konyo, Yuichi Ambe, Hikaru Nagano, Yu Yamauchi, Satoshi Tadokoro, Yoshiaki Bando, Katsutoshi Itoyama, Hiroshi G. Okuno, Takayuki Okatani, Kanta Shimizu, Eisuke Ito:
ImPACT-TRC Thin Serpentine Robot Platform for Urban Search and Rescue. Disaster Robotics 2019: 25-76 - [p1]Fumitoshi Matsuno, Tetsushi Kamegawa, Wei Qi, Tatsuya Takemori, Motoyasu Tanaka, Mizuki Nakajima, Kenjiro Tadakuma, Masahiro Fujita, Yosuke Suzuki, Katsutoshi Itoyama, Hiroshi G. Okuno, Yoshiaki Bando, Tomofumi Fujiwara, Satoshi Tadokoro:
Development of Tough Snake Robot Systems. Disaster Robotics 2019: 267-326 - [i4]Kazuki Shimada, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Unsupervised Speech Enhancement Based on Multichannel NMF-Informed Beamforming for Noise-Robust Automatic Speech Recognition. CoRR abs/1903.09341 (2019) - 2018
- [j17]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Tatsuya Kawahara, Hiroshi G. Okuno:
Speech Enhancement Based on Bayesian Low-Rank and Sparse Decomposition of Multichannel Magnitude Spectrograms. IEEE ACM Trans. Audio Speech Lang. Process. 26(2): 215-230 (2018) - [j16]Kousuke Itakura, Yoshiaki Bando, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Bayesian Multichannel Audio Source Separation Based on Integrated Source and Spatial Models. IEEE ACM Trans. Audio Speech Lang. Process. 26(4): 831-846 (2018) - [c59]Yusuke Wada, Ryo Nishikimi, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Sequential Generation of Singing F0 Contours from Musical Note Sequences Based on WaveNet. APSIPA 2018: 983-989 - [c58]Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Statistical Speech Enhancement Based on Probabilistic Integration of Variational Autoencoder and Non-Negative Matrix Factorization. ICASSP 2018: 716-720 - [c57]Kazuki Shimada, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Unsupervised Beamforming Based on Multichannel Nonnegative Matrix Factorization for Noisy Speech Recognition. ICASSP 2018: 5734-5738 - [c56]Hiroaki Tsushima, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Interactive Arrangement of Chords and Melodies Based on a Tree-Structured Generative Model. ISMIR 2018: 145-151 - [c55]Ryosuke Taniguchi, Kotaro Hoshiba, Katsutoshi Itoyama, Kenji Nishida, Kazuhiro Nakadai:
Signal Restoration based on Bi-directional LSTM with Spectral Filtering for Robot Audition. RO-MAN 2018: 955-960 - 2017
- [j15]Karim Youssef, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Simultaneous Identification and Localization of Still and Mobile Speakers Based on Binaural Robot Audition. J. Robotics Mechatronics 29(1): 59-71 (2017) - [j14]Kouhei Sekiguchi, Yoshiaki Bando, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Layout Optimization of Cooperative Distributed Microphone Arrays Based on Estimation of Source Separation Performance. J. Robotics Mechatronics 29(1): 83-93 (2017) - [j13]Misato Ohkita, Yoshiaki Bando, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Audio-Visual Beat Tracking Based on a State-Space Model for a Robot Dancer Performing with a Human Dancer. J. Robotics Mechatronics 29(1): 125-136 (2017) - [j12]Yoshiaki Bando, Hiroshi Saruwatari, Nobutaka Ono, Shoji Makino, Katsutoshi Itoyama, Daichi Kitamura, Masaru Ishimura, Moe Takakusaki, Narumi Mae, Kouei Yamaoka, Yutaro Matsui, Yuichi Ambe, Masashi Konyo, Satoshi Tadokoro, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Low Latency and High Quality Two-Stage Human-Voice-Enhancement System for a Hose-Shaped Rescue Robot. J. Robotics Mechatronics 29(1): 198-212 (2017) - [c54]Kousuke Itakura, Yoshiaki Bando, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Bayesian multichannel nonnegative matrix factorization for audio source separation and localization. ICASSP 2017: 551-555 - [c53]Ryo Nishikimi, Eita Nakamura, Masataka Goto, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Scale- and Rhythm-Aware Musical Note Estimation for Vocal F0 Trajectories Based on a Semi-Tatum-Synchronous Hierarchical Hidden Semi-Markov Model. ISMIR 2017: 376-382 - [c52]Hiroaki Tsushima, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Function- and Rhythm-Aware Melody Harmonization Based on Tree-Structured Parsing and Split-Merge Sampling of Chord Sequences. ISMIR 2017: 502-508 - [c51]Masaya Wake, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Semi-Blind speech enhancement basedon recurrent neural network for source separation and dereverberation. MLSP 2017: 1-6 - [c50]Kazuyoshi Yoshii, Eita Nakamura, Katsutoshi Itoyama, Masataka Goto:
Infinite probabilistic latent component analysis for audio source separation. MLSP 2017: 1-6 - [i3]Hiroaki Tsushima, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Generative Statistical Models with Self-Emergent Grammar of Chord Sequences. CoRR abs/1708.02255 (2017) - [i2]Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Statistical Speech Enhancement Based on Probabilistic Integration of Variational Autoencoder and Non-Negative Matrix Factorization. CoRR abs/1710.11439 (2017) - 2016
- [j11]Yukara Ikemiya, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Singing Voice Separation and Vocal F0 Estimation Based on Mutual Combination of Robust Principal Component Analysis and Subharmonic Summation. IEEE ACM Trans. Audio Speech Lang. Process. 24(11): 2084-2095 (2016) - [c49]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Variational Bayesian multi-channel robust NMF for human-voice enhancement with a deformable and partially-occluded microphone array. EUSIPCO 2016: 1018-1022 - [c48]Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Rhythm transcription of MIDI performances based on hierarchical Bayesian modelling of repetition and modification of musical note patterns. EUSIPCO 2016: 1946-1950 - [c47]Kousuke Itakura, Yoshiaki Bando, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
A unified Bayesian model of time-frequency clustering and low-rank approximation for multi-channel source separation. EUSIPCO 2016: 2280-2284 - [c46]Kazuyoshi Yoshii, Katsutoshi Itoyama, Masataka Goto:
Student's T nonnegative matrix factorization and positive semidefinite tensor factorization for single-channel audio source separation. ICASSP 2016: 51-55 - [c45]Kouhei Sekiguchi, Yoshiaki Bando, Keisuke Nakamura, Kazuhiro Nakadai, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Online simultaneous localization and mapping of multiple sound sources and asynchronous microphone arrays. IROS 2016: 1973-1979 - [c44]Yuta Ojima, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
A Hierarchical Bayesian Model of Chords, Pitches, and Spectrograms for Multipitch Analysis. ISMIR 2016: 309-315 - [c43]Ryo Nishikimi, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Musical Note Estimation for F0 Trajectories of Singing Voices Based on a Bayesian Semi-Beat-Synchronous HMM. ISMIR 2016: 461-467 - [c42]Koichi Kitamura, Yoshiaki Bando, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Student's t multichannel nonnegative matrix factorization for blind source separation. IWAENC 2016: 1-5 - [c41]Koichiro Yoshino, Naoki Hirayama, Shinsuke Mori, Fumihiko Takahashi, Katsutoshi Itoyama, Hiroshi G. Okuno:
Parallel Speech Corpora of Japanese Dialects. LREC 2016 - [c40]Yoshiaki Bando, Hiroki Suhara, Motoyasu Tanaka, Tetsushi Kamegawa, Katsutoshi Itoyama, Kazuyoshi Yoshii, Fumitoshi Matsuno, Hiroshi G. Okuno:
Sound-based online localization for an in-pipe snake robot. SSRR 2016: 207-213 - [i1]Yukara Ikemiya, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Singing Voice Separation and Vocal F0 Estimation based on Mutual Combination of Robust Principal Component Analysis and Subharmonic Summation. CoRR abs/1604.00192 (2016) - 2015
- [j10]Yoshiaki Bando, Takuma Otsuka, Takeshi Mizumoto, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Hiroshi G. Okuno:
Posture estimation of hose-shaped robot by using active microphone array. Adv. Robotics 29(1): 35-49 (2015) - [j9]Izaya Nishimuta, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Toward a quizmaster robot for speech-based multiparty interaction. Adv. Robotics 29(18): 1205-1219 (2015) - [j8]Shotaro Sano, Takuma Otsuka, Katsutoshi Itoyama, Hiroshi G. Okuno:
HMM-based Attacks on Google's ReCAPTCHA with Continuous Visual and Audio Symbols. J. Inf. Process. 23(6): 814-826 (2015) - [j7]Naoki Hirayama, Koichiro Yoshino, Katsutoshi Itoyama, Shinsuke Mori, Hiroshi G. Okuno:
Automatic Speech Recognition for Mixed Dialect Utterances by Mixing Dialect Language Models. IEEE ACM Trans. Audio Speech Lang. Process. 23(2): 373-382 (2015) - [c39]Yoshiaki Bando, Takuma Otsuka, Ikkyu Aihara, Hiromitsu Awano, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi Gitchang Okuno:
Recognition of In-Field Frog Chorusing Using Bayesian Nonparametric Microphone Array Processing. AAAI Workshop: Computational Sustainability 2015 - [c38]Satoshi Maruo, Kazuyoshi Yoshii, Katsutoshi Itoyama, Matthias Mauch, Masataka Goto:
A feedback framework for improved chord recognition based on NMF-based approximate note transcription. ICASSP 2015: 196-200 - [c37]Yukara Ikemiya, Kazuyoshi Yoshii, Katsutoshi Itoyama:
Singing voice analysis and editing based on mutually dependent F0 estimation and source separation. ICASSP 2015: 574-578 - [c36]Yoshiaki Bando, Takuma Otsuka, Katsutoshi Itoyama, Kazuyoshi Yoshii, Yoko Sasaki, Satoshi Kagami, Hiroshi G. Okuno:
Challenges in deploying a microphone array to localize and separate sound sources in real auditory scenes. ICASSP 2015: 723-727 - [c35]Kousuke Itakura, Izaya Nishimuta, Yoshiaki Bando, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Bayesian integration of sound source separation and speech recognition: a new approach to simultaneous speech recognition. INTERSPEECH 2015: 736-740 - [c34]Kouhei Sekiguchi, Yoshiaki Bando, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Optimizing the layout of multiple mobile robots for cooperative sound source separation. IROS 2015: 5548-5554 - [c33]Misato Ohkita, Yoshiaki Bando, Yukara Ikemiya, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Audio-visual beat tracking based on a state-space model for a music robot dancing with humans. IROS 2015: 5555-5560 - [c32]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Microphone-accelerometer based 3D posture estimation for a hose-shaped rescue robot. IROS 2015: 5580-5586 - [c31]Kazuyoshi Yoshii, Katsutoshi Itoyama, Masataka Goto:
Infinite Superimposed Discrete All-Pole Modeling for Multipitch Analysis of Wavelet Spectrograms. ISMIR 2015: 86-92 - [c30]Karim Youssef, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Identification and Localization of One or Two Concurrent Speakers in a Binaural Robotic Context. SMC 2015: 407-412 - [c29]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Human-voice enhancement based on online RPCA for a hose-shaped rescue robot with a microphone array. SSRR 2015: 1-6 - [c28]Akira Maezawa, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Unified inter- and intra-recording duration model for multiple music audio alignment. WASPAA 2015: 1-5 - 2014
- [j6]Akira Maezawa, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Nonparametric Bayesian dereverberation of power spectrograms based on infinite-order autoregressive processes. IEEE ACM Trans. Audio Speech Lang. Process. 22(12): 1918-1930 (2014) - [c27]Izaya Nishimuta, Naoki Hirayama, Kazuyoshi Yoshii, Katsutoshi Itoyama, Hiroshi G. Okuno:
A robot quizmaster that can localize, separate, and recognize simultaneous utterances for a fastest-voice-first quiz game. Humanoids 2014: 967-972 - [c26]Kazuki Yazawa, Katsutoshi Itoyama, Hiroshi G. Okuno:
Automatic transcription of guitar tablature from audio signals in accordance with player's proficiency. ICASSP 2014: 3122-3126 - [c25]Yukara Ikemiya, Katsutoshi Itoyama, Hiroshi G. Okuno:
Transcribing vocal expression from polyphonic music. ICASSP 2014: 3127-3131 - [c24]Katsutoshi Itoyama, Hiroshi G. Okuno:
Parameter Estimation of Virtual Musical Instrument Synthesizers. ICMC 2014 - [c23]Yukara Ikemiya, Katsutoshi Itoyama, Hiroshi G. Okuno:
Transferring Vocal Expression of F0 Contour Using Singing Voice Synthesizer. IEA/AIE (2) 2014: 250-259 - [c22]Takahiro Iyama, Osamu Sugiyama, Takuma Otsuka, Katsutoshi Itoyama, Hiroshi G. Okuno:
Visualization of auditory awareness based on sound source positions estimated by depth sensor and microphone array. IROS 2014: 1908-1913 - [c21]Akira Maezawa, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Bayesian Audio Alignment based on a Unified Model of Music Composition and Performance. ISMIR 2014: 233-238 - [c20]Osamu Sugiyama, Katsutoshi Itoyama, Kazuhiro Nakadai, Hiroshi G. Okuno:
Sound annotation tool for multidirectional sounds based on spatial information extracted by HARK robot audition software. SMC 2014: 2335-2340 - [c19]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Hiroshi G. Okuno:
A sound-based online method for estimating the time-varying posture of a hose-shaped robot. SSRR 2014: 1-6 - 2013
- [j5]Daichi Sakaue, Katsutoshi Itoyama, Tetsuya Ogata, Hiroshi G. Okuno:
Robust Multipitch Analyzer against Initialization based on Latent Harmonic Allocation using Overtone Corpus. Inf. Media Technol. 8(2): 467-476 (2013) - [j4]Daichi Sakaue, Katsutoshi Itoyama, Tetsuya Ogata, Hiroshi G. Okuno:
Robust Multipitch Analyzer against Initialization based on Latent Harmonic Allocation using Overtone Corpus. J. Inf. Process. 21(2): 246-255 (2013) - [c18]Kazuki Yazawa, Daichi Sakaue, Kohei Nagira, Katsutoshi Itoyama, Hiroshi G. Okuno:
Audio-based guitar tablature transcription using multipitch analysis and playability constraints. ICASSP 2013: 196-200 - [c17]Daichi Sakaue, Takuma Otsuka, Katsutoshi Itoyama, Hiroshi G. Okuno:
Initialization-robust Bayesian multipitch analyzer based on psychoacoustical and musical criteria. ICASSP 2013: 226-230 - [c16]Naoyuki Kanda, Katsutoshi Itoyama, Hiroshi G. Okuno:
Multiple index combination for Japanese spoken term detection with optimum index selection based on OOV-region classifier. ICASSP 2013: 8540-8544 - [c15]Naoki Hirayama, Koichiro Yoshino, Katsutoshi Itoyama, Shinsuke Mori, Hiroshi G. Okuno:
Automatic estimation of dialect mixing ratio for dialect speech recognition. INTERSPEECH 2013: 1492-1496 - [c14]Yoshiaki Bando, Takeshi Mizumoto, Katsutoshi Itoyama, Kazuhiro Nakadai, Hiroshi G. Okuno:
Posture estimation of hose-shaped robot using microphone array localization. IROS 2013: 3446-3451 - [c13]Koutarou Furukawa, Keita Okutani, Kohei Nagira, Takuma Otsuka, Katsutoshi Itoyama, Kazuhiro Nakadai, Hiroshi G. Okuno:
Noise correlation matrix estimation for improving sound source localization by multirotor UAV. IROS 2013: 3943-3948 - 2012
- [j3]Akira Maezawa, Katsutoshi Itoyama, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Automated Violin Fingering Transcription Through Analysis of an Audio Recording. Comput. Music. J. 36(3): 57-72 (2012) - [c12]Daichi Sakaue, Katsutoshi Itoyama, Tetsuya Ogata, Hiroshi G. Okuno:
Initialization-robust multipitch estimation based on latent harmonic allocation using overtone corpus. ICASSP 2012: 425-428 - [c11]Katsutoshi Itoyama, Tetsuya Ogata, Hiroshi G. Okuno:
Automatic Chord Recognition Based on Probabilistic Integration of Acoustic Features, Bass Sounds, and Chord Transition. IEA/AIE 2012: 58-67 - [c10]Daichi Sakaue, Takuma Otsuka, Katsutoshi Itoyama, Hiroshi G. Okuno:
Bayesian Nonnegative Harmonic-Temporal Factorization and Its Application to Multipitch Analysis. ISMIR 2012: 91-96 - 2011
- [c9]Katsutoshi Itoyama, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Simultaneous processing of sound source separation and musical instrument identification using Bayesian spectral modeling. ICASSP 2011: 3816-3819 - [c8]Naoki Nishikawa, Katsutoshi Itoyama, Hiromasa Fujihara, Masataka Goto, Tetsuya Ogata, Hiroshi G. Okuno:
A musical mood trajectory estimation method using lyrics and acoustic features. MIRUM 2011: 51-56 - 2010
- [c7]Akira Maezawa, Katsutoshi Itoyama, Toru Takahashi, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Violin Fingering Estimation Based on Violin Pedagogical Fingering Model Constrained by Bowed Sequence Estimation from Audio Input. IEA/AIE (3) 2010: 249-259
2000 – 2009
- 2009
- [j2]Katsutoshi Itoyama, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Parameter Estimation for Harmonic and Inharmonic Models by Using Timbre Feature Distributions. Inf. Media Technol. 4(3): 672-682 (2009) - [j1]Katsutoshi Itoyama, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Parameter Estimation for Harmonic and Inharmonic Models by Using Timbre Feature Distributions. J. Inf. Process. 17: 191-201 (2009) - [c6]Akira Maezawa, Katsutoshi Itoyama, Toru Takahashi, Tetsuya Ogata, Hiroshi G. Okuno:
Bowed String Sequence Estimation of a Violin Based on Adaptive Audio Signal Classification and Context-Dependent Error Correction. ISM 2009: 9-16 - [c5]Naoki Yasuraoka, Takehiro Abe, Katsutoshi Itoyama, Toru Takahashi, Tetsuya Ogata, Hiroshi G. Okuno:
Changing timbre and phrase in existing musical performances as you like: manipulations of single part using harmonic and inharmonic models. ACM Multimedia 2009: 203-212 - 2008
- [c4]Kouhei Sumi, Katsutoshi Itoyama, Kazuyoshi Yoshii, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Automatic Chord Recognition Based on Probabilistic Integration of Chord Transition and Bass Pitch Estimation. ISMIR 2008: 39-44 - [c3]Katsutoshi Itoyama, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Instrument Equalizer for Query-by-Example Retrieval: Improving Sound Source Separation Based on Integrated Harmonic and Inharmonic Models. ISMIR 2008: 133-138 - 2007
- [c2]Katsutoshi Itoyama, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Integration and Adaptation of Harmonic and Inharmonic Models for Separating Polyphonic Musical Signals. ICASSP (1) 2007: 57-60 - 2006
- [c1]Katsutoshi Itoyama, Tetsuro Kitahara, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Automatic Feature Weighting in Automatic Transcription of Specified Part in Polyphonic Music. ISMIR 2006: 172-175
Coauthor Index
aka: Hiroshi Gitchang Okuno
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-12-15 02:19 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint