{"id":537,"date":"2023-06-13T14:29:56","date_gmt":"2023-06-13T05:29:56","guid":{"rendered":"https:\/\/slp.cs.tut.ac.jp\/?page_id=537"},"modified":"2026-03-04T16:04:25","modified_gmt":"2026-03-04T07:04:25","slug":"activity","status":"publish","type":"page","link":"https:\/\/slp.cs.tut.ac.jp\/en\/activity\/","title":{"rendered":"PUBLICATION"},"content":{"rendered":"\n<p class=\"has-text-align-center\"><a rel=\"noreferrer noopener\" href=\"http:\/\/www.google.com\/url?q=http%3A%2F%2Fsites.google.com%2Fsite%2Fnorihidekitaokashome%2Fpublication-list&amp;sa=D&amp;sntz=1&amp;usg=AFQjCNGBP8U1oOfry4oWP6rc87aaeyA6cw\" target=\"_blank\">Click here to Kitaoka&#8217;s publication list.<\/a><\/p>\n\n\n\n<p class=\"has-text-align-center\"><a href=\"#2024\">2024<\/a>  \/  <a href=\"#2023\">2023<\/a>  \/  <a href=\"#2022\">2022<\/a>  \/  <a href=\"#2021\">2021<\/a>  \/  <a href=\"#2020\">2020<\/a>  <br>\u3000<a href=\"#2019\">2019<\/a>  \/  <a href=\"#2018\">2018<\/a>  \/  <a href=\"#2017\">2017<\/a>  \/  <a href=\"#2016\">2016<\/a>  \/  <a href=\"#2015\">2015<\/a>  \/  <a href=\"#2014\">2014<\/a><\/p>\n\n\n\n<div style=\"height:24px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2025\">2025<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Journal Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Koharu Horii<\/strong>, <strong>Meiko Fukuda<\/strong>, Kengo Ohta, <strong>Ryota Nishimura<\/strong>, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;End-to-End Spontaneous Speech Recognition Based on Disfluency Labeling,&#8221; IEICE Trans. Inf. &amp; Syst., (accepted), 2026.<\/li>\n\n\n\n<li><strong>Takahiro Kinouchi<\/strong>, Atsunori Ogawa,<strong> Yukoh Wakabayashi<\/strong>, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, &#8220;<a href=\"https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0167639325001189\">Domain adaptation using non-parallel target domain corpus for self-supervised learning-based automatic speech recognition<\/a>,&#8221; SPEECH COMMUNICATION, Vol. 174, 103303, (8 pages) Oct., 2025.<\/li>\n\n\n\n<li>Shuming Luan, <strong>Yukoh Wakabayashi<\/strong>, and Tomoki Toda, &#8220;<a href=\"https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0003682X25001781\">Generalized sound field interpolation for freely spaced microphone arrays in rotation-robust beamforming<\/a>,&#8221; Applied Acoustics, vol. 236, pp. 110706, Apr., 2025.<\/li>\n\n\n\n<li>Kazuya Tsubokura, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, &#8220;<a href=\"https:\/\/search.ieice.org\/bin\/summary.php?id=j108-d_4_182&amp;category=-D&amp;year=2025&amp;lang=J&amp;abst=\">Relationship between User Reactions and Individual Characteristics during Dialogue Breakdown in Spoken Dialogue Systems<\/a>,&#8221; IEICE Transactions on Information and Communication D, Vol.J108-D, No.04, pp.182-191, Apr., 2025.(DOI: 10.14923\/transinfj.2024PDP0035)<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Meiko Fukuda<\/strong>, <strong>Ryota Nishimura<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;<a href=\"https:\/\/ieeexplore.ieee.org\/document\/11384874\">A corpus-based investigation of acoustic features influencing intelligibility of super-elderly Japanese speech<\/a>,&#8221; Proc. Oriental-COCOSDA 2025, pp.145-150, Nov., 2025.<\/li>\n\n\n\n<li><strong>Rikuto Yamanaka<\/strong>, <strong>Tsubasa Saito<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;<a href=\"https:\/\/ieeexplore.ieee.org\/document\/11385109\">Speech input interface for electronic medical record supporting automatic SOAP generation using large language models<\/a>,&#8221; pp.258-263, Nov., 2025.<\/li>\n\n\n\n<li><strong>Tsubasa Saito<\/strong>, <strong>Rikuto Yamanaka<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;<a href=\"https:\/\/ieeexplore.ieee.org\/document\/11385094\">Generation and automatic evaluation of SOAP notes from medical dialogue using large language models<\/a>.&#8221; , pp.312-317, Nov., 2025.<\/li>\n\n\n\n<li>Jotaro Emoto, <strong>Ryota Nishimura<\/strong>, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, \u201cReal-time VAD-less speech recognition by fine-tuning SSL model with data containing tagged non-speech segments,\u201d Proc. APSIPA ASC 2025, pp. 980-985, Oct., 2025.<\/li>\n\n\n\n<li>Ryota Uematsu, Chee Siang Leow, <strong>Norihide Kitaoka<\/strong>, Hiromitsu Nishizaki, \u201cImproving Automatic Speech Recognition Model for Super-Elderly Voice Using Speech Synthesis Model,\u201d Proc. APSIPA ASC 2025, pp. 986-988, Oct., 2025.<\/li>\n\n\n\n<li><strong>Tamon Mikawa<\/strong>, Yasuhisa Fujii, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, <strong>Ryota Nishimura<\/strong>, <strong>Norihide Kitaoka<\/strong>, \u201cImproving Listening Head Generation Performance Using Speech Representations from Self-Supervised Learning,\u201d Proc. APSIPA ASC 2025, pp. 956-961, Oct., 2025.<\/li>\n\n\n\n<li>Kazuya Tsubokura, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, \u201cToward Natural System Repair: An Analysis of Human Other-Initiated Self-Repair Patterns in Japanese Casual Conversations,\u201d Proc. APSIPA ASC 2025, 1004-1009, Oct., 2025.<\/li>\n\n\n\n<li>Yoshinori Fukunaga, <strong>Ryota Nishimura<\/strong>, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, &#8220;Backchannel prediction for natural spoken dialog systems using general speaker and listener information,&#8221; Proc. INTERSPEECH2025, pp.1078-1082, Aug., 2025. <\/li>\n\n\n\n<li><strong>Kaito Takahashi<\/strong>, <strong>Keigo Hojo<\/strong>, <strong>Toshimitsu Sakai<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Fine-tuning Parakeet-TDT for Dysarthric Speech Recognition in the Speech Accessibility Project Challenge,&#8221; Proc. INTERSPEECH2025, pp.3304-3308, Aug., 2025.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Keigo Hojo<\/strong>, <strong>Kaito Takahashi<\/strong>, <strong>Toshimitsu Sakai<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, \u201cImproving Accuracy of Speech Recognition for Articulation Disorders Using Parakeet-TDT,\u201d Japan Acoustical Society Fall Research Presentation Meeting, 1-Q-37, pp. 1-4, Sep., 2025.<\/li>\n\n\n\n<li><strong>Tatsunari Takagi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, \u201cConstruction of a Speech Recognition System for Articulation Disorders Based on Generative Error Correction Using Whisper and LLM,\u201d Japan Acoustical Society Fall Research Presentation Meeting, 1-Q-38, pp. 1-2, Sep., 2025.<\/li>\n\n\n\n<li><strong>Tamon Mikawa<\/strong>, Yasuhisa Fujii, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta , <strong>Ryota Nishimura<\/strong> , <strong>Norihide Kitaoka<\/strong>, \u201cGeneration of Listener Head Motions Using Speech Expressions via Self-Supervised Learning,\u201d Autumn Meeting of the Acoustical Society of Japan, 1-Q-38, pp. 1-4, Sep., 2025.<\/li>\n\n\n\n<li><strong>Rikuto Yamanaka<\/strong>, <strong>Tsubasa Saito<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, \u201cElectronic Medical Record Voice Input Interface Supporting Automatic Generation of SOAP Elements Using Large-Scale Language Models,\u201d FIT2025, E-008, pp. 1-2, Sep., 2025.<\/li>\n\n\n\n<li><strong>Keigo Hojo<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Accuracy improvement of CTC speech recognition using intermediate features based on attention mechanism,&#8221; Spring Meeting of Acoustical Society of Japan, pp.1-4, Mar., 2025.<\/li>\n\n\n\n<li><strong>Kaito Takahashi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Akio Kobayashi, <strong>Norihide Kitaoka<\/strong>, &#8220;Accurate Speech Recognition by Alternate Acoustic and Linguistic Adaptation in Speech of Deaf People,&#8221; Spring Meetings of Acoustical Society of Japan, 1-R-15, pp.1-4, Mar., 2025.<\/li>\n\n\n\n<li><strong>Tatsunari Takagi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Evaluation of Domain Adaptation by Internal Language Information Estimation in a Non-Autoregressive Speech Recognition Model,&#8221; Spring Meeting of Acoustical Society of Japan, 1-2-13, pp.1-4, Mar., 2025.<\/li>\n\n\n\n<li>Doh Kato, ZHOU QIHANG, <strong>Norihide Kitaoka<\/strong>, Takatoshi Jitsuhiro, &#8220;Japanese speech error detection for non-native speakers using convolutional autoencoder based on anomaly detection technique,&#8221; Spring Meetings of Acoustical Society of Japan, 1-R-11, pp.1-2, Mar., 2025.<\/li>\n\n\n\n<li>Tohma Sugimoto, <strong>Ryota Nishimura<\/strong>, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, &#8220;Prediction of Aizuchi Generation Timing Using HuBERT&#8217;s Intermediate Stratified Features,&#8221; Spring Meeting of Acoustical Society of Japan, 3-P-11, pp.1-4, Mar., 2025.<\/li>\n\n\n\n<li>Jotaro Emoto, <strong>Ryota Nishimura<\/strong>, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of a Real-Time VAD-less Speech Recognition Model under Noisy Environment and Comparison with Other Models,&#8221; 1-R-2, pp.1-4, Mar., 2025.<\/li>\n\n\n\n<li><strong>Rikuto Yamanaka<\/strong>, <strong>Tsubasa Saito<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Automatic Input Interface for Electronic Medical Records Using Speech Recognition and Multiple Large-scale Language Models,&#8221; The 87th National Convention of Information Processing Society of Japan, pp.1-2, Mar., 2025.<\/li>\n\n\n\n<li><strong>Tsubasa Saito<\/strong>, <strong>Rikuto Yamanaka<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Development of SOAP creation support system for electronic medical records using large-scale language models,&#8221; The 31st Annual Conference of the Association for Natural Language Processing, pp.1-6, Mar., 2025.<\/li>\n\n\n\n<li>Kazuya Tsubokura, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, &#8220;Analysis of Other-initiation Repair in Japanese Daily Conversation,&#8221; The 31st Annual Meeting of the Association for Natural Language Processing, Mar., 2025.<\/li>\n\n\n\n<li>Yoshiaki Kuwayama, Kazuya Tsubokura, Yurie Iribe, Katsunori Yokoi, Akinori Nakamura, <strong>Norihide Kitaoka<\/strong>, Masao Katsuno, &#8220;Prediction of early Alzheimer&#8217;s disease from free conversational speech using self-supervised learning,&#8221; The 31st Annual Conference of the Association for Natural Language Processing, Mar., 2025.<\/li>\n\n\n\n<li>Tatsuaki Yamada, Kazuya Tsubokura, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, &#8220;Bankruptcy estimation in dialogue systems considering individual characteristics,&#8221; The 31st Annual Meeting of the Association for Natural Language Processing, Mar., 2025.<\/li>\n\n\n\n<li><strong>Keigo Hojo<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of Speech Recognition Model Based on Self-Supervised Learning with Intermediate Layer Output,&#8221; SPEASIP Workshop 2025, pp.1-6, Mar., 2025.<\/li>\n\n\n\n<li><strong>Takanori Kanai<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Ryota Nishimura<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Accurate Estimation of Speech Termination Time for Spoken Dialogue Systems and Its Evaluation,&#8221; SPEASIP Workshop 2025, pp.1-6, Mar., 2025.<\/li>\n\n\n\n<li><strong>Tamon Mikawa<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, <strong>Ryota Nishimura<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Multimodal Generation of Listener&#8217;s Head Reaction Behaviors in Japanese Chat Dialogues Using Speech Self-Supervised Model,&#8221; SPEASIP Workshop 2025, pp.1-6, Mar., 2025.<\/li>\n\n\n\n<li><strong>Kaito Takahashi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, &#8220;Performance improvement of elderly speech recognition based on alternating acoustic and language adaptation,&#8221; SPEASIP Workshop 2025, pp.1-6, Mar., 2025.<\/li>\n\n\n\n<li><strong>Shintaro Uchiyama<\/strong>, <strong>Tatsunari Takagi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Effects of temperature parameters on domain adaptation of non-autoregressive speech recognition models,&#8221; SPEASIP Workshop 2025, pp.1-6, Mar., 2025.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2024\">2024<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Journal Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Shuming Luan,&nbsp;<strong>Yukoh Wakabayashi<\/strong>, Tomoki Toda, &#8220;<a href=\"https:\/\/ieeexplore.ieee.org\/document\/10551423\">Unequally Spaced Sound Field Interpolation for Rotation-Robust Beamforming<\/a>,&#8221; IEEE\/ACM Transactions on Audio, Speech, and Language Processing, vol. 32, pp. 3185\u20153199, Jun., 2024. DOI: 10.1109\/TASLP.2024.3410879<\/li>\n\n\n\n<li><strong>Daiki Mori<\/strong>, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;<a href=\"https:\/\/asmp-eurasipjournals.springeropen.com\/articles\/10.1186\/s13636-024-00360-8\">Recognition of target domain Japanese speech using language model replacement<\/a>,&#8221; EURASIP Journal on Audio, Speech and Music Processing, Article number: 40 (2024), 14 pages, 2024. (DOI: 10.1186\/s13636-024-00360-8)<\/li>\n\n\n\n<li>Ryota Nishimura, Takaaki Uno, Taiki Yamamoto, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, &#8220;<a href=\"http:\/\/Detection of Arbitrary Wake Words by Coupling a Phoneme Predictor and a Phoneme Sequence Detector\">Detection of Arbitrary Wake Words by Coupling a Phoneme Predictor and a Phoneme Sequence Detector<\/a>,&#8221; APSIPA Transactions on Signal and Information Processing,Vol. 13: No. 1, e14, pp. 1-21, Aug., 2024.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Kaito Takahashi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Akio Kobayashi, <strong>Norihide Kitaoka<\/strong>, &#8220;Domain Adaptation by Alternating Learning of Acoustic and Linguistic Information for Japanese Deaf and Hard-of-Hearing People,&#8221; Proc. APSIPA ASC 2024, (7 pages), Dec., 2024.<\/li>\n\n\n\n<li><strong>Takanori Kanai<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, \u201cPredicting Utterance-final Timing Considering Linguistic Features Using Wav2vec 2.0\u201d, ICAICTA2024, (5 pages) Sep., 2024.<\/li>\n\n\n\n<li><strong>Tamon Mikawa<\/strong>, Yasuhisa Fujii, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;Listening Head Motion Generation for Multimodal Dialog System,&#8221; ICAICTA2024, (6 pages) Sep., 2024.<\/li>\n\n\n\n<li><strong>Kaito Takahashi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Akio Kobayashi, <strong>Norihide Kitaoka<\/strong>, \u201cImproving Speech Recognition for Japanese Deaf and Hard-of-Hearing People by Replacing Encoder Layers\u201d, ICAICTA2024, (5 pages) Sep., 2024.<\/li>\n\n\n\n<li>Makoto Hotta, Chee Siang Leow, <strong>Norihide Kitaoka<\/strong>, Hiromitsu Nishizaki, &#8220;Evaluation of Speech Translation Subtitles Generated by ASR with Unnecessary Word Detection,&#8221; Proc. GCCE, pp. 835-839, Oct., 2024.<\/li>\n\n\n\n<li><strong>Tatsunari Takagi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Text-only Domain Adaptation for CTC-based Speech Recognition through Substitution of Implicit Linguistic Information in the Search Space,&#8221; Proc. INTERSPEECH, pp. 287-291, Sep., 2024.<\/li>\n\n\n\n<li><strong>Keigo Hojo<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Boosting CTC-based ASR using inter-layer attention-based CTC loss,&#8221; Proc. INTERSPEECH, pp. 2860-2864, Sep., 2024.<\/li>\n\n\n\n<li>Kazuya Tsubokura, Takuya Takeda, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, &#8220;Dialog Breakdown Recovery Strategies Based on User Personality,&#8221; Proc. of The 14th International Workshop on Spoken Dialogue Systems Technology (IWSDS2024), Mar., 2024.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Tamon Mikawa<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, \u201cConstruction and Analysis of a Multimodal Japanese Chat Dialogue Dataset for Building a Multimodal Dialogue System,\u201d Symposium on Interactive Systems, Nov. 2024.<\/li>\n\n\n\n<li><strong>Toshimitsu Sakai<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Speech recognition without the need for speech segment detection with noise and silence labelling,&#8221; Acoustical Society of Japan Autumn Research Conference, 2 pages, Sep. 2024.<\/li>\n\n\n\n<li><strong>Takanori Kanai<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;Estimating the end time of input utterances to a spoken dialogue system considering linguistic features using wav2vec 2.0,&#8221; Acoustical Society of Japan Autumn Research Conference, 2 pages, Sep. 2024.<\/li>\n\n\n\n<li><strong>Kaito Takahashi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Akio Kobayashi, <strong>Norihide Kitaoka<\/strong>, &#8220;Improving Speech Recognition Accuracy with Encoder Layer Substitution in Deaf Speech,&#8221; Acoustical Society of Japan Autumn Research Conference, 2 pages, Sep. 2024.<\/li>\n\n\n\n<li><strong>Keigo Hojo<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ota, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Improving the performance of CTC speech recognition models by weighting the encoder layer using attention mechanisms,&#8221; Symposium on Sound Science, Jun. 2024.<\/li>\n\n\n\n<li><strong>YANG TINGCHENG<\/strong>, Yuya Hosoda, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Japanese Pronunciation Scoring of L2 Learners Based on LSTM,&#8221; IPSJ 86th National Conference, 4R-08, Mar. 2024.<\/li>\n\n\n\n<li>Kazuya Tsubokura, Mai Okada, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, &#8220;Collection and Analysis of Dialogue Break Repair Corpus &#8211; Towards Repair Sentence Generation Considering User&#8217;s Individual Characteristics and Relationship with the System,&#8221; 30th Annual Conference of the Association for Natural Language Processing, pp. 1436-1440 (P5-18), Mar. 2024.<\/li>\n\n\n\n<li>Tomoya Okada, Yurie Iribe, Katsunori Yokoi, Akinori Nakamura, <strong>Norihide Kitaoka<\/strong>, Masao Katsuno, &#8220;Analysis of the effects of dementia etiologic agents on conversational content and prediction of pre-onset Alzheimer&#8217;s disease,&#8221; 30th Annual Conference of the Association for Language Processing, pp. 571-575 (P2-22), Mar. 2024.<\/li>\n\n\n\n<li>Yuki Nagae, Tomoya Okada, Yurie Iribe, Katsunori Yokoi, Akinori Nakamura, <strong>Norihide Kitaoka<\/strong>, Masao Katsuno, &#8220;Detecting mild cognitive impairment based on a topic model of free conversation,&#8221; 30th Annual Conference of the Association for Language Processing, pp. 472-476 (P2-4), Mar. 2024.<\/li>\n\n\n\n<li>Yuka Maruyama, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>,  Katsunori Yokoi, Masao Katsuno, &#8220;Comparative analysis of phonemes and syllables in conversational speech of Parkinson&#8217;s disease patients,&#8221; Spring Meeting of the Acoustical Society of Japan, 2-P-18, Mar. 2024.<\/li>\n\n\n\n<li><strong>Tatsunari Takagi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Domain adaptation by substituting linguistic information in streaming speech recognition using CTC,&#8221; in Spring Meeting of the Acoustical Society of Japan, 1-Q-22, Mar. 2024.<\/li>\n\n\n\n<li><strong>Takanori Kanai<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;Advance Estimation of Speech Termination Time for Smooth Spoken Dialogue Systems,&#8221; Spring Meeting of the Acoustical Society of Japan, 2-P-7, Mar. 2024.<\/li>\n\n\n\n<li><strong>Kaito Takahashi<\/strong>, <strong>Takahiro Kinouchi<\/strong>, <strong>Tatsunari Takagi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ota, Akio Kobayashi, <strong>Norihide Kitaoka<\/strong>, &#8220;Evaluation of Speech Recognition Based on Self-Supervised Learning in Speech of the Deaf,&#8221; Spring Meeting of the Acoustical Society of Japan, 1-Q-23, Mar. 2024.<\/li>\n\n\n\n<li><strong>Tamon Mikawa<\/strong>, Yasushi Fujii, Kengo Ota, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Analysis of human head movements in response to a dialogue partner&#8217;s voice in a multimodal chat dialogue dataset,&#8221; in Spring Meeting of the Acoustical Society of Japan, 2-P-8, Mar. 2024.<\/li>\n\n\n\n<li><strong>Li Chengfeng<\/strong>, <strong>Tatsunari Takagi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Building an Adaptive Speech Recognition Model for Electronic Medical Record Entry Based on Data Extension with ChatGPT,&#8221; Spring Meeting of the Acoustical Society of Japan, 1-Q-24, Mar. 2024.<\/li>\n\n\n\n<li><strong>Takahiro Kinouchi<\/strong>, Atsunori Ogawa, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ota, <strong>Norihide Kitaoka<\/strong>, &#8220;Domain adaptation using only large-scale speech data for speech recognition based on multilingual SSL models,&#8221; in Spring Meeting of the Acoustical Society of Japan, 1-2-2, Mar. 2024.<\/li>\n\n\n\n<li><strong>Takumi Shine<\/strong>, <strong>Takahiro Kiuchi<\/strong>, <strong>Yuko Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Improving the accuracy of speech recognition for the elderly by combining an age estimation task,&#8221; Spring Meetings of the Acoustical Society of Japan, 1-2-5, Mar. 2024.<\/li>\n\n\n\n<li>Rintaro Imamoto, Ryota Nishimura, Kengo Ota, <strong>Norihide Kitaoka<\/strong>, &#8220;Construction and evaluation of a real-time spoken dialogue system incorporating a model of aizuchi generation and speaker alternation,&#8221; Spring Meeting of Acoustical Society of Japan, 2-P-6, Mar. 2024.<\/li>\n\n\n\n<li>Jotaro Emoto, Ryota Nishimura, Kengo Ota, <strong>Norihide Kitaoka<\/strong>, &#8220;Development of a real-time VAD-less speech recognition model with noise and silence rejection,&#8221; Spring Meeting of Acoustical Society of Japan, &nbsp;1-Q-14, Mar. 2024.<\/li>\n\n\n\n<li>Yoshinori Fukunaga, Ryota Nishimura, Kengo Ota, <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of a phase selection model for natural spoken dialogue system using deep learning,&#8221; Spring Meeting of Acoustical Society of Japan, &nbsp;2-P-4, Mar. 2024.<\/li>\n\n\n\n<li><strong>Meiko Fukuda<\/strong>, Ryota Nishimura, Yurie Iribe, Kazuhiro Yamamoto, <strong>Norihide Kitaoka<\/strong>, &#8220;EARS: Construction of a corpus of Japanese very elderly people&#8217;s speech,&#8221; Spring Meeting of the Acoustical Society of Japan, 1-2-4, Mar. 2024.<\/li>\n\n\n\n<li><strong>Tatsunari Takagi<\/strong> , <strong>Yukoh Wakabayashi<\/strong>, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Replacement of implicit linguistic information within beam-search decoding in CTC speech recognition models,&#8221; SPEASIP Workshop 2024, pp. 1-6, Mar. 2024.<\/li>\n\n\n\n<li><strong>Sota Hosoi<\/strong>, <strong>Takahiro Kinouchi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Intermediate Speech Synthesis between Two Speakers Using x-vector Speaker Space,&#8221; SPEASIP Workshop 2024, pp. 1-6, Mar. 2024.<\/li>\n\n\n\n<li><strong>Takahiro Kinouchi<\/strong>, Atsunori Ogawa, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ota, <strong>Norihide Kitaoka<\/strong>, &#8220;Domain adaptation of speech recognition models using only SSL-based speech data,&#8221; SPEASIP Workshop 2024, pp. 1-6, Mar. 2024.<\/li>\n\n\n\n<li><strong>Kaito Takahashi<\/strong>, <strong>Takahiro Kinouchi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ota, Akio Kobayashi, <strong>Norihide Kitaoka<\/strong>, &#8220;Evaluation of Speech Recognition for the Deaf and Hard of Hearing by Speaker Adaptation,&#8221; SPEASIP Workshop 2024, pp. 1-6, Mar. 2024.<\/li>\n\n\n\n<li><strong>Keigo Hojo<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ota, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Integrating Multiple Speech Recognition Models for High Accuracy in Speech Recognition Systems,&#8221; SPEASIP Workshop 2024, pp. 1-6, Mar. 2024.<\/li>\n\n\n\n<li><strong>Takumi Shine<\/strong>, <strong>Takahiro Kinouchi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Improving the accuracy of elderly speech recognition by multitask learning with age information,&#8221; SPEASIP Workshop 2024, pp. 1-6, Mar. 2024.<\/li>\n\n\n\n<li><strong>Ryo Maejima<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Construction and Evaluation of a Batch Speech Input Interface for Electronic Medical Records Using Large-scale Language Models,&#8221; SPEASIP Workshop 2024, pp. 1-6, Mar. 2024.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2023\">2023<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Journal Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Kazuya Tsubokura, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, <a href=\"https:\/\/www.tandfonline.com\/doi\/full\/10.1080\/01691864.2023.2279610\">\u201cAnalysis of the Relationship between User Response to Dialog Breakdown and Personality Traits,\u201d Advanced Robotics, Vol. 37, Issue 21, pp.1-10 ,Nov., 2023. (DOI: 10.1080\/01691864.2023.2279610)<\/a><\/li>\n\n\n\n<li><strong>Yukoh Wakabayashi<\/strong>, Kouei Yamaoka, and Nobutaka Ono, <a href=\"https:\/\/ieeexplore.ieee.org\/document\/10141851\">&#8220;Sound field interpolation for rotation-invariant multichannel array signal processing,&#8221; IEEE Transactions on Audio, Speech, and Language Processing, vol. 31, pp. 2286\u20152298, Jun. 2023. DOI: 10.1109\/TASLP.2023.3282098<\/a><\/li>\n\n\n\n<li>Katsunori Yokoi, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, Takashi Tsuboi, Keita Hiraga, Yuki Satake, Makoto Hattori, Yasuhiro Tanaka, Maki Sato, Akihiro Hori, Masahisa Katsuno, <a href=\"https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1353802023001347\">&#8220;Analysis of spontaneous speech in Parkinson\u2019s disease by natural language processing,&#8221; Parkinsonism and Related Disorders, Vol. 112, pp. 1-6<\/a>, <a href=\"https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1353802023001347\">April, 2023. (DOI: 10.1016\/j.parkreldis.2023.105411)<\/a><\/li>\n\n\n\n<li>Binh Thien Nguyen,&nbsp;<strong>Yukoh Wakabayashi<\/strong>, Kenta Iwai, and Takanobu Nishiura, <a href=\"https:\/\/ieeexplore.ieee.org\/document\/10106067\">&#8220;Inter-frequency phase difference for phase reconstruction using deep neural networks and maximum likelihood,&#8221; IEEE Transactions on Audio, Speech, and Language Processing,&nbsp;vol. 31, pp. 1667\u20151680, Apr. 2023. DOI: 10.1109\/TASLP.2023.3268577<\/a><\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Koharu Horii<\/strong>, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Language Modeling for Spontaneous Speech Recognition Based on Disfluency Labeling and Generation of Disfluent Text,&#8221; APSIPA ASC 2023, pp. 1867-1872, Nov. 2023.<\/li>\n\n\n\n<li><strong>Keigo Hojo<\/strong>, <strong>Daiki Mori<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Kengo Ohta, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Combining Multiple End-To-End Speech Recognition Models Based on Density Ratio Approach,&#8221; APSIPA ASC 2023, pp. 2250-2255, Nov. 2023.<\/li>\n\n\n\n<li><strong>Nagito Shione<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of Automatic Speech Recognition Model That Recognizes Linguistic Information and Verbal\/Non-Verbal Phenomena,&#8221; APSIPA ASC 2023, pp. 2282-2287, Nov. 2023.<\/li>\n\n\n\n<li><strong>Tatsunari Takagi<\/strong>, <strong>Norihide Kitaoka<\/strong>, Atsunori Ogawa, <strong>Yukoh Wakabayashi<\/strong>, &#8220;Streaming End-To-End ASR Using CTC Decoder and DRA for Linguistic Information Substitution,&#8221; APSIPA ASC 2023, pp. 1768-1772, Nov. 2023.<\/li>\n\n\n\n<li><strong>Ryo Maejima<\/strong> and <strong>Norihide Kitaoka<\/strong>, &#8220;Speech recognition interface for updating electronic medical records with automatic itemization,&#8221; ICAICTA2023, (5 pages) Oct., 2023.<\/li>\n\n\n\n<li><strong>Takahiro Kinouchi<\/strong>, Atsunori Ogawa, <strong>Yukoh Wakabayashi<\/strong> and <strong>Norihide Kitaoka<\/strong>, &#8220;Domain adaptation with a non-parallel target domain corpus,&#8221; ICAICTA2023, (6 pages) Oct., 2023.<\/li>\n\n\n\n<li><strong>Tatsunari Takagi<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Atsunori Ogawa and <strong>Norihide Kitaoka<\/strong>, &#8220;Domain Adaptation Using Density Ratio Approach and CTC Decoder for Streaming Speech Recognition,&#8221; ICAICTA2023, (5 pages) Oct., 2023.<\/li>\n\n\n\n<li><strong>Shione Nagito<\/strong>, <strong>Yukoh Wakabayashi<\/strong> and <strong>Norihide Kitaoka<\/strong>, &#8220;Automatic Speech Recognition Using Linguistic and Verbal\/Non-verbal Information,&#8221; ICAICTA2023, (6 pages) Oct., 2023.<\/li>\n\n\n\n<li>Aito Nakata, Ryota Nishimura, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, &#8220;Development of a Model for Predicting Timing of Back-Channel in a Real-Time Spoken Dialog System,&#8221; GCCE2023, (to appear), Oct., 2023.<\/li>\n\n\n\n<li>Kazuya Tsubokura, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, &#8220;Relationships Between Gender, Personality Traits and Features of Multi-Modal Data to Responses to Spoken Dialog Systems Breakdown,&#8221; INTERSPEECH2023, pp. 2713-2717, Oct., 2023. (DOI: 10.21437\/Interspeech.2023-1267)<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Nagito Shione<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, and <strong>Norihide Kitaoka<\/strong>, &#8220;Proposal of a Speech Recognition System for Detecting Verbal and Nonverbal Phenomena, Acoustical Society of Japan Autumn Meeting, 2-Q-3, Sep. 2023.<\/li>\n\n\n\n<li><strong>Takahiro Kinouchi<\/strong>, Atsunori Ogawa, <strong>Yukoh Wakabayashi<\/strong>, <strong>Norihide Kitaoka<\/strong>, &#8220;Domain adaptation of speech recognition models using only SSL-based speech data,&#8221; Acoustical Society of Japan Autumn Meeting, 2-Q-9, Sep. 2023.<\/li>\n\n\n\n<li><strong>Keigo Hojo<\/strong>, <strong>Daiki Mori<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Atsunori Ogawa, and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of Robust Speech Recognition System by Integration of Multiple Speech Recognition Models Based on Density Ratio Approach,&#8221; Acoustical Society of Japan Autumn Meeting, 2-Q-10, Sep. 2023.<\/li>\n\n\n\n<li>Kaito Kofuji, Ryota Nishimura, Kengo Ota, and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction and evaluation of a multilingual speech synthesis model using monolingual speakers,&#8221; Acoustical Society of Japan Autumn Meeting, 2-Q-37, Sep. 2023.<\/li>\n\n\n\n<li><strong>Tamon Mikawa<\/strong>, Daishi Yamaoka, and <strong>Norihide Kitaoka<\/strong>, &#8220;Evaluation of Robust Speech Recognition Models for Overlap Using End-to-end Models,&#8221; Acoustical Society of Japan Autumn Meeting, 3-Q-1, Sep. 2023.<\/li>\n\n\n\n<li><strong>Keigo Hojo<\/strong>, Sousuke Kawahigashi, and <strong>Norihide Kitaoka<\/strong>, &#8220;Data expansion for speech recognition by sentence generation and speech synthesis focusing on words that are difficult to recognise,&#8221; Acoustical Society of Japan Autumn Meeting, 3-Q-2, Sep. 2023.<\/li>\n\n\n\n<li><strong>Tatsunari Takagi<\/strong>, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, &#8220;Domain adaptation based on Density Ratio Approach for streamable speech recognition using CTC decoder,&#8221; Acoustical Society of Japan Autumn Meeting, 3-Q-6, Sep. 2023.<\/li>\n\n\n\n<li><strong>Ryo Maejima<\/strong>, <strong>Daiki Mori<\/strong>, <strong>Yukoh Wakabayashi<\/strong> and <strong>Norihide Kitaoka<\/strong>, &#8220;Building a medical electronic health record item-specific automatic entry interface using speech recognition,&#8221; FIT2023, Sep. 2023.<\/li>\n\n\n\n<li><strong>Ryo Maejima<\/strong> and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of an Automatic Input Interface for Medical Electronic Medical Record Items Using Continuous Speech Recognition and ChatGPT,&#8221; Tokai Section Joint Conference, Aug, 2023.<\/li>\n\n\n\n<li>Yuki Nagae, Tomoya Okada, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, Katsumi Yokoi and Masahiro Katsuno, &#8220;Analysis of linguistic features extracted from free conversation speech of dementia patients,&#8221; Tokai Section Joint Conference, Aug, 2023.<\/li>\n\n\n\n<li><strong>Tatsunari Takagi<\/strong>, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, &#8220;Streaming End-to-End Speech Recognition Using a CTC Decoder to Replace Implicit Linguistic Information,&#8221; Acoustics Symposium, Jun. 2023.<\/li>\n\n\n\n<li><strong>Takahiro Kinouchi<\/strong>, Atsunori Ogawa, <strong>Yukoh Wakabayashi<\/strong>, and <strong>Norihide Kitaoka<\/strong>, &#8220;Domain adaptation of speech recognition models based on self-supervised learning using target domain speech,&#8221; Acoustics Symposium, Jun. 2023.<\/li>\n\n\n\n<li><strong>Nagito Shione<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of a speech recognition model for simultaneous recognition of linguistic information and verbal and non-verbal phenomena,&#8221; Acoustics Symposium, Jun. 2023.<\/li>\n\n\n\n<li><strong>Ryo Maejima<\/strong>, <strong>Daiki Mori<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, and <strong>Norihide Kitaoka, <\/strong>&#8220;Construction of a Language Model for Speech Recognition Based on Sentence Generation for Small Training Data Domains,&#8221; SPEASIP Workshop, Mar. 2023.<\/li>\n\n\n\n<li><strong>Nagito Shione<\/strong>, <strong>Yukoh Wakabayashi <\/strong>and <strong>Norihide Kitaoka, <\/strong>&#8220;A study of speech recognition models with linguistic and non-linguistic information tags,&#8221; SPEASIP Workshop, Mar. 2023.<\/li>\n\n\n\n<li>Tomohiro Takahashi, Hiruma Kinoshita, <strong>Yukoh Wakabayashi, <\/strong>Junki Ono, Jun Honda, Seiji Fukuma and Hiroshi Nakagawa, &#8220;Traffic Monitoring by Sound Based on Learning Data Obtained by Traffic Counter,&#8221; Proceedings of the Acoustical Society of Japan, 1-1-12, Mar., 2023.<\/li>\n\n\n\n<li>Kanato Uesaka, Shuto Kawauchi, Kouei Yamaoka, <strong>Yukoh Wakabayashi<\/strong>, Yuma Kinoshita, Junki Ono, Jun Noguchi, Kei Watanabe, Noritaka Ichido, Seiko Benner and Hidenori Yamasue, &#8220;Vocal classification of marmosets using machine learning and analysis of developmental vocal change based on it,&#8221; Proceedings of the Acoustical Society of Japan, 3-4-5, Mar., 2023.<\/li>\n\n\n\n<li><strong>Koharu Horii<\/strong>, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa and <strong>Norihide Kitaoka<\/strong>, &#8220;Language modelling based on non-fluent sentence generation by BERT for spontaneous speech recognition,&#8221; Proceedings of the Acoustical Society of Japan, 1-3-2, Mar. 2023.<\/li>\n\n\n\n<li>Ryuto Date, Kengo Ohta, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;Improvement of speech recognition accuracy in noise using lip information by deep learning,&#8221; Proceedings of the Acoustical Society of Japan, 1-3P-3, Mar., 2023.<\/li>\n\n\n\n<li><strong>Keigo Hojo<\/strong>, <strong>Daiki Mori<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Atsunori Ogawa and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of a robust speech recognition system by integration of multiple Encoder-Decoder speech recognition models,&#8221; Proceedings of the Acoustical Society of Japan, 1-3Q-3, Mar., 2023.<\/li>\n\n\n\n<li>Masato Sugiyama, Kengo Ohta, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;A real-time speaker alternation system for interrupted speech,&#8221; Proc. of the Acoustical Society of Japan, 2-3P-1, Mar. 2023.<\/li>\n\n\n\n<li>Kazuya Tsubokura, Takuya Takeda, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Relationship between user responses to dialogue breakdowns in spoken dialogue systems and individual characteristics,&#8221; 29th Annual Conference of the Association for Language Processing, pp. 2002-2006, Mar. 2023.<\/li>\n\n\n\n<li>Makoto Hotta, <strong>Koharu. Horii<\/strong>, <strong>Norihide Kitaoka<\/strong>, Hiromitsu Nishizaki, &#8220;Generation of easy-to-understand English subtitles based on shaping of Japanese speech recognition results,&#8221; IPSJ 85th National Convention, 1W-01, Mar. 2023.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2022\">2022<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Journal Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Meiko Fukuda, Ryota Nishimura, Hiromitsu Nishizaki, <strong>Koharu Horii<\/strong>, Yurie Iribe, Kazumasa Yamamoto, <strong>Norihide Kitaoka<\/strong>. &#8220;A new speech corpus of super-elderly Japanese for acoustic modeling,&#8221; Computer Speech &amp; Language, Vol. 77, pp. 1-22, 2022 (DOI: 10.1016\/j.csl.2022.101424)<\/li>\n\n\n\n<li>Ryota Nishimura, Raita Mori, Kengo Ohta, and <strong>Norihide Kitaoka<\/strong>, &#8220;A Topic Complementation Method to Input Speech by Matching Analysis Corresponding to Free Speech for Spoken Dialogue Systems,&#8221; Transactions of the Japanese Society for Artificial Intelligence, Vol. <em>37<\/em>, No. 3, pp. 1-13,. 2022.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h4 class=\"wp-block-heading marker fl\"><strong>Explanation<\/strong><\/h4>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Norihide Kitaoka<\/strong>, Ryota Nishimura, and Kengo Ohta, &#8220;Multimodal dialogue with photorealistic CG agents,&#8221; Journal of the Acoustical Society of Japan, Vol. 78, No. 5, pp. 257-264, May, 2022.<\/li>\n\n\n\n<li>Ikkoh Yamamoto, Hideki Sakano, and <strong>Norihide Kitaoka, <\/strong>&#8220;On the &#8216;uncanny valley&#8217; in spoken dialogue systems,&#8221; Journal of the Acoustical Society of Japan, Vol. 78, No. 5, pp. 245-248. , May, 2022.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Binh Thien Nguyen,&nbsp;<strong>Yukoh Wakabayashi<\/strong>, Geng Yuting, Kenta Iwai, and Takanobu Nishiura, &#8220;Von Mises mixture model-based DNN for sign indetermination problem in phase reconstruction,&#8221; Proc. APSIPA ASC 2022, pp. 958\u2015962, Chiang Mai, Nov., 2022.<\/li>\n\n\n\n<li>Yui Kuriki, Taishi Nakashima, Kouei Yamaoka, Natsuki Ueno,&nbsp;<strong>Yukoh Wakabayashi<\/strong>, Nobutaka Ono, and Ryo Sato, &#8220;Efficient low-latency convolution with uniform filter partition and its evaluation on real-time blind source separation,&#8221; Proc. APSIPA ASC 2022, pp. 766\u2015770, Chiang Mai, Nov., 2022.<\/li>\n\n\n\n<li>Kenta Yamada, Yoshiki Masuyama,&nbsp;<strong>Yukoh Wakabayashi<\/strong>, and Nobutaka Ono, &#8220;Simultaneous frequency estimation for three or more sinusoids based on sinusoidal constraint differential equation,&#8221; Proc. APSIPA ASC 2022, pp. 976\u2015979, Chiang Mai, Nov., 2022.<\/li>\n\n\n\n<li>Meiko Fukuda, Masakazu Sugiyama, Ryota Nishimura, Yurie Iribe, Kazumasa Yamamoto, <strong>Norihide Kitaoka<\/strong>, &#8220;A corpus-based analysis of age-related change in the acoustic features of elderly to super elderly speech,&#8221; Proc. Oriental-COCOSDA, (6 pages), Nov., 2022.<\/li>\n\n\n\n<li>Haruki Nammoku, Kouei Yamaoka, Taishi Nakashima,&nbsp;<strong>Yukoh Wakabayashi<\/strong>, and Nobutaka Ono, &#8220;Analysis and source separation of overlapping speech using corpus of everyday Japanese conversation,&#8221; Proc. ICA, Gyeongju, Oct., 2022.<\/li>\n\n\n\n<li>Kazuya Tsubokura, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, &#8220;Dialog Breakdown Detection Using Multimodal Features for Non-Task-Oriented Dialog Systems,&#8221; GCCE2022, pp. 359-363, Oct., 2022.<\/li>\n\n\n\n<li>Shuming Luan,&nbsp;<strong>Yukoh Wakabayashi<\/strong>, and Tomoki Toda, &#8220;Modified sound field interpolation method for rotation-robust beamforming with unequally spaced circular microphone array,&#8221; Proc. EUSIPCO 2022, pp. 344\u2015348, Belgrade, Sep., 2022.<\/li>\n\n\n\n<li><strong>Daiki Mori<\/strong>, Kengo Ohta, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, \u201cImplicit language information replace method in Japanese encoder-decode ASR model,&#8221; ICAICTA-2022, Sep., 2022.<\/li>\n\n\n\n<li><strong>Takahiro Kinouchi<\/strong>, <strong>Norihide Kitaoka<\/strong>, \u201cA response generation method of chat-bot system using input formatting and reference resolution,&#8221; ICAICTA-2022, Sep., 2022.<\/li>\n\n\n\n<li><strong>Koharu Horii<\/strong>, Meiko Fukuda, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>. &#8220;End-to-End Spontaneous Speech Recognition Using Disfluency Labeling,&#8221; (5 pages), Proc. INTERSPEECH 2022, Sep., 2022.<\/li>\n\n\n\n<li>Meiko Fukuda, Maina Umezawa, Ryota Nishimura, Yurie Iribe, Kazumasa Yamamoto, <strong>Norihide Kitaoka<\/strong>, &#8220;Elderly Conversational Speech Corpus with Cognitive Impairment Test and Pilot Dementia Detection Experiment Using Acoustic Characteristics of Speech in Japanese Dialects,&#8221; Proc. LREC2022. pp. 1016-1022, Jun, 2022.<\/li>\n\n\n\n<li>Akio Kobayashi, Junji Onishi, Hiromitsu Nishizaki, <strong>Norihide Kitaoka<\/strong>, &#8220;End-to-End Speech to Braille Translation in Japanese,&#8221; ICCE2021, 2 pages, Jan., 2022.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Keigo Hojo<\/strong>, <strong>Daiki Mori<\/strong>, <strong>Yukoh Wakabayashi<\/strong>, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Multiple Encoder-Decoder Speech Recognition Model Integration Method Based on Density Ratio Approach,&#8221; 24th Symposium on Spoken Language and 9th Natural Language Processing. Dec. 2022.<\/li>\n\n\n\n<li>Akihiro Torii, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of Dialogue Failure Detector in Spoken Dialogue System,&#8221; Proceedings of 2022 Shikoku Branch Joint Conference of Institutes of Electrical, Electronics and Information Engineers, vol. 15?8, pp. 145?145, 2022.<\/li>\n\n\n\n<li>Akihiro Torii, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of Dialogue Failure Detector in Spoken Dialogue System,&#8221; Proceedings of 2022 Shikoku Branch Joint Conference of Institutes of Electrical, Electronics and Information Engineers, vol. 15?8, pp. 145?145, 2022.<\/li>\n\n\n\n<li>Kohyo Fukumura, Ryota Nishimura, and <strong>Norihide Kitaoka, &#8220;<\/strong>Extension of Chat Dialogue Topics by BERT,&#8221; Proceedings of the 2022 Shikoku Branch Joint Conference of Institutes of Electrical, Electronics and Information Engineers, vol. 15?9, pp. 146?146, 2022.<\/li>\n\n\n\n<li>Binh Thien Nguyen, <strong>Yukoh Wakabayashi<\/strong>, Yuting GENG, Kenta Iwai, and Takanobu Nishiura, &#8220;Two-stage phase reconstruction using inter-frequency phase difference,&#8221; Proceedings of the Acoustical Society of Japan, 1-Q-11, Sep. 2022.<\/li>\n\n\n\n<li>Guanzang Ren, Daishi Nakajima, <strong>Yukoh Wakabayashi<\/strong>, and Junki Ono, &#8220;Self-rotation angle estimation of circular microphone array based on auxiliary function method,&#8221; Proceedings of the Acoustical Society of Japan, 1-R-29, Sep. 2022.<\/li>\n\n\n\n<li>Daishi Nakajima, <strong>Yukoh Wakabayashi<\/strong>, and Junki Ono, &#8220;Rotationally robust blind source separation of circular microphone arrays using sound field interpolation,&#8221; Proceedings of the Acoustical Society of Japan, 1-Q-23, Sep. 2022.<\/li>\n\n\n\n<li><strong>Koharu Horii<\/strong>, Meiko Fukuda, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, and <strong>Norihide Kitaoka<\/strong>, &#8220;Evaluation of an end-to-end non-fluent shaped speech recognition system by dialogue speech,&#8221; Proceedings of the Acoustical Society of Japan, 2-8-5, Sep. 2022.<\/li>\n\n\n\n<li><strong>Daiki Mori<\/strong>, Kengo Ohta, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;Design of an Encoder-Decoder speech recognition model augmented with out-of-domain acoustic information,&#8221; Proceedings of the Acoustical Society of Japan, 2-Q-26, Sep. 2022.<\/li>\n\n\n\n<li>Kazuya Tsubokura, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Individual user differences during dialogue breakdown in a multimodal dialogue system,&#8221; Proceedings of the Acoustical Society of Japan, 3-Q-13, Sep. 2022.<\/li>\n\n\n\n<li>Tomoya Okada, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Detection of suspected dementia from chat dialog speech using BERT,&#8221; Proceedings of the Acoustical Society of Japan, 3-Q-29, Sep. 2022.<\/li>\n\n\n\n<li>Meiko Fukuda, Masakazu Sugiyama, Ryota Nishimura, Yurie Iribe, Kazunori Yamamoto and <strong>Norihide Kitaoka<\/strong>, &#8220;Analysis of acoustic features of elderly speech using a corpus of very elderly people and S-JNAS,&#8221; Proceedings of the Acoustical Society of Japan, 3-Q-32, Sep. 2022.<\/li>\n\n\n\n<li>Yuka Maruyama, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, Katsunori Yokoi, and Masao Katsuno, &#8220;Analysis of acoustic features based on severity of Parkinson&#8217;s disease,&#8221; Proceedings of the Acoustical Society of Japan, 3-Q-43, Sep. 2022.<\/li>\n\n\n\n<li>Yuka Maruyama, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, Katsunori Yokoi, Masao Katsuno, &#8220;Parkinson&#8217;s disease detection from short speech utterances using acoustic information,&#8221; Proceedings of the Acoustical Society of Japan, 2-3P-10, Mar. 2022.<\/li>\n\n\n\n<li><strong>Mori, Daiki<\/strong>, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, and <strong>Norihide Kitaoka, <\/strong>&#8220;Design of an End-to-End Speech Recognition Model with Extra-Task Acoustic Information,&#8221; Proceedings of the Acoustical Society of Japan, 2-3Q-2, Mar. 2022.<\/li>\n\n\n\n<li><strong>Koharu Horii<\/strong>, Meiko Fukuda, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa and <strong>Norihide Kitaoka<\/strong>, &#8220;Yodomi-shaping end-to-end speech recognition using non-fluent labels,&#8221; Proceedings of the Acoustical Society of Japan, 1-3-5, Mar. 2022.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2021\">2021<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Journal Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Zolzaya Byambadorj\uff0cRyota Nishimura\uff0cAltangerel Ayush, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, &#8220;Text-to-speech system for low-resource language using cross-lingual transfer learning and data augmentation,&#8221; EURASIP Journal on Audio, Speech, and Music Processing, 2021:42, 20 pages, Dec., 2021. (DOI: 10.1186\/s13636-021-00225-4) .<\/li>\n\n\n\n<li>Zolzaya Byambadorj\uff0cRyota Nishimura\uff0cAltangerel Ayush, <strong>Norihide Kitaoka<\/strong>, &#8220;Normalization of Transliterated Mongolian Words Using Seq2Seq Model with Limited Data,&#8221; ACM Transactions on Asian and Low-Resource Language Information Processing, No. 103,, pp. 1-19, Nov. , 2021.<\/li>\n\n\n\n<li>Kego Ohta\uff0cRyota Nishimura\uff0c<strong>Norihide Kitaoka<\/strong>, &#8220;Response Type Selection for Chat-like Spoken Dialog Systems Based on LSTM and Multi-task Learning,&#8221; SPEECH COMMUNICAGTION, vol. 133, pp. 23-30, Oct., 2021.<\/li>\n\n\n\n<li>Hayato Ishihara,  Yurie Iribe and <strong>Norihide Kitaoka<\/strong>, &#8220;Detection of Suspected Dementia from Chat Dialogues Focusing on Engagement Distance,&#8221; IEICE Transactions D, Vol. J104-D,No. 04, pp. 357-367, Apr. 2021.<\/li>\n\n\n\n<li><strong>Norihide Kitaoka<\/strong>; Bohan Chen; Yuya Obashi, &#8220;Dynamic out-of-vocabulary word registration to language model for speech recognition,&#8221; EURASIP Journal on Audio, Speech, and Music Processing, 2021:4, (8 pages), 2021. (DOI: 10.1186\/s13636-020-00193-1)<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Zolzaya Byambadorj, Ryota Nishimura, Altangerel Ayush, Kengo Ohta, <strong>Norihide Kitaoka<\/strong>, &#8220;Multi-speaker TTS system for low-resource language using cross-lingual transfer learning and data augmentation,&#8221; Proc. APSIPA ASC 2021, pp. 849-853, 2021.<\/li>\n\n\n\n<li>Daiki Mori, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Advanced language model fusion method for encoder-decoder model in Japanese speech,&#8221; Proc. APSIPA ASC 2021, pp. 503-510, 2021.<\/li>\n\n\n\n<li><strong>Koharu Horii<\/strong>, Meiko Fukuda, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, <strong>Norihide Kitaoka<\/strong>, &#8220;End-to-end spontaneous speech recognition using hesitation labeling,&#8221; Proc. APSIPA ASC 2021, pp. 1077-1081, 2021.<\/li>\n\n\n\n<li>Akio Kobayashi, Keiichi Yasu, Hiromitsu Nishizaki, <strong>Norihide Kitaoka<\/strong>, &#8220;Corpus Design and Automatic Speech Recognition for Deaf and Hard-Of-Hearing People,&#8221; GCCE2021, pp. 17-18, Oct., 2021.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h4 class=\"wp-block-heading marker fl\"><strong>Explanation<\/strong><\/h4>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Susumu Ohsuga, Godai Tanaka, Ayana Nabekura, Hiroyuki Fujii, Ryota Nakano, Ryota Watanabe, TELYUKA, Kengo Ohta, Ryota Nishimura, <strong>Norihide Kitaoka, &#8220;<\/strong>Multimodal Agent &#8220;Saya&#8221; for Next Generation Mobility. ,&#8221; Automotive Technology, Vol. 75, No. 9, pp. 109-109, Sep. 2021.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Koharu Horii<\/strong>, Meiko Fukuda, Kengo Ohta, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;End-to-End Speech Recognition Considering Stammering,&#8221; 19th Workshop on Informatics (WiNF2021), S-5-2, Nov. 2021.<\/li>\n\n\n\n<li><strong>Takahiro Kiuchi <\/strong>and <strong>Norihide Kitaoka<\/strong>, &#8220;A chat response generation system using speech-formatted dialogue history,&#8221; 19th Workshop on Informatics (WiNF2021), S-5-3, Nov. 2021.<\/li>\n\n\n\n<li><strong>Daiki Mori<\/strong>, Kengo Ohta, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;A replacement method for implicit linguistic information in the Encoder-Decoder speech recognition model,&#8221; 19th Workshop on Informatics (WiNF2021), S-5-5, Nov. 2021.<\/li>\n\n\n\n<li><strong>Daiki<\/strong> <strong>Mori<\/strong>, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa and <strong>Norihide Kitaoka<\/strong>, &#8220;A method for replacing implicit linguistic information in the Encoder-Decoder speech recognition model,&#8221; Proc.<\/li>\n\n\n\n<li><strong>Koharu Horii<\/strong>, Meiko Fukuda, Kengo Ohta, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;End-to-End Speech Recognition of Free Speech Considering Stuttering,&#8221; Proceedings of the Acoustical Society of Japan, 1-3-1, Sep. 2021.<\/li>\n\n\n\n<li>Zolzaya Byambadorj, Ryota Nishimura, Altangerel Ayush, Kengo Ohta and <strong>Norihide Kitaoka<\/strong>, &#8220;Cross-lingual, multi-speaker text-to-speech synthesis for low resource languages,&#8221; Proceedings of the Acoustical Society of Japan, 1-3-7, Sep. 2021.<\/li>\n\n\n\n<li>Narangerel Purevdorj, Ryota Nishimura, Altangerel Ayush, Kengo Ohta and <strong>Norihide Kitaoka<\/strong>, &#8220;How language similarity affects the Mongolian ASR using cross-lingual transfer learning,&#8221; Proceedings of the Acoustical Society of Japan, 2-3-7, Sep. 2021.<\/li>\n\n\n\n<li>Akio Kobayashi, Junji Ohnishi, Hiromitsu Nishizaki, and <strong>Norihide Kitaoka<\/strong>, &#8220;End-to-End Spoken Braille Translation of Readout Sentences,&#8221; Proceedings of the Acoustical Society of Japan, 2-3P-3, Sep. 2021.<\/li>\n\n\n\n<li>Meiko Fukuda, Ryota Nishimura, Hiromitsu Nishizaki, Yurie Iribe, Kazunori Yamamoto and <strong>Norihide Kitaoka<\/strong>, &#8220;Acoustic features of the very elderly in the EARS speech corpus of the very elderly,&#8221; Proceedings of the Acoustical Society of Japan, 2-3P-11, Sep. 2021.<\/li>\n\n\n\n<li>Ryota Nishimura, Takahiro Mori and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of a spoken dialogue system with real-time control using ROS,&#8221; Proceedings of the Acoustical Society of Japan, 2-3Q-4, Sep. 2021.<\/li>\n\n\n\n<li><strong>Norihide Kitaoka<\/strong>, Ryota Nishimura, Kengo Ohta, Teruyuki Ishikawa, Yuka Ishikawa, Ryota Nakano, Godai Tanaka, Ayana Nabekura, Tatsuya Sato, Ryota Watanabe and Susumu Ohsuga, &#8220;Response control in dialogue with the 3D CG agent Saya,&#8221; Proceedings of the Acoustical Society of Japan, 3-3-14, Sep . ., 2021.<\/li>\n\n\n\n<li>Katsunori Yokoi, Takashi Tsuboi, Makoto Hattori, Yuki Satake, Keita Hiraga, Yasuhiro Tanaka, Maki Sato, Akihiro Hori, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, Masao Katsuno, &#8220;Natural language processing of oral reading and conversation in Parkinson&#8217;s disease patients,&#8221; Parkinson&#8217;s and Movement Disorders Congress Programme, Abstracts. 15th p. 81, Jul. 2021.<\/li>\n\n\n\n<li><strong>Daiki<\/strong> <strong>Mori<\/strong>, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa and <strong>Norihide Kitaoka, <\/strong>&#8220;A replacement method for implicit linguistic information in end-to-end speech recognition models,&#8221; Acoustics Symposium, Jun. 2021.<\/li>\n\n\n\n<li><strong>Koharu Horii<\/strong>, Meiko Fukuda, Kengo Ohta, Ryota Nishimura and <strong>Norihide Kitaoka, <\/strong>&#8220;End-to-End Speech Recognition of Non-Fluent Speech by Stamina Labelling,&#8221; Acoustics Symposium, Jun. 2021.<\/li>\n\n\n\n<li><strong>Norihide Kitaoka<\/strong>, Ryota Nishimura, Kengo Ohta, Teruyuki Ishikawa, Yuka Ishikawa (TELYUKA), Ryota Nakano, Godai Tanaka, Ayana Nabekura, Tatsuya Sato, Ryota Watanabe, Susumu Osuga, &#8220;Construction of a multimodal dialogue system with photorealistic CG agents,&#8221; Proceedings, 1-2-6, Mar. 2021.<\/li>\n\n\n\n<li>Kengo Ohta, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of a Japanese End-to-End Speech Synthesis Server Considering Accented Phrases,&#8221; Proceedings of the Acoustical Society of Japan, 1-2-7, Mar. 2021.<\/li>\n\n\n\n<li>Akio Kobayashi, Keiichi Yasu, Hiromitsu Nishizaki, <strong>Norihide Kitaoka<\/strong>, &#8220;Collection of speech data of hearing-impaired people and evaluation by phoneme recognition,&#8221; Proceedings of the Acoustical Society of Japan, 2-2-4, Mar. 2021.<\/li>\n\n\n\n<li>Motoki Shimogasa, Hiromitsu Nishizaki and <strong>Norihide Kitaoka<\/strong>, &#8220;Data expansion using CycleGAN for speech recognition of very elderly people,&#8221; Proceedings of the Acoustical Society of Japan, 2-2P-6, Mar., 2021.<\/li>\n\n\n\n<li>Narangerel Purevdorj, Ryota Nishimura, Altangerel Ayush, Kengo Ohta and <strong>Norihide Kitaoka<\/strong>, &#8220;Building a low resource speech recogniser: Transfer learning and data augmentation,&#8221; Proceedings of the Acoustical Society of Japan, 3-2-9, Mar. 2021.<\/li>\n\n\n\n<li>Zolzaya Byambadorj, Ryota Nishimura, Altangerel Ayush, Kengo <strong>Ohta<\/strong> and Norihide <strong>Kitaoka<\/strong>, &#8220;Text to speech system for low resource languages by cross-lingual transfer learning and data augmentation,&#8221; Proceedings of the Acoustical Society of Japan, 3-2-10, Mar., 2021.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2020\">2020<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Journal Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Jiahao Chen, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;End-to-End Recognition of Streaming Japanese Speech Using CTC and Local Attention,&#8221; APSIPA Transactions on Signal and Information Processing, vol. 9, e 25, pp. 1-7, 2020.<\/li>\n\n\n\n<li><strong>Norihide Kitaoka<\/strong>, Eichi Seto, Ryota Nishimura, &#8220;Example phrase adaptation method for customized, example-based sialog system using user data and disributed word representations,&#8221; IEICE Trans. Inf. &amp; Syst., Vol. E103-D, No. 11, pp. 2332-2339, Nov., 2020.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Chee Siang Leow, Tomoaki Hayakawa, Hiromitsu Nishizaki, <strong>Norihide Kitaoka<\/strong>, &#8220;Development of a Low-Latency and Real-Time Automatic Speech Recognition System,&#8221; GCCE2020, pp. 464-467, Oct., 2020.<\/li>\n\n\n\n<li>Meiko Fukuda, Hiromitsu Nishizaki, Yurie Iribe, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;Improving speech recognition for the elderly: A new corpus of elderly Japanese speech and investigation of acoustic modeling for speech recognition,&#8221; Proc. LREC2020, 9 pages, Jun, 2020.<\/li>\n\n\n\n<li>Jiahao Chen, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;E2E Streaming Speech Recognition Using CTC and Local Attention,&#8221; Proc. NCSP&#8217;20, 4 pages, Mar. 2020.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Maina Umezawa, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Discrimination of elderly people with dementia based on spoken language information,&#8221; Shinrigaku Giho (SP2020-12, WIT2020-12), 6 pages, Oct. 2020.<\/li>\n\n\n\n<li>Meiko Fukuda, Yurie Iribe, Hiromitsu Nishizaki, Kazuhiro Yamamoto, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of EARS, a speech corpus for very elderly people, and preliminary study of its use for speech recognition,&#8221; IJI-KENPO, Vol. 2020-SLP-133 No. 6, pp. 1-6, Oct. 2020.<\/li>\n\n\n\n<li>Leo Qishan, Hiromitsu Nishizaki and <strong>Norihide Kitaoka, &#8220;<\/strong>Development and evaluation of a Kaldi-based low-latency real-time speech recognition system,&#8221; in Proc. of the Acoust.? -?????? , Sep&#8230; , 2020.<\/li>\n\n\n\n<li>Kaito Suzuki, Yurie Iribe, and <strong>Norihide Kitaoka, &#8220;<\/strong>Dialogue Breakdown Detection Using Facial Expression and Acoustic Information,&#8221; Proceedings of the Acoustical Society of Japan, 2-P1-4, pp. ????? -?????? , Sep&#8230; , 2020.<\/li>\n\n\n\n<li>Yamazaki, Taiga, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of an End-to-End Japanese Speech Synthesis System Capable of Expressing Emotions,&#8221; Proceedings of the Acoustical Society of Japan, 2-P1-2, pp. ????? -?????? , Sep&#8230; , 2020.<\/li>\n\n\n\n<li>Hayato Ishihara, Yurie Iribe, and <strong>Norirhide Kitaoka,<\/strong> &#8220;Detection of Dementia Tendency from Chat Dialogue Speech Considering Sentence Complexity,&#8221; Proceedings of the Acoustical Society of Japan, 2-P1-2, pp. ???? -?????? , Sep.. , 2020.<\/li>\n\n\n\n<li>Byambadorj Zolzaya, Ryota Nishimura, Ayush Altangerel, <strong>Norihide Kitaoka<\/strong>, &#8220;Normalisation of transliterated words using seq2seq model with spell checker,&#8221; 26th Annual Conference of the Association for Language Processing, E5-3, pp. 1133-1136, Mar. 2020.<\/li>\n\n\n\n<li>Chen, Jiahao, Ryota Nishimura and <strong>Norihide Kitaoka, &#8220;<\/strong>Streaming Speech Recognition Using Uni-directional LSTM and Local Attention,&#8221; Proceedings of the Acoustical Society of Japan, 2-Q-12, pp. 943-946, Mar. 2020.<\/li>\n\n\n\n<li>Meiko Fukuda, Hiromitsu Nishizaki, Yurie Iribe, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of a speech corpus for the elderly and analysis of age and dialect effects on speech recognition,&#8221; Proceedings of the Acoustical Society of Japan, 2-Q-13, pp. 947-950, Mar. 2020.<\/li>\n\n\n\n<li>Yuya Kobashi, Ryota Nishimura and <strong>Norihide Kitaoka<\/strong>, &#8220;Evaluation of a Language Model for Spoken Language Recognition Using Text Conversion from Written to Spoken Language,&#8221; Proc. of the Acoustical Society of Japan, 2-Q-13, pp. 951-954, Mar. 2020.<\/li>\n\n\n\n<li>Raita Mori, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;Spoken dialogue system with collocational analysis for free speech,&#8221; Proceedings of the Acoustical Society of Japan, 3-P-13, pp. 1023-1026, Mar. 2020.<\/li>\n\n\n\n<li>Kanta Kiyohara, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, <strong>&#8220;<\/strong>Construction of a Multimodal Geometric Problem Solving System by Integrated Understanding of Speech and Pointing,&#8221; IPSJ 82nd National Convention, 5F-03, pp. 4-5 &#8211; 4-6, Mar. 2020.<\/li>\n\n\n\n<li>Hayato Ishihara, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Dementia tendency detection from chat dialogues focusing on lexical and engaged structures,&#8221; The 82nd National Convention of Information Processing Society of Japan, 5ZE-03, pp. 4-459 &#8211; 4-460, Mar. 2020.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Book Chapters<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Norihide Kitaoka<\/strong>, Takuma Nakagawa, Ryota Nishimura, Yoshio Ishiguro, Shin&#8217;ichi Kojima and Shin Ohsuga, &#8220;A multimodal control system for autonomous vehicles using speech, gesture and gaze recognition,&#8221; pp. 101-111, in Vehicles, Drivers, and Safety, De Gruyter, 2020.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2019\">2019<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Invited talk<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Norihide Kitaoka<\/strong>, &#8220;Spoken and multimodal interfaces: Interaction systems with machines,&#8221; ICAICTA2019 (Keynote speech), Sep 2019.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Meiko Fukuda, Ryota Nishimura, Hiromitsu Nishizaki, Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, &#8220;A new corpus of elderly Japanese speech for acoustic modeling, and a preliminary investigation of dialect-dependent speech recognition,&#8221; Proc. Oriental-COCOSDA2019, 6 pages, Oct., 2019.<strong> (Best paper award)<\/strong><\/li>\n\n\n\n<li>Akihira Komatsu, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;Environmental sounds recognition with convolutional-LSTM,&#8221; GCCE2019, pp. 717-719, 2018.<\/li>\n\n\n\n<li>Yuya Obashi, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;Automatic conversion of written language into spoken language using a sequence-to-sequence model trained with a parallel corpus,&#8221; Proc. Oriental-COCOSDA2019, 5 pages, Oct., 2019.<\/li>\n\n\n\n<li>Taiki Yamamoto, Ryota Nishimura, Masayuki Misaki, <strong>Norihide Kitaoka<\/strong>, &#8220;Small-footprint magic word detection method using convolutional LSTM neural network,&#8221; Proc. INTERSPEECH2019, pp. 2035-2039, Sep. 2019.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Yuya Kobashi, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;Text conversion from written to spoken language for a language model for spoken speech recognition using a sequence-to-sequence model,&#8221; in Proc. 807-810, Sep. 2019.<\/li>\n\n\n\n<li>Chen Jiahao, Ryota Nishimura, and <strong>Norihide Kitaoka, &#8220;<\/strong>End-to-end streaming speech recognition using CTC and Attention,&#8221; Proceedings of the Acoustical Society of Japan, 1-P-16, pp. 871-874, Sep. 2019.<\/li>\n\n\n\n<li>Meiko Fukuda, Ryota Nishimura, Hiromitsu Nishizaki, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Speech corpus construction for elderly speech recognition and the effect of adaptation to dialects,&#8221; in Proceedings of the Acoustical Society of Japan, 1-P-17, pp. 875-878, Sep. 2019.<\/li>\n\n\n\n<li>Taiki Yamamoto, Ryota Nishimura, Masayuki Misaki, and <strong>Norihide Kitaoka, &#8220;<\/strong>Memory-saving Magic Word detection using Convolutional LSTM,&#8221; Proceedings of the Acoustical Society of Japan, 2-3-4, pp. 819-822, Sep. 2019.<\/li>\n\n\n\n<li>Akihisa Komatsu, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;Environmental sound recognition using CNN and CLSTM,&#8221; in Proceedings of the Acoustical Society of Japan, 2-Q-17 pp. 925-928, Sep. 2019.<\/li>\n\n\n\n<li>Shion Akimizu, Yurie Iribe, and <strong>Norihide Kitaoka, &#8220;<\/strong>Detecting dialogue breakdowns in dialogue systems using non-verbal information,&#8221; IPSJ 81st National Conference, 2T-08, pp. 2-365-2-366, Mar. 2019.<\/li>\n\n\n\n<li>Maina Umezawa, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Detection of Dementia Tendency in the Elderly Based on Spoken Language Information Considering Dialect,&#8221; 81st National Conference of Information Processing Society, 4ZE-07, pp. 4-463-4-464, Mar. 2019.<\/li>\n\n\n\n<li>Yasuyuki Umehara, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;A method for constructing a spoken dialogue system integrating various dialogue strategies,&#8221; Proceedings of the Acoustical Society of Japan, 2-P-1, pp. 945-948, Mar. 2019.<\/li>\n\n\n\n<li>Kazuaki Kajinami, Ryota Nishimura, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;A Spoken Dialogue Data Recording System for the Development of a Spoken Dialogue Failure Detection Method,&#8221; Proc. of the Acoustical Society of Japan, 2-P-2, pp. 949-952, Mar. 2019.<\/li>\n\n\n\n<li>Kengo Ohta, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;Response type selection for a chatting spoken dialogue system using multi-task learning with LSTM,&#8221; Proceedings of the Acoustical Society of Japan, 2-P-3, 953-956, Mar. 2019.<\/li>\n\n\n\n<li>Kanta Kiyohara, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction and evaluation of a learning support system using voice and pointing in geometry problems,&#8221; Proc. of the IAAC, 2-P-17, pp. 989-992, Mar. 2019.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2018\">2018<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Journal Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Ryota Nishimura, Daisuke Yamamoto, Takahiro Uchiya, Ichi Takumi, \u201cWeb-based environment for user generation of spoken dialog for virtual assistants,\u201d EURASIP Journal Audio, Speech, Music Process., pp. 1-13, 2018.<\/li>\n\n\n\n<li>Ryota Nishimura, Daisuke Yamamotob, Takahiro Uchiya, Ichi Takumi, \u201cMMDAE: Dialog scenario editor for MMDAgent on the web browser,\u201d ICT Express, 1-5, 2018. (In Press)<\/li>\n\n\n\n<li>Tomoki Hayashi, Masafumi Nishida, <strong>Norihide Kitaoka<\/strong>, Tomoki Toda, Kazuya Takeda, &#8220;Daily Activity Recognition with Large-scaled Real-life Recording Datasets Based on Deep Neural Network using Multi-modal Signals,&#8221; IEICE Trans. Fundamentals, Vol.E101-A,No.1, pp. 199-210,Jan. 2018.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Letters<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Ryota Nishimura, Takumi Nagao, Ikujin Ichimanda, and <strong>Norihide Kitaoka<\/strong>, &#8220;A study of speech intelligibility processing method based on speech perception characteristics of elderly people,&#8221; Journal of the Japanese Society for Fuzzy Intelligent Information, Vol. 30, No. 6, pp. 840-845, Dec. 2018.<\/li>\n\n\n\n<li>Ryota Nishimura, Miho Higaki, and <strong>Norihide Kitaoka, &#8220;<\/strong>Mapping between Acoustic Vector Space and Document Vector Space by RNN-LSTM,&#8221; Journal of Japanese Society for Fuzzy Intelligent Information, Vol. 30, No. 4, pp. 628-633, Aug. 2018.<\/li>\n\n\n\n<li><strong>Norihide Kitaoka<\/strong>, Shuhei Segawa, Ryota Nishimura, Kazuya Takeda, &#8220;Recognizing emotions from speech using a physical model,&#8221; Acoustical Science and Technology, Vol. 39, Issue 2, pp. 167-170, Feb., 2018. (doi: 10.1250\/ast.39.167)<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Invited talk<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Norihide Kitaoka<\/strong>, Yurie Iribe, Hiromitsu Nishizaki, &#8220;Construction of a corpus of elderly Japanese spech for analysis and recognition,&#8221; LREC2018, May 2018.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Eichi Seto, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;Customization of an example-based dialog system with user data and distributed word representations,&#8221; Proc. APSIPA2018, 7 pages, Nov. 2018.<\/li>\n\n\n\n<li>Ryota Nishimura, Miho Higaki, <strong>Norihide Kitaoka<\/strong>, \u201cMapping acoustic vector space and document vector space by RNN-LSTM,\u201d 2018 IEEE 7th Global Conference on Consumer Electronics, GCCE 2018, pp.296-297, 2018.<\/li>\n\n\n\n<li>Meiko Fukuda, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, Hiromitsu Nishizaki, Yurie Iribe, \u201cConstruction of a corpus for elderly Japanese speech recognition,\u201d 2018 IEEE 7th Global Conference on Consumer Electronics, GCCE 2018, pp.652-653, 2018.<\/li>\n\n\n\n<li>Kanta Kiyohara, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, \u201cMulti-modal geometry tutoring system using speech and touchscreen figure tracing,\u201d 2018 IEEE 7th Global Conference on Consumer Electronics, GCCE 2018, pp.225-226, 2018.<\/li>\n\n\n\n<li><strong>Norihide Kitaoka<\/strong>, Takuma Nakagawa, Ryota Nishimura, Yoshio Ishiguro, Shin\u2019ichi Kojima, Shin Ohsuga, \u201cA multimodal control system for autonomous vehicles using speech, gesture, and gaze recognition,\u201d DSP in Vehicles 2018, (no paper), 2018.<\/li>\n\n\n\n<li>Kazuaki Kajinami, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, \u201cConstruction of dialog database for development of spoken dialog breakdown detection methods,\u201d in ICAICTA-2018, pp.1-5, 2018.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Kengo Ohta, Ryota Nishimura, and <strong>Norihide Kitaoka, <\/strong>&#8220;Response type selection for a chatting spoken dialogue system based on distributed representation of utterances,&#8221; Spoken Language Symposium, Technical Journal of Spoken Language, SP2017-55, pp. 1-5, Dec. 2018.<\/li>\n\n\n\n<li>Motohki Shimogasa, Hiromitsu Nishizaki, Meiko Fukuda, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;A study of speech recognition models for spontaneous speech of very elderly people,&#8221; Proceedings of the Acoustical Society of Japan, 1-R-10, pp. 977-978, Mar. 2018.<\/li>\n\n\n\n<li>Taiki Yamamoto, Ryota Nishimura, Masayuki Misaki, and <strong>Norihide Kitaoka, &#8220;<\/strong>Magic Word detection method in continuous speech using LSTM Neural Network,&#8221; Proceedings of the Acoustical Society of Japan, 1-R-21, pp. 1009-1012, Mar. 2018.<\/li>\n\n\n\n<li>Yuya Kobashi, Ryota Nishimura and <strong>Norihide Kitaoka, &#8220;<\/strong>Discovery of unknown words based on changes in words used in Twitter and adaptation of language models for speech recognition based on them,&#8221; in Proceedings of the Acoustical Society of Japan, 1-R-24, pp. 1017-1020, Mar. 2018.<\/li>\n\n\n\n<li>Ryota Nishimura, Miho Higaki, and <strong>Norihide Kitaoka<\/strong>, &#8220;Mapping of Acoustic Vector Time Series to Document Vectors Based on RNNs,&#8221; in Shinagaku Giho (PRMU2018-32, SP2018-12), 2018.<\/li>\n\n\n\n<li>Yoshiki Tabata and Akinori Kawachi, &#8220;A probabilistic circuit indistinguishability obfuscator for worst-case input samplers,&#8221; Symposium on Cryptography and Information Security (SCIS), 1B1-5, Jan. 2018.<\/li>\n\n\n\n<li>Kanta Kiyohara, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;A Geometry Problem Learning Support System for Understanding Pointing and Oral Explanation,&#8221; FIT-2018, J-011, (2 pages), Mar. 2018.<\/li>\n\n\n\n<li>Ryota Nishimura, Chen Bohan, and <strong>Norihide Kitaoka<\/strong>, &#8220;A method for registering unknown words to language models in speech recognition,&#8221; in Proceedings of the Acoustical Society of Japan, 1-Q-23, pp. 127-130, Mar. 2018.<\/li>\n\n\n\n<li>Kengo Ohta, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;Response Type Selection for a Chatting Spoken Dialogue System Based on LSTM-RNN Considering Word Order,&#8221; Proceedings of the Acoustical Society of Japan, 2-8-7, pp. 45-48, Mar. 2018.<\/li>\n\n\n\n<li>Eiji Seto, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;User adaptation of an example-based chatter spoken dialogue system based on distributed representation of words,&#8221; Proceedings of the Acoustical Society of Japan, 2-8-8, pp. 49-52, Mar. 2018.<\/li>\n\n\n\n<li>Takuma Nakagawa, Ryota Nishimura, Yurie Iribe, Sachio Ishiguro, Susumu Osuga, and <strong>Norihide Kitaoka, &#8220;<\/strong>Multimodal interaction in the operation of automated vehicles,&#8221; in Proceedings of the Acoustic Society of Japan, 2-8-10, pp. 57-60, Mar. 2018.<\/li>\n\n\n\n<li>Kazuaki Kajinami, Ryota Nishimura, and <strong>Norihide Kitaoka<\/strong>, &#8220;Construction of a dialogue database for the development of a spoken dialogue failure detection method,&#8221; Proceedings of the Acoustical Society of Japan, 2-Q-14, pp. 177-180, Mar. 2018.<\/li>\n\n\n\n<li>Manami Kawashima, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Discrimination of dementia tendency based on linguistic and acoustic features extracted from elderly people&#8217;s dialogue speech,&#8221; in Proceedings of the Acoustical Society of Japan, 2-Q-36, pp. 369-370, Mar. 2018.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2017\">2017<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Tutorial Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Yurie Iribe and <strong>Norihide Kitaoka<\/strong>, &#8220;Corpus construction of very elderly people&#8217;s speech for speech recognition,&#8221; Minor Special Issue &#8211; Sound environment for elderly and visually impaired people -, J. Acoust. -310, May, 2017.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Kengo Ohta, Rikito Marumoto, Ryota Nishimura, <strong>Norihide Kitaoka<\/strong>, &#8220;Selecting type of response for chat-like spoken dialogue systems based on acoustic features of user utterances,&#8221; Proc. APSIPA2017, 5 pages, Dec. 2017.<\/li>\n\n\n\n<li>Takahiro Uchiya, Ryota Nishimura, Takahiro Hirano, Masaru Sakurai, &#8221; Design of Reminiscence Therapy System for Elderly People with Dementia,&#8221; Proc. BWCCA2017-Workshop-RI3C-2017, pp. 844-853, Nov., 2017.<\/li>\n\n\n\n<li>Takuma Nakagawa, Ryota Nishimura, Yurie iribe, Yoshio Ishiguro, Shin Osuga, <strong>Norihide Kitaoka<\/strong>, &#8220;A human machine interface framework for autonomous vehicle control,&#8221; Proc. GCCE 2017, pp. 411-413, Oct., 2017.<\/li>\n\n\n\n<li>Takahiro Uchiya, Satoshi Otake, Ryota Nishimura, Daisuke Yamamoto, Ichi Takumi, &#8220;Extraction of User Preferences based on Voice Interaction ,&#8221; Proc. GCCE 2017, pp. 416-417, Oct., 2017.<\/li>\n\n\n\n<li>Ryota Nishimura, Takahiro Uchiya, Takahiro Hirano, Masaru Sakurai, &#8220;Proposal of Reminiscence Therapy System using Spoken Dialog to Suppress Dementia ,&#8221; Proc. GCCE 2017, pp. 418-419, Occt., 2017.<\/li>\n\n\n\n<li>Eichi Seto, <strong>Norihide Kitaoka<\/strong>, &#8220;User adaptation of input-response pairs in an example-based dialog system using distributed reporesentations of words,&#8221; Proc. ICAICTA2017, Aug., 2017.<\/li>\n\n\n\n<li>Akinori Kawachi, Kenichi Kawano, Fran\u00e7ois Le Gall, and Suguru Tamaki, &#8220;Quantum query complexity of unitary operator discrimination,&#8221; Proc. COCOON&#8217;17, SESSION 2, Aug., 2017.<\/li>\n\n\n\n<li>Akinori Kawachi and Yoshiki Tabata, \u201cOn indistinguishability obfuscation of probabilistic circuits for worst-case-input subexponentially indistinguishable samplers,\u201d The 12th International Workshop on Security (IWSEC\u201917), Poster Session, Aug., 2017.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Invited talk<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Norihide Kitaoka<\/strong>, Yurie Iribe, &#8220;Recording, analysis and recognition of elderly speech,&#8221; Symposium on Speech Resource Utilisation (invited talk), Sep. 2017.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences and Research Meetings<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Takuma Nakagawa, Susumu Ohsuga, and <strong>Norihide Kitaoka<\/strong>, &#8220;Multimodal interaction with self-driving cars using voice, pointing and gaze recognition,&#8221; IEICE General Conference, D-14-4 (1 page), Mar. 2017.<\/li>\n\n\n\n<li>Yuki Kurokawa, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Analysis of dementia tendency in the elderly using acoustic features,&#8221; Proceedings of the Acoustical Society of Japan, 1-Q-36, pp. 313-314, Mar. 2017.<\/li>\n\n\n\n<li>Yuki Sawada, Yurie Iribe, and <strong>Norihide Kitaoka<\/strong>, &#8220;Estimation of speech for systems during driving using multimodal information,&#8221; in Proceedings of the Acoustical Society of Japan, 2-P-6, pp. 149-150, Mar. 2017.<\/li>\n\n\n\n<li>Yuzo Fuyuno, <strong>Norihide Kitaoka<\/strong>, and Peng Zhiyuan, &#8220;An analysis of feature-preference relationships in task-oriented and non-task-oriented dialogues with dialogue systems,&#8221; in Proceedings of the Acoustic Society of Japan, 2-P-13, pp. 171-172, Mar. 2017.<\/li>\n\n\n\n<li>Kengo Ohta, Rikito Marumoto, and <strong>Norihide Kitaoka<\/strong>, &#8220;Response type selection for a chat dialogue system based on acoustic information of user utterances,&#8221; Proceedings of the Acoustical Society of Japan, 3-5-4, pp. 71-74, Mar. 2017.<\/li>\n\n\n\n<li>Akinori Kawachi, Kenichi Kawano, Rugal Francois, and Taku Tamaki, &#8220;Question Computational Complexity for Unitary Arithmetic Identification Problems,&#8221; RIMS Research Conference on Advanced Theoretical Computer Science (Winter LA Symposium), S3, Feb. 2017.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Book<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Kenji Mase and <strong>Norihide Kitaoka<\/strong>, Encyclopaedia of Artificial Intelligence (Chapter 9, General introduction), ed. by the Japanese Association for Artificial Intelligence, pp. 696-705, ISBN 978-4320124202, Jul. 2017.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2016\">2016<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Journal Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Bohan Chen, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Impact of acoustic similarity on efficiency of verbal information transmission via subtle prosodic cues,&#8221; EURASIP Journal on Audio, Speech, and Music Processing, 2016:19, 2016. (DOI: 10.1186\/s13636-016-0097-6)<\/li>\n\n\n\n<li>Satoshi Tamura, Hiroshi Ninomiya, <strong>Norihide Kitaoka<\/strong>, Shin Osuga, Yurie Iribe, Kazuya Takeda, &#8220;Investigation of DNN-based audio-visual speech recognition,&#8221; IEICE Trans. Inf. &amp; Syst., pp. 2444-2451, Oct., 2016.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Foreword<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Norihide Kitaoka<\/strong>, &#8220;FOREWORD: Special section on recent advances in machine learning in spoken language processing,&#8221; IEICE Trans. Inf. &amp; Syst., Vol. E-99-D, No. 10, p. 2422, Oct., 2016.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Takuma Nakagawa, <strong>Norihide Kitaoka<\/strong>, &#8220;Multimodal control system for autonomous vehicles using speec and gesture recognition,&#8221; 5th ASA\/ASJ Joint Meeting, Nov., 2016.<\/li>\n\n\n\n<li>Eichi Seto, <strong>Norihide Kitaoka<\/strong>, &#8220;Example-based spoken chat system which can be cutomized for each user,&#8221; 5th ASA\/ASJ Joint Meeting, Nov., 2016.<\/li>\n\n\n\n<li><strong>Norihide Kitaoka<\/strong>, Shuhei Segawa, Kazuya Takeda, &#8220;Emotion recognition from speech using a physical model,&#8221; Proc. ICA2016, ICA2016-714 (8 pages), Sep., 2016.<\/li>\n\n\n\n<li>Yurie Iribe, <strong>Norihide Kitaoka<\/strong> and Shuhei Segawa, &#8220;Speech Corpus Spoken by Young-old, Old-old and Oldest-old Japanese,&#8221; Proc. LREC 2016, pp. 4674-4677, May, 2016.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences and Research Meetings<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Eiji Seto and <strong>Norihide Kitaoka<\/strong>, &#8220;Case Adaptation of a Chat Dialogue System Using Distributed Representation of Words,&#8221; Proc. of the IAAC, 2-Q-8, (4 pages), Sep. 2016.<\/li>\n\n\n\n<li>Nakagawa, Takuma, and <strong>Norihide Kitaoka, &#8220;<\/strong>Multimodal Interaction with Automatic Driving Vehicles Using Speech and Pointing,&#8221; Proc. of the NACS, 3-Q-11, (4 pages), Sep. 2016.<\/li>\n\n\n\n<li>Tomoki Hayashi, <strong>Norihide Kitaoka<\/strong>, Tomoki Toda, Kazuya Takeda, &#8220;An Adaptation Method in Recognition of Everyday Activities Based on Deep Neural Network,&#8221; IEICE Technical Report, SP2016-27, pp. 1-6, Aug. 2016.<\/li>\n\n\n\n<li>Bokhan Chen, <strong>Norihide Kitaoka <\/strong>and Kazuya Takeda, &#8220;Difference of prosodic information transmission efficiency casued by verbally meaningless acoustic difference : An experimental study,&#8221; Proceedings of the Acoustical Society of Japan, 2-Q-32, (2 pages), Mar. 2016.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Book<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Norihide Kitaoka <\/strong>(Editorial Board Member (Field Secretary, Speech)), Acoustic Keyword Book, ISBN 978-4-339-00880-7, Mar. 2016.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2015\">2015<\/h2>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Journal Papers<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Ken Ichikawa, <strong>Norihide Kitaoka<\/strong>, Satoru Tsuge, Kazuya Takeda, Kenji Kita, &#8220;Improving the Robustness of Various Text Retrieval Models for Speech Document Retrieval,&#8221; Transactions of Information Processing Society of Japan, Vol. 56, No. 3, 1003-1012, Mar. 2015.<\/li>\n\n\n\n<li>Yiyang Li, Chiyomi Miyajima, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;An evaluation method of aggressiveness of driving behaviour using drive recorders,&#8221; IEEJ Journal of Industry Applications, Vol. 4, No. 1, pp. 59-66, 2015.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Letters<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Xiao Yao, Takatoshi Jitsuhiro, Chiyomi Miyajima, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Modelling of Physical Characteristics of Speech under Stress,&#8221; IEEE Signal Processing Letters, (accepted), 2015.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Shuhei Segawa, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Elderly person&#8217;s emotional state estimation in conversation based on speech features for spoken dialogue systems,&#8221; 12th Western pacific Acoustics Conference 2015 (WESPAC2015), pp. 299-301, Dec., 2015.<\/li>\n\n\n\n<li>Bohan Chen, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Relationship between Speaker\/Listener Similarity and Information Transmission Quality in Speech Communication,&#8221; APSIPA ASC 2015, pp. 1190-1193, Dec., 2015.<\/li>\n\n\n\n<li>Masafumi Nishida, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Daily activity recognition based on acoustic signals and acceleration signals estimated with Gaussian process,&#8221; APSIPA ASC 2015, pp. 279-282, Dec., 2015.<\/li>\n\n\n\n<li>Satoshi Tamura, Hiroshi Ninomiya, <strong>Norihide Kitaoka<\/strong>, Shin Osuga, Yurie Iribe, Kazuya Takeda, Satoru Hayamizu, &#8220;Audio-visual speech recognition using deep bottleneck features and high-perfromanc lipreading,&#8221; APSIPA ASC 2015, pp. 575-582, Dec., 2015.<\/li>\n\n\n\n<li>Yurie Iribe, <strong>Norihide Kitaoka<\/strong>, Shuhei Segawa, &#8220;Development of new speech corpus for elderly Japanese speech recognition,&#8221; Oriental-COCOSDA\/CASLRE, pp. 27-31, Oct., 2015.<\/li>\n\n\n\n<li>Bohan Chen, Norihide Kitaoka, Kazuya Takeda, &#8220;Effect of speaking rate and speech complexity on transmission quality during driving navigation task,&#8221; 7th Biennial Workshop on DSP for In-Vehicle Systems and Safety, 4 pages, Oct., 2015.<\/li>\n\n\n\n<li>Satoshi Tamura, Hiroshi Ninomiya, <strong>Norihide Kitaoka<\/strong>, Shin Osuga, Yurie Iribe, Kazuya Takeda, Satoru Hayamizu, &#8220;Audio-visual processing toward robust speech recognition in cars,&#8221; 7th Biennial Workshop on DSP for In-Vehicle Systems and Safety, 4 pages, Oct., 2015.<\/li>\n\n\n\n<li>Satoshi Tamura, Hiroshi Ninomiya, <strong>Norihide Kitaoka<\/strong>, Shin Osuga, Yurie Iribe, Kazuya Takeda, Satoru Hayamizu, &#8220;Investigation of DNN-based modeling for audio-visual speech recognition,&#8221; 2015 First International Workshop on Spoken Language Processing (MLSLP2015), (6 pages), Oct., 2015.<\/li>\n\n\n\n<li>Hiroshi Ninomiya, <strong>Norihide Kitaoka<\/strong>, Satoshi Tamura, Yurie Iribe, Kazuya Takeda, &#8220;Integration of Deep Bottleneck Features for Audio-Visual Speech Recognition,&#8221; Proc. INTERSPEECH, pp. 563-566, Sep., 2015.<\/li>\n\n\n\n<li>Tomoki Hayashi, Masafumi Nishida, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Dayly activity recogntion based on DNN using environmental sound and acceleration signals,&#8221; Proc. EUSIPCO 2015, pp. 2351-2355, Sep. 2015.<\/li>\n\n\n\n<li>Yuto Dekiura, Tetsuya Matsumoto, Yoshinori Takeuchi, Hiroaki Kudo, Noboru Onishi, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Fast Separation and Accurate Recognition of Overlapped Speech &#8212; Separation by Spectral Subtraction and Acoustic Model Training using Separated Speeches&#8212;,&#8221; 2015 RISP International Workshop on Nonlinear Circuits, Communications and Signal Processing (NCSP&#8217;15), pp. 1-4, Mar., 2015.<\/li>\n\n\n\n<li>Katsuya Sakoyama, Chiyomi Miyajima, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Tracking Roadside Signage Observed by Drivers,&#8221; 2014 RISP International Workshop on Nonlinear Circuits, Communications and Signal Processing (NCSP&#8217;15), pp. 429-432, Mar., 2015.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences and Research Meetings<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Tetsushi Tamura, Hiroshi Ninomiya, <strong>Norihide Kitaoka<\/strong>, Susumu Osuga, Yurie Iribe, Kazuya Takeda, Satoru Hayami, &#8220;Multimodal speech recognition using deep learning &#8211; A survey of deep learning applications,&#8221; 2nd Silent Speech Recognition Workshop, ID 16, p . 8, Oct. 2015.<\/li>\n\n\n\n<li>Tetsushi Tamura, Hiroshi Ninomiya, <strong>Norihide Kitaoka<\/strong>, Susumu Osuga, Yurie Iribe, Kazuya Takeda, Satoru Hayami, &#8220;Multimodal speech recognition using deep learning &#8211; improvement of image features,&#8221; 2nd Silent Speech Recognition Workshop, ID 15, p. 8. Oct. 2015.<\/li>\n\n\n\n<li>Tetsushi Tamura, Hiroshi Ninomiya, <strong>Norihide Kitaoka<\/strong>, Susumu Osuga, Yurie Iribe, Kazuya Takeda, and Satoru Hayami, &#8220;Multimodal speech recognition using bottleneck features by deep learning,&#8221; IEICE Technical Report, SP2015-69, vol. 115, no. 253, pp. 57-62, Oct. 2015.<\/li>\n\n\n\n<li>Tetsushi Tamura, Hiroshi Ninomiya, <strong>Norihide Kitaoka<\/strong>, Susumu Osuga, Yurie Iribe, Kazuya Takeda and Satoru Hayami, &#8220;Multimodal speech recognition using acoustic and image features by deep learning,&#8221; Proceedings of the Acoustical Society of Japan, 3-2-5, (2 pages), Sep. 2015.<\/li>\n\n\n\n<li>Shuhei Segawa, <strong>Norihide Kitaoka <\/strong>and Kazuya Takeda, &#8220;Emotion Recognition of Elderly People from Speech for Application to Dialogue Strategies in Spoken Dialogue Systems,&#8221; Proceedings of the Acoustical Society of Japan,  3-Q-19, (2 pages), Sep., 2015.<\/li>\n\n\n\n<li>Chen, Bohan, <strong>Norihide Kitaoka<\/strong>, Mihoko Otake and Kazuya Takeda, &#8220;Probabilistic modelling of speaker alternation and evaluation of speaker activity using information content,&#8221; Proceedings of the Acoustical Society of Japan, 1-Q-35, (4 pages), Sep. 2015.<\/li>\n\n\n\n<li>Bohan Chen, <strong>Norihide Kitaoka<\/strong>, Mihoko Otake, Kazuya Takeda, &#8220;Evaluation of speaker engagement using turn-taking behaviour entropy. ,&#8221; IEICE Technical Report SP, SP2015-52, pp. 13-17, Jun. 2014.<\/li>\n\n\n\n<li>Nitoto Kawai, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;A method for correcting English pronunciation by presenting approximate pronunciation based on prosody-corrected learners&#8217; speech and Japanese syllables,&#8221; Proceedings of the Acoustical Society of Japan, 1-2-10, (4 pages), Mar. 2015.<\/li>\n\n\n\n<li>Chen, H., <strong>N. Kitaoka<\/strong>, and K. Takeda, &#8220;Rational Speech Feature Control in Speech Information Transfer and Its Effect on Transfer Efficiency,&#8221; Proceedings of the Acoustical Society of Japan, 1-R-20, (4 pages), Mar. 2015.<\/li>\n\n\n\n<li>Tomoki Hayashi, Masashi Nishida, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Recognition of daily life behaviour using environmental sound and acceleration signals by DNN,&#8221; Proceedings of the Acoustical Society of Japan, 2-1-16, (4 pages), Mar. 2015.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Book<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Norihide Kitaoka<\/strong>, Evolving Speech Communication between Humans and Machines (Part 4, Chapter 2: Task-oriented Dialogue), Nikkei Printing, ISBN 978-4-86469-065-2, Sep. 2015.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading ribbon\" id=\"2014\">2014 (Oct.~)<\/h2>\n\n\n\n<h3 class=\"wp-block-heading marker fl\">International Conferences<\/h3>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<ol class=\"wp-block-list\">\n<li><strong>Norihide Kitaoka<\/strong>, Tomoki Hayashi, Kazuya Takeda, &#8220;Noisy speech recognition using blind spatial subtraction array technique and deep bottleneck features,&#8221; APSIPA ASC 2014, (5 pages), Oct., 2014<\/li>\n\n\n\n<li>Masafumi Nishida, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Development and preliminary analysis of sensor signal database of contiuous daily living activity over the long term,&#8221; APSIPA ASC 2014, (6 pages), Oct., 2014<\/li>\n\n\n\n<li>Panikos Heracleous, Pongtep Angkititrakul, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Unsupervised energy disaggregation using conditional random fields,&#8221; IEEE ISGT Europe 2014, (6 pages), Oct., 2014.<\/li>\n\n\n\n<li>Yiyang Li, Chiyomi Miyajima, <strong>Norihide Kitaoka<\/strong>, Kazuya Takeda, &#8220;Measuring Aggressive Driving Behavior Using,&#8221; IEEE ITSC14, 1886-1887, Oct., 2014.<\/li>\n<\/ol>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-layout-constrained wp-block-group-is-layout-constrained\">\n<h3 class=\"wp-block-heading marker fl\">Domestic Conferences and Research Meetings<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Kazuki Morita, Chiyomi Miyajima, <strong>Norihide Kitaoka<\/strong> and Kazuya Takeda, &#8220;Evaluation of Arranged Song Retrieval Performance Focusing on Differences in Song Structure,&#8221; IEICE General Conference, D-12-4, (1 page), Mar. 2015. Chen, H., Kitaoka, N., and Takeda, K. &#8220;Relationship between speech feature similarity among interlocutors and the information transfer effect of dialogue,&#8221; Spoken Language Symposium, SP2014-124, pp. 147-152, Dec. 2014.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>Click here to Kitaoka&#8217;s publication list. 2024 \/ 2023 \/ 2022 \/ 2021 \/ 2020 \u30002019 \/ 2018 \/ 2017 \/ 2016 \/ 2015 \/ 2014 2025 Journal Papers International Conferences Domestic Conferences 2024 Journal Papers International Conferences Domestic Conferences 2023 Journal Papers International Conferences Domestic Conferences 2022 Journal Papers Explanation International Conferences Domestic Conferences [&hellip;]<\/p>\n","protected":false},"author":10,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"wp-custom-template-en%e5%9b%ba%e5%ae%9a%e3%83%9a%e3%83%bc%e3%82%b8","meta":{"_locale":"en_US","_original_post":"https:\/\/slp.cs.tut.ac.jp\/?page_id=63","footnotes":""},"class_list":["post-537","page","type-page","status-publish","hentry","en-US"],"_links":{"self":[{"href":"https:\/\/slp.cs.tut.ac.jp\/wp-json\/wp\/v2\/pages\/537","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/slp.cs.tut.ac.jp\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/slp.cs.tut.ac.jp\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/slp.cs.tut.ac.jp\/wp-json\/wp\/v2\/users\/10"}],"replies":[{"embeddable":true,"href":"https:\/\/slp.cs.tut.ac.jp\/wp-json\/wp\/v2\/comments?post=537"}],"version-history":[{"count":79,"href":"https:\/\/slp.cs.tut.ac.jp\/wp-json\/wp\/v2\/pages\/537\/revisions"}],"predecessor-version":[{"id":2256,"href":"https:\/\/slp.cs.tut.ac.jp\/wp-json\/wp\/v2\/pages\/537\/revisions\/2256"}],"wp:attachment":[{"href":"https:\/\/slp.cs.tut.ac.jp\/wp-json\/wp\/v2\/media?parent=537"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}