default search action
19th Interspeech 2018: Hyderabad, India
- B. Yegnanarayana:
19th Annual Conference of the International Speech Communication Association, Interspeech 2018, Hyderabad, India, September 2-6, 2018. ISCA 2018
ISCA Medal Talk
- Bishnu S. Atal:
From Vocoders to Code-Excited Linear Prediction: Learning How We Hear What We Hear. 1
End-to-End Speech Recognition
- Shigeki Karita, Shinji Watanabe, Tomoharu Iwata, Atsunori Ogawa, Marc Delcroix:
Semi-Supervised End-to-End Speech Recognition. 2-6 - Albert Zeyer, Kazuki Irie, Ralf Schlüter, Hermann Ney:
Improved Training of End-to-end Attention Models for Speech Recognition. 7-11 - Hossein Hadian, Hossein Sameti, Daniel Povey, Sanjeev Khudanpur:
End-to-end Speech Recognition Using Lattice-free MMI. 12-16 - Stefan Braun, Daniel Neil, Jithendar Anumula, Enea Ceolini, Shih-Chii Liu:
Multi-channel Attention for End-to-End Speech Recognition. 17-21 - Titouan Parcollet, Ying Zhang, Mohamed Morchid, Chiheb Trabelsi, Georges Linarès, Renato de Mori, Yoshua Bengio:
Quaternion Convolutional Neural Networks for End-to-End Automatic Speech Recognition. 22-26 - Ruoming Pang, Tara N. Sainath, Rohit Prabhavalkar, Suyog Gupta, Yonghui Wu, Shuyuan Zhang, Chung-Cheng Chiu:
Compression of End-to-End Models. 27-31
Prosody Modeling and Generation
- Zack Hodari, Oliver Watts, Srikanth Ronanki, Simon King:
Learning Interpretable Control Dimensions for Speech Synthesis by Using External Data. 32-36 - Hieu-Thi Luong, Xin Wang, Junichi Yamagishi, Nobuyuki Nishizawa:
Investigating Accuracy of Pitch-accent Annotations in Neural Network-based Speech Synthesis and Denoising Effects. 37-41 - Guan-Ting Liou, Chen-Yu Chiang, Yih-Ru Wang, Sin-Horng Chen:
An Exploration of Local Speaking Rate Variations in Mandarin Read Speech. 42-46 - Yibin Zheng, Jianhua Tao, Zhengqi Wen, Ya Li:
BLSTM-CRF Based End-to-End Prosodic Boundary Prediction with Context Sensitive Embeddings in a Text-to-Speech Front-End. 47-51 - Berrak Sisman, Haizhou Li:
Wavelet Analysis of Speaker Dependent and Independent Prosody for Voice Conversion. 52-56 - Rui Liu, Feilong Bao, Guanglai Gao, Hui Zhang, Yonghe Wang:
Improving Mongolian Phrase Break Prediction by Using Syllable and Morphological Embeddings with BiLSTM Model. 57-61
Speaker Verification I
- Lanhua You, Wu Guo, Yan Song, Sheng Zhang:
Improved Supervised Locality Preserving Projection for I-vector Based Speaker Verification. 62-66 - Ziqiang Shi, Huibin Lin, Liu Liu, Rujie Liu:
Double Joint Bayesian Modeling of DNN Local I-Vector for Text Dependent Speaker Verification with Random Digit Strings. 67-71 - Anna Silnova, Niko Brümmer, Daniel Garcia-Romero, David Snyder, Lukás Burget:
Fast Variational Bayes for Heavy-tailed PLDA Applied to i-vectors and x-vectors. 72-76 - Massimiliano Todisco, Héctor Delgado, Kong-Aik Lee, Md. Sahidullah, Nicholas W. D. Evans, Tomi Kinnunen, Junichi Yamagishi:
Integrated Presentation Attack Detection and Automatic Speaker Verification: Common Features and Gaussian Back-end Fusion. 77-81 - Luciana Ferrer, Mitchell McLaren:
A Generalization of PLDA for Joint Modeling of Speaker Identity and Multiple Nuisance Conditions. 82-86 - Nanxin Chen, Jesús Villalba, Najim Dehak:
An Investigation of Non-linear i-vectors for Speaker Verification. 87-91
Spoken Term Detection
- Dhananjay Ram, Lesly Miculicich, Hervé Bourlard:
CNN Based Query by Example Spoken Term Detection. 92-96 - Yougen Yuan, Cheung-Chi Leung, Lei Xie, Hongjie Chen, Bin Ma, Haizhou Li:
Learning Acoustic Word Embeddings with Temporal Context for Query-by-Example Speech Search. 97-101 - Ziwei Zhu, Zhiyong Wu, Runnan Li, Helen Meng, Lianhong Cai:
Siamese Recurrent Auto-Encoder Representation for Query-by-Example Spoken Term Detection. 102-106 - Wei Li, Brian Mak:
Fast Derivation of Cross-lingual Document Vectors from Self-attentive Neural Machine Translation Model. 107-111 - Laxmi Pandey, Karan Nathwani:
LSTM Based Attentive Fusion of Spectral and Prosodic Information for Keyword Spotting in Hindi Language. 112-116 - Ravi Shankar, Vikram C. M., S. R. Mahadeva Prasanna:
Spoken Keyword Detection Using Joint DTW-CNN. 117-121
The INTERSPEECH 2018 Computational Paralinguistics ChallengE (ComParE): Atypical & Self-Assessed Affect, Crying & Heart Beats 1
- Björn W. Schuller, Stefan Steidl, Anton Batliner, Peter B. Marschik, Harald Baumeister, Fengquan Dong, Simone Hantke, Florian B. Pokorny, Eva-Maria Rathner, Katrin D. Bartl-Pokorny, Christa Einspieler, Dajie Zhang, Alice Baird, Shahin Amiriparian, Kun Qian, Zhao Ren, Maximilian Schmitt, Panagiotis Tzirakis, Stefanos Zafeiriou:
The INTERSPEECH 2018 Computational Paralinguistics Challenge: Atypical & Self-Assessed Affect, Crying & Heart Beats. 122-126 - Ahmed Imtiaz Humayun, Md. Tauhiduzzaman Khan, Shabnam Ghaffarzadegan, Zhe Feng, Taufiq Hasan:
An Ensemble of Transfer, Semi-supervised and Supervised Learning Methods for Pathological Heart Sound Classification. 127-131 - Mehmet Ali Tugtekin Turan, Engin Erzin:
Monitoring Infant's Emotional Cry in Domestic Environments Using the Capsule Network Architecture. 132-136 - Mark A. Huckvale:
Neural Network Architecture That Combines Temporal and Summative Features for Infant Cry Classification in the Interspeech 2018 Computational Paralinguistics Challenge. 137-141 - Zixing Zhang, Jing Han, Kun Qian, Björn W. Schuller:
Evolving Learning for Analysing Mood-Related Infant Vocalisation. 142-146 - Johannes Wagner, Dominik Schiller, Andreas Seiderer, Elisabeth André:
Deep Learning in Paralinguistic Recognition Tasks: Are Hand-crafted Features Still Relevant? 147-151 - Danqing Luo, Yuexian Zou, Dongyan Huang:
Investigation on Joint Representation Learning for Robust Feature Extraction in Speech Emotion Recognition. 152-156 - Soo Jin Park, Amber Afshan, Zhi Ming Chua, Abeer Alwan:
Using Voice Quality Supervectors for Affect Identification. 157-161 - Dengke Tang, Junlin Zeng, Ming Li:
An End-to-End Deep Learning Framework for Speech Emotion Recognition of Atypical Individuals. 162-166
Show and Tell 1
- Alexander Koller, Timo Baumann, Arne Köhn:
DialogOS: Simple and Extensible Dialogue Modeling. 167-168 - Franck Dernoncourt, Trung Bui, Walter Chang:
A Framework for Speech Recognition Benchmarking. 169-170 - Takayuki Arai:
Flexible Tongue Housed in a Static Model of the Vocal Tract With Jaws, Lips and Teeth. 171-172 - Lani Mathew, K. Gopakumar:
Voice Analysis Using Acoustic and Throat Microphones for Speech Therapy. 173-174 - Manny Rayner, Nikos Tsourakis, Jan Stanek:
A Robust Context-Dependent Speech-to-Speech Phraselator Toolkit for Alexa. 175-176
Speech Segments and Voice Quality
- RaviShankar Prasad, Sudarsana Reddy Kadiri, Suryakanth V. Gangashetty, Bayya Yegnanarayana:
Discriminating Nasals and Approximants in English Language Using Zero Time Windowing. 177-181 - Phil Howson, Alexei Kochetov:
Gestural Lenition of Rhotics Captures Variation in Brazilian Portuguese. 182-186 - RaviShankar Prasad, Bayya Yegnanarayana:
Identification and Classification of Fricatives in Speech Using Zero Time Windowing Method. 187-191 - Nattanun Chanchaochai, Christopher Cieri, Japhet Debrah, Hongwei Ding, Yue Jiang, Sishi Liao, Mark Liberman, Jonathan Wright, Jiahong Yuan, Juhong Zhan, Yuqing Zhan:
GlobalTIMIT: Acoustic-Phonetic Datasets for the World's Languages. 192-196 - Anne Hermes, Doris Mücke, Bastian Auris, Rachid Ridouane:
Structural Effects on Properties of Consonantal Gestures in Tashlhiyt. 197-201 - Alexei Kochetov, Matthew Faytak, Kiranpreet Nara:
The Retroflex-dental Contrast in Punjabi Stops and Nasals: A Principal Component Analysis of Ultrasound Images. 202-206 - Yang Yue, Fang Hu:
Vowels and Diphthongs in Hangzhou Wu Chinese Dialect. 207-211 - Mahesh M, Jeena J. Prakash, Hema A. Murthy:
Resyllabification in Indian Languages and Its Implications in Text-to-speech Systems. 212-216 - Andy Murphy, Irena Yanushevskaya, Ailbhe Ní Chasaide, Christer Gobl:
Voice Source Contribution to Prominence Perception: Rd Implementation. 217-221 - Christer Gobl, Andy Murphy, Irena Yanushevskaya, Ailbhe Ní Chasaide:
On the Relationship between Glottal Pulse Shape and Its Spectrum: Correlations of Open Quotient, Pulse Skew and Peak Flow with Source Harmonic Amplitudes. 222-226 - Vincent Hughes, Philip Harrison, Paul Foulkes, Peter French, Colleen Kavanagh, Eugenia San Segundo Fernández:
The Individual and the System: Assessing the Stability of the Output of a Semi-automatic Forensic Voice Comparison System. 227-231 - Sudarsana Reddy Kadiri, Bayya Yegnanarayana:
Breathy to Tense Voice Discrimination using Zero-Time Windowing Cepstral Coefficients (ZTWCCs). 232-236 - Pamir Gogoi, Sishir Kalita, Parismita Gogoi, Ratree Wayland, Priyankoo Sarmah, S. R. Mahadeva Prasanna:
Analysis of Breathiness in Contextual Vowel of Voiceless Nasals in Mizo. 237-241
Speaker State and Trait
- Yijia Xu, Mark Hasegawa-Johnson, Nancy McElwain:
Infant Emotional Outbursts Detection in Infant-parent Spoken Interactions. 242-246 - Jaejin Cho, Raghavendra Pappagari, Purva Kulkarni, Jesús Villalba, Yishay Carmiel, Najim Dehak:
Deep Neural Networks for Emotion Recognition Combining Audio and Transcripts. 247-251 - Srinivas Parthasarathy, Carlos Busso:
Preference-Learning with Qualitative Agreement for Sentence Level Emotional Annotations. 252-256 - Siddique Latif, Rajib Rana, Shahzad Younis, Junaid Qadir, Julien Epps:
Transfer Learning for Improving Speech Emotion Classification Accuracy. 257-261 - Patrick Meyer, Eric Buschermöhle, Tim Fingscheidt:
What Do Classifiers Actually Learn? a Case Study on Emotion Recognition Datasets. 262-266 - Eva-Maria Rathner, Yannik Terhorst, Nicholas Cummins, Björn W. Schuller, Harald Baumeister:
State of Mind: Classification through Self-reported Affect and Word Use in Speech. 267-271 - Ziping Zhao, Yu Zheng, Zixing Zhang, Haishuai Wang, Yiqin Zhao, Chao Li:
Exploring Spatio-Temporal Representations by Integrating Attention-based Bidirectional-LSTM-RNNs and FCNs for Speech Emotion Recognition. 272-276 - Pegah Ghahremani, Phani Sankar Nidadavolu, Nanxin Chen, Jesús Villalba, Daniel Povey, Sanjeev Khudanpur, Najim Dehak:
End-to-end Deep Neural Network Age Estimation. 277-281 - Rajat Hebbar, Krishna Somandepalli, Shrikanth S. Narayanan:
Improving Gender Identification in Movie Audio Using Cross-Domain Data. 282-286 - Selen Hande Kabil, Hannah Muckenhirn, Mathew Magimai-Doss:
On Learning to Identify Genders from Raw Speech Signal Using CNNs. 287-291 - Jilt Sebastian, Manoj Kumar, Pavan Kumar D. S., Mathew Magimai-Doss, Hema A. Murthy, Shrikanth S. Narayanan:
Denoising and Raw-waveform Networks for Weakly-Supervised Gender Identification on Noisy Speech. 292-296 - James R. Williamson, Thomas F. Quatieri, Adam C. Lammert, Katherine Mitchell, Katherine Finkelstein, Nicole Ekon, Caitlin Dillon, Robert Kenefick, Kristin Heaton:
The Effect of Exposure to High Altitude and Heat on Speech Articulatory Coordination. 297-301
Deep Learning for Source Separation and Pitch Tracking
- Lianwu Chen, Meng Yu, Yanmin Qian, Dan Su, Dong Yu:
Permutation Invariant Training of Generative Adversarial Network for Monaural Speech Separation. 302-306 - Jun Wang, Jie Chen, Dan Su, Lianwu Chen, Meng Yu, Yanmin Qian, Dong Yu:
Deep Extractor Network for Target Speaker Recovery from Single Channel Speech Mixtures. 307-311 - Weipeng He, Petr Motlícek, Jean-Marc Odobez:
Joint Localization and Classification of Multiple Sound Sources Using a Multi-task Neural Network. 312-316 - Shuai Yang, Zhiyong Wu, Binbin Shen, Helen Meng:
Detection of Glottal Closure Instants from Speech Signals: A Convolutional Neural Network Based Method. 317-321 - Zhong-Qiu Wang, Xueliang Zhang, DeLiang Wang:
Robust TDOA Estimation Based on Time-Frequency Masking and Deep Neural Networks. 322-326 - Akihiro Kato, Tomi Kinnunen:
Waveform to Single Sinusoid Regression to Estimate the F0 Contour from Noisy Speech Using Recurrent Deep Neural Networks. 327-331 - Paul Magron, Konstantinos Drossos, Stylianos Ioannis Mimilakis, Tuomas Virtanen:
Reducing Interference with Phase Recovery in DNN-based Monaural Singing Voice Separation. 332-336 - Kanru Hua:
Nebula: F0 Estimation and Voicing Detection by Modeling the Statistical Properties of Feature Extractors. 337-341 - Yi Luo, Nima Mesgarani:
Real-time Single-channel Dereverberation and Separation with Time-domain Audio Separation Network. 342-346 - Rajath Kumar, Yi Luo, Nima Mesgarani:
Music Source Activity Detection and Separation Using Deep Attractor Network. 347-351 - Longfei Yang, Yanlu Xie, Jinsong Zhang:
Improving Mandarin Tone Recognition Using Convolutional Bidirectional Long Short-Term Memory with Attention. 352-356
Acoustic Analysis-Synthesis of Speech Disorders
- Rob van Son, Catherine Middag, Kris Demuynck:
Vowel Space as a Tool to Evaluate Articulation Problems. 357-361 - Véronique Delvaux, Kathy Huet, Myriam Piccaluga, Sophie van Malderen, Bernard Harmegnies:
Towards a Better Characterization of Parkinsonian Speech: A Multidimensional Acoustic Study. 362-366 - Sishir Kalita, S. R. Mahadeva Prasanna, Samarendra Dandapat:
Self-similarity Matrix Based Intelligibility Assessment of Cleft Lip and Palate Speech. 367-371 - Akhilesh Kumar Dubey, S. R. Mahadeva Prasanna, Samarendra Dandapat:
Pitch-Adaptive Front-end Feature for Hypernasality Detection. 372-376 - Raquel Norel, Mary Pietrowicz, Carla Agurto, Shay Rishoni, Guillermo A. Cecchi:
Detection of Amyotrophic Lateral Sclerosis (ALS) via Acoustic Analysis. 377-381 - Vikram C. M., S. R. Mahadeva Prasanna, Ajish K. Abraham, Pushpavathi M, Girish K. S:
Detection of Glottal Activity Errors in Production of Stop Consonants in Children with Cleft Lip and Palate. 382-386
ASR Systems and Technologies
- Anuroop Sriram, Heewoo Jun, Sanjeev Satheesh, Adam Coates:
Cold Fusion: Training Seq2Seq Models Together with Language Models. 387-391 - Kazuki Irie, Zhihong Lei, Liuhui Deng, Ralf Schlüter, Hermann Ney:
Investigation on Estimation of Sentence Probability by Combining Forward, Backward and Bi-directional LSTM-RNNs. 392-395 - Thomas Zenkel, Ramon Sanabria, Florian Metze, Alex Waibel:
Subword and Crossword Units for CTC Acoustic Models. 396-400 - Tomohiro Tanaka, Ryo Masumura, Hirokazu Masataki, Yushi Aono:
Neural Error Corrective Language Models for Automatic Speech Recognition. 401-405 - Mohammad Sadegh Rasooli, Sarangarajan Parthasarathy:
Entity-Aware Language Model as an Unsupervised Reranker. 406-410 - Iksoo Choi, Jinhwan Park, Wonyong Sung:
Character-level Language Modeling with Gated Hierarchical Recurrent Neural Networks. 411-415
Deception, Personality, and Culture Attribute
- Sarah Ita Levitan, Angel Maredia, Julia Hirschberg:
Acoustic-Prosodic Indicators of Deception and Trust in Interview Dialogues. 416-420 - Guozhen An, Sarah Ita Levitan, Julia Hirschberg, Rivka Levitan:
Deep Personality Recognition for Deception Detection. 421-425 - Hansjörg Mixdorff, Albert Rilliard, Tan Lee, Matthew K. H. Ma, Angelika Hönemann:
Cross-cultural (A)symmetries in Audio-visual Attitude Perception. 426-430 - Fasih Haider, Fahim A. Salim, Owen Conlan, Saturnino Luz:
An Active Feature Transformation Method for Attitude Recognition of Video Bloggers. 431-435 - Fu-Sheng Tsai, Hao-Chun Yang, Wei-Wen Chang, Chi-Chun Lee:
Automatic Assessment of Individual Culture Attribute of Power Distance Using a Social Context-Enhanced Prosodic Network Representation. 436-440 - Sudarsana Reddy Kadiri, Bayya Yegnanarayana:
Analysis and Detection of Phonation Modes in Singing Voice using Excitation Source Features and Single Frequency Filtering Cepstral Coefficients (SFFCC). 441-445
Automatic Detection and Recognition of Voice and Speech Disorders
- Huiyi Wu, John J. Soraghan, Anja Lowit, Gaetano Di Caterina:
A Deep Learning Method for Pathological Voice Detection Using Convolutional Deep Belief Networks. 446-450 - Chitralekha Bhat, Biswajit Das, Bhavik Vachhani, Sunil Kumar Kopparapu:
Dysarthric Speech Recognition Using Time-delay Neural Network Based Denoising Autoencoder. 451-455 - Juan Camilo Vásquez-Correa, Tomás Arias-Vergara, Juan Rafael Orozco-Arroyave, Elmar Nöth:
A Multitask Learning Approach to Assess the Dysarthria Severity in Patients with Parkinson's Disease. 456-460 - Jason Lilley, Erin L. Crowgey, H. Timothy Bunnell:
The Use of Machine Learning and Phonetic Endophenotypes to Discover Genetic Variants Associated with Speech Sound Disorder. 461-465 - Meredith Moore, Hemanth Venkateswara, Sethuraman Panchanathan:
Whistle-blowing ASRs: Evaluating the Need for More Inclusive Speech Recognition Systems. 466-470 - Bhavik Vachhani, Chitralekha Bhat, Sunil Kumar Kopparapu:
Data Augmentation Using Healthy Speech for Dysarthric Speech Recognition. 471-475
Voice Conversion
- Shaojin Ding, Guanlong Zhao, Christopher Liberatore, Ricardo Gutierrez-Osuna:
Improving Sparse Representations in Exemplar-Based Voice Conversion with a Phoneme-Selective Objective Function. 476-480 - Shaojin Ding, Christopher Liberatore, Ricardo Gutierrez-Osuna:
Learning Structured Dictionaries for Exemplar-based Voice Conversion. 481-485 - Yu-Huai Peng, Hsin-Te Hwang, Yi-Chiao Wu, Yu Tsao, Hsin-Min Wang:
Exemplar-Based Spectral Detail Compensation for Voice Conversion. 486-490 - G. Nisha Meenakshi, Prasanta Kumar Ghosh:
Whispered Speech to Neutral Speech Conversion Using Bidirectional LSTMs. 491-495 - Songxiang Liu, Jinghua Zhong, Lifa Sun, Xixin Wu, Xunying Liu, Helen Meng:
Voice Conversion Across Arbitrary Speakers Based on a Single Target-Speaker Utterance. 496-500 - Ju-Chieh Chou, Cheng-chieh Yeh, Hung-yi Lee, Lin-Shan Lee:
Multi-target Voice Conversion without Parallel Data by Adversarially Learning Disentangled Audio Representations. 501-505
The INTERSPEECH 2018 Computational Paralinguistics ChallengE (ComParE): Atypical & Self-Assessed Affect, Crying & Heart Beats 2
- Cristina Gorrostieta, Richard Brutti, Kye Taylor, Avi Shapiro, Joseph Moran, Ali Azarbayejani, John Kane:
Attention-based Sequence Classification for Affect Detection. 506-510 - Zafi Sherhan Syed, Julien Schroeter, Kirill A. Sidorov, A. David Marshall:
Computational Paralinguistics: Automatic Assessment of Emotions, Mood and Behavioural State from Acoustics of Speech. 511-515 - Sai Krishna Rallabandi, Bhavya Karki, Carla Viegas, Eric Nyberg, Alan W. Black:
Investigating Utterance Level Representations for Detecting Intent from Acoustics. 516-520 - Heysem Kaya, Dmitrii Fedotov, Ali Yesilkanat, Oxana Verkholyak, Yang Zhang, Alexey Karpov:
LSTM Based Cross-corpus and Cross-task Acoustic Emotion Recognition. 521-525 - Bogdan Vlasenko, Jilt Sebastian, Pavan Kumar D. S., Mathew Magimai-Doss:
Implementing Fusion Techniques for the Classification of Paralinguistic Information. 526-530 - Gábor Gosztolya, Tamás Grósz, László Tóth:
General Utterance-Level Feature Extraction for Classifying Crying Sounds, Atypical & Self-Assessed Affect and Heart Beats. 531-535 - Bo-Hao Su, Sung-Lin Yeh, Ming-Ya Ko, Huan-Yu Chen, Shun-Chang Zhong, Jeng-Lin Li, Chi-Chun Lee:
Self-Assessed Affect Recognition Using Fusion of Attentional BLSTM and Static Acoustic Features. 536-540 - Claude Montacié, Marie-José Caraty:
Vocalic, Lexical and Prosodic Cues for the INTERSPEECH 2018 Self-Assessed Affect Challenge. 541-545
Show and Tell 2
- Anand P. A, Chiranjeevi Yarra, N. K. Kausthubha, Prasanta Kumar Ghosh:
Intonation tutor by SPIRE (In-SPIRE): An Online Tool for an Automatic Feedback to the Second Language Learners in Learning Intonation. 546-547 - Keelan Evanini, Veronika Timpe-Laughlin, Eugene Tsuprun, Ian Blood, Jeremy Lee, James V. Bruno, Vikram Ramanarayanan, Patrick L. Lange, David Suendermann-Oeft:
Game-based Spoken Dialog Language Learning Applications for Young Students. 548-549 - Alexander Sorin, Slava Shechtman, Zvi Kons, Ron Hoory, Shay Ben-David, Joe Pavitt, Shai Rozenberg, Carmel Rabinovitz, Tal Drory:
The IBM Virtual Voice Creator. 550-551 - Gayathri G, N. Mohana, Radhika Pal, Hema A. Murthy:
Mobile Application for Learning Languages for the Unlettered. 552-553 - Haihua Xu, Van Tung Pham, Zin Tun Kyaw, Zhi Hao Lim, Eng Siong Chng, Haizhou Li:
Mandarin-English Code-switching Speech Recognition. 554-555
Spoken Dialogue Systems and Conversational Analysis
- Joo-Kyung Kim, Young-Bum Kim:
Joint Learning of Domain Classification and Out-of-Domain Detection with Dynamic Class Weighting for Satisficing False Acceptance Rates. 556-560 - Sankar Mukherjee, Thierry Legou, Leonardo Lancia, Pauline M. Hilt, Alice Tomassini, Luciano Fadiga, Alessandro D'Ausilio, Leonardo Badino, Noël Nguyen:
Analyzing Vocal Tract Movements During Speech Accommodation. 561-565 - Yujiang Li, Xuemin Zhao, Weiqun Xu, Yonghong Yan:
Cross-Lingual Multi-Task Neural Architecture for Spoken Language Understanding. 566-570 - Grant P. Strimel, Kanthashree Mysore Sathyendra, Stanislav Peshterliev:
Statistical Model Compression for Small-Footprint Natural Language Understanding. 571-575 - Norbert Braunschweiler, Alexandros Papangelis:
Comparison of an End-to-end Trainable Dialogue System with a Modular Statistical Dialogue System. 576-580 - Megan M. Willi, Stephanie A. Borrie, Tyson S. Barrett, Ming Tu, Visar Berisha:
A Discriminative Acoustic-Prosodic Approach for Measuring Local Entrainment. 581-585 - Matthew Roddy, Gabriel Skantze, Naomi Harte:
Investigating Speech Features for Continuous Turn-Taking Prediction Using LSTMs. 586-590 - Ivan Kraljevski, Diane Hirschfeld:
Classification of Correction Turns in Multilingual Dialogue Corpus. 591-595 - Chetan Naik, Arpit Gupta, Hancheng Ge, Lambert Mathias, Ruhi Sarikaya:
Contextual Slot Carryover for Disparate Schemas. 596-600 - Vincent Renkens, Hugo Van hamme:
Capsule Networks for Low Resource Spoken Language Understanding. 601-605 - Padmasundari, Srinivas Bangalore:
Intent Discovery Through Unsupervised Semantic Text Clustering. 606-610 - Yulun Du, Alan W. Black, Louis-Philippe Morency, Maxine Eskénazi:
Multimodal Polynomial Fusion for Detecting Driver Distraction. 611-615 - Koji Inoue, Divesh Lala, Katsuya Takanashi, Tatsuya Kawahara:
Engagement Recognition in Spoken Dialogue via Neural Network by Aggregating Different Annotators' Models. 616-620 - Tuarik Buanzur, Margaret Zellers, Saudah Namyalo, Alena Witzlack-Makarevich:
A First Investigation of the Timing of Turn-taking in Ruuli. 621-625
Spoofing Detection
- Yuanjun Zhao, Roberto Togneri, Victor Sreeram:
Spoofing Detection Using Adaptive Weighting Framework and Clustering Analysis. 626-630 - Sarfaraz Jelil, Sishir Kalita, S. R. Mahadeva Prasanna, Rohit Sinha:
Exploration of Compressed ILPR Features for Replay Attack Detection. 631-635 - Tharshini Gunendradasan, Buddhi Wickramasinghe, Phu Ngoc Le, Eliathamby Ambikairajah, Julien Epps:
Detection of Replay-Spoofing Attacks Using Frequency Modulation Features. 636-640 - Madhu R. Kamble, Hemlata Tak, Hemant A. Patil:
Effectiveness of Speech Demodulation-Based Features for Replay Detection. 641-645 - Madhu R. Kamble, Hemant A. Patil:
Novel Variable Length Energy Separation Algorithm Using Instantaneous Amplitude Features for Replay Detection. 646-650 - Ji-Chen Yang, Changhuai You, Qianhua He:
Feature with Complementarity of Statistics and Principal Information for Spoofing Detection. 651-655 - Dongbo Li, Longbiao Wang, Jianwu Dang, Meng Liu, Zeyan Oo, Seiichi Nakagawa, Haotian Guan, Xiangang Li:
Multiple Phase Information Combination for Replay Attacks Detection. 656-660 - Buddhi Wickramasinghe, Saad Irtza, Eliathamby Ambikairajah, Julien Epps:
Frequency Domain Linear Prediction Features for Replay Spoofing Attack Detection. 661-665 - Hardik B. Sailor, Madhu R. Kamble, Hemant A. Patil:
Auditory Filterbank Learning for Temporal Modulation Features in Replay Spoof Speech Detection. 666-670 - Kaavya Sriskandaraja, Vidhyasaharan Sethu, Eliathamby Ambikairajah:
Deep Siamese Architecture Based Replay Detection for Secure Voice Biometric. 671-675 - Alejandro Gómez Alanís, Antonio M. Peinado, José A. González, Ángel M. Gómez:
A Deep Identity Representation for Noise Robust Spoofing Detection. 676-680 - Francis Tom, Mohit Jain, Prasenjit Dey:
End-To-End Audio Replay Attack Detection Using Deep Convolutional Networks with Attention. 681-685 - M. S. Saranya, Hema A. Murthy:
Decision-level Feature Switching as a Paradigm for Replay Attack Detection. 686-690 - Gajan Suthokumar, Vidhyasaharan Sethu, Chamith Wijenayake, Eliathamby Ambikairajah:
Modulation Dynamic Features for the Detection of Replay Attacks. 691-695
Speech Analysis and Representation
- Erfan Loweimi, Jon Barker, Thomas Hain:
On the Usefulness of the Speech Phase Spectrum for Pitch Extraction. 696-700 - Manu Airaksinen, Lauri Juvela, Okko Räsänen, Paavo Alku:
Time-regularized Linear Prediction for Noise-robust Extraction of the Spectral Envelope of Speech. 701-705 - Hardik B. Sailor, Hemant A. Patil:
Auditory Filterbank Learning Using ConvRBM for Infant Cry Classification. 706-710 - Nirmesh J. Shah, Hemant A. Patil:
Effectiveness of Dynamic Features in INCA and Temporal Context-INCA. 711-715 - Rong Gong, Xavier Serra:
Singing Voice Phoneme Segmentation by Hierarchically Inferring Syllable and Phoneme Onset Positions. 716-720 - Prasad Tapkir, Hemant A. Patil:
Novel Empirical Mode Decomposition Cepstral Features for Replay Spoof Detection. 721-725 - Hemlata Tak, Hemant A. Patil:
Novel Linear Frequency Residual Cepstral Features for Replay Attack Detection. 726-730 - Kumud Tripathi, K. Sreenivasa Rao:
Analysis of sparse representation based feature on speech mode classification. 731-735 - Jitendra Kumar Dhiman, Neeraj Sharma, Chandra Sekhar Seelamantula:
Multicomponent 2-D AM-FM Modeling of Speech Spectrograms. 736-740 - Abhilash Sainathan, Sunil Rudresh, Chandra Sekhar Seelamantula:
An Optimization Framework for Recovery of Speech from Phase-Encoded Spectrograms. 741-745 - Wei Xia, John H. L. Hansen:
Speaker Recognition with Nonlinear Distortion: Clipping Analysis and Impact. 746-750 - Madhusudan Singh, Debadatta Pati:
Linear Prediction Residual based Short-term Cepstral Features for Replay Attacks Detection. 751-755 - Surbhi Sakshi, Avinash Kumar, Gayadhar Pradhan:
Analysis of Variational Mode Functions for Robust Detection of Vowels. 756-760
Sequence Models for ASR
- Chao Weng, Jia Cui, Guangsen Wang, Jun Wang, Chengzhu Yu, Dan Su, Dong Yu:
Improving Attention Based Sequence-to-Sequence Models for End-to-End English Conversational Speech Recognition. 761-765 - Eugen Beck, Mirko Hannemann, Patrick Dötsch, Ralf Schlüter, Hermann Ney:
Segmental Encoder-Decoder Models for Large Vocabulary Automatic Speech Recognition. 766-770 - Shiliang Zhang, Ming Lei:
Acoustic Modeling with DFSMN-CTC and Joint CTC-CE Learning. 771-775 - Jaesung Bae, Dae-Shik Kim:
End-to-End Speech Command Recognition with Capsule Network. 776-780 - Neil Zeghidour, Nicolas Usunier, Gabriel Synnaeve, Ronan Collobert, Emmanuel Dupoux:
End-to-End Speech Recognition from the Raw Waveform. 781-785 - Chengzhu Yu, Chunlei Zhang, Chao Weng, Jia Cui, Dong Yu:
A Multistage Training Framework for Acoustic-to-Word Model. 786-790 - Shiyu Zhou, Linhao Dong, Shuang Xu, Bo Xu:
Syllable-Based Sequence-to-Sequence Speech Recognition with the Transformer in Mandarin Chinese. 791-795 - Kyu Jeong Han, Akshay Chandrashekaran, Jungsuk Kim, Ian R. Lane:
Densely Connected Networks for Conversational Speech Recognition. 796-800 - Tomoki Hayashi, Shinji Watanabe, Tomoki Toda, Kazuya Takeda:
Multi-Head Decoder for End-to-End Speech Recognition. 801-805 - Takuma Mori, Andros Tjandra, Sakriani Sakti, Satoshi Nakamura:
Compressing End-to-end ASR Networks by Tensor-Train Decomposition. 806-810 - Yu-An Chung, James R. Glass:
Speech2Vec: A Sequence-to-Sequence Framework for Learning Word Embeddings from Speech. 811-815 - Linhao Dong, Shiyu Zhou, Wei Chen, Bo Xu:
Extending Recurrent Neural Aligner for Streaming End-to-End Speech Recognition in Mandarin. 816-820
Source Separation and Spatial Analysis
- Disong Wang, Yuexian Zou:
Joint Noise and Reverberation Adaptive Learning for Robust Speaker DOA Estimation with an Acoustic Vector Sensor. 821-825 - Hong Liu, Haipeng Lan, Bing Yang, Cheng Pang:
Multiple Concurrent Sound Source Tracking Based on Observation-Guided Adaptive Particle Filter. 826-830 - Gurunath Reddy M., K. Sreenivasa Rao, Partha Pratim Das:
Harmonic-Percussive Source Separation of Polyphonic Music by Suppressing Impulsive Noise Events. 831-835 - Enea Ceolini, Jithendar Anumula, Adrian E. G. Huber, Ilya Kiselev, Shih-Chii Liu:
Speaker Activity Detection and Minimum Variance Beamforming for Source Separation. 836-840 - Xiaoke Qi, Jianhua Tao:
Sparsity-Constrained Weight Mapping for Head-Related Transfer Functions Individualization from Anthropometric Features. 841-845 - Dheeraj Sai D. V. L. N, Kishor K. S, Sri Rama Murty Kodukula:
Speech Source Separation Using ICA in Constant Q Transform Domain. 846-850 - Lu Yin, Ziteng Wang, Risheng Xia, Junfeng Li, Yonghong Yan:
Multi-talker Speech Separation Based on Permutation Invariant Training and Beamforming. 851-855 - Paul Magron, Tuomas Virtanen:
Expectation-Maximization Algorithms for Itakura-Saito Nonnegative Matrix Factorization. 856-860 - Girija Ramesan Karthik, Parth Suresh, Prasanta Kumar Ghosh:
Subband Weighting for Binaural Speech Source Localization. 861-865
Plenary Talk-1
- Jacqueline Vaissière:
Universal Tendencies for Cross-Linguistic Prosodic Tendencies: A Review and Some New Proposals. 866
Acoustic Model Adaptation
- Ondrej Klejch, Joachim Fainberg, Peter Bell:
Learning to Adapt: A Meta-learning Approach for Speaker Adaptation. 867-871 - Yu Wang, Chao Zhang, Mark J. F. Gales, Philip C. Woodland:
Speaker Adaptation and Adaptive Training for Jointly Optimised Tandem Systems. 872-876 - Markus Kitza, Ralf Schlüter, Hermann Ney:
Comparison of BLSTM-Layer-Specific Affine Transformations for Speaker Adaptation. 877-881 - Rini A. Sharon, Sandeep Reddy Kothinti, Srinivasan Umesh:
Correlational Networks for Speaker Normalization in Automatic Speech Recognition. 882-886 - Andros Tjandra, Sakriani Sakti, Satoshi Nakamura:
Machine Speech Chain with One-shot Speaker Adaptation. 887-891 - Khe Chai Sim, Arun Narayanan, Ananya Misra, Anshuman Tripathi, Golan Pundak, Tara N. Sainath, Parisa Haghani, Bo Li, Michiel Bacchiani:
Domain Adaptation Using Factorized Hidden Layer for Robust Automatic Speech Recognition. 892-896
Statistical Parametric Speech Synthesis
- Moquan Wan, Gilles Degottex, Mark J. F. Gales:
Waveform-Based Speaker Representations for Speech Synthesis. 897-901 - Tomoya Yanagita, Sakriani Sakti, Satoshi Nakamura:
Incremental TTS for Japanese Language. 902-906 - Ruibo Fu, Jianhua Tao, Yibin Zheng, Zhengqi Wen:
Transfer Learning Based Progressive Neural Networks for Acoustic Modeling in Statistical Parametric Speech Synthesis. 907-911 - Min-Jae Hwang, Eunwoo Song, Jin-Seob Kim, Hong-Goo Kang:
A Unified Framework for the Generation of Glottal Signals in Deep Learning-based Parametric Speech Synthesis Systems. 912-916 - Joun Yeop Lee, Sung Jun Cheon, Byoung Jin Choi, Nam Soo Kim, Eunwoo Song:
Acoustic Modeling Using Adversarially Trained Variational Recurrent Neural Network for Speech Synthesis. 917-921 - Yibin Zheng, Jianhua Tao, Zhengqi Wen, Ruibo Fu:
On the Application and Compression of Deep Time Delay Neural Network for Embedded Statistical Parametric Speech Synthesis. 922-926
Emotion Modeling
- Efthymios Tzinis, Georgios Paraskevopoulos, Christos Baziotis, Alexandros Potamianos:
Integrating Recurrence Dynamics for Speech Emotion Recognition. 927-931 - Wenjing Han, Huabin Ruan, Xiaomin Chen, Zhixiang Wang, Haifeng Li, Björn W. Schuller:
Towards Temporal Modelling of Categorical Speech Emotion Recognition. 932-936 - John W. Kim, Rif A. Saurous:
Emotion Recognition from Human Speech Using Temporal Information and Deep Learning. 937-940 - Kusha Sridhar, Srinivas Parthasarathy, Carlos Busso:
Role of Regularization in the Prediction of Valence from Speech. 941-945 - Karttikeya Mangalam, Tanaya Guha:
Learning Spontaneity to Improve Emotion Recognition in Speech. 946-950 - Reza Lotfian, Carlos Busso:
Predicting Categorical Emotions by Jointly Learning Primary and Secondary Emotions through Multitask Learning. 951-955
Models of Speech Perception
- Tiphaine Caudrelier, Pascal Perrier, Jean-Luc Schwartz, Amélie Rochet-Capellan:
Picture Naming or Word Reading: Does the Modality Affect Speech Motor Adaptation and Its Transfer? 956-960 - Yufan Du, Yi Shen, Hongying Yang, Xihong Wu, Jing Chen:
Measuring the Band Importance Function for Mandarin Chinese with a Bayesian Adaptive Procedure. 961-965 - Elnaz Shafaei-Bajestan, R. Harald Baayen:
Wide Learning for Auditory Comprehension. 966-970 - Louis ten Bosch, Mirjam Ernestus, Lou Boves:
Analyzing Reaction Time Sequences from Human Participants in Auditory Experiments. 971-975 - Jasper Ooster, Rainer Huber, Bernd T. Meyer:
Prediction of Perceived Speech Quality Using Deep Machine Listening. 976-980 - Paul Kranzusch, Rainer Huber, Melanie Krüger, Birger Kollmeier, Bernd T. Meyer:
Prediction of Subjective Listening Effort from Acoustic Data with Non-Intrusive Deep Models. 981-985
Multimodal Dialogue Systems
- Margarita Kotti, Vassilios Diakoloukas, Alexandros Papangelis, Michail Lagoudakis, Yannis Stylianou:
A Case Study on the Importance of Belief State Representation for Dialogue Policy Management. 986-990 - Kohei Hara, Koji Inoue, Katsuya Takanashi, Tatsuya Kawahara:
Prediction of Turn-taking Using Multitask Learning with Prediction of Backchannels and Fillers. 991-995 - Chandrakant Bothe, Sven Magg, Cornelius Weber, Stefan Wermter:
Conversational Analysis Using Utterance-level Attention-based Bidirectional Recurrent Neural Networks. 996-1000 - Yasuhito Ohsugi, Daisuke Saito, Nobuaki Minematsu:
A Comparative Study of Statistical Conversion of Face to Voice Based on Their Subjective Impressions. 1001-1005 - Ming-Hsiang Su, Chung-Hsien Wu, Kun-Yi Huang, Qian-Bei Hong, Huai-Hung Huang:
Follow-up Question Generation Using Pattern-based Seq2seq with a Small Corpus for Interview Coaching. 1006-1010 - Alessandra Cervone, Evgeny A. Stepanov, Giuseppe Riccardi:
Coherence Models for Dialogue. 1011-1015
Speech Recognition for Indian Languages
- K. E. Manjunath, K. Sreenivasa Rao, Dinesh Babu Jayagopi, V. Ramasubramanian:
Indian Languages ASR: A Multilingual Phone Recognition Framework with IPA Based Common Phone-set, Predicted Articulatory Features and Feature fusion. 1016-1020 - Agha Ali Raza, Awais Athar, Shan Randhawa, Zain Tariq, Muhammad Bilal Saleem, Haris Bin Zia, Umar Saif, Roni Rosenfeld:
Rapid Collection of Spontaneous Speech Corpora Using Telephonic Community Forums. 1021-1025 - Savitha Murthy, Dinkar Sitaram, Sunayana Sitaram:
Effect of TTS Generated Audio on OOV Detection and Word Error Rate in ASR for Low-resource Languages. 1026-1030 - Tanvina Patel, Krishna D. N, Noor Fathima, Nisar Shah, Mahima C, Deepak Kumar, Anuroop Iyengar:
Development of Large Vocabulary Speech Recognition System with Keyword Search for Manipuri. 1031-1035 - Abhishek Dey, Biswajit Dev Sarma, Wendy Lalhminghlui, Lalnunsiami Ngente, Parismita Gogoi, Priyankoo Sarmah, S. R. Mahadeva Prasanna, Rohit Sinha, S. R. Nirmala:
Robust Mizo Continuous Speech Recognition. 1036-1040 - Maharajan Chellapriyadharshini, Anoop Toffy, Srinivasa Raghavan K. M., V. Ramasubramanian:
Semi-supervised and Active-learning Scenarios: Efficient Acoustic Model Refinement for a Low Resource Indian Language. 1041-1045 - Debadatta Dash, Myung Jong Kim, Kristin Teplansky, Jun Wang:
Automatic Speech Recognition with Articulatory Information and a Unified Dictionary for Hindi, Marathi, Bengali and Oriya. 1046-1050
Show and Tell 3
- Aku Rouhe, Reima Karhila, Aija Elg, Minnaleena Toivola, Peter Smit, Anna-Riikka Smolander, Mikko Kurimo:
Captaina: Integrated Pronunciation Practice and Data Collection Portal. 1051-1052 - Umesh Sachdev, Rajagopal Jayaraman, Zainab Millwala:
auMina™ - Enterprise Speech Analytics. 1053-1054 - Annam Naresh, Rushabh Gandhi, Mallikarjuna Rao Bellamkonda, Mithun Das Gupta:
HoloCompanion: An MR Friend for EveryOne. 1055-1056 - Umesh Sachdev, Rajagopal Jayaraman, Zainab Millwala:
akeira™ - Virtual Assistant. 1057-1058 - Srihari Maruthachalam, Sidharth Aggarwal, Mari Ganesh Kumar, Mriganka Sur, Hema A. Murthy:
Brain-Computer Interface using Electroencephalogram Signatures of Eye Blinks. 1059-1060
Speaker Verification II
- Moez Ajili, Jean-François Bonastre, Solange Rossato:
Voice Comparison and Rhythm: Behavioral Differences between Target and Non-target Comparisons. 1061-1065 - Longting Xu, Kong-Aik Lee, Haizhou Li, Zhen Yang:
Co-whitening of I-vectors for Short and Long Duration Speaker Verification. 1066-1070 - Fahimeh Bahmaninezhad, John H. L. Hansen:
Compensation for Domain Mismatch in Text-independent Speaker Recognition. 1071-1075 - Ziqiang Shi, Liu Liu, Huibin Lin, Rujie Liu:
Joint Learning of J-Vector Extractor and Joint Bayesian Model for Text Dependent Speaker Verification. 1076-1080 - Ziqiang Shi, Huibin Lin, Liu Liu, Rujie Liu:
Latent Factor Analysis of Deep Bottleneck Features for Speaker Verification with Random Digit Strings. 1081-1085 - Joon Son Chung, Arsha Nagrani, Andrew Zisserman:
VoxCeleb2: Deep Speaker Recognition. 1086-1090 - Shreyas Ramoji, Sriram Ganapathy:
Supervised I-vector Modeling - Theory and Applications. 1091-1095 - Evgeny Dmitriev, Yulia Kim, Anastasia Matveeva, Claude Montacié, Yannick Boulard, Yadviga Sinyavskaya, Yulia Zhukova, Adam Zarazinski, Egor Akhanov, Ilya I. Viksnin, Andrei A. Shlykov, Maria Usova:
LOCUST - Longitudinal Corpus and Toolset for Speaker Verification. 1096-1100 - Srikanth R. Madikeri, Subhadeep Dey, Petr Motlícek:
Analysis of Language Dependent Front-End for Speaker Recognition. 1101-1105 - Mahesh Kumar Nandwana, Julien van Hout, Mitchell McLaren, Allen R. Stauffer, Colleen Richey, Aaron Lawson, Martin Graciarena:
Robust Speaker Recognition from Distant Speech under Real Reverberant Environments Using Speaker Embeddings. 1106-1110 - Phani Sankar Nidadavolu, Cheng-I Lai, Jesús Villalba, Najim Dehak:
Investigation on Bandwidth Extension for Speaker Recognition. 1111-1115 - Hannah Muckenhirn, Mathew Magimai-Doss, Sébastien Marcel:
On Learning Vocal Tract System Related Speaker Discriminative Information from Raw Signal Using CNNs. 1116-1120 - Rajath Kumar, Vaishnavi Yeruva, Sriram Ganapathy:
On Convolutional LSTM Modeling for Joint Wake-Word Detection and Text Dependent Speaker Verification. 1121-1125 - Zhongxin Bai, Xiao-Lei Zhang, Jingdong Chen:
Cosine Metric Learning for Speaker Verification in the I-vector Space. 1126-1130 - Arindam Jati, Panayiotis G. Georgiou:
An Unsupervised Neural Prediction Framework for Learning Speaker Embeddings Using Recurrent Neural Networks. 1131-1135
Novel Approaches to Enhancement
- Ashutosh Pandey, DeLiang Wang:
A New Framework for Supervised Speech Enhancement in the Time Domain. 1136-1140 - Jishnu Sadasivan, Subhadip Mukherjee, Chandra Sekhar Seelamantula:
Speech Enhancement Using the Minimum-probability-of-error Criterion. 1141-1145 - Pavlos Papadopoulos, Colin Vaz, Shrikanth S. Narayanan:
Exploring the Relationship between Conic Affinity of NMF Dictionaries and Speech Enhancement Metrics. 1146-1150 - Yun Liu, Hui Zhang, Xueliang Zhang:
Using Shifted Real Spectrum Mask as Training Target for Supervised Speech Separation. 1151-1155 - Nagapuri Srinivas, Gayadhar Pradhan, Syed Shahnawazuddin:
Enhancement of Noisy Speech Signal by Non-Local Means Estimation of Variational Mode Functions. 1156-1160 - Priya Pallavi, Ch. V. Rama Rao:
Phase-locked Loop (PLL) Based Phase Estimation in Single Channel Speech Enhancement. 1161-1164 - Zhong Meng, Jinyu Li, Yifan Gong, Biing-Hwang Fred Juang:
Cycle-Consistent Speech Enhancement. 1165-1169 - Aviv Gabbay, Asaph Shamir, Shmuel Peleg:
Visual Speech Enhancement. 1170-1174 - Saketh Sharma, Nitya Tiwari, Prem C. Pandey:
Implementation of Digital Hearing Aid as a Smartphone Application. 1175-1179 - Ching Hua Lee, Bhaskar D. Rao, Harinath Garudadri:
Bone-Conduction Sensor Assisted Noise Estimation for Improved Speech Enhancement. 1180-1184 - Pramod B. Bachhav, Massimiliano Todisco, Nicholas W. D. Evans:
Artificial Bandwidth Extension with Memory Inclusion Using Semi-supervised Stacked Auto-encoders. 1185-1189 - Soumi Maiti, Joey Ching, Michael I. Mandel:
Large Vocabulary Concatenative Resynthesis. 1190-1194 - Ali Raza Syed, Viet Anh Trinh, Michael I. Mandel:
Concatenative Resynthesis with Improved Training Signals for Speech Enhancement. 1195-1199
Syllabification, Rhythm, and Voice Activity Detection
- Okko Räsänen, Shreyas Seshadri, Marisa Casillas:
Comparison of Syllabification Algorithms and Training Strategies for Robust Word Count Estimation across Different Languages and Recording Conditions. 1200-1204 - Matthew C. Kelley, Benjamin V. Tucker:
A Comparison of Input Types to a Deep Neural Network-based Forced Aligner. 1205-1209 - Youngmoon Jung, Younggwan Kim, Yeunju Choi, Hoirin Kim:
Joint Learning Using Denoising Variational Autoencoders for Voice Activity Detection. 1210-1214 - Nauman Dawalatabad, Jom Kuriakose, Chellu Chandra Sekhar, Hema A. Murthy:
Information Bottleneck Based Percussion Instrument Diarization System for Taniavartanam Segments of Carnatic Music Concerts. 1215-1219 - Debayan Ghosh, R. Muralishankar, Sanjeev Gurugopinath:
Robust Voice Activity Detection Using Frequency Domain Long-Term Differential Entropy. 1220-1224 - Sri Harish Reddy Mallidi, Roland Maas, Kyle Goehner, Ariya Rastrow, Spyros Matsoukas, Björn Hoffmeister:
Device-directed Utterance Detection. 1225-1228 - Rohit M. A., Preeti Rao:
Acoustic-Prosodic Features of Tabla Bol Recitation and Correspondence with the Tabla Imitation. 1229-1233 - Teun F. Krikke, Frank Broz, David Lane:
Who Said That? a Comparative Study of Non-negative Matrix Factorization Techniques. 1234-1238 - Sourish Chaudhuri, Joseph Roth, Daniel P. W. Ellis, Andrew C. Gallagher, Liat Kaver, Radhika Marvin, Caroline Pantofaru, Nathan Reale, Loretta Guarino Reid, Kevin W. Wilson, Zhonghua Xi:
AVA-Speech: A Densely Labeled Dataset of Speech Activity in Movies. 1239-1243 - Fei Tao, Carlos Busso:
Audiovisual Speech Activity Detection with Advanced Long Short-Term Memory. 1244-1248 - Pramit Saha, Praneeth Srungarapu, Sidney S. Fels:
Towards Automatic Speech Identification from Vocal Tract Shape Dynamics in Real-time MRI. 1249-1253
Selected Topics in Neural Speech Processing
- Kaiyu Shi, Kai Yu:
Structured Word Embedding for Low Memory Neural Network Language Model. 1254-1258 - Ryo Masumura, Tomohiro Tanaka, Atsushi Ando, Hirokazu Masataki, Yushi Aono:
Role Play Dialogue Aware Language Models Based on Conditional Hierarchical Recurrent Encoder-Decoder. 1259-1263 - Samuel Myer, Vikrant Singh Tomar:
Efficient Keyword Spotting Using Time Delay Neural Networks. 1264-1268 - Tsukasa Yoshida, Takafumi Moriya, Kazuho Watanabe, Yusuke Shinohara, Yoshikazu Yamaguchi, Yushi Aono:
Automatic DNN Node Pruning Using Mixture Distribution-based Group Regularization. 1269-1273 - Raffaele Tavarone, Leonardo Badino:
Conditional-Computation-Based Recurrent Neural Networks for Computationally Efficient Acoustic Modelling. 1274-1278 - Antonios Anastasopoulos, David Chiang:
Leveraging Translations for Speech Transcription in Low-resource Settings. 1279-1283 - Antoine Bruguier, Heiga Zen, Arkady Arkhangorodsky:
Sequence-to-sequence Neural Network Model with 2D Attention for Learning Japanese Pitch Accents. 1284-1287 - Sahar Ghannay, Yannick Estève, Nathalie Camelin:
Task Specific Sentence Embeddings for ASR Error Detection. 1288-1292 - Jan Niehues, Ngoc-Quan Pham, Thanh-Le Ha, Matthias Sperber, Alex Waibel:
Low-Latency Neural Speech Translation. 1293-1297 - Sameer Bansal, Herman Kamper, Karen Livescu, Adam Lopez, Sharon Goldwater:
Low-Resource Speech-to-Text Translation. 1298-1302 - Ferdinand Brasser, Tommaso Frassetto, Korbinian Riedhammer, Ahmad-Reza Sadeghi, Thomas Schneider, Christian Weinert:
VoiceGuard: Secure and Private Speech Processing. 1303-1307
Perspective Talk-1
- Dilek Hakkani-Tür:
Deep Learning based Situated Goal-oriented Dialogue Systems. 1308
Dereverberation
- Chenxing Li, Tieqiang Wang, Shuang Xu, Bo Xu:
Single-channel Speech Dereverberation via Generative Adversarial Training. 1309-1313 - Wolfgang Mack, Soumitro Chakrabarty, Fabian-Robert Stöter, Sebastian Braun, Bernd Edler, Emanuël A. P. Habets:
Single-Channel Dereverberation Using Direct MMSE Optimization and Bidirectional LSTM Networks. 1314-1318 - Ina Kodrasi, Hervé Bourlard:
Single-channel Late Reverberation Power Spectral Density Estimation Using Denoising Autoencoders. 1319-1323 - Nikhil Mohanan, Rajbabu Velmurugan, Preeti Rao:
A Non-convolutive NMF Model for Speech Dereverberation. 1324-1328 - Peter Guzewich, Stephen A. Zahorian, Xiao Chen, Hao Zhang:
Cross-Corpora Convolutional Deep Neural Network Dereverberation Preprocessing for Speaker Verification and Speech Enhancement. 1329-1333 - Ladislav Mosner, Oldrich Plchot, Pavel Matejka, Ondrej Novotný, Jan Cernocký:
Dereverberation and Beamforming in Robust Far-Field Speaker Recognition. 1334-1338
Audio Events and Acoustic Scenes
- Yun Wang, Juncheng Li, Florian Metze:
Comparing the Max and Noisy-Or Pooling Functions in Multiple Instance Learning for Weakly Supervised Sequence Learning Tasks. 1339-1343 - Weiran Wang, Chieh-Chi Kao, Chao Wang:
A Simple Model for Detection of Rare Sound Events. 1344-1348 - Teng Zhang, Kailai Zhang, Ji Wu:
Temporal Transformer Networks for Acoustic Scene Classification. 1349-1353 - Xugang Lu, Peng Shen, Sheng Li, Yu Tsao, Hisashi Kawai:
Temporal Attentive Pooling for Acoustic Event Detection. 1354-1357 - Chieh-Chi Kao, Weiran Wang, Ming Sun, Chao Wang:
R-CRNN: Region-based Convolutional Recurrent Neural Network for Audio Event Detection. 1358-1362 - Constantinos Papayiannis, Justice Amoh, Viktor Rozgic, Shiva Sundaram, Chao Wang:
Detecting Media Sound Presence in Acoustic Scenes. 1363-1367
Speaker Diarization
- Pierre-Alexandre Broux, Florent Desnous, Anthony Larcher, Simon Petitrenaud, Jean Carrive, Sylvain Meignier:
S4D: Speaker Diarization Toolkit in Python. 1368-1372 - Tae Jin Park, Panayiotis G. Georgiou:
Multimodal Speaker Segmentation and Diarization Using Lexical and Acoustic Cues via Sequence to Sequence Neural Networks. 1373-1377 - Nikolaos Flemotomos, Pavlos Papadopoulos, James Gibson, Shrikanth S. Narayanan:
Combined Speaker Clustering and Role Recognition in Conversational Speech. 1378-1382 - Adrien Le Franc, Eric Riebling, Julien Karadayi, Yun Wang, Camila Scaff, Florian Metze, Alejandrina Cristià:
The ACLEW DiViMe: An Easy-to-use Diarization Tool. 1383-1387 - Evdokia Kazimirova, Andrey Belyaev:
Automatic Detection of Multi-speaker Fragments with High Time Resolution. 1388-1392 - Ruiqing Yin, Hervé Bredin, Claude Barras:
Neural Speech Turn Segmentation and Affinity Propagation for Speaker Diarization. 1393-1397
Phonation
- Minghui Zhang, Fang Hu:
Pitch or Phonation: on the Glottalization in Tone Productions in the Ruokeng Hui Chinese Dialect. 1398-1402 - Marc Antony Hullebus, Stephen J. Tobin, Adamantios I. Gafos:
Speaker-specific Structure in German Voiceless Stop Voice Onset Times. 1403-1407 - Kätlin Aare, Pärtel Lippus, Marcin Wlodarczak, Mattias Heldner:
Creak in the Respiratory Cycle. 1408-1412 - Cuiling Zhang, Bin Li, Si Chen, Yike Yang:
Acoustic Analysis of Whispery Voice Disguise in Mandarin Chinese. 1413-1416 - Dieter Maurer, Christian d'Heureuse, Heidy Suter, Volker Dellwo, Daniel Friedrichs, Thayabaran Kathiresan:
The Zurich Corpus of Vowel and Voice Quality, Version 1.0. 1417-1421 - Joshua Penney, Felicity Cox, Anita Szakay:
Weighting of Coda Voicing Cues: Glottalisation and Vowel Duration. 1422-1426
Cognition and Brain Studies
- Bin Zhao, Jinfeng Huang, Gaoyan Zhang, Jianwu Dang, Minbo Chen, YingjianFu, Longbiao Wang:
Revealing Spatiotemporal Brain Dynamics of Speech Production Based on EEG and Eye Movement. 1427-1431 - Natalie Boll-Avetisyan, Jessie S. Nixon, Tomas O. Lentz, Liquan Liu, Sandrien van Ommen, Çagri Çöltekin, Jacolien van Rij:
Neural Response Development During Distributional Learning. 1432-1436 - Akshay Raj Maggu, Wenqing Zong, Vina Law, Patrick C. M. Wong:
Learning Two Tone Languages Enhances the Brainstem Encoding of Lexical Tones. 1437-1441 - Daniel Williams, Paola Escudero, Adamantios I. Gafos:
Perceptual Sensitivity to Spectral Change in Australian English Close Front Vowels: An Electroencephalographic Investigation. 1442-1446 - Jessie S. Nixon:
Effective Acoustic Cue Learning Is Not Just Statistical, It Is Discriminative. 1447-1451 - Kimberley Mulder, Louis ten Bosch, Lou Boves:
Analyzing EEG Signals in Auditory Speech Comprehension Using Temporal Response Functions and Generalized Additive Models. 1452-1456
Deep Neural Networks: How Can We Interpret What They Learned?
- Louis ten Bosch, Lou Boves:
Information Encoding by Deep Neural Networks: What Can We Learn? 1457-1461 - Wei-Ning Hsu, James R. Glass:
Scalable Factorized Hierarchical Variational Autoencoder Training. 1462-1466 - Lyan Verwimp, Hugo Van hamme, Vincent Renkens, Patrick Wambacq:
State Gradients for RNN Memory Analysis. 1467-1471 - Linxue Bai, Philip Weber, Peter Jancovic, Martin J. Russell:
Exploring How Phone Classification Neural Networks Learn Phonetic Information by Visualising and Interpreting Bottleneck Features. 1472-1476 - Jeroen Zegers, Hugo Van hamme:
Memory Time Span in LSTMs for Multi-Speaker Source Separation. 1477-1481 - Odette Scharenborg, Sebastian Tiesmeyer, Mark Hasegawa-Johnson, Najim Dehak:
Visualizing Phoneme Category Adaptation in Deep Neural Networks. 1482-1486
Show and Tell 4
- G. R. Kasthuri, Prabha Ramanathan, Hema A. Murthy, Namita Jacob, Anil Prabhakar:
Early Vocabulary Development Through Picture-based Software Solutions. 1487-1488 - Kamini Sabu, Kanhaiya Kumar, Preeti Rao:
Automatic Detection of Expressiveness in Oral Reading. 1489-1490 - Madhab Pal, Rajib Roy, Soma Khan, Milton Samirakshma Bepari, Joyanta Basu:
PannoMulloKathan: Voice Enabled Mobile App for Agricultural Commodity Price Dissemination in Bengali Language. 1491-1492 - Alp Öktem, Mireia Farrús, Antonio Bonafonte:
Visualizing Punctuation Restoration in Speech Transcripts with Prosograph. 1493-1494 - Mithul Mathivanan, Kinnera Saranu, Abhishek Pandey, Jithendra Vepa:
CACTAS - Collaborative Audio Categorization and Transcription for ASR Systems. 1495-1496
Speech and Singing Production
- Benjamin Parrell, Vikram Ramanarayanan, Srikantan S. Nagarajan, John F. Houde:
FACTS: A Hierarchical Task-based Control Model of Speech Incorporating Sensory Feedback. 1497-1501 - William F. Katz, Patrick Reidy, Divya Prabhakaran:
Sensorimotor Response to Tongue Displacement Imagery by Talkers with Parkinson's Disease. 1502-1506 - Chitralekha Gupta, Haizhou Li, Ye Wang:
Automatic Pronunciation Evaluation of Singing. 1507-1511 - Rachel E. Bouserhal, Philippe Chabot, Milton Sarria Paja, Patrick Cardinal, Jérémie Voix:
Classification of Nonverbal Human Produced Audio Events: A Pilot Study. 1512-1516 - Lorenzo Spreafico, Michael Pucher, Anna Matosova:
UltraFit: A Speaker-friendly Headset for Ultrasound Recordings in Speech Science. 1517-1520 - Elísabet Eir Cortes, Marcin Wlodarczak, Juraj Simko:
Articulatory Consequences of Vocal Effort Elicitation Method. 1521-1525 - Anne Hermes, Jane Mertens, Doris Mücke:
Age-related Effects on Sensorimotor Control of Speech Production. 1526-1530 - Maida Percival, Alexei Kochetov, Yoonjung Kang:
An Ultrasound Study of Gemination in Coronal Stops in Eastern Oromo. 1531-1535 - Protima Nomo Sudro, Sishir Kalita, S. R. Mahadeva Prasanna:
Processing Transition Regions of Glottal Stop Substituted /S/ for Intelligibility Enhancement of Cleft Palate Speech. 1536-1540 - Abinay Reddy Naini, M. V. Achuth Rao, G. Nisha Meenakshi, Prasanta Kumar Ghosh:
Reconstructing Neutral Speech from Tracheoesophageal Speech. 1541-1545 - Keiko Ochi, Koichi Mori, Naomi Sakai:
Automatic Evaluation of Soft Articulatory Contact for Stuttering Treatment. 1546-1550 - Juntae Kim, Heejin Choi, Jinuk Park, Minsoo Hahn, Sang-Jin Kim, Jong-Jin Kim:
Korean Singing Voice Synthesis Based on an LSTM Recurrent Neural Network. 1551-1555 - Xuanda Chen, Ziyu Xiong, Jian Hu:
The Trajectory of Voice Onset Time with Vocal Aging. 1556-1560
Robust Speech Recognition
- Jon Barker, Shinji Watanabe, Emmanuel Vincent, Jan Trmal:
The Fifth 'CHiME' Speech Separation and Recognition Challenge: Dataset, Task and Baselines. 1561-1565 - Colleen Richey, María Auxiliadora Barrios, Zeb Armstrong, Chris Bartels, Horacio Franco, Martin Graciarena, Aaron Lawson, Mahesh Kumar Nandwana, Allen R. Stauffer, Julien van Hout, Paul Gamble, Jeffrey Hetherly, Cory Stephenson, Karl Ni:
Voices Obscured in Complex Environmental Settings (VOiCES) Corpus. 1566-1570 - Szu-Jui Chen, Aswin Shanmugam Subramanian, Hainan Xu, Shinji Watanabe:
Building State-of-the-art Distant Speech Recognition Using the CHiME-4 Challenge with a Setup of Speech Enhancement Baseline. 1571-1575 - Wei-Ning Hsu, Hao Tang, James R. Glass:
Unsupervised Adaptation with Interpretable Disentangled Representations for Distant Conversational Speech Recognition. 1576-1580 - Ke Wang, Junbo Zhang, Sining Sun, Yujun Wang, Fei Xiang, Lei Xie:
Investigating Generative Adversarial Networks Based Speech Dereverberation for Robust Speech Recognition. 1581-1585 - Xuankai Chang, Yanmin Qian, Dong Yu:
Monaural Multi-Talker Speech Recognition with Attention Mechanism and Gated Convolutional Networks. 1586-1590 - Cong-Thanh Do, Yannis Stylianou:
Weighting Time-Frequency Representation of Speech Using Auditory Saliency for Automatic Speech Recognition. 1591-1595 - Pegah Ghahremani, Hossein Hadian, Hang Lv, Daniel Povey, Sanjeev Khudanpur:
Acoustic Modeling from Frequency Domain Representations of Speech. 1596-1600 - Ishwar Chandra Yadav, Avinash Kumar, Syed Shahnawazuddin, Gayadhar Pradhan:
Non-Uniform Spectral Smoothing for Robust Children's Speech Recognition. 1601-1605 - Aaron Nicolson, Kuldip K. Paliwal:
Bidirectional Long-Short Term Memory Network-based Estimation of Reliable Spectral Component Locations. 1606-1610 - Lili Guo, Longbiao Wang, Jianwu Dang, Linjuan Zhang, Haotian Guan, Xiangang Li:
Speech Emotion Recognition by Combining Amplitude and Phase Information Using Convolutional Neural Network. 1611-1615 - Viet Anh Trinh, Brian McFee, Michael I. Mandel:
Bubble Cooperative Networks for Identifying Important Speech Cues. 1616-1620
Applications in Education and Learning
- Jian Cheng:
Real-Time Scoring of an Oral Reading Assessment on Mobile Devices. 1621-1625 - Konstantinos Kyriakopoulos, Kate M. Knill, Mark J. F. Gales:
A Deep Learning Approach to Assessing Non-native Pronunciation of English Using Phone Distances. 1626-1630 - Yujia Xiao, Frank K. Soong, Wenping Hu:
Paired Phone-Posteriors Approach to ESL Pronunciation Quality Assessment. 1631-1635 - Ming Tu, Anna Grabek, Julie Liss, Visar Berisha:
Investigating the Role of L1 in Automatic Pronunciation Evaluation of L2 Speech. 1636-1640 - Kate M. Knill, Mark J. F. Gales, Konstantinos Kyriakopoulos, Andrey Malinin, Anton Ragni, Yu Wang, Andrew Caines:
Impact of ASR Performance on Free Speaking Language Assessment. 1641-1645 - Yoon Seok Hong, Kyung Seo Ki, Gahgene Gweon:
Automatic Miscue Detection Using RNN Based Models with Data Augmentation. 1646-1650 - Yusuke Inoue, Suguru Kabashima, Daisuke Saito, Nobuaki Minematsu, Kumi Kanamura, Yutaka Yamauchi:
A Study of Objective Measurement of Comprehensibility through Native Speakers' Shadowing of Learners' Utterances. 1651-1655 - Dean Luo, Chunxiao Zhang, Linzhong Xia, Lixin Wang:
Factorized Deep Neural Network Adaptation for Automatic Scoring of L2 Speech in English Speaking Tests. 1656-1660 - Gary Yeung, Abeer Alwan:
On the Difficulties of Automatic Speech Recognition for Kindergarten-Aged Children. 1661-1665 - Mauro Nicolao, Michiel Sanders, Thomas Hain:
Improved Acoustic Modelling for Automatic Literacy Assessment of Children. 1666-1670
Integrating Speech Science and Technology for Clinical Applications
- Mostafa Ali Shahin, Beena Ahmed, Jim X. Ji, Kirrie J. Ballard:
Anomaly Detection Approach for Pronunciation Verification of Disordered Speech Using Speech Attribute Features. 1671-1675 - Amber Afshan, Jinxi Guo, Soo Jin Park, Vijay Ravi, Jonathan Flint, Abeer Alwan:
Effectiveness of Voice Quality Features in Detecting Depression. 1676-1680 - Prasanna V. Kothalkar, Johanna Rudolph, Christine Dollaghan, Jennifer McGlothlin, Thomas F. Campbell, John H. L. Hansen:
Fusing Text-dependent Word-level i-Vector Models to Screen 'at Risk' Child Speech. 1681-1685 - Ram Charan Chandra Shekar, Hussnain Ali, John H. L. Hansen:
Testing Paradigms for Assistive Hearing Devices in Diverse Acoustic Environments. 1686-1690 - Tsuyoki Ujiro, Hiroki Tanaka, Hiroyoshi Adachi, Hiroaki Kazui, Manabu Ikeda, Takashi Kudo, Satoshi Nakamura:
Detection of Dementia from Responses to Atypical Questions Asked by Embodied Conversational Agents. 1691-1695 - Wang Zhang, Xiangquan Gui, Tianqi Wang, Manwa L. Ng, Feng Yang, Lan Wang, Nan Yan:
Acoustic Features Associated with Sustained Vowel and Continuous Speech Productions by Chinese Children with Functional Articulation Disorders. 1696-1700 - Vikram C. M., Ayush Tripathi, Sishir Kalita, S. R. Mahadeva Prasanna:
Estimation of Hypernasality Scores from Cleft Lip and Palate Speech. 1701-1705 - Tifani Warnita, Nakamasa Inoue, Koichi Shinoda:
Detecting Alzheimer's Disease Using Gated Convolutional Neural Network from Audio Data. 1706-1710 - Andrea Bandini, Jordan R. Green, Brian Richburg, Yana Yunusova:
Automatic Detection of Orofacial Impairment in Stroke. 1711-1715 - Tuka Al Hanai, Mohammad M. Ghassemi, James R. Glass:
Detecting Depression with Audio/Text Sequence Modeling of Interviews. 1716-1720
Speaker Characterization and Analysis
- Yu-Wun Wang, Hen-Hsen Huang, Kuan-Yu Chen, Hsin-Hsi Chen:
Discourse Marker Detection for Hesitation Events on Mandarin Conversation. 1721-1725 - Puyang Geng, Wentao Gu, Hiroya Fujisaki:
Acoustic and Perceptual Characteristics of Mandarin Speech in Homosexual and Heterosexual Male Speakers. 1726-1730 - Atsushi Ando, Reine Asakawa, Ryo Masumura, Hosana Kamiyama, Satoshi Kobashikawa, Yushi Aono:
Automatic Question Detection from Acoustic and Phonetic Features Using Feature-wise Pre-training. 1731-1735 - Fasih Haider, Saturnino Luz, Carl Vogel, Nick Campbell:
Improving Response Time of Active Speaker Detection Using Visual Prosody Information Prior to Articulation. 1736-1740 - Bekir Berker Türker, Engin Erzin, Yücel Yemez, T. Metin Sezgin:
Audio-Visual Prediction of Head-Nod and Turn-Taking Events in Dyadic Interactions. 1741-1745 - Haoran Wu, Yuya Chiba, Takashi Nose, Akinori Ito:
Analyzing Effect of Physical Expression on English Proficiency for Multimodal Computer-Assisted Language Learning. 1746-1750 - Sri Harsha Dumpala, Ashish Panda, Sunil Kumar Kopparapu:
Analysis of the Effect of Speech-Laugh on Speaker Recognition System. 1751-1755 - Jennifer Sloboda, Adam C. Lammert, James R. Williamson, Christopher J. Smalt, Daryush D. Mehta, C. O. L. Ian Curry, Kristin Heaton, Jeff Palmer, Thomas F. Quatieri:
Vocal Biomarkers for Cognitive Performance Estimation in a Working Memory Task. 1756-1760 - Guozhen An, Rivka Levitan:
Lexical and Acoustic Deep Learning Model for Personality Recognition. 1761-1765
Perspective Talk-2
- Bhuvana Ramabhadran:
Open Problems in Speech Recognition. 1766
Plenary Talk-2
- Hervé Bourlard:
Evolution of Neural Network Architectures for Speech Recognition. 1767
Novel Neural Network Architectures for Acoustic Modelling
- Jinyu Li, Changliang Liu, Yifan Gong:
Layer Trajectory LSTM. 1768-1772 - Chao Zhang, Philip C. Woodland:
Semi-tied Units for Efficient Gating in LSTM and Highway Networks. 1773-1777 - Max W. Y. Lam, Shoukang Hu, Xurong Xie, Shansong Liu, Jianwei Yu, Rongfeng Su, Xunying Liu, Helen Meng:
Gaussian Process Neural Networks for Speech Recognition. 1778-1782 - Jian Tang, Yan Song, Lirong Dai, Ian McLoughlin:
Acoustic Modeling with Densely Connected Residual Network for Multichannel Speech Recognition. 1783-1787 - Jie Li, Xiaorui Wang, Yuanyuan Zhao, Yan Li:
Gated Recurrent Unit Based Acoustic Modeling with Future Context. 1788-1792 - Gaofeng Cheng, Daniel Povey, Lu Huang, Ji Xu, Sanjeev Khudanpur, Yonghong Yan:
Output-Gate Projected Gated Recurrent Unit for Speech Recognition. 1793-1797
Language Identification
- Seyed Omid Sadjadi, Timothée Kheyrkhah, Craig S. Greenberg, Elliot Singer, Douglas A. Reynolds, Lisa P. Mason, Jaime Hernandez-Cordero:
Performance Analysis of the 2017 NIST Language Recognition Evaluation. 1798-1802 - Lukás Mateju, Petr Cerva, Jindrich Zdánský, Radek Safarík:
Using Deep Neural Networks for Identification of Slavic Languages from Acoustic Signal. 1803-1807 - Hagai Taitelbaum, Ehud Ben-Reuven, Jacob Goldberger:
Adding New Classes without Access to the Original Training Data with Applications to Language Identification. 1808-1812 - Peng Shen, Xugang Lu, Sheng Li, Hisashi Kawai:
Feature Representation of Short Utterances Based on Knowledge Distillation for Spoken Language Identification. 1813-1817 - Sarith Fernando, Vidhyasaharan Sethu, Eliathamby Ambikairajah:
Sub-band Envelope Features Using Frequency Domain Linear Prediction for Short Duration Language Identification. 1818-1822 - Peter Sibbern Frederiksen, Jesús Villalba, Shinji Watanabe, Zheng-Hua Tan, Najim Dehak:
Effectiveness of Single-Channel BLSTM Enhancement for Language Identification. 1823-1827
Production of Prosody
- Erica Gold:
Articulation Rate as a Speaker Discriminant in British English. 1828-1832 - Jenny Yu, Katharina Zahner:
Truncation and Compression in Southern German and Australian English. 1833-1837 - Heini Kallio, Antti Suni, Päivi Virkkunen, Juraj Simko:
Prominence-based Evaluation of L2 Prosody. 1838-1842 - Rachid Ridouane, Giuseppina Turco, Julien Meyer:
Length Contrast and Covarying Features: Whistled Speech as a Case Study. 1843-1847 - Eleanor Chodroff, Jennifer S. Cole:
Information Structure, Affect and Prenuclear Prominence in American English. 1848-1852 - John S. Novak III, Robert V. Kenyon:
Effects of User Controlled Speech Rate on Intelligibility in Noisy Environments. 1853-1857
Speech Intelligibility and Quality
- Kazuhiro Kondo, Kazuya Taira, Yosuke Kobayashi:
Binaural Speech Intelligibility Estimation Using Deep Neural Networks. 1858-1862 - Katsuhiko Yamamoto, Toshio Irino, Narumi Ohashi, Shoko Araki, Keisuke Kinoshita, Tomohiro Nakatani:
Multi-resolution Gammachirp Envelope Distortion Index for Intelligibility Prediction of Noisy Speech. 1863-1867 - P. V. Muhammed Shifas, Vassilis Tsiaras, Yannis Stylianou:
Speech Intelligibility Enhancement Based on a Non-causal Wavenet-like Model. 1868-1872 - Szu-Wei Fu, Yu Tsao, Hsin-Te Hwang, Hsin-Min Wang:
Quality-Net: An End-to-End Non-intrusive Speech Quality Assessment Model Based on BLSTM. 1873-1877 - Rohith Aralikatti, Dilip Kumar Margam, Tanay Sharma, Abhinav Thanda, Shankar M. Venkatesan:
Global SNR Estimation of Speech Signals Using Entropy and Uncertainty Estimates from Dropout Networks. 1878-1882 - Gabriel Mittag, Sebastian Möller:
Detecting Packet-Loss Concealment Using Formant Features and Decision Tree Learning. 1883-1887
Integrating Speech Science and Technology for Clinical Applications
- Aciel Eshky, Manuel Sam Ribeiro, Joanne Cleland, Korin Richmond, Zoe Roxburgh, James M. Scobbie, Alan Wrench:
UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions. 1888-1892 - Bahman Mirheidari, Daniel Blackburn, Traci Walker, Annalena Venneri, Markus Reuber, Heidi Christensen:
Detecting Signs of Dementia Using Word Vector Representations. 1893-1897 - Matthew Perez, Wenyu Jin, Duc Le, Noelle Carlozzi, Praveen Dayalu, Angela Roberts, Emily Mower Provost:
Classification of Huntington Disease Using Acoustic and Lexical Features. 1898-1902 - Soheil Khorram, Mimansa Jaiswal, John Gideon, Melvin G. McInnis, Emily Mower Provost:
The PRIORI Emotion Dataset: Linking Mood to Emotion Detected In-the-Wild. 1903-1907 - Nikolaos Flemotomos, Victor R. Martinez, James Gibson, David C. Atkins, Torrey A. Creed, Shrikanth S. Narayanan:
Language Features for Automated Evaluation of Cognitive Behavior Psychotherapy Sessions. 1908-1912 - Kwanghoon An, Myung Jong Kim, Kristin Teplansky, Jordan R. Green, Thomas F. Campbell, Yana Yunusova, Daragh Heitzman, Jun Wang:
Automatic Early Detection of Amyotrophic Lateral Sclerosis from Intelligible Speech Using Convolutional Neural Networks. 1913-1917
Speech Technologies for Code-Switching in Multilingual Communities
- Preeti Rao, Mugdha Pandya, Kamini Sabu, Kanhaiya Kumar, Nandini Bondale:
A Study of Lexical and Prosodic Cues to Segmentation in a Hindi-English Code-switched Discourse. 1918-1922 - Emre Yilmaz, Astik Biswas, Ewald van der Westhuizen, Febe de Wet, Thomas Niesler:
Building a Unified Code-Switching ASR System for South African Languages. 1923-1927 - Pengcheng Guo, Haihua Xu, Lei Xie, Eng Siong Chng:
Study of Semi-supervised Approaches to Improving English-Mandarin Code-Switching Speech Recognition. 1928-1932 - Emre Yilmaz, Henk van den Heuvel, David A. van Leeuwen:
Acoustic and Textual Data Augmentation for Improved ASR of Code-Switching Speech. 1933-1937 - Victor Soto, Nishmar Cestero, Julia Hirschberg:
The Role of Cognate Words, POS Tags and Entrainment in Code-Switching. 1938-1942 - Brij Mohan Lal Srivastava, Sunayana Sitaram:
Homophone Identification and Merging for Code-switched Speech Recognition. 1943-1947 - Anju Leela Thomas, Anusha Prakash, Arun Baby, Hema A. Murthy:
Code-switching in Indic Speech Synthesisers. 1948-1952 - Ganji Sreeram, Rohit Sinha:
A Novel Approach for Effective Recognition of the Code-Switched Data on Monolingual Language Model. 1953-1957
Show and Tell 5
- Ramya Viswanathan, Periyasamy Paramasivam, Jithendra Vepa:
Hierarchical Accent Determination and Application in a Large Scale ASR System. 1958-1959 - Vikram Ramanarayanan, David Pautler, Patrick L. Lange, Eugene Tsuprun, Rutuja Ubale, Keelan Evanini, David Suendermann-Oeft:
Toward Scalable Dialog Technology for Conversational Language Learning: Case Study of the TOEFL® MOOC. 1960-1961 - João Freitas, Jorge Ribeiro, Daan Baldewijns, Sara Oliveira, Daniela Braga:
Machine Learning Powered Data Platform for High-Quality Speech and NLP Workflows. 1962-1963 - Raphael Cohen, Orgad Keller, Jason Levy, Russell Levy, Micha Breakstone, Amit Ashkenazi:
Fully Automatic Speaker Separation System, with Automatic Enrolling of Recurrent Speakers. 1964-1965 - Madhavaraj Ayyavu, Shiva Kumar H. R., A. G. Ramakrishnan:
Online Speech Translation System for Tamil. 1966-1967
Voice Conversion and Speech Synthesis
- Nirmesh J. Shah, Maulik C. Madhavi, Hemant A. Patil:
Unsupervised Vocal Tract Length Warped Posterior Features for Non-Parallel Voice Conversion. 1968-1972 - Cong Zhou, Michael Horgan, Vivek Kumar, Cristina Vasco, Dan Darcy:
Voice Conversion with Conditional SampleRNN. 1973-1977 - Berrak Sisman, Mingyang Zhang, Haizhou Li:
A Voice Conversion Framework with Tandem Feature Sparse Representation and Speaker-Adapted WaveNet Vocoder. 1978-1982 - Li-Juan Liu, Zhen-Hua Ling, Yuan Jiang, Ming Zhou, Li-Rong Dai:
WaveNet Vocoder with Limited Training Data for Voice Conversion. 1983-1987 - Yi-Chiao Wu, Kazuhiro Kobayashi, Tomoki Hayashi, Patrick Lumban Tobing, Tomoki Toda:
Collapsed Speech Segment Detection and Suppression for WaveNet Vocoder. 1988-1992 - Kuan Chen, Bo Chen, Jiahao Lai, Kai Yu:
High-quality Voice Conversion Using Spectrogram-Based WaveNet Vocoder. 1993-1997 - Antonio Bonafonte, Santiago Pascual, Georgina Dorca:
Spanish Statistical Parametric Speech Synthesis Using a Neural Vocoder. 1998-2001 - Monika Podsiadlo, Victor Ungureanu:
Experiments with Training Corpora for Statistical Text-to-speech Systems. 2002-2006 - Yu Gu, Yongguo Kang:
Multi-task WaveNet: A Multi-task Generative Model for Statistical Parametric Speech Synthesis without Fundamental Frequency Conditions. 2007-2011 - Lauri Juvela, Vassilis Tsiaras, Bajibabu Bollepalli, Manu Airaksinen, Junichi Yamagishi, Paavo Alku:
Speaker-independent Raw Waveform Model for Glottal Excitation. 2012-2016 - Yang Cui, Xi Wang, Lei He, Frank K. Soong:
A New Glottal Neural Vocoder for Speech Synthesis. 2017-2021 - Oliver Watts, Cassia Valentini-Botinhao, Felipe Espic, Simon King:
Exemplar-based Speech Waveform Generation. 2022-2026 - Hideki Kawahara, Ken-Ichi Sakakibara, Masanori Morise, Hideki Banno, Tomoki Toda, Toshio Irino:
Frequency Domain Variants of Velvet Noise and Their Application to Speech Processing and Synthesis. 2027-2031
Extracting Information from Audio
- Pei-Hung Chung, Kuan Tung, Ching-Lun Tai, Hung-yi Lee:
Joint Learning of Interactive Spoken Content Retrieval and Trainable User Simulator. 2032-2036 - Changhao Shan, Junbo Zhang, Yujun Wang, Lei Xie:
Attention-based End-to-End Models for Small-Footprint Keyword Spotting. 2037-2041 - Ragesh Rajan M, Ashwin Vijayakumar, Deepu Vijayasenan:
Prediction of Aesthetic Elements in Karnatic Music: A Machine Learning Approach. 2042-2046 - Wenda Chen, Mark Hasegawa-Johnson, Nancy F. Chen:
Topic and Keyword Identification for Low-resourced Speech Using Cross-Language Transfer Learning. 2047-2051 - Matthew Wiesner, Chunxi Liu, Lucas Ondel, Craig Harman, Vimal Manohar, Jan Trmal, Zhongqiang Huang, Najim Dehak, Sanjeev Khudanpur:
Automatic Speech Recognition and Topic Identification from Speech for Almost-Zero-Resource Languages. 2052-2056 - Bo Xiao, Nicholas Monath, Shankar Ananthakrishnan, Abishek Ravi:
Play Duration Based User-Entity Affinity Modeling in Spoken Dialog System. 2057-2061 - Shi-wook Lee, Kazuyo Tanaka, Yoshiaki Itoh:
Empirical Analysis of Score Fusion Application to Combined Neural Networks for Open Vocabulary Spoken Term Detection. 2062-2066 - Afsaneh Asaei, Dhananjay Ram, Hervé Bourlard:
Phonological Posterior Hashing for Query by Example Spoken Term Detection. 2067-2071 - Maren Kucza, Jan Niehues, Thomas Zenkel, Alex Waibel, Sebastian Stüker:
Term Extraction via Neural Sequence Labeling a Comparative Evaluation of Strategies Using Recurrent Neural Networks. 2072-2076 - Anjuli Kannan, Kai Chen, Diana Jaunzeikare, Alvin Rajkomar:
Semi-supervised Learning for Information Extraction from Dialogue. 2077-2081 - Youhyun Shin, Kang Min Yoo, Sang-goo Lee:
Slot Filling with Delexicalized Sentence Generation. 2082-2086 - Deepanway Ghosal, Maheshkumar H. Kolekar:
Music Genre Recognition Using Deep Neural Networks and Transfer Learning. 2087-2091 - Siddharth Sigtia, Rob Haynes, Hywel Richards, Erik Marchi, John Bridle:
Efficient Voice Trigger Detection for Low Resource Hardware. 2092-2096
Signal Analysis for the Natural, Biological and Social Sciences
- Qiguang Lin, Yiwen Shao:
A Novel Normalization Method for Autocorrelation Function for Pitch Detection and for Speech Activity Detection. 2097-2101 - T. V. Ananthapadmanabha, A. G. Ramakrishnan:
Estimation of the Vocal Tract Length of Vowel Sounds Based on the Frequency of the Significant Spectral Valley. 2102-2106 - Ivan Himawan, Michael Towsey, Bradley Law, Paul Roe:
Deep Learning Techniques for Koala Activity Detection. 2107-2111 - Jindrich Matousek, Daniel Tihelka:
Glottal Closure Instant Detection from Speech Signal Using Voting Classifier and Recursive Feature Elimination. 2112-2116 - Midia Yousefi, Navid Shokouhi, John H. L. Hansen:
Assessing Speaker Engagement in 2-Person Debates: Overlap Detection in United States Presidential Debates. 2117-2121 - Arjun Pankajakshan, Anshul Thakur, Daksh Thapar, Padmanabhan Rajan, Aditya Nigam:
All-Conv Net for Bird Activity Detection: Significance of Learned Pooling. 2122-2126 - Anshul Thakur, Vinayak Abrol, Pulkit Sharma, Padmanabhan Rajan:
Deep Convex Representations: Feature Representations for Bioacoustics Classification. 2127-2131 - Hirak Dasgupta, Prem C. Pandey, K. S. Nataraj:
Detection of Glottal Excitation Epochs in Speech Signal Using Hilbert Envelope. 2132-2136 - Hong Zhang:
Analyzing Thai Tone Distribution through Functional Data Analysis. 2137-2141 - Danny Merkx, Odette Scharenborg:
Articulatory Feature Classification Using Convolutional Neural Networks. 2142-2146 - Shoufeng Lin:
A New Frequency Coverage Metric and a New Subband Encoding Model, with an Application in Pitch Estimation. 2147-2151 - B. Ganga Gowri, Soman K. P, D. Govind:
Improved Epoch Extraction from Telephonic Speech Using Chebfun and Zero Frequency Filtering. 2152-2156
Speech Prosody
- Arne Köhn, Timo Baumann, Oskar Dörfler:
An Empirical Analysis of the Correlation of Syntax and Prosody. 2157-2161 - Timo Baumann, Hussein Hussein, Burkhard Meyer-Sickendiek:
Analysing the Focus of a Hierarchical Attention Network: the Importance of Enjambments When Classifying Post-modern Poetry. 2162-2166 - Daniil Kocharov, Alla Menshikova:
Language-Dependent Melody Embeddings. 2167-2170 - Yuan Jia, Xiaoxiao Ma:
Stress Distribution of Given Information in Chinese Reading Texts. 2171-2175 - Vera Cabarrão, Fernando Batista, Helena Moniz, Isabel Trancoso, Ana Isabel Mata:
Acoustic-prosodic Entrainment in Structural Metadata Events. 2176-2180 - Marija Tabain, Richard Beare, Andrew Butcher:
Formant Measures of Vowels Adjacent to Alveolar and Retroflex Consonants in Arrernte: Stressed and Unstressed Position. 2181-2185 - Quy-Thao Truong, Tsuneo Kato, Seiichi Yamamoto:
Automatic Assessment of L2 English Word Prosody Using Weighted Distances of F0 and Intensity Contours. 2186-2190 - Olga Maxwell, Elinor Payne, Rosey Billington:
Homogeneity vs Heterogeneity in Indian English: Investigating Influences of L1 on f0 Range. 2191-2195 - Yixin Zhang, Tianzhu Geng, Jinsong Zhang:
Emotional Prosody Perception in Mandarin-speaking Congenital Amusics. 2196-2200 - Takaaki Shochi, Jean-Luc Rouas, Marine Guerry, Donna Erickson:
Cultural Differences in Pattern Matching: Multisensory Recognition of Socio-affective Prosody. 2201-2205
Perspective Talk-3
- Nima Mesgarani:
Speech Processing in the Human Brain Meets Deep Learning. 2206
Recurrent Neural Models for ASR
- Shinji Watanabe, Takaaki Hori, Shigeki Karita, Tomoki Hayashi, Jiro Nishitoba, Yuya Unno, Nelson Enrique Yalta Soplin, Jahn Heymann, Matthew Wiesner, Nanxin Chen, Adithya Renduchintala, Tsubasa Ochiai:
ESPnet: End-to-End Speech Processing Toolkit. 2207-2211 - Zhehuai Chen, Justin Luitjens, Hainan Xu, Yiming Wang, Daniel Povey, Sanjeev Khudanpur:
A GPU-based WFST Decoder with Exact Lattice Generation. 2212-2216 - Anton Ragni, Mark J. F. Gales:
Automatic Speech Recognition System Development in the "Wild". 2217-2221 - Leonid Velikovich, Ian Williams, Justin Scheiner, Petar S. Aleksic, Pedro J. Moreno, Michael Riley:
Semantic Lattice Processing in Contextual Automatic Speech Recognition for Google Assistant. 2222-2226 - Ian Williams, Anjuli Kannan, Petar S. Aleksic, David Rybach, Tara N. Sainath:
Contextual Speech Recognition in End-to-end Neural Network Systems Using Beam Search. 2227-2231 - Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Forward-Backward Attention Decoder. 2232-2236
Speaker Verification Using Neural Network Methods I
- Sarthak Yadav, Atul Rai:
Learning Discriminative Features for Speaker Identification and Verification. 2237-2241 - Sergey Novoselov, Vadim Shchemelinin, Andrey Shulipa, Alexander Kozlov, Ivan Kremnev:
Triplet Loss Based Cosine Similarity Metric Learning for Text-independent Speaker Recognition. 2242-2246 - Yi Liu, Liang He, Jia Liu, Michael T. Johnson:
Speaker Embedding Extraction with Phonetic Information. 2247-2251 - Koji Okabe, Takafumi Koshinaka, Koichi Shinoda:
Attentive Statistics Pooling for Deep Speaker Embedding. 2252-2256 - Nam Le, Jean-Marc Odobez:
Robust and Discriminative Speaker Embedding via Intra-Class Distance Variance Regularization. 2257-2261 - Na Li, Deyi Tuo, Dan Su, Zhifeng Li, Dong Yu:
Deep Discriminative Embeddings for Duration Robust Speaker Verification. 2262-2266
Speech Perception in Adverse Conditions
- Olympia Simantiraki, Martin Cooke, Simon King:
Impact of Different Speech Types on Listening Effort. 2267-2271 - Moïra-Phoebé Huet, Christophe Micheyl, Etienne Gaudrain, Etienne Parizet:
Who Are You Listening to? Towards a Dynamic Measure of Auditory Attention to Speech-on-speech. 2272-2275 - Jeesun Kim, Sonya Karisma, Vincent Aubanel, Chris Davis:
Investigating the Role of Familiar Face and Voice Cues in Speech Processing in Noise. 2276-2279 - Odette Scharenborg, Martha A. Larson:
The Conversation Continues: the Effect of Lyrics and Music Complexity of Background Music on Spoken-Word Recognition. 2280-2284 - Julien Meyer, Fanny Meunier, Laure Dentel, Noelia Do Carmo Blanco, Frédéric Sèbe:
Loud and Shouted Speech Perception at Variable Distances in a Forest. 2285-2289 - Noelia Do Carmo Blanco, Julien Meyer, Michel Hoen, Fanny Meunier:
Phoneme Resistance and Phoneme Confusion in Noise: Impact of Dyslexia. 2290-2294
Measuring Pitch and Articulation
- Albert Haque, Michelle Guo, Prateek Verma:
Conditional End-to-End Audio Transforms. 2295-2299 - Gunnam Aneeja, Sudarsana Reddy Kadiri, Bayya Yegnanarayana:
Detection of Glottal Closure Instants in Degraded Speech Using Single Frequency Filtering Analysis. 2300-2304 - Loren Lugosch, Vikrant Singh Tomar:
Tone Recognition Using Lifters and CTC. 2305-2309 - Vikram C. M., S. R. Mahadeva Prasanna:
Epoch Extraction from Pathological Children Speech Using Single Pole Filtering Approach. 2310-2314 - Balamurali B. T., Jer-Ming Chen:
Automated Classification of Vowel-Gesture Parameters Using External Broadband Excitation. 2315-2318 - Sudarsana Reddy Kadiri, Bayya Yegnanarayana:
Estimation of Fundamental Frequency from Singing Voice Using Harmonics of Impulse-like Excitation Source. 2319-2323
Speech and Language Analytics for Mental Health
- Jochen Weiner, Miguel Angrick, Srinivasan Umesh, Tanja Schultz:
Investigating the Effect of Audio Duration on Dementia Detection Using Acoustic Features. 2324-2328 - Yun-Shao Lin, Susan Shur-Fen Gau, Chi-Chun Lee:
An Interlocutor-Modulated Attentional LSTM for Differentiating between Subgroups of Autism Spectrum Disorder. 2329-2333 - Shahin Amiriparian, Alice Baird, Sahib Julka, Alyssa Alcorn, Sandra Ottl, Suncica Petrovic, Eloise Ainger, Nicholas Cummins, Björn W. Schuller:
Recognition of Echolalic Autistic Child Vocalisations Utilising Convolutional Recurrent Neural Networks. 2334-2338 - Sandeep Nallan Chakravarthula, Brian R. Baucom, Panayiotis G. Georgiou:
Modeling Interpersonal Influence of Verbal Behavior in Couples Therapy Dyadic Interactions. 2339-2343 - Anil Ramakrishna, Timothy Greer, David C. Atkins, Shrikanth S. Narayanan:
Computational Modeling of Conversational Humor in Psychotherapy. 2344-2348 - Nicanor García, Juan Camilo Vásquez-Correa, Juan Rafael Orozco-Arroyave, Elmar Nöth:
Multimodal I-vectors to Detect and Evaluate Parkinson's Disease. 2349-2353
Spoken CALL Shared Task, Second Edition
- Claudia Baur, Andrew Caines, Cathy Chua, Johanna Gerlach, Mengjie Qian, Manny Rayner, Martin J. Russell, Helmer Strik, Xizi Wei:
Overview of the 2018 Spoken CALL Shared Task. 2354-2358 - Dominik Jülg, Mario Kunstek, Cem Philipp Freimoser, Kay Berkling, Mengjie Qian:
The CSU-K Rule-Based System for the 2nd Edition Spoken CALL Shared Task. 2359-2363 - Huy Nguyen, Lei Chen, Ramon Prieto, Chuan Wang, Yang Liu:
Liulishuo's System for the Spoken CALL Shared Task 2018. 2364-2368 - Mohammad A. Ateeq, Abualsoud Hanani, Aziz Qaroush:
An Optimization Based Approach for Solving Spoken CALL Shared Task. 2369-2373 - Mengjie Qian, Xizi Wei, Peter Jancovic, Martin J. Russell:
The University of Birmingham 2018 Spoken CALL Shared Task Systems. 2374-2378 - Keelan Evanini, Matthew Mulholland, Rutuja Ubale, Yao Qian, Robert A. Pugh, Vikram Ramanarayanan, Aoife Cahill:
Improvements to an Automated Content Scoring System for Spoken CALL Responses: the ETS Submission to the Second Spoken CALL Shared Task. 2379-2383
Show and Tell 6
- Nagendra Kumar Goel, Mousmita Sarma, Tejendra Kushwah, Dharmesh Agarwal, Zikra Iqbal, Surbhi Chauhan:
Extracting Speaker's Gender, Accent, Age and Emotional State from Speech. 2384-2385 - B. H. V. S. Narayanamurthy, J. V. Satyanarayana, Bayya Yegnanarayana:
Determining Speaker Location from Speech in a Practical Environment. 2386-2387 - Tanvina Patel, Krishna D. N, Noor Fathima, Nisar Shah, Mahima C, Deepak Kumar, Anuroop Iyengar:
An Automatic Speech Transcription System for Manipuri Language. 2388-2389 - Chiranjeevi Yarra, Anand P. A, N. K. Kausthubha, Prasanta Kumar Ghosh:
SPIRE-SST: An Automatic Web-based Self-learning Tool for Syllable Stress Tutoring (SST) to the Second Language Learners. 2390-2391 - Kishalay Chakraborty, Senjam Shantirani Devi, Sanjeevan Devnath, S. R. Mahadeva Prasanna, Priyankoo Sarmah:
Glotto Vibrato Graph: A Device and Method for Recording, Analysis and Visualization of Glottal Activity. 2392-2393
Adjusting to Speaker, Accent, and Domain
- Adithya Renduchintala, Shuoyang Ding, Matthew Wiesner, Shinji Watanabe:
Multi-Modal Data Augmentation for End-to-end ASR. 2394-2398 - Takafumi Moriya, Sei Ueno, Yusuke Shinohara, Marc Delcroix, Yoshikazu Yamaguchi, Yushi Aono:
Multi-task Learning with Augmentation Strategy for Acoustic-to-word Attention-based Encoder-decoder Speech Recognition. 2399-2403 - Sining Sun, Ching-Feng Yeh, Mari Ostendorf, Mei-Yuh Hwang, Lei Xie:
Training Augmentation with Adversarial Examples for Robust Speech Recognition. 2404-2408 - Takashi Fukuda, Raul Fernandez, Andrew Rosenberg, Samuel Thomas, Bhuvana Ramabhadran, Alexander Sorin, Gakuto Kurata:
Data Augmentation Improves Recognition of Foreign Accented Speech. 2409-2413 - Natalia A. Tomashenko, Yuri Y. Khokhlov, Yannick Estève:
Speaker Adaptive Training and Mixup Regularization for Neural Network Acoustic Models in Automatic Speech Recognition. 2414-2418 - Markus Müller, Sebastian Stüker, Alex Waibel:
Neural Language Codes for Multilingual Acoustic Models. 2419-2423 - Sei Ueno, Takafumi Moriya, Masato Mimura, Shinsuke Sakai, Yusuke Shinohara, Yoshikazu Yamaguchi, Yushi Aono, Tatsuya Kawahara:
Encoder Transfer for Attention-based Acoustic-to-word Speech Recognition. 2424-2428 - Ke Wang, Junbo Zhang, Yujun Wang, Lei Xie:
Empirical Evaluation of Speaker Adaptation on DNN Based Acoustic Model. 2429-2433 - Amit Das, Mark Hasegawa-Johnson:
Improving DNNs Trained with Non-Native Transcriptions Using Knowledge Distillation and Target Interpolation. 2434-2438 - Siyuan Feng, Tan Lee:
Improving Cross-Lingual Knowledge Transferability Using Multilingual TDNN-BLSTM with Language-Dependent Pre-Final Layer. 2439-2443 - Marc Delcroix, Shinji Watanabe, Atsunori Ogawa, Shigeki Karita, Tomohiro Nakatani:
Auxiliary Feature Based Adaptation of End-to-end ASR Systems. 2444-2448 - Shahram Ghorbani, John H. L. Hansen:
Leveraging Native Language Information for Improved Accented Speech Recognition. 2449-2453 - Abhinav Jain, Minali Upreti, Preethi Jyothi:
Improved Accented Speech Recognition Using Accent Embeddings and Multi-task Learning. 2454-2458 - Sibo Tong, Philip N. Garner, Hervé Bourlard:
Fast Language Adaptation Using Phonological Information. 2459-2463
Speech Synthesis Paradigms and Methods
- Hiroki Murakami, Sunao Hara, Masanobu Abe, Masaaki Sato, Shogo Minagi:
Naturalness Improvement Algorithm for Reconstructed Glossectomy Patient's Speech Using Spectral Differential Modification in Voice Conversion. 2464-2468 - Satoshi Tamura, Kento Horio, Hajime Endo, Satoru Hayamizu, Tomoki Toda:
Audio-visual Voice Conversion Using Deep Canonical Correlation Analysis for Deep Bottleneck Features. 2469-2473 - Pallavi Baljekar, Sai Krishna Rallabandi, Alan W. Black:
An Investigation of Convolution Attention Based Models for Multilingual Speech Synthesis of Indian Languages. 2474-2478 - Danny Websdale, Sarah Taylor, Ben Milner:
The Effect of Real-Time Constraints on Automatic Speech Animation. 2479-2483 - David Greenwood, Iain A. Matthews, Stephen D. Laycock:
Joint Learning of Facial Expression and Head Pose from Speech. 2484-2488 - Kévin Vythelingum, Yannick Estève, Olivier Rosec:
Acoustic-dependent Phonemic Transcription for Text-to-speech Synthesis. 2489-2493 - Hieu-Thi Luong, Junichi Yamagishi:
Multimodal Speech Synthesis Architecture for Unsupervised Speaker Adaptation. 2494-2498 - Fumiaki Taguchi, Tokihiko Kaburagi:
Articulatory-to-speech Conversion Using Bi-directional Long Short-term Memory. 2499-2503 - Keisuke Tanihara, Shogo Yonekura, Yasuo Kuniyoshi:
Implementation of Respiration in Articulatory Synthesis Using a Pressure-Volume Lung Model. 2504-2508 - Xiao Zhou, Zhen-Hua Ling, Zhi-Ping Zhou, Li-Rong Dai:
Learning and Modeling Unit Embeddings for Improving HMM-based Unit Selection Speech Synthesis. 2509-2513 - Ruibo Fu, Jianhua Tao, Yibin Zheng, Zhengqi Wen:
Deep Metric Learning for the Target Cost in Unit-Selection Speech Synthesizer. 2514-2518 - Kentaro Sone, Toru Nakashika:
DNN-based Speech Synthesis for Small Data Sets Considering Bidirectional Speech-Text Conversion. 2519-2523 - Branislav Gerazov, Gérard Bailly, Yi Xu:
A Weighted Superposition of Functional Contours Model for Modelling Contextual Prominence of Elementary Prosodic Contours. 2524-2528 - Toru Nakashika:
LSTBM: A Novel Sequence Representation of Speech Spectra Using Restricted Boltzmann Machine with Long Short-Term Memory. 2529-2533
Second Language Acquisition and Code-switching
- Barbara E. Bullock, Gualberto A. Guzmán, Jacqueline Serigos, Almeida Jacqueline Toribio:
Should Code-switching Models Be Asymmetric? 2534-2538 - Kimiko Tsukada, Yu Rong:
Cross-language Perception of Mandarin Lexical Tones by Mongolian-speaking Bilinguals in the Inner Mongolia Autonomous Region, China. 2539-2543 - Lionel Fontan, Maxime Le Coz, Sylvain Detey:
Automatically Measuring L2 Speech Fluency without the Need of ASR: A Proof-of-concept Study with Japanese Learners of French. 2544-2548 - Yue Sun, Win Thuzar Kyaw, Jinsong Zhang, Yoshinori Sagisaka:
Analysis of L2 Learners' Progress of Distinguishing Mandarin Tone 2 and Tone 3. 2549-2553 - Xu Li, Shaoguang Mao, Xixin Wu, Kun Li, Xunying Liu, Helen Meng:
Unsupervised Discovery of Non-native Phonetic Patterns in L2 English Speech for Mispronunciation Detection and Diagnosis. 2554-2558 - Lei Wang, Jie Cui, Ying Chen:
Wuxi Speakers' Production and Perception of Coda Nasals in Mandarin. 2559-2562 - Natalia Dyrenko, Robert Fuchs:
The Diphthongs of Formal Nigerian English: A Preliminary Acoustic Analysis. 2563-2567 - Chris Davis, Jeesun Kim:
Characterizing Rhythm Differences between Strong and Weak Accented L2 Speech. 2568-2572 - Eva Fringi, Martin J. Russell:
Analysis of Phone Errors Attributable to Phonological Effects Associated With Language Acquisition Through Bottleneck Feature Visualisations. 2573-2577 - Jacques C. Koreman:
Category Similarity in Multilingual Pronunciation Training. 2578-2582 - Alejandrina Cristià, Shobhana Ganesh, Marisa Casillas, Sriram Ganapathy:
Talker Diarization in the Wild: the Case of Child-centered Daylong Audio-recordings. 2583-2587 - Zixing Zhang, Alejandrina Cristià, Anne S. Warlaumont, Björn W. Schuller:
Automated Classification of Children's Linguistic versus Non-Linguistic Vocalisations. 2588-2592 - Jiahong Yuan, Qiusi Dong, Fei Wu, Huan Luan, Xiaofei Yang, Hui Lin, Yang Liu:
Pitch Characteristics of L2 English Speech by Chinese Speakers: A Large-scale Study. 2593-2597
Topics in Speech Recognition
- Saurabh Garg, Tanmay Parekh, Preethi Jyothi:
Dual Language Models for Code Switched Speech Recognition. 2598-2602 - Astik Biswas, Febe de Wet, Ewald van der Westhuizen, Emre Yilmaz, Thomas Niesler:
Multilingual Neural Network Acoustic Modelling for ASR of Under-Resourced English-isiZulu Code-Switched Speech. 2603-2607 - Raghav Menon, Herman Kamper, John A. Quinn, Thomas Niesler:
Fast ASR-free and Almost Zero-resource Keyword Spotting Using DTW and CNNs for Humanitarian Monitoring. 2608-2612 - Meng Yu, Xuan Ji, Yi Gao, Lianwu Chen, Jie Chen, Jimeng Zheng, Dan Su, Dong Yu:
Text-Dependent Speech Enhancement for Small-Footprint Robust Keyword Detection. 2613-2617 - Di He, Boon Pang Lim, Xuesong Yang, Mark Hasegawa-Johnson, Deming Chen:
Improved ASR for Under-resourced Languages through Multi-task Learning with Acoustic Landmarks. 2618-2622 - Nick K. Chibuye, Todd Rosenstock, Brian DeRenzi:
Cross-language Phoneme Mapping for Low-resource Languages: An Exploration of Benefits and Trade-offs. 2623-2627 - Máté Ákos Tündik, György Szaszák, Gábor Gosztolya, András Beke:
User-centric Evaluation of Automatic Punctuation in ASR Closed Captioning. 2628-2632 - Piotr Zelasko, Piotr Szymanski, Jan Mizgajski, Adrian Szymczak, Yishay Carmiel, Najim Dehak:
Punctuation Prediction Model for Conversational Speech. 2633-2637 - Martin Karafiát, Murali Karthick Baskar, Igor Szöke, Vladimír Malenovský, Karel Veselý, Frantisek Grézl, Lukás Burget, Jan Cernocký:
BUT OpenSAT 2017 Speech Recognition System. 2638-2642 - Li Liu, Thomas Hueber, Gang Feng, Denis Beautemps:
Visual Recognition of Continuous Cued Speech Using a Tandem CNN-HMM Approach. 2643-2647 - Kwanchiva Thangthai, Richard W. Harvey:
Building Large-vocabulary Speaker-independent Lipreading Systems. 2648-2652 - Vishwa Gupta, Gilles Boulianne:
CRIM's System for the MGB-3 English Multi-Genre Broadcast Media Transcription. 2653-2657 - Rachid Riad, Corentin Dancette, Julien Karadayi, Neil Zeghidour, Thomas Schatz, Emmanuel Dupoux:
Sampling Strategies in Siamese Networks for Unsupervised Speech Representation Learning. 2658-2662 - Mengzhe Chen, Shiliang Zhang, Ming Lei, Yong Liu, Haitao Yao, Jie Gao:
Compact Feedforward Sequential Memory Networks for Small-footprint Keyword Spotting. 2663-2667
Zero-resource Speech Recognition
- Enno Hermann, Sharon Goldwater:
Multilingual Bottleneck Features for Subword Modeling in Zero-resource Languages. 2668-2672 - Siyuan Feng, Tan Lee:
Exploiting Speaker and Phonetic Diversity of Mismatched Language Resources for Unsupervised Subword Modeling. 2673-2677 - Pierre Godard, Marcely Zanon Boito, Lucas Ondel, Alexandre Berard, François Yvon, Aline Villavicencio, Laurent Besacier:
Unsupervised Word Segmentation from Speech with Attention. 2678-2682 - Nils Holzenberger, Mingxing Du, Julien Karadayi, Rachid Riad, Emmanuel Dupoux:
Learning Word Embeddings: Unsupervised Methods for Fixed-size Representations of Variable-length Speech Segments. 2683-2687 - Thomas Glarner, Patrick Hanebrink, Janek Ebbers, Reinhold Haeb-Umbach:
Full Bayesian Hidden Markov Model Variational Autoencoder for Acoustic Unit Discovery. 2688-2692 - Benjamin Milde, Chris Biemann:
Unspeech: Unsupervised Speech Context Embeddings. 2693-2697
Spatial and Phase Cues for Source Separation and Speech Recognition
- Yuan Gong, Christian Poellabauer:
Impact of Aliasing on Deep CNN-Based End-to-End Acoustic Models. 2698-2702 - Sunit Sivasankaran, Emmanuel Vincent, Dominique Fohr:
Keyword Based Speaker Localization: Localizing a Target Speaker in a Multi-speaker Environment. 2703-2707 - Zhong-Qiu Wang, Jonathan Le Roux, DeLiang Wang, John R. Hershey:
End-to-End Speech Separation with Unfolded Iterative Phase Reconstruction. 2708-2712 - Naoya Takahashi, Purvi Agrawal, Nabarun Goswami, Yuki Mitsufuji:
PhaseNet: Discretized Phase Modeling with Deep Neural Networks for Audio Source Separation. 2713-2717 - Zhong-Qiu Wang, DeLiang Wang:
Integrating Spectral and Spatial Features for Multi-Channel Speaker Separation. 2718-2722 - Mandar Gogate, Ahsan Adeel, Ricard Marxer, Jon Barker, Amir Hussain:
DNN Driven Speaker Independent Audio-Visual Mask Estimation for Speech Separation. 2723-2727
Dialectal Variation
- Ioana Vasilescu, Nidia Hernández, Bianca Vieru, Lori Lamel:
Exploring Temporal Reduction in Dialectal Spanish: A Large-scale Study of Lenition of Voiced Stops and Coda-s. 2728-2732 - Phil Rose:
Dialect-geographical Acoustic-Tonetics: Five Disyllabic Tone Sandhi Patterns in Cognate Words from the Wu Dialects of ZhèJiāNg Province. 2733-2737 - Adrian Leemann, Stephan Schmid, Dieter Studer-Joho, Marie-José Kolly:
Regional Variation of /r/ in Swiss German Dialects. 2738-2742 - Kate Earnshaw, Erica Gold:
Variation in the FACE Vowel across West Yorkshire: Implications for Forensic Speaker Comparisons. 2743-2747 - Erica Gold, Sula Ross, Kate Earnshaw:
The 'West Yorkshire Regional English Database': Investigations into the Generalizability of Reference Populations for Forensic Speaker Comparison Casework. 2748-2752 - Jane Wottawa, Djegdjiga Amazouz, Martine Adda-Decker, Lori Lamel:
Studying Vowel Variation in French-Algerian Arabic Code-switched Speech. 2753-2757
Spoken Corpora and Annotation
- John H. L. Hansen, Abhijeet Sangwan, Aditya Joglekar, Ahmet Emin Bulut, Lakshmish Kaushik, Chengzhu Yu:
Fearless Steps: Apollo-11 Corpus Advancements for Speech Technologies from Earth to the Moon. 2758-2762 - Manoj Kumar, Pooja Chebolu, So Hyun Kim, Kassandra Martinez, Catherine Lord, Shrikanth S. Narayanan:
A Knowledge Driven Structural Segmentation Approach for Play-Talk Classification During Autism Assessment. 2763-2767 - Jesin James, Li Tian, Catherine Inez Watson:
An Open Source Emotional Speech Corpus for Human Robot Interaction Applications. 2768-2772 - Itshak Lapidot, Héctor Delgado, Massimiliano Todisco, Nicholas W. D. Evans, Jean-François Bonastre:
Speech Database and Protocol Validation Using Waveform Entropy. 2773-2777 - Lucas D. Terissi, Gonzalo D. Sad, Mauricio Cerda, Slim Ouni, Rodrigo Galvez, Juan Carlos Gómez, Bernard Girau, Nancy Hitschfeld-Kahler:
A French-Spanish Multimodal Speech Communication Corpus Incorporating Acoustic Data, Facial, Hands and Arms Gestures Information. 2778-2782 - Guanlong Zhao, Sinem Sonsaat, Alif Silpachai, Ivana Lucic, Evgeny Chukharev-Hudilainen, John Levis, Ricardo Gutierrez-Osuna:
L2-ARCTIC: A Non-native English Speech Corpus. 2783-2787
The First DIHARD Speech Diarization Challenge
- Zbynek Zajíc, Marie Kunesová, Jan Zelinka, Marek Hrúz:
ZCU-NTIS Speaker Diarization System for the DIHARD 2018 Challenge. 2788-2792 - Lei Sun, Jun Du, Chao Jiang, Xueyang Zhang, Shan He, Bing Yin, Chin-Hui Lee:
Speaker Diarization with Enhancing Speech for the First DIHARD Challenge. 2793-2797 - Mireia Díez, Federico Landini, Lukás Burget, Johan Rohdin, Anna Silnova, Katerina Zmolíková, Ondrej Novotný, Karel Veselý, Ondrej Glembek, Oldrich Plchot, Ladislav Mosner, Pavel Matejka:
BUT System for DIHARD Speech Diarization Challenge 2018. 2798-2802 - Ignacio Viñals, Pablo Gimeno, Alfonso Ortega, Antonio Miguel, Eduardo Lleida:
Estimation of the Number of Speakers with Variational Bayesian PLDA in the DIHARD Diarization Challenge. 2803-2807 - Gregory Sell, David Snyder, Alan McCree, Daniel Garcia-Romero, Jesús Villalba, Matthew Maciejewski, Vimal Manohar, Najim Dehak, Daniel Povey, Shinji Watanabe, Sanjeev Khudanpur:
Diarization is Hard: Some Experiences and Lessons Learned for the JHU Team in the Inaugural DIHARD Challenge. 2808-2812 - Jose Patino, Héctor Delgado, Nicholas W. D. Evans:
The EURECOM Submission to the First DIHARD Challenge. 2813-2817 - Valter Akira Miasato Filho, Diego Augusto Silva, Luis Gustavo Depra Cuozzo:
Joint Discriminative Embedding Learning, Speech Activity and Overlap Detection for the DIHARD Speaker Diarization Challenge. 2818-2822
Text Analysis, Multilingual Issues and Evaluation in Speech Synthesis
- Jinfu Ni, Yoshinori Shiga, Hisashi Kawai:
Multilingual Grapheme-to-Phoneme Conversion with Global Character Vectors. 2823-2827 - Somnath Roy, Shakuntala Mahanta:
A Hybrid Approach to Grapheme to Phoneme Conversion in Assamese. 2828-2832 - Seyed Hamidreza Mohammadi, Taehwan Kim:
Investigation of Using Disentangled and Interpretable Representations for One-shot Cross-lingual Voice Conversion. 2833-2837 - Avashna Govender, Simon King:
Using Pupillometry to Measure the Cognitive Load of Synthetic Speech. 2838-2842 - Avashna Govender, Simon King:
Measuring the Cognitive Load of Synthetic Speech Using a Dual Task Paradigm. 2843-2847 - Iroro Orife:
Attentive Sequence-to-Sequence Learning for Diacritic Restoration of YorùBá Language Text. 2848-2852 - Peixin Chen, Wu Guo, Zhi Chen, Jian Sun, Lanhua You:
Gated Convolutional Neural Network for Sentence Matching. 2853-2857 - Dravyansh Sharma:
On Training and Evaluation of Grapheme-to-Phoneme Mappings with Limited Data. 2858-2862 - Alice Baird, Emilia Parada-Cabaleiro, Simone Hantke, Felix Burkhardt, Nicholas Cummins, Björn W. Schuller:
The Perception and Analysis of the Likeability and Human Likeness of Synthesized Speech. 2863-2867 - Yosi Mass, Slava Shechtman, Moran Mordechay, Ron Hoory, Oren Sar Shalom, Guy Lev, David Konopnicki:
Word Emphasis Prediction for Expressive Text to Speech. 2868-2872 - Kai-Zhan Lee, Erica Cooper, Julia Hirschberg:
A Comparison of Speaker-based and Utterance-based Data Selection for Text-to-Speech Synthesis. 2873-2877 - Markus Toman, Geoffrey S. Meltzner, Rupal Patel:
Data Requirements, Selection and Augmentation for DNN-based Speech Synthesis from Crowdsourced Data. 2878-2882
Neural Network Training Strategies for ASR
- Karel Veselý, Carlos Segura, Igor Szöke, Jordi Luque, Jan Cernocký:
Lightly Supervised vs. Semi-supervised Training of Acoustic Model on Luxembourgish for Low-resource Automatic Speech Recognition. 2883-2887 - Wenjie Li, Gaofeng Cheng, Fengpei Ge, Pengyuan Zhang, Yonghong Yan:
Investigation on the Combination of Batch Normalization and Dropout in BLSTM-based Acoustic Modeling for ASR. 2888-2892 - Masayuki Suzuki, Tohru Nagano, Gakuto Kurata, Samuel Thomas:
Inference-Invariant Transformation of Batch Normalization for Domain Adaptation of Acoustic Models. 2893-2897 - Yanhua Long, Hong Ye, Yijie Li, Jiaen Liang:
Active Learning for LF-MMI Trained Neural Networks in ASR. 2898-2902 - Ivan Medennikov, Yuri Y. Khokhlov, Aleksei Romanenko, Dmitry Popov, Natalia A. Tomashenko, Ivan Sorokin, Alexander Zatvornitskiy:
An Investigation of Mixup Training Strategies for Acoustic Models in ASR. 2903-2907 - Purvi Agrawal, Sriram Ganapathy:
Comparison of Unsupervised Modulation Filter Learning Methods for ASR. 2908-2912 - Suyoun Kim, Michael L. Seltzer, Jinyu Li, Rui Zhao:
Improved Training for Online End-to-end Speech Recognition Systems. 2913-2917 - Adnan Haider, Philip C. Woodland:
Combining Natural Gradient with Hessian Free Methods for Sequence Training. 2918-2922 - Naoyuki Kanda, Yusuke Fujita, Kenji Nagamatsu:
Lattice-free State-level Minimum Bayes Risk Training of Acoustic Models. 2923-2927 - Hao Tang, Wei-Ning Hsu, François Grondin, James R. Glass:
A Study of Enhancement, Augmentation and Autoencoder Methods for Domain Adaptation in Distant Speech Recognition. 2928-2932 - Andreas Søeborg Kirkedal, Yeon-Jun Kim:
Multilingual Deep Neural Network Training Using Cyclical Learning Rate. 2933-2937
Application of ASR in Medical Practice
- Jianwei Yu, Xurong Xie, Shansong Liu, Shoukang Hu, Max W. Y. Lam, Xixin Wu, Ka Ho Wong, Xunying Liu, Helen Meng:
Development of the CUHK Dysarthric Speech Recognition System for the UA Speech Corpus. 2938-2942 - Imed Laaridh, Corinne Fredouille, Alain Ghio, Muriel Lalain, Virginie Woisard:
Automatic Evaluation of Speech Intelligibility Based on I-vectors in the Context of Head and Neck Cancers. 2943-2947 - Myung Jong Kim, Beiming Cao, Kwanghoon An, Jun Wang:
Dysarthric Speech Recognition Using Convolutional LSTM Neural Network. 2948-2952 - Imed Laaridh, Julien Tardieu, Cynthia Magnen, Pascal Gaillard, Jérôme Farinas, Julien Pinquier:
Perceptual and Automatic Evaluations of the Intelligibility of Speech Degraded by Noise Induced Hearing Loss Simulation. 2953-2957 - Emre Yilmaz, Vikramjit Mitra, Chris Bartels, Horacio Franco:
Articulatory Features for ASR of Pathological Speech. 2958-2962 - M. Joana Correia, Bhiksha Raj, Isabel Trancoso, Francisco Teixeira:
Mining Multimodal Repositories for Speech Affecting Diseases. 2963-2967 - Zhen Qin, Tom Ko, Guangjian Tian:
Long Distance Voice Channel Diagnosis Using Deep Neural Networks. 2968-2971 - Chung-Cheng Chiu, Anshuman Tripathi, Katherine Chou, Chris Co, Navdeep Jaitly, Diana Jaunzeikare, Anjuli Kannan, Patrick Nguyen, Hasim Sak, Ananth Sankar, Justin Tansuwan, Nathan Wan, Yonghui Wu, Xuedong Zhang:
Speech Recognition for Medical Conversations. 2972-2976
Source and Supra-segmentals
- Chadi Farah, Stephane Roman, Mariapaola D'Imperio:
Prosodic Focus Acquisition in French Early Cochlear Implanted Children. 2977-2981 - Nassima Fezza:
The Role of Temporal Variation in Narrative Organization. 2982-2986 - Tiina Murtola, Jarmo Malinen:
Interaction Mechanisms between Glottal Source and Vocal Tract in Pitch Glides. 2987-2991 - Astha Singh, G. Nisha Meenakshi, Prasanta Kumar Ghosh:
Relating Articulatory Motions in Different Speaking Rates. 2992-2996 - João Cabral:
Estimation of the Asymmetry Parameter of the Glottal Flow Waveform Using the Electroglottographic Signal. 2997-3001 - Tanumay Mandal, K. Sreenivasa Rao, Sanjay Kumar Gupta:
Classification of Disorders in Vocal Folds Using Electroglottographic Signal. 3002-3006 - M. V. Achuth Rao, Rahul Krishnamurthy, Pebbili Gopikishore, Veeramani Priyadharshini, Prasanta Kumar Ghosh:
Automatic Glottis Localization and Segmentation in Stroboscopic Videos Using Deep Neural Network. 3007-3011 - Toshiko Isei-Jaakkola, Keiko Ochi, Keikichi Hirose:
Respiratory and Respiratory Muscular Control in JL1's and JL2's Text Reading Utilizing 4-RSTs and a Soft Respiratory Mask with a Two-Way Bulb. 3012-3016 - Lixia Hao, Wei Zhang, Yanlu Xie, Jinsong Zhang:
A Preliminary Study on Tonal Coarticulation in Continuous Speech. 3017-3021
Plenary Talk-3
- Helen Meng:
Speech and Language Processing for Learning and Wellbeing. 3022
Distant ASR
- Sriram Ganapathy, Madhumita Harish:
Far-Field Speech Recognition Using Multivariate Autoregressive Models. 3023-3027 - Chanwoo Kim, Ehsan Variani, Arun Narayanan, Michiel Bacchiani:
Efficient Implementation of the Room Simulator for Training Deep Neural Network Acoustic Models. 3028-3032 - Xiaofei Wang, Ruizhi Li, Hynek Hermansky:
Stream Attention for Distributed Multi-Microphone Speech Recognition. 3033-3037 - Takuya Yoshioka, Hakan Erdogan, Zhuo Chen, Xiong Xiao, Fil Alleva:
Recognizing Overlapped Speech in Meetings: A Multichannel Separation Approach Using Neural Networks. 3038-3042 - Lukas Drude, Christoph Böddeker, Jahn Heymann, Reinhold Haeb-Umbach, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani:
Integrating Neural Network Based Beamforming and Weighted Prediction Error Dereverberation. 3043-3047 - Suliang Bu, Yunxin Zhao, Mei-Yuh Hwang, Sining Sun:
A Probability Weighted Beamformer for Noise Robust ASR. 3048-3052
Expressive Speech Synthesis
- Masaki Yokoyama, Tomohiro Nagata, Hiroki Mori:
Effects of Dimensional Input on Paralinguistic Information Perceived from Synthesized Dialogue Speech with Neural Network. 3053-3056 - Shereen Oraby, Lena Reed, Sharath T. S., Shubhangi Tandon, Marilyn A. Walker:
Neural MultiVoice Models for Expressing Novel Personalities in Dialog. 3057-3061 - Igor Jauk, Jaime Lorenzo-Trueba, Junichi Yamagishi, Antonio Bonafonte:
Expressive Speech Synthesis Using Sentiment Embeddings. 3062-3066 - Kei Akuzawa, Yusuke Iwasawa, Yutaka Matsuo:
Expressive Speech Synthesis via Modeling Expressions with Variational Autoencoder. 3067-3071 - Xixin Wu, Yuewen Cao, Mu Wang, Songxiang Liu, Shiyin Kang, Zhiyong Wu, Xunying Liu, Dan Su, Dong Yu, Helen Meng:
Rapid Style Adaptation Using Residual Error Embedding for Expressive Speech Synthesis. 3072-3076 - Hao Li, Yongguo Kang, Zhenyu Wang:
EMPHASIS: An Emotional Phoneme-based Acoustic Model for Speech Synthesis System. 3077-3081
Representation Learning for Emotion
- Jing Han, Zixing Zhang, Maximilian Schmitt, Zhao Ren, Fabien Ringeval, Björn W. Schuller:
Bags in Bag: Generating Context-Aware Bags for Tracking Emotions from Speech. 3082-3086 - Pengcheng Li, Yan Song, Ian McLoughlin, Wu Guo, Lirong Dai:
An Attention Pooling Based Representation Learning Method for Speech Emotion Recognition. 3087-3091 - Zixiaofan Yang, Julia Hirschberg:
Predicting Arousal and Valence from Waveforms and Spectrograms Using Deep Neural Networks. 3092-3096 - Mousmita Sarma, Pegah Ghahremani, Daniel Povey, Nagendra Kumar Goel, Kandarpa Kumar Sarma, Najim Dehak:
Emotion Identification from Raw Speech Signals Using DNNs. 3097-3101 - Jeng-Lin Li, Chi-Chun Lee:
Encoding Individual Acoustic Features Using Dyad-Augmented Deep Variational Representations for Dialog-level Emotion Recognition. 3102-3106 - Siddique Latif, Rajib Rana, Junaid Qadir, Julien Epps:
Variational Autoencoders for Learning Latent Representations of Speech Emotion: A Preliminary Study. 3107-3111
Articulatory Information, Modeling and Inversion
- Théo Biasutto-Lervat, Slim Ouni:
Phoneme-to-Articulatory Mapping Using Bidirectional Gated RNN. 3112-3116 - Zhihua Su, Jianguo Wei, Qiang Fang, Jianrong Wang, Kiyoshi Honda:
Tongue Segmentation with Geometrically Constrained Snake Model. 3117-3121 - Aravind Illa, Prasanta Kumar Ghosh:
Low Resource Acoustic-to-articulatory Inversion Using Bi-directional Long Short Term Memory. 3122-3126 - Chandana Srinivasan, Chiranjeevi Yarra, Ritu Aggarwal, Sanjeev Kumar Mittal, N. K. Kausthubha, Raseena K. T, Astha Singh, Prasanta Kumar Ghosh:
Automatic Visual Augmentation for Concatenation Based Synthesized Articulatory Videos from Real-time MRI Data for Spoken Language Training. 3127-3131 - C. A. Valliappan, Renuka Mannem, Prasanta Kumar Ghosh:
Air-Tissue Boundary Segmentation in Real-Time Magnetic Resonance Imaging Video Using Semantic Segmentation with Fully Convolutional Networks. 3132-3136 - Nadee Seneviratne, Ganesh Sivaraman, Vikramjit Mitra, Carol Y. Espy-Wilson:
Noise Robust Acoustic to Articulatory Speech Inversion. 3137-3141
Novel Paradigms for Direct Synthesis Based on Speech-Related Biosignals
- Farzaneh Ahmadi, Tomoki Toda:
Designing a Pneumatic Bionic Voice Prosthesis - A Statistical Approach for Source Excitation Generation. 3142-3146 - Bastian Schnell, Philip N. Garner:
A Neural Model to Predict Parameters for a Generalized Command Response Model of Intonation. 3147-3151 - Beiming Cao, Myung Jong Kim, Jun R. Wang, Jan P. H. van Santen, Ted Mau, Jun Wang:
Articulation-to-Speech Synthesis Using Articulatory Flesh Point Sensors' Orientation Information. 3152-3156 - Neil Shah, Nirmesh J. Shah, Hemant A. Patil:
Effectiveness of Generative Adversarial Network for Non-Audible Murmur-to-Whisper Speech Conversion. 3157-3161 - Lorenz Diener, Tanja Schultz:
Investigating Objective Intelligibility in Real-Time EMG-to-Speech Conversion. 3162-3166 - Michael Wand, Tanja Schultz, Jürgen Schmidhuber:
Domain-Adversarial Training for Session Independent EMG-based Speech Recognition. 3167-3171 - László Tóth, Gábor Gosztolya, Tamás Grósz, Alexandra Markó, Tamás Gábor Csapó:
Multi-Task Learning of Speech Recognition and Speech Synthesis Parameters for Ultrasound-based Silent Speech Interfaces. 3172-3176
Low Resource Speech Recognition Challenge for Indian Languages
- Jeena J. Prakash, Rajan Golda Brunet, Hema A. Murthy:
Transcription Correction for Indian Languages Using Acoustic Signatures. 3177-3181 - Bhargav Pulugundla, Murali Karthick Baskar, Santosh Kesiraju, Ekaterina Egorova, Martin Karafiát, Lukás Burget, Jan Cernocký:
BUT System for Low Resource Indian Language ASR. 3182-3186 - Hardik B. Sailor, Maddala Venkata Siva Krishna, Diksha Chhabra, Ankur T. Patil, Madhu R. Kamble, Hemant A. Patil:
DA-IICT/IIITV System for Low Resource Speech Recognition Challenge 2018. 3187-3191 - Hari Krishna Vydana, Krishna Gurugubelli, Vishnu Vidyadhara Raju Vegesna, Anil Kumar Vuppala:
An Exploration towards Joint Acoustic Modeling for Indian Languages: IIIT-H Submission for Low Resource Speech Recognition Challenge for Indian Languages, INTERSPEECH 2018. 3192-3196 - Noor Fathima, Tanvina Patel, Mahima C, Anuroop Iyengar:
TDNN-based Multilingual Speech Recognition System for Low Resource Indian Languages. 3197-3201 - Vishwas M. Shetty, Rini A. Sharon, Basil Abraham, Tejaswi Seeram, Anusha Prakash, Nithya Ravi, Srinivasan Umesh:
Articulatory and Stacked Bottleneck Features for Low Resource Speech Recognition. 3202-3206 - Jayadev Billa:
ISI ASR System for the Low Resource Speech Recognition Challenge for Indian Languages. 3207-3211
Show and Tell 7
- Gregory P. Finley, Erik Edwards, Amanda Robinson, Najmeh Sadoughi, James Fone, Mark Miller, David Suendermann-Oeft, Michael Brenndoerfer, Nico Axtmann:
An Automated Assistant for Medical Scribes. 3212-3213 - Abhishek Dey, Abhash Deka, Siddika Imani, Barsha Deka, Rohit Sinha, S. R. Mahadeva Prasanna, Priyankoo Sarmah, K. Samudravijaya, S. R. Nirmala:
AGROASSAM: A Web Based Assamese Speech Recognition Application for Retrieving Agricultural Commodity Price and Weather Information. 3214-3215 - Dan Aharon:
Voice-powered Solutions with Cloud AI. 3216 - Ganesh Sivaraman, Parav Nagarsheth, Elie Khoury:
Speech Synthesis in the Wild. 3217-3218
Deep Enhancement
- Shuai Nie, Shan Liang, Bin Liu, Yaping Zhang, Wenju Liu, Jianhua Tao:
Deep Noise Tracking Network: A Hybrid Signal Processing/Deep Learning Approach to Speech Enhancement. 3219-3223 - Zhiheng Ouyang, Hongjiang Yu, Wei-Ping Zhu, Benoît Champagne:
A Deep Neural Network Based Harmonic Noise Model for Speech Enhancement. 3224-3228 - Ke Tan, DeLiang Wang:
A Convolutional Recurrent Neural Network for Real-Time Speech Enhancement. 3229-3233 - Zhong-Qiu Wang, DeLiang Wang:
All-Neural Multi-Channel Speech Enhancement. 3234-3238 - Hao Zhang, DeLiang Wang:
Deep Learning for Acoustic Echo Cancellation in Noisy and Double-Talk Scenarios. 3239-3243 - Triantafyllos Afouras, Joon Son Chung, Andrew Zisserman:
The Conversation: Deep Audio-Visual Speech Enhancement. 3244-3248 - Aswin Shanmugam Subramanian, Szu-Jui Chen, Shinji Watanabe:
Student-Teacher Learning for BLSTM Mask-based Speech Enhancement. 3249-3253 - Pavan Karjol, Prasanta Kumar Ghosh:
Speech Enhancement Using Deep Mixture of Experts Based on Hard Expectation Maximization. 3254-3258 - Zhong Meng, Jinyu Li, Yifan Gong, Biing-Hwang Fred Juang:
Adversarial Feature-Mapping for Speech Enhancement. 3259-3263 - Deepak Baby, Sarah Verhulst:
Biophysically-inspired Features Improve the Generalizability of Neural Network-based Speech Enhancement Systems. 3264-3268 - Li Chai, Jun Du, Chin-Hui Lee:
Error Modeling via Asymmetric Laplace Distribution for Deep Neural Network Based Single-Channel Speech Enhancement. 3269-3273 - Yangyang Xia, Richard M. Stern:
A Priori SNR Estimation Based on a Recurrent Neural Network for Robust Speech Enhancement. 3274-3278
Acoustic Scenes and Rare Events
- Shao-Yen Tseng, Juncheng Li, Yun Wang, Florian Metze, Joseph Szurley, Samarjit Das:
Multiple Instance Deep Learning for Weakly Supervised Small-Footprint Audio Event Detection. 3279-3283 - Liwen Zhang, Jiqing Han, Shiwen Deng:
Unsupervised Temporal Feature Learning Based on Sparse Coding Embedded BoAW for Acoustic Event Recognition. 3284-3288 - Teng Zhang, Kailai Zhang, Ji Wu:
Data Independent Sequence Augmentation Method for Acoustic Scene Classification. 3289-3293 - Hongwei Song, Jiqing Han, Shiwen Deng:
A Compact and Discriminative Feature Based on Auditory Summary Statistics for Acoustic Scene Classification. 3294-3298 - Pulkit Sharma, Vinayak Abrol, Anshul Thakur:
ASe: Acoustic Scene Embedding Using Deep Archetypal Analysis and GMM. 3299-3303 - Hangting Chen, Pengyuan Zhang, Haichuan Bai, Qingsheng Yuan, Xiuguo Bao, Yonghong Yan:
Deep Convolutional Neural Network with Scalogram for Audio Scene Modeling. 3304-3308 - Pankaj Joshi, Digvijaysingh Gautam, Ganesh Ramakrishnan, Preethi Jyothi:
Time Aggregation Operators for Multi-label Audio Event Detection. 3309-3313 - Ian McLoughlin, Yan Song, Lam Dang Pham, Ramaswamy Palaniappan, Huy Phan, Yue Lang:
Early Detection of Continuous and Partial Audio Events Using CNN. 3314-3318 - Manjunath Mulimani, Shashidhar G. Koolagudi:
Robust Acoustic Event Classification Using Bag-of-Visual-Words. 3319-3322 - Shefali Waldekar, Goutam Saha:
Wavelet Transform Based Mel-scaled Features for Acoustic Scene Classification. 3323-3327 - Teng Zhang, Kailai Zhang, Ji Wu:
Multi-modal Attention Mechanisms in LSTM and Its Application to Acoustic Scene Classification. 3328-3332
Language Modeling
- Anirudh Raju, Behnam Hedayatnia, Linda Liu, Ankur Gandhe, Chandra Khatri, Angeliki Metallinou, Anu Venkatesh, Ariya Rastrow:
Contextual Language Model Adaptation for Conversational Agents. 3333-3337 - Oscar Chen, Anton Ragni, Mark J. F. Gales, Xie Chen:
Active Memory Networks for Language Modeling. 3338-3342 - Yerbolat Khassanov, Eng Siong Chng:
Unsupervised and Efficient Vocabulary Expansion for Recurrent Neural Network Language Models in ASR. 3343-3347 - Yike Zhang, Pengyuan Zhang, Yonghong Yan:
Improving Language Modeling with an Adversarial Critic for Automatic Speech Recognition. 3348-3352 - Yue Deng, Yilin Shen, KaWai Chen, Hongxia Jin:
Training Recurrent Neural Network through Moment Matching for NLP Applications. 3353-3357 - Zoltán Tüske, Ralf Schlüter, Hermann Ney:
Investigation on LSTM Recurrent N-gram Language Models for Speech Recognition. 3358-3362 - Chih Chi Hu, Bing Liu, John Shen, Ian R. Lane:
Online Incremental Learning for Speaker-Adaptive Language Models. 3363-3367 - Jesús Andrés-Ferrer, Nathan Bodenstab, Paul Vozila:
Efficient Language Model Adaptation with Noise Contrastive Estimation and Kullback-Leibler Regularization. 3368-3372 - Ke Li, Hainan Xu, Yiming Wang, Daniel Povey, Sanjeev Khudanpur:
Recurrent Neural Network Language Model Adaptation for Conversational Speech Recognition. 3373-3377 - Michael Levit, Sarangarajan Parthasarathy, Shuangyu Chang:
What to Expect from Expected Kneser-Ney Smoothing. 3378-3382 - Karel Benes, Santosh Kesiraju, Lukás Burget:
i-Vectors in Language Modeling: An Efficient Way of Domain Adaptation for Feed-Forward Models. 3383-3387
Speech Pathology, Depression, and Medical Applications
- Eva-Maria Rathner, Julia Djamali, Yannik Terhorst, Björn W. Schuller, Nicholas Cummins, Gudrun Salamon, Christina Hunger-Schoppe, Harald Baumeister:
How Did You like 2017? Detection of Language Markers of Depression and Narcissism in Personal Narratives. 3388-3392 - Zhaocheng Huang, Julien Epps, Dale Joachim, Michael Chen:
Depression Detection from Short Utterances via Diverse Smartphones in Natural Environmental Conditions. 3393-3397 - Yasin Özkanca, Cenk Demiroglu, Asli Besirli, Selime Celik:
Multi-Lingual Depression-Level Assessment from Conversational Speech Using Acoustic and Text Features. 3398-3402 - N. P. Narendra, Paavo Alku:
Dysarthric Speech Classification Using Glottal Features Computed from Non-words, Words and Sentences. 3403-3407 - Gábor Gosztolya, Anita Bagi, Szilvia Szalóki, István Szendi, Ildikó Hoffmann:
Identifying Schizophrenia Based on Temporal Parameters in Spontaneous Speech. 3408-3412 - Karan Singla, Zhuohao Chen, Nikolaos Flemotomos, James Gibson, Dogan Can, David C. Atkins, Shrikanth S. Narayanan:
Using Prosodic and Lexical Information for Learning Utterance-level Behaviors in Psychotherapy. 3413-3417 - Ying Qin, Tan Lee, Siyuan Feng, Anthony Pak-Hin Kong:
Automatic Speech Assessment for People with Aphasia Using TDNN-BLSTM with Multi-Task Learning. 3418-3422 - Md. Nasir, Brian R. Baucom, Shrikanth S. Narayanan, Panayiotis G. Georgiou:
Towards an Unsupervised Entrainment Distance in Conversational Speech Using Deep Neural Networks. 3423-3427 - Francisco Teixeira, Alberto Abad, Isabel Trancoso:
Patient Privacy in Paralinguistic Tasks. 3428-3432 - Sadeen Alharbi, Madina Hasan, Anthony J. H. Simons, Shelagh Brumfitt, Phil D. Green:
A Lightly Supervised Approach to Detect Stuttering in Children's Speech. 3433-3437 - Jeng-Lin Li, Yi-Ming Weng, Chip-Jin Ng, Chi-Chun Lee:
Learning Conditional Acoustic Latent Representation with Gender and Age Attributes for Automatic Pain Level Recognition. 3438-3442
Perspective Talk-4
- Sriram Ganapathy:
Speaker and Language Recognition - From Laboratory Technologies to the Wild. 3443
Spoken Language Understanding
- Yu Wang, Abhishek Patel, Yilin Shen, Hongxia Jin:
A Deep Reinforcement Learning Based Multimodal Coaching Model (DCM) for Slot Filling in Spoken Language Understanding(SLU). 3444-3448 - Frédéric Béchet, Christian Raymond:
Is ATIS Too Shallow to Go Deeper for Benchmarking Spoken Language Understanding Models? 3449-3453 - Avik Ray, Yilin Shen, Hongxia Jin:
Robust Spoken Language Understanding via Paraphrasing. 3454-3458 - Chia-Hsuan Li, Szu-Lin Wu, Chi-Liang Liu, Hung-yi Lee:
Spoken SQuAD: A Study of Mitigating the Impact of Speech Recognition Errors on Listening Comprehension. 3459-3463 - Yilin Shen, Xiangyu Zeng, Yu Wang, Hongxia Jin:
User Information Augmented Semantic Frame Parsing Using Progressive Neural Networks. 3464-3468 - Raghav Gupta, Abhinav Rastogi, Dilek Hakkani-Tür:
An Efficient Approach to Encoding Context for Spoken Language Understanding. 3469-3473
Source Separation from Monaural Input
- Jeffrey Hetherly, Paul Gamble, Maria Alejandra Barrios, Cory Stephenson, Karl Ni:
Deep Speech Denoising with Vector Space Projections. 3474-3478 - Chenglin Xu, Wei Rao, Eng Siong Chng, Haizhou Li:
A Shifted Delta Coefficient Objective for Monaural Speech Separation Using Multi-task Learning. 3479-3483 - Ke Tan, DeLiang Wang:
A Two-Stage Approach to Noisy Cochannel Speech Separation with Gated Residual Networks. 3484-3488 - Laxmi Pandey, Anurendra Kumar, Vinay P. Namboodiri:
Monoaural Audio Source Separation Using Variational Autoencoders. 3489-3493 - Arpita Gang, Pravesh Biyani, Akshay Soni:
Towards Automated Single Channel Source Separation Using Neural Networks. 3494-3498 - Hakan Erdogan, Takuya Yoshioka:
Investigations on Data Augmentation and Loss Functions for Deep Learning Based Speech-Background Separation. 3499-3503
Multimodal Systems
- Simone Hantke, Christoph Stemp, Björn W. Schuller:
Annotator Trustability-based Cooperative Learning Solutions for Intelligent Audio Analysis. 3504-3508 - Rongfeng Su, Xunying Liu, Lan Wang:
Semi-supervised Cross-domain Visual Feature Learning for Audio-Visual Broadcast Speech Transcription. 3509-3513 - Triantafyllos Afouras, Joon Son Chung, Andrew Zisserman:
Deep Lip Reading: A Comparison of Models and an Online Application. 3514-3518 - Ajay Srinivasamurthy, Petr Motlícek, Mittul Singh, Youssef Oualil, Matthias Kleinert, Heiko Ehr, Hartmut Helmke:
Iterative Learning of Speech Recognition Models for Air Traffic Control. 3519-3523 - Leda Sari, Mark Hasegawa-Johnson, Kumaran S, Georg Stemmer, Krishnakumar N. Nair:
Speaker Adaptive Audio-Visual Fusion for the Open-Vocabulary Section of AVICAR. 3524-3528 - Marek Hrúz, Ales Prazák, Michal Busta:
Multimodal Name Recognition in Live TV Subtitling. 3529-3532
Coding
- Tom Bäckström, Johannes Fischer, Sneha Das:
Dithered Quantization for Frequency-Domain Speech and Audio Coding. 3533-3537 - Sneha Das, Tom Bäckström:
Postfiltering with Complex Spectral Correlations for Speech and Audio Coding. 3538-3542 - Sneha Das, Tom Bäckström:
Postfiltering Using Log-Magnitude Spectrum for Speech and Audio Coding. 3543-3547 - Arijit Biswas, Per Hedelin, Lars F. Villemoes, Vinay Melkote:
Temporal Noise Shaping with Companding. 3548-3552 - Yaxing Li, Eshete Derb Emiru, Shengwu Xiong, Anna Zhu, Pengfei Duan, Yichang Li:
Multi-frame Quantization of LSF Parameters Using a Deep Autoencoder and Pyramid Vector Quantizer. 3553-3557 - Yaxing Li, Shan Xu, Shengwu Xiong, Anna Zhu, Pengfei Duan, Yueming Ding:
Multi-frame Coding of LSF Parameters Using Block-Constrained Trellis Coded Vector Quantization. 3558-3562
Speaker Verification Using Neural Network Methods II
- Heewoong Park, Sukhyun Cho, Kyubyong Park, Namju Kim, Jonghun Park:
Training Utterance-level Embedding Networks for Speaker Identification and Verification. 3563-3567 - Mahesh Kumar Nandwana, Mitchell McLaren, Diego Castán, Julien van Hout, Aaron Lawson:
Analysis of Complementary Information Sources in the Speaker Embeddings Framework. 3568-3572 - Yingke Zhu, Tom Ko, David Snyder, Brian Mak, Daniel Povey:
Self-Attentive Speaker Embeddings for Text-Independent Speaker Verification. 3573-3577 - Zhifu Gao, Yan Song, Ian McLoughlin, Wu Guo, Lirong Dai:
An Improved Deep Embedding Learning Method for Short Duration Speaker Verification. 3578-3582 - Jee-weon Jung, Hee-Soo Heo, Il-Ho Yang, Hye-jin Shim, Ha-Jin Yu:
Avoiding Speaker Overfitting in End-to-End DNNs Using Raw Waveform for Text-Independent Speaker Verification. 3583-3587 - Gautam Bhattacharya, Jahangir Alam, Vishwa Gupta, Patrick Kenny:
Deeply Fused Speaker Embeddings for Text-Independent Speaker Verification. 3588-3592 - Md. Hafizur Rahman, Ivan Himawan, Mitchell McLaren, Clinton Fookes, Sridha Sridharan:
Employing Phonetic Information in DNN Speaker Embeddings to Improve Speaker Recognition Performance. 3593-3597 - Subhadeep Dey, Srikanth R. Madikeri, Petr Motlícek:
End-to-end Text-dependent Speaker Verification Using Novel Distance Measures. 3598-3602 - Harishchandra Dubey, Abhijeet Sangwan, John H. L. Hansen:
Robust Speaker Clustering using Mixtures of von Mises-Fisher Distributions for Naturalistic Audio Streams. 3603-3607 - Huan Song, Megan M. Willi, Jayaraman J. Thiagarajan, Visar Berisha, Andreas Spanias:
Triplet Network with Attention for Speaker Diarization. 3608-3612 - Jiacen Zhang, Nakamasa Inoue, Koichi Shinoda:
I-vector Transformation Using Conditional Generative Adversarial Networks for Short Utterance Speaker Verification. 3613-3617 - Weicheng Cai, Jinkun Chen, Ming Li:
Analysis of Length Normalization in End-to-End Speaker Verification System. 3618-3622 - Zili Huang, Shuai Wang, Kai Yu:
Angular Softmax for Short-Duration Text-independent Speaker Verification. 3623-3627 - Ruifang Ji, Xinyuan Cai, Bo Xu:
An End-to-End Text-Independent Speaker Identification System on Short Utterances. 3628-3632 - Wenhao Ding, Liang He:
MTGAN: Speaker Verification through Multitasking Triplet Generative Adversarial Networks. 3633-3637
Emotion Recognition and Analysis
- Emilia Parada-Cabaleiro, Giovanni Costantini, Anton Batliner, Alice Baird, Björn W. Schuller:
Categorical vs Dimensional Perception of Italian Emotional Speech. 3638-3642 - Xingfeng Li, Masato Akagi:
A Three-Layer Emotion Perception Model for Valence and Arousal-Based Detection from Multilingual Speech. 3643-3647 - Brecht Desplanques, Kris Demuynck:
Cross-lingual Speech Emotion Recognition through Factor Analysis. 3648-3652 - Jian Cheng, Jared Bernstein, Elizabeth Rosenfeld, Peter W. Foltz, Alex S. Cohen, Terje B. Holmlund, Brita Elvevåg:
Modeling Self-Reported and Observed Affect from Speech. 3653-3657 - Che-Wei Huang, Shrikanth S. Narayanan:
Stochastic Shake-Shake Regularization for Affective Learning from Speech. 3658-3662 - Anderson R. Avila, Md. Jahangir Alam, Douglas D. O'Shaughnessy, Tiago H. Falk:
Investigating Speech Enhancement and Perceptual Quality for Speech Emotion Recognition. 3663-3667 - Mia Atcheson, Vidhyasaharan Sethu, Julien Epps:
Demonstrating and Modelling Systematic Time-varying Annotator Disagreement in Continuous Emotion Annotation. 3668-3672 - Jian Huang, Ya Li, Jianhua Tao, Zhen Lian:
Speech Emotion Recognition from Variable-Length Inputs with Triplet Loss Function. 3673-3677 - Xiaotong Zhang, Xingliang Cheng, Mingxing Xu, Thomas Fang Zheng:
Imbalance Learning-based Framework for Fear Recognition in the MediaEval Emotional Impact of Movies Task. 3678-3682 - Xi Ma, Zhiyong Wu, Jia Jia, Mingxing Xu, Helen Meng, Lianhong Cai:
Emotion Recognition from Variable-Length Speech Segments Using Deep Learning on Spectrograms. 3683-3687 - Promod Yenigalla, Abhay Kumar, Suraj Tripathi, Chirag Singh, Sibsambhu Kar, Jithendra Vepa:
Speech Emotion Recognition Using Spectrogram & Phoneme Embedding. 3688-3692 - Saurabh Sahu, Rahul Gupta, Carol Y. Espy-Wilson:
On Enhancing Speech Emotion Recognition Using Generative Adversarial Networks. 3693-3697 - Srinivas Parthasarathy, Carlos Busso:
Ladder Networks for Emotion Recognition: Using Unsupervised Auxiliary Tasks to Improve Predictions of Emotional Attributes. 3698-3702
Acoustic Modelling
- Mingkun Huang, Yongbin You, Zhehuai Chen, Yanmin Qian, Kai Yu:
Knowledge Distillation for Sequence Model. 3703-3707 - Sheng Li, Xugang Lu, Ryoichi Takashima, Peng Shen, Tatsuya Kawahara, Hisashi Kawai:
Improving CTC-based Acoustic Model with Very Deep Residual Time-delay Neural Networks. 3708-3712 - Jinxi Guo, Ning Xu, Xin Chen, Yang Shi, Kaiyuan Xu, Abeer Alwan:
Filter Sampling and Combination CNN (FSC-CNN): A Compact CNN Model for Small-footprint ASR Acoustic Modeling Using Raw Waveforms. 3713-3717 - Mirco Ravanelli, Dmitriy Serdyuk, Yoshua Bengio:
Twin Regularization for Online Speech Recognition. 3718-3722 - Matthias Sperber, Jan Niehues, Graham Neubig, Sebastian Stüker, Alex Waibel:
Self-Attentional Acoustic Models. 3723-3727 - Jinhwan Park, Iksoo Choi, Yoonho Boo, Wonyong Sung:
Hierarchical Recurrent Neural Networks for Acoustic Modeling. 3728-3732 - Antoine Bruguier, Anton Bakhtin, Dravyansh Sharma:
Dictionary Augmented Sequence-to-Sequence Neural Network for Grapheme to Phoneme Prediction. 3733-3737 - Ankit Raj, Shakti P. Rath, Jithendra Vepa:
Leveraging Second-Order Log-Linear Model for Improved Deep Learning Based ASR Performance. 3738-3742 - Daniel Povey, Gaofeng Cheng, Yiming Wang, Ke Li, Hainan Xu, Mahsa Yarmohammadi, Sanjeev Khudanpur:
Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks. 3743-3747 - Da-Rong Liu, Kuan-Yu Chen, Hung-yi Lee, Lin-Shan Lee:
Completely Unsupervised Phoneme Recognition by Adversarially Learning Mapping Relationships from Audio Embeddings. 3748-3752 - Mengjie Qian, Linxue Bai, Peter Jancovic, Martin J. Russell:
Phone Recognition Using a Non-Linear Manifold with Broad Phone Class Dependent DNNs. 3753-3757 - Ehsan Hosseini-Asl, Yingbo Zhou, Caiming Xiong, Richard Socher:
A Multi-Discriminator CycleGAN for Unsupervised Non-Parallel Speech Domain Adaptation. 3758-3762
Speech and Speaker Perception
- Chong Cao, Wei Wei, Wei Wang, Yanlu Xie, Jinsong Zhang:
Interactions between Vowels and Nasal Codas in Mandarin Speakers' Perception of Nasal Finals. 3763-3767 - Qinglin Meng, Nengheng Zheng, Ambika Prasad Mishra, Jacinta Dan Luo, Jan W. H. Schnupp:
Weighting Pitch Contour and Loudness Contour in Mandarin Tone Perception in Cochlear Implant Listeners. 3768-3771 - Filip Nenadic, Louis ten Bosch, Benjamin V. Tucker:
Implementing DIANA to Model Isolated Auditory Word Recognition in English. 3772-3776 - Bhamini Sharma:
Effects of Homophone Density on Spoken Word Recognition in Mandarin Chinese. 3777-3780 - Hui Xie, Biao Zeng, Rui Wang:
Visual Timing Information in Audiovisual Speech Perception: Evidence from Lexical Tone Contour. 3781-3785 - Marie-Lou Barnaud, Julien Diard, Pierre Bessière, Jean-Luc Schwartz:
COSMO SylPhon: A Bayesian Perceptuo-motor Model to Assess Phonological Learning. 3786-3790 - Akshay Raj Maggu, Patrick C. M. Wong, Hanjun Liu, Francis C. K. Wong:
Experience-dependent Influence of Music and Language on Lexical Pitch Learning Is Not Additive. 3791-3794 - Volker Dellwo, Thayabaran Kathiresan, Elisa Pellegrino, Lei He, Sandra Schwab, Dieter Maurer:
Influences of Fundamental Oscillation on Speaker Identification in Vocalic Utterances by Humans and Computers. 3795-3799
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.