{"id":605979,"date":"2019-09-03T17:48:28","date_gmt":"2019-09-04T00:48:28","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-event&p=605979"},"modified":"2025-08-06T11:56:17","modified_gmt":"2025-08-06T18:56:17","slug":"interspeech-2019","status":"publish","type":"msr-event","link":"https:\/\/www.microsoft.com\/en-us\/research\/event\/interspeech-2019\/","title":{"rendered":"Microsoft at Interspeech 2019"},"content":{"rendered":"\n\n

Venue:<\/strong> Messecongress Graz (opens in new tab)<\/span><\/a><\/p>\n

Website:<\/strong> Interspeech 2019 (opens in new tab)<\/span><\/a>Opens in a new tab<\/span><\/p>\n

Interspeech is the world\u2018s largest and most comprehensive conference on the science and technology of spoken language processing. Microsoft joins the conference as a proud gold sponsor. Stop by our booth to chat with our experts, see demos of our latest research and find out about career opportunities (opens in new tab)<\/span><\/a>\u00a0with Microsoft.Opens in a new tab<\/span><\/p>\n

Monday, September 16<\/h3>\n

15:30-15:50 | Hall 1 | Oral
\nSpeaker Adaptation for Attention-Based End-to-End Speech Recognition<\/strong>
\n
Zhong Meng (opens in new tab)<\/span><\/a>, Yashesh Gaur (opens in new tab)<\/span><\/a>, Jinyu Li<\/a>, Yifan Gong (opens in new tab)<\/span><\/a><\/p>\n

14:30-16:30 | Gallery C | Poster
\nZero Shot Intent Classification Using Long-Short Term Memory Networks<\/strong>
\n<\/strong>
Kyle Williams<\/a><\/p>\n

14:30 \u2013 16:30 | Hall 4 | Show & Tell
\nSpeech Based Web Navigation for Movement Impaired Users<\/strong>
\n
Vasiliy Radostev (opens in new tab)<\/span><\/a>, Serge Berger (opens in new tab)<\/span><\/a>, Justin Tabrizi (opens in new tab)<\/span><\/a>, Pasha Kamyshev (opens in new tab)<\/span><\/a>, Hisami Suzuki (opens in new tab)<\/span><\/a><\/p>\n

Tuesday, September 17<\/h3>\n

10:00-12:00 | Hall 10\/E | Poster
\nA Scalable Noisy Speech Dataset and Online Subjective Test Framework<\/strong>\u00a0<\/strong>
\n
Ebrahim Beyrami (opens in new tab)<\/span><\/a>, Chandan Karadagur Ananda Reddy (opens in new tab)<\/span><\/a>, Jamie Pool (opens in new tab)<\/span><\/a>, Ross Cutler (opens in new tab)<\/span><\/a>, Sriram Srinivasan (opens in new tab)<\/span><\/a>, Johannes Gehrke <\/a><\/p>\n

13:30-15:30 | Hall 10\/E | Poster
\nSpeech Signal Characterization 3\/Vocal Pitch Extraction in Polyphonic Music using Convolutional Residual Network<\/strong>
\nMingye Dong, <\/em>
Jie Wu (opens in new tab)<\/span><\/a>, Jian Luan (opens in new tab)<\/span><\/a><\/p>\n

13:30-13:50 | Hall 1 | Oral
\nForward-Backward Decoding for Regularizing End-to-End TTS<\/strong>
\nYibin Zheng,
Xi Wang (opens in new tab)<\/span><\/a>, Lei He (opens in new tab)<\/span><\/a>, Shifeng Pan (opens in new tab)<\/span><\/a>, Frank Soong<\/a>, Zhengqi Wen, Jianhua Tao (opens in new tab)<\/span><\/a><\/p>\n

13:50-14:10 | Hall 2 | Oral
\nA New GAN-based End-to-End TTS Training Algorithm<\/strong>\u00a0<\/strong>
\nHaohan Guo,
Frank Soong<\/a>, Lei He (opens in new tab)<\/span><\/a>, Lei Xie<\/p>\n

14:10-14:30 | Hall 2 | Oral
\nRobust Sequence-to-Sequence Acoustic Modeling with Stepwise Monotonic Attention for Neural TTS<\/strong>
\n<\/strong>Mutian He,
Yan Deng (opens in new tab)<\/span><\/a>, Lei He (opens in new tab)<\/span><\/a><\/p>\n

16:00-18:00 | Gallery A | Poster
\nToken-Level Ensemble Distillation for Grapheme-to-Phoneme Conversion<\/strong>\u00a0 <\/strong>
\nHao Sun,
Xu Tan<\/a>, Jun-Wei Gan (opens in new tab)<\/span><\/a>, Hongzhi Liu, Sheng Zhao (opens in new tab)<\/span><\/a>, Tao Qin (opens in new tab)<\/span><\/a>, Tie-Yan Liu<\/a><\/p>\n

16:00-18:00 | Gallery B | Poster
\nExploiting Monolingual Speech Corpora for Code-mixed Speech Recognition<\/strong>
\nKaran Taneja,
Satarupa Guha (opens in new tab)<\/span><\/a>, Preethi Jyothi, Basil Abraham (opens in new tab)<\/span><\/a><\/p>\n

16:40-17:00 | Hall 1 | Oral
\nLayer Trajectory BLSTM<\/strong>
\nEric Sun<\/strong>,
Jinyu Li<\/a>, Yifan Gong (opens in new tab)<\/span><\/a><\/p>\n

16:00-18:00 | Gallery C | Poster
\nAcoustic-to-Phrase Models for Speech Recognition<\/strong>\u00a0 <\/strong>
\n
Yashesh Gaur (opens in new tab)<\/span><\/a>, Jinyu Li<\/a>, Zhong Meng (opens in new tab)<\/span><\/a>, Yifan Gong (opens in new tab)<\/span><\/a><\/p>\n

Wednesday, September 18<\/h3>\n

11:20-11:40 | Hall 1 | Oral
\nSupervised Classifiers for Audio Impairments with Noisy Labels<\/strong>\u00a0<\/strong>
\n
Chandan Karadagur Ananda Reddy (opens in new tab)<\/span><\/a>, Ross Cutler (opens in new tab)<\/span><\/a>, Johannes Gehrke<\/a><\/p>\n

10:00-12:00 | Gallery B | Poster
\nMeeting Transcription Using Asynchronous Distant Microphones
\n
Takuya Yoshioka<\/a>, Dimitrios Dimitriadis<\/a>, Andreas Stolcke<\/a>, William Hinthorn<\/a>, Zhuo Chen (opens in new tab)<\/span><\/a>, Michael Zeng<\/a>, Xuedong Huang<\/a><\/p>\n

13:30-15:30 | Gallery B | Poster
\nCompression of CTC-Trained Acoustic Models by Dynamic Frame-Wise Distillation or Segment-Wise N-Best Hypotheses Imitation<\/strong>
\nHaisong Ding,
Kai Chen<\/a>, Qiang Huo<\/a><\/p>\n

13:30-15:30 | Gallery B | Poster
\nLatent Dirichlet Allocation based Acoustic Data Selection for Automatic Speech Recognition<\/strong>
\n<\/strong>
Mortaza (Morrie) Doulaty (opens in new tab)<\/span><\/a>, Thomas Hain<\/p>\n

17:40-18:00 | Hall 1| Oral
\nSelf-Teaching Networks<\/strong>\u00a0<\/strong>
\n
Liang Lu (opens in new tab)<\/span><\/a>, Eric Sun<\/strong>, Yifan Gong (opens in new tab)<\/span><\/a><\/p>\n

16:00-18:00 | Hall 10\/E | Poster
\nSound Event Detection in Multichannel Audio Using Convolutional Time-Frequency Channel Squeeze and Excitation<\/strong>
\nWei Xia,
Kazuhito Koishida<\/a><\/p>\n

Thursday, September 19<\/h3>\n

13:30-15:30 | Gallery C | Poster
\nExploiting Syntactic<\/strong>
\nFeatures in a Parsed Tree to Improve End-to-End TTS<\/strong>\u00a0<\/strong>
\nHaohan Guo,
Frank Soong<\/a>, Lei He (opens in new tab)<\/span><\/a>, Lei Xie<\/p>\n

13:30-15:30 | Hall 12 | Special Session
\nSpeech Technologies for Code-Switching in Multilingual Communities<\/strong>
\nOrganizers:
Kalika Bali<\/a>, Alan W Black, Julia Hirschberg, Sunayana Sitaram<\/a>, Thamar SolorioOpens in a new tab<\/span><\/p>\n

\t\t\t

\n\t\t\t
\n\t\t\t\t\n
\n\t
\n\t\t

\n\t\t\t\t\t\t\tCognition and Speech Scientist<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t
<\/div>

Type<\/strong>: Full-time<\/p>

Lab\/Location<\/strong>: Redmond, Washington<\/p>

We are looking for a motivated, self-driven software development engineer\/scientist to join our mission to change the world with TTS technology.<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>

\n

\n\t
\n\t\t

\n\t\t\t\t\t\t\tCognition and Speech Scientist Intern<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t<\/p>
<\/div>

Type<\/strong>: Internship<\/p>

Lab\/Location<\/strong>: Redmond, Washington<\/p>

We are looking for a motivated, self-driven software development engineer\/scientist intern to join our mission to change the world with TTS technology.<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>

\n

\n\t
\n\t\t

\n\t\t\t\t\t\t\tApplied Scientist<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t<\/p>
<\/div>

Type<\/strong>: Full-time<\/p>

Lab\/Location<\/strong>: Redmond, Washington<\/p>

We are hiring Scientists\/Engineers with outstanding machine learning (ML) and speech recognition (SR) technology development skills to advance Microsoft’s core speech technology.<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>\t\t\t<\/div>\n\t\t<\/div>\n\t\t
\n\t\t\t

\n\t\t\t
\n\t\t\t\t\n
\n\t
\n\t\t

\n\t\t\t\t\t\t\tSr. Applied Scientist<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t
<\/div>

Type<\/strong>: Full-time<\/p>

Lab\/Location<\/strong>: Redmond, Washington<\/p>

The Speech Group develops speech recognition features in Enterprise, Entertainment and Desktop and Mobile products and particularly in the voice platform that powers Microsoft 365 Search and Assistant…<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>

\n

\n\t
\n\t\t

\n\t\t\t\t\t\t\tApplied Scientist II<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t<\/p>
<\/div>

Type<\/strong>: Full-time<\/p>

Lab\/Location<\/strong>: Bellevue, Washington<\/p>

Are you interested in AI and machine learning technology, especially involving speech and language? Are you an expert in deep learning or willing to learn those advance techniques used in Cloud+AI products…<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>\t\t\t<\/div>\n\t\t<\/div>\n\t\tOpens in a new tab<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"

Interspeech is the world\u2018s largest and most comprehensive conference on the science and technology of spoken language processing. Microsoft joins the conference as a proud gold sponsor. Stop by our booth to chat with our experts, see demos of our latest research and find out about career opportunities\u00a0with Microsoft.<\/p>\n","protected":false},"featured_media":606480,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_startdate":"2019-09-15","msr_enddate":"2019-09-19","msr_location":"Graz, Austria","msr_expirationdate":"","msr_event_recording_link":"","msr_event_link":"https:\/\/www.interspeech2019.org\/registration\/registration_overview_and_fees\/","msr_event_link_redirect":false,"msr_event_time":"","msr_hide_region":false,"msr_private_event":false,"msr_hide_image_in_river":0,"footnotes":""},"research-area":[13545],"msr-region":[239178],"msr-event-type":[197941],"msr-video-type":[],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[],"msr-impact-theme":[],"class_list":["post-605979","msr-event","type-msr-event","status-publish","has-post-thumbnail","hentry","msr-research-area-human-language-technologies","msr-region-europe","msr-event-type-conferences","msr-locale-en_us"],"msr_about":"\n\n

Venue:<\/strong> Messecongress Graz (opens in new tab)<\/span><\/a><\/p>\n

Website:<\/strong> Interspeech 2019 (opens in new tab)<\/span><\/a>Opens in a new tab<\/span><\/p>\n

Interspeech is the world\u2018s largest and most comprehensive conference on the science and technology of spoken language processing. Microsoft joins the conference as a proud gold sponsor. Stop by our booth to chat with our experts, see demos of our latest research and find out about career opportunities (opens in new tab)<\/span><\/a>\u00a0with Microsoft.Opens in a new tab<\/span><\/p>\n

Monday, September 16<\/h3>\n

15:30-15:50 | Hall 1 | Oral
\nSpeaker Adaptation for Attention-Based End-to-End Speech Recognition<\/strong>
\n
Zhong Meng<\/a>, Yashesh Gaur<\/a>, Jinyu Li<\/a>, Yifan Gong<\/a><\/p>\n

14:30-16:30 | Gallery C | Poster
\nZero Shot Intent Classification Using Long-Short Term Memory Networks<\/strong>
\n<\/strong>
Kyle Williams<\/a><\/p>\n

14:30 \u2013 16:30 | Hall 4 | Show & Tell
\nSpeech Based Web Navigation for Movement Impaired Users<\/strong>
\n
Vasiliy Radostev<\/a>, Serge Berger<\/a>, Justin Tabrizi<\/a>, Pasha Kamyshev<\/a>, Hisami Suzuki<\/a><\/p>\n

Tuesday, September 17<\/h3>\n

10:00-12:00 | Hall 10\/E | Poster
\nA Scalable Noisy Speech Dataset and Online Subjective Test Framework<\/strong>\u00a0<\/strong>
\n
Ebrahim Beyrami<\/a>, Chandan Karadagur Ananda Reddy<\/a>, Jamie Pool<\/a>, Ross Cutler<\/a>, Sriram Srinivasan<\/a>, Johannes Gehrke <\/a><\/p>\n

13:30-15:30 | Hall 10\/E | Poster
\nSpeech Signal Characterization 3\/Vocal Pitch Extraction in Polyphonic Music using Convolutional Residual Network<\/strong>
\nMingye Dong, <\/em>
Jie Wu<\/a>, Jian Luan<\/a><\/p>\n

13:30-13:50 | Hall 1 | Oral
\nForward-Backward Decoding for Regularizing End-to-End TTS<\/strong>
\nYibin Zheng,
Xi Wang<\/a>, Lei He<\/a>, Shifeng Pan<\/a>, Frank Soong<\/a>, Zhengqi Wen, Jianhua Tao<\/a><\/p>\n

13:50-14:10 | Hall 2 | Oral
\nA New GAN-based End-to-End TTS Training Algorithm<\/strong>\u00a0<\/strong>
\nHaohan Guo,
Frank Soong<\/a>, Lei He<\/a>, Lei Xie<\/p>\n

14:10-14:30 | Hall 2 | Oral
\nRobust Sequence-to-Sequence Acoustic Modeling with Stepwise Monotonic Attention for Neural TTS<\/strong>
\n<\/strong>Mutian He,
Yan Deng<\/a>, Lei He<\/a><\/p>\n

16:00-18:00 | Gallery A | Poster
\nToken-Level Ensemble Distillation for Grapheme-to-Phoneme Conversion<\/strong>\u00a0 <\/strong>
\nHao Sun,
Xu Tan<\/a>, Jun-Wei Gan<\/a>, Hongzhi Liu, Sheng Zhao<\/a>, Tao Qin<\/a>, Tie-Yan Liu<\/a><\/p>\n

16:00-18:00 | Gallery B | Poster
\nExploiting Monolingual Speech Corpora for Code-mixed Speech Recognition<\/strong>
\nKaran Taneja,
Satarupa Guha<\/a>, Preethi Jyothi, Basil Abraham<\/a><\/p>\n

16:40-17:00 | Hall 1 | Oral
\nLayer Trajectory BLSTM<\/strong>
\nEric Sun<\/strong>,
Jinyu Li<\/a>, Yifan Gong<\/a><\/p>\n

16:00-18:00 | Gallery C | Poster
\nAcoustic-to-Phrase Models for Speech Recognition<\/strong>\u00a0 <\/strong>
\n
Yashesh Gaur<\/a>, Jinyu Li<\/a>, Zhong Meng<\/a>, Yifan Gong<\/a><\/p>\n

Wednesday, September 18<\/h3>\n

11:20-11:40 | Hall 1 | Oral
\nSupervised Classifiers for Audio Impairments with Noisy Labels<\/strong>\u00a0<\/strong>
\n
Chandan Karadagur Ananda Reddy<\/a>, Ross Cutler<\/a>, Johannes Gehrke<\/a><\/p>\n

10:00-12:00 | Gallery B | Poster
\nMeeting Transcription Using Asynchronous Distant Microphones
\n
Takuya Yoshioka<\/a>, Dimitrios Dimitriadis<\/a>, Andreas Stolcke<\/a>, William Hinthorn<\/a>, Zhuo Chen<\/a>, Michael Zeng<\/a>, Xuedong Huang<\/a><\/p>\n

13:30-15:30 | Gallery B | Poster
\nCompression of CTC-Trained Acoustic Models by Dynamic Frame-Wise Distillation or Segment-Wise N-Best Hypotheses Imitation<\/strong>
\nHaisong Ding,
Kai Chen<\/a>, Qiang Huo<\/a><\/p>\n

13:30-15:30 | Gallery B | Poster
\nLatent Dirichlet Allocation based Acoustic Data Selection for Automatic Speech Recognition<\/strong>
\n<\/strong>
Mortaza (Morrie) Doulaty<\/a>, Thomas Hain<\/p>\n

17:40-18:00 | Hall 1| Oral
\nSelf-Teaching Networks<\/strong>\u00a0<\/strong>
\n
Liang Lu<\/a>, Eric Sun<\/strong>, Yifan Gong<\/a><\/p>\n

16:00-18:00 | Hall 10\/E | Poster
\nSound Event Detection in Multichannel Audio Using Convolutional Time-Frequency Channel Squeeze and Excitation<\/strong>
\nWei Xia,
Kazuhito Koishida<\/a><\/p>\n

Thursday, September 19<\/h3>\n

13:30-15:30 | Gallery C | Poster
\nExploiting Syntactic<\/strong>
\nFeatures in a Parsed Tree to Improve End-to-End TTS<\/strong>\u00a0<\/strong>
\nHaohan Guo,
Frank Soong<\/a>, Lei He<\/a>, Lei Xie<\/p>\n

13:30-15:30 | Hall 12 | Special Session
\nSpeech Technologies for Code-Switching in Multilingual Communities<\/strong>
\nOrganizers:
Kalika Bali<\/a>, Alan W Black, Julia Hirschberg, Sunayana Sitaram<\/a>, Thamar SolorioOpens in a new tab<\/span><\/p>\n

\t\t\t

\n\t\t\t
\n\t\t\t\t\n
\n\t
\n\t\t

\n\t\t\t\t\t\t\tCognition and Speech Scientist<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t
<\/div>

Type<\/strong>: Full-time<\/p>

Lab\/Location<\/strong>: Redmond, Washington<\/p>

We are looking for a motivated, self-driven software development engineer\/scientist to join our mission to change the world with TTS technology.<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>

\n

\n\t
\n\t\t

\n\t\t\t\t\t\t\tCognition and Speech Scientist Intern<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t<\/p>
<\/div>

Type<\/strong>: Internship<\/p>

Lab\/Location<\/strong>: Redmond, Washington<\/p>

We are looking for a motivated, self-driven software development engineer\/scientist intern to join our mission to change the world with TTS technology.<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>

\n

\n\t
\n\t\t

\n\t\t\t\t\t\t\tApplied Scientist<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t<\/p>
<\/div>

Type<\/strong>: Full-time<\/p>

Lab\/Location<\/strong>: Redmond, Washington<\/p>

We are hiring Scientists\/Engineers with outstanding machine learning (ML) and speech recognition (SR) technology development skills to advance Microsoft’s core speech technology.<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>\t\t\t<\/div>\n\t\t<\/div>\n\t\t
\n\t\t\t

\n\t\t\t
\n\t\t\t\t\n
\n\t
\n\t\t

\n\t\t\t\t\t\t\tSr. Applied Scientist<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t
<\/div>

Type<\/strong>: Full-time<\/p>

Lab\/Location<\/strong>: Redmond, Washington<\/p>

The Speech Group develops speech recognition features in Enterprise, Entertainment and Desktop and Mobile products and particularly in the voice platform that powers Microsoft 365 Search and Assistant…<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>

\n

\n\t
\n\t\t

\n\t\t\t\t\t\t\tApplied Scientist II<\/a>\n\t\t\t\t\t<\/h3>\n\n\t\t
\n\t\t\t
\n\t\t\t\t\t\t\t\t<\/p>
<\/div>

Type<\/strong>: Full-time<\/p>

Lab\/Location<\/strong>: Bellevue, Washington<\/p>

Are you interested in AI and machine learning technology, especially involving speech and language? Are you an expert in deep learning or willing to learn those advance techniques used in Cloud+AI products…<\/p>

\t\t\t<\/div>\n\t\t<\/div>\n\n\t<\/div>\n<\/article>\n<\/p>\t\t\t<\/div>\n\t\t<\/div>\n\t\tOpens in a new tab<\/span><\/p>\n","tab-content":[{"id":0,"name":"About","content":"Interspeech is the world\u2018s largest and most comprehensive conference on the science and technology of spoken language processing. Microsoft joins the conference as a proud gold sponsor. Stop by our booth to chat with our experts, see demos of our latest research and find out about career opportunities<\/a>\u00a0with Microsoft."},{"id":1,"name":"Schedule","content":"

Monday, September 16<\/h3>\r\n15:30-15:50 | Hall 1 | Oral\r\nSpeaker Adaptation for Attention-Based End-to-End Speech Recognition<\/strong>\r\nZhong Meng<\/a>, Yashesh Gaur<\/a>, Jinyu Li<\/a>, Yifan Gong<\/a>\r\n\r\n14:30-16:30 | Gallery C | Poster\r\nZero Shot Intent Classification Using Long-Short Term Memory Networks<\/strong>\r\n<\/strong>Kyle Williams<\/a>\r\n\r\n14:30 \u2013 16:30 | Hall 4 | Show & Tell\r\nSpeech Based Web Navigation for Movement Impaired Users<\/strong>\r\nVasiliy Radostev<\/a>, Serge Berger<\/a>, Justin Tabrizi<\/a>, Pasha Kamyshev<\/a>, Hisami Suzuki<\/a>\r\n

Tuesday, September 17<\/h3>\r\n10:00-12:00 | Hall 10\/E | Poster\r\nA Scalable Noisy Speech Dataset and Online Subjective Test Framework<\/strong>\u00a0<\/strong>\r\nEbrahim Beyrami<\/a>, Chandan Karadagur Ananda Reddy<\/a>, Jamie Pool<\/a>, Ross Cutler<\/a>, Sriram Srinivasan<\/a>, Johannes Gehrke <\/a>\r\n\r\n13:30-15:30 | Hall 10\/E | Poster\r\nSpeech Signal Characterization 3\/Vocal Pitch Extraction in Polyphonic Music using Convolutional Residual Network<\/strong>\r\nMingye Dong, <\/em>Jie Wu<\/a>, Jian Luan<\/a>\r\n\r\n13:30-13:50 | Hall 1 | Oral\r\nForward-Backward Decoding for Regularizing End-to-End TTS<\/strong>\r\nYibin Zheng, Xi Wang<\/a>, Lei He<\/a>, Shifeng Pan<\/a>, Frank Soong<\/a>, Zhengqi Wen, Jianhua Tao<\/a>\r\n\r\n13:50-14:10 | Hall 2 | Oral\r\nA New GAN-based End-to-End TTS Training Algorithm<\/strong>\u00a0<\/strong>\r\nHaohan Guo, Frank Soong<\/a>, Lei He<\/a>, Lei Xie\r\n\r\n14:10-14:30 | Hall 2 | Oral\r\nRobust Sequence-to-Sequence Acoustic Modeling with Stepwise Monotonic Attention for Neural TTS<\/strong>\r\n<\/strong>Mutian He, Yan Deng<\/a>, Lei He<\/a>\r\n\r\n16:00-18:00 | Gallery A | Poster\r\nToken-Level Ensemble Distillation for Grapheme-to-Phoneme Conversion<\/strong>\u00a0 <\/strong>\r\nHao Sun, Xu Tan<\/a>, Jun-Wei Gan<\/a>, Hongzhi Liu, Sheng Zhao<\/a>, Tao Qin<\/a>, Tie-Yan Liu<\/a>\r\n\r\n16:00-18:00 | Gallery B | Poster\r\nExploiting Monolingual Speech Corpora for Code-mixed Speech Recognition<\/strong>\r\nKaran Taneja, Satarupa Guha<\/a>, Preethi Jyothi, Basil Abraham<\/a>\r\n\r\n16:40-17:00 | Hall 1 | Oral\r\nLayer Trajectory BLSTM<\/strong>\r\nEric Sun<\/strong>, Jinyu Li<\/a>, Yifan Gong<\/a>\r\n\r\n16:00-18:00 | Gallery C | Poster\r\nAcoustic-to-Phrase Models for Speech Recognition<\/strong>\u00a0 <\/strong>\r\nYashesh Gaur<\/a>, Jinyu Li<\/a>, Zhong Meng<\/a>, Yifan Gong<\/a>\r\n

Wednesday, September 18<\/h3>\r\n11:20-11:40 | Hall 1 | Oral\r\nSupervised Classifiers for Audio Impairments with Noisy Labels<\/strong>\u00a0<\/strong>\r\nChandan Karadagur Ananda Reddy<\/a>, Ross Cutler<\/a>, Johannes Gehrke<\/a>\r\n\r\n10:00-12:00 | Gallery B | Poster\r\nMeeting Transcription Using Asynchronous Distant Microphones\r\nTakuya Yoshioka<\/a>, Dimitrios Dimitriadis<\/a>, Andreas Stolcke<\/a>, William Hinthorn<\/a>, Zhuo Chen<\/a>, Michael Zeng<\/a>, Xuedong Huang<\/a>\r\n\r\n13:30-15:30 | Gallery B | Poster\r\nCompression of CTC-Trained Acoustic Models by Dynamic Frame-Wise Distillation or Segment-Wise N-Best Hypotheses Imitation<\/strong>\r\nHaisong Ding, Kai Chen<\/a>, Qiang Huo<\/a>\r\n\r\n13:30-15:30 | Gallery B | Poster\r\nLatent Dirichlet Allocation based Acoustic Data Selection for Automatic Speech Recognition<\/strong>\r\n<\/strong>Mortaza (Morrie) Doulaty<\/a>, Thomas Hain\r\n\r\n17:40-18:00 | Hall 1| Oral\r\nSelf-Teaching Networks<\/strong>\u00a0<\/strong>\r\nLiang Lu<\/a>, Eric Sun<\/strong>, Yifan Gong<\/a>\r\n\r\n16:00-18:00 | Hall 10\/E | Poster\r\nSound Event Detection in Multichannel Audio Using Convolutional Time-Frequency Channel Squeeze and Excitation<\/strong>\r\nWei Xia, Kazuhito Koishida<\/a>\r\n

Thursday, September 19<\/h3>\r\n13:30-15:30 | Gallery C | Poster\r\nExploiting Syntactic<\/strong>\r\nFeatures in a Parsed Tree to Improve End-to-End TTS<\/strong>\u00a0<\/strong>\r\nHaohan Guo, Frank Soong<\/a>, Lei He<\/a>, Lei Xie\r\n\r\n13:30-15:30 | Hall 12 | Special Session\r\nSpeech Technologies for Code-Switching in Multilingual Communities<\/strong>\r\nOrganizers: Kalika Bali<\/a>, Alan W Black, Julia Hirschberg, Sunayana Sitaram<\/a>, Thamar Solorio"},{"id":2,"name":"Career Opportunities","content":"[row]\r\n[card title=\"Cognition and Speech Scientist\" url=\"https:\/\/careers.microsoft.com\/students\/us\/en\/job\/653143\/Full-Time-Opportunities-for-PhD-Students-or-Recent-Graduates-Cognition-and-Speech-Scientist\" ]\r\n
<\/div>\r\n

Type<\/strong>: Full-time<\/p>\r\n

Lab\/Location<\/strong>: Redmond, Washington<\/p>\r\n

We are looking for a motivated, self-driven software development engineer\/scientist to join our mission to change the world with TTS technology.<\/p>\r\n[\/card]\r\n\r\n[card title=\"Cognition and Speech Scientist Intern\" url=\"https:\/\/careers.microsoft.com\/students\/us\/en\/job\/653144\/Internship-Opportunities-for-PhD-Students-Cognition-and-Speech-Scientist\" ]\r\n

<\/div>\r\n

Type<\/strong>: Internship<\/p>\r\n

Lab\/Location<\/strong>: Redmond, Washington<\/p>\r\n

We are looking for a motivated, self-driven software development engineer\/scientist intern to join our mission to change the world with TTS technology.<\/p>\r\n[\/card]\r\n\r\n[card title=\"Applied Scientist\" url=\"https:\/\/careers.microsoft.com\/us\/en\/job\/676849\/Applied-Scientist\" ]\r\n

<\/div>\r\n

Type<\/strong>: Full-time<\/p>\r\n

Lab\/Location<\/strong>: Redmond, Washington<\/p>\r\n

We are hiring Scientists\/Engineers with outstanding machine learning (ML) and speech recognition (SR) technology development skills to advance Microsoft's core speech technology.<\/p>\r\n[\/card]\r\n[\/row]\r\n[row]\r\n[card title=\"Sr. Applied Scientist\" url=\"https:\/\/careers.microsoft.com\/us\/en\/job\/676850\/Sr-Applied-Scientist\" ]\r\n

<\/div>\r\n

Type<\/strong>: Full-time<\/p>\r\n

Lab\/Location<\/strong>: Redmond, Washington<\/p>\r\n

The Speech Group develops speech recognition features in Enterprise, Entertainment and Desktop and Mobile products and particularly in the voice platform that powers Microsoft 365 Search and Assistant...<\/p>\r\n[\/card]\r\n\r\n[card title=\"Applied Scientist II\" url=\"https:\/\/careers.microsoft.com\/us\/en\/job\/615317\/Applied-Scientist-II\" ]\r\n

<\/div>\r\n

Type<\/strong>: Full-time<\/p>\r\n

Lab\/Location<\/strong>: Bellevue, Washington<\/p>\r\n

Are you interested in AI and machine learning technology, especially involving speech and language? Are you an expert in deep learning or willing to learn those advance techniques used in Cloud+AI products...<\/p>\r\n[\/card]\r\n[\/row]"}],"msr_startdate":"2019-09-15","msr_enddate":"2019-09-19","msr_event_time":"","msr_location":"Graz, Austria","msr_event_link":"https:\/\/www.interspeech2019.org\/registration\/registration_overview_and_fees\/","msr_event_recording_link":"","msr_startdate_formatted":"September 15, 2019","msr_register_text":"Watch now","msr_cta_link":"https:\/\/www.interspeech2019.org\/registration\/registration_overview_and_fees\/","msr_cta_text":"Watch now","msr_cta_bi_name":"Event Register","featured_image_thumbnail":"\"a","event_excerpt":"Interspeech is the world\u2018s largest and most comprehensive conference on the science and technology of spoken language processing. Microsoft joins the conference as a proud gold sponsor. Stop by our booth to chat with our experts, see demos of our latest research and find out about career opportunities\u00a0with Microsoft.","msr_research_lab":[],"related-researchers":[],"msr_impact_theme":[],"related-academic-programs":[],"related-groups":[],"related-projects":[],"related-opportunities":[],"related-publications":[],"related-videos":[],"related-posts":[606492,607386],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/605979","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-event"}],"version-history":[{"count":3,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/605979\/revisions"}],"predecessor-version":[{"id":1147026,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/605979\/revisions\/1147026"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/606480"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=605979"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=605979"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=605979"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=605979"},{"taxonomy":"msr-video-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video-type?post=605979"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=605979"},{"taxonomy":"msr-program-audience","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-program-audience?post=605979"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=605979"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=605979"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}