default search action
Jasha Droppo
Person information
- affiliation: Microsoft Research
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [c83]Haoyu Wang, Ruirui Li, Zhengyang Wang, Xianfeng Tang, Danqing Zhang, Monica Xiao Cheng, Bing Yin, Jasha Droppo, Suhang Wang, Jing Gao:
LightLT: A Lightweight Representation Quantization Framework for Long-Tail Data. ICDE 2024: 1380-1393 - 2023
- [c82]Milind Rao, Gopinath Chennupati, Gautam Tiwari, Anit Kumar Sahu, Anirudh Raju, Ariya Rastrow, Jasha Droppo:
Federated Self-Learning with Weak Supervision for Speech Recognition. ICASSP 2023: 1-5 - [c81]Kamil Deja, Georgi Tinchev, Marta Czarnowska, Marius Cotescu, Jasha Droppo:
Diffusion-based accent modelling in speech synthesis. INTERSPEECH 2023: 5516-5520 - [i28]Milind Rao, Gopinath Chennupati, Gautam Tiwari, Anit Kumar Sahu, Anirudh Raju, Ariya Rastrow, Jasha Droppo:
Federated Self-Learning with Weak Supervision for Speech Recognition. CoRR abs/2306.12015 (2023) - [i27]Guruprasad V. Ramesh, Gopinath Chennupati, Milind Rao, Anit Kumar Sahu, Ariya Rastrow, Jasha Droppo:
Federated Representation Learning for Automatic Speech Recognition. CoRR abs/2308.02013 (2023) - 2022
- [c80]Arman Zharmagambetov, Qingming Tang, Chieh-Chi Kao, Qin Zhang, Ming Sun, Viktor Rozgic, Jasha Droppo, Chao Wang:
Improved Representation Learning For Acoustic Event Classification Using Tree-Structured Ontology. ICASSP 2022: 321-325 - [c79]Hua Shen, Yuguang Yang, Guoli Sun, Ryan Langman, Eunjung Han, Jasha Droppo, Andreas Stolcke:
Improving Fairness in Speaker Verification via Group-Adapted Fusion Network. ICASSP 2022: 7077-7081 - [c78]Viet Anh Trinh, Pegah Ghahremani, Brian John King, Jasha Droppo, Andreas Stolcke, Roland Maas:
Reducing Geographic Disparities in Automatic Speech Recognition via Elastic Weight Consolidation. INTERSPEECH 2022: 1298-1302 - [c77]Minho Jin, Chelsea Ju, Zeya Chen, Yi-Chieh Liu, Jasha Droppo, Andreas Stolcke:
Adversarial Reweighting for Speaker Verification Fairness. INTERSPEECH 2022: 4800-4804 - [c76]Gokce Keskin, Minhua Wu, Brian John King, Sri Harish Mallidi, Yang Gao, Jasha Droppo, Ariya Rastrow, Roland Maas:
Do You Listen with one or two Microphones? A Unified ASR Model for Single and Multi-Channel Audio. IWAENC 2022: 1-5 - [c75]Gopinath Chennupati, Milind Rao, Gurpreet Chadha, Aaron Eakin, Anirudh Raju, Gautam Tiwari, Anit Kumar Sahu, Ariya Rastrow, Jasha Droppo, Andy Oberlin, Buddha Nandanoor, Prahalad Venkataramanan, Zheng Wu, Pankaj Sitpure:
ILASR: Privacy-Preserving Incremental Learning for Automatic Speech Recognition at Production Scale. KDD 2022: 2780-2788 - [c74]Aparna Khare, Minhua Wu, Saurabhchand Bhati, Jasha Droppo, Roland Maas:
Guided Contrastive Self-Supervised Pre-Training for Automatic Speech Recognition. SLT 2022: 174-181 - [i26]Hua Shen, Yuguang Yang, Guoli Sun, Ryan Langman, Eunjung Han, Jasha Droppo, Andreas Stolcke:
Improving fairness in speaker verification via Group-adapted Fusion Network. CoRR abs/2202.11323 (2022) - [i25]Minho Jin, Chelsea J.-T. Ju, Zeya Chen, Yi-Chieh Liu, Jasha Droppo, Andreas Stolcke:
Adversarial Reweighting for Speaker Verification Fairness. CoRR abs/2207.07776 (2022) - [i24]Viet Anh Trinh, Pegah Ghahremani, Brian John King, Jasha Droppo, Andreas Stolcke, Roland Maas:
Reducing Geographic Disparities in Automatic Speech Recognition via Elastic Weight Consolidation. CoRR abs/2207.07850 (2022) - [i23]Gopinath Chennupati, Milind Rao, Gurpreet Chadha, Aaron Eakin, Anirudh Raju, Gautam Tiwari, Anit Kumar Sahu, Ariya Rastrow, Jasha Droppo, Andy Oberlin, Buddha Nandanoor, Prahalad Venkataramanan, Zheng Wu, Pankaj Sitpure:
ILASR: Privacy-Preserving Incremental Learning for Automatic Speech Recognition at Production Scale. CoRR abs/2207.09078 (2022) - [i22]Aparna Khare, Minhua Wu, Saurabhchand Bhati, Jasha Droppo, Roland Maas:
Guided contrastive self-supervised pre-training for automatic speech recognition. CoRR abs/2210.12335 (2022) - [i21]Xin Zhang, Iván Vallés-Pérez, Andreas Stolcke, Chengzhu Yu, Jasha Droppo, Olabanji Shonibare, Roberto Barra-Chicote, Venkatesh Ravichandran:
Stutter-TTS: Controlled Synthesis and Improved Recognition of Stuttered Speech. CoRR abs/2211.09731 (2022) - 2021
- [c73]Yixin Chen, Weiyi Lu, Alejandro Mottini, Li Erran Li, Jasha Droppo, Zheng Du, Belinda Zeng:
Top-Down Attention in End-to-End Spoken Language Understanding. ICASSP 2021: 6199-6203 - [c72]Surabhi Punjabi, Harish Arsikere, Zeynab Raeesy, Chander Chandak, Nikhil Bhave, Ankish Bansal, Markus Müller, Sergio Murillo, Ariya Rastrow, Andreas Stolcke, Jasha Droppo, Sri Garimella, Roland Maas, Mat Hans, Athanasios Mouchtaris, Siegfried Kunzmann:
Joint ASR and Language Identification Using RNN-T: An Efficient Approach to Dynamic Language Switching. ICASSP 2021: 7218-7222 - [c71]Milind Rao, Pranav Dheram, Gautam Tiwari, Anirudh Raju, Jasha Droppo, Ariya Rastrow, Andreas Stolcke:
DO as I Mean, Not as I Say: Sequence Loss Training for Spoken Language Understanding. ICASSP 2021: 7473-7477 - [c70]Andrew Werchniak, Roberto Barra-Chicote, Yuriy Mishchenko, Jasha Droppo, Jeff Condal, Peng Liu, Anish Shah:
Exploring the application of synthetic audio in training keyword spotters. ICASSP 2021: 7993-7996 - [c69]Samik Sadhu, Di He, Che-Wei Huang, Sri Harish Mallidi, Minhua Wu, Ariya Rastrow, Andreas Stolcke, Jasha Droppo, Roland Maas:
wav2vec-C: A Self-Supervised Model for Speech Representation Learning. Interspeech 2021: 711-715 - [c68]Muhammad A. Shah, Joseph Szurley, Markus Müller, Athanasios Mouchtaris, Jasha Droppo:
Evaluating the Vulnerability of End-to-End Automatic Speech Recognition Models to Membership Inference Attacks. Interspeech 2021: 891-895 - [c67]Amin Fazel, Wei Yang, Yulan Liu, Roberto Barra-Chicote, Yixiong Meng, Roland Maas, Jasha Droppo:
SynthASR: Unlocking Synthetic Data for Speech Recognition. Interspeech 2021: 896-900 - [c66]Jie Pu, Yuguang Yang, Ruirui Li, Oguz Elibol, Jasha Droppo:
Scaling Effect of Self-Supervised Speech Models. Interspeech 2021: 1084-1088 - [c65]Jasha Droppo, Oguz Elibol:
Scaling Laws for Acoustic Models. Interspeech 2021: 2576-2580 - [c64]Iván Vallés-Pérez, Julian Roth, Grzegorz Beringer, Roberto Barra-Chicote, Jasha Droppo:
Improving Multi-Speaker TTS Prosody Variance with a Residual Encoder and Normalizing Flows. Interspeech 2021: 3131-3135 - [c63]Swayambhu Nath Ray, Minhua Wu, Anirudh Raju, Pegah Ghahremani, Raghavendra Bilgi, Milind Rao, Harish Arsikere, Ariya Rastrow, Andreas Stolcke, Jasha Droppo:
Listen with Intent: Improving Speech Recognition with Audio-to-Intent Front-End. Interspeech 2021: 3455-3459 - [c62]Daniel Korzekwa, Roberto Barra-Chicote, Szymon Zaporowski, Grzegorz Beringer, Jaime Lorenzo-Trueba, Alicja Serafinowicz, Jasha Droppo, Thomas Drugman, Bozena Kostek:
Detection of Lexical Stress Errors in Non-Native (L2) English with Data Augmentation and Attention. Interspeech 2021: 3915-3919 - [c61]Rupak Vignesh Swaminathan, Brian John King, Grant P. Strimel, Jasha Droppo, Athanasios Mouchtaris:
CoDERT: Distilling Encoder Representations with Co-Learning for Transducer-Based Speech Recognition. Interspeech 2021: 4543-4547 - [i20]Milind Rao, Pranav Dheram, Gautam Tiwari, Anirudh Raju, Jasha Droppo, Ariya Rastrow, Andreas Stolcke:
Do as I mean, not as I say: Sequence Loss Training for Spoken Language Understanding. CoRR abs/2102.06750 (2021) - [i19]Samik Sadhu, Di He, Che-Wei Huang, Sri Harish Mallidi, Minhua Wu, Ariya Rastrow, Andreas Stolcke, Jasha Droppo, Roland Maas:
Wav2vec-C: A Self-supervised Model for Speech Representation Learning. CoRR abs/2103.08393 (2021) - [i18]Bhargav Pulugundla, Yang Gao, Brian John King, Gokce Keskin, Sri Harish Mallidi, Minhua Wu, Jasha Droppo, Roland Maas:
Attention-based Neural Beamforming Layers for Multi-channel Speech Recognition. CoRR abs/2105.05920 (2021) - [i17]Swayambhu Nath Ray, Minhua Wu, Anirudh Raju, Pegah Ghahremani, Raghavendra Bilgi, Milind Rao, Harish Arsikere, Ariya Rastrow, Andreas Stolcke, Jasha Droppo:
Listen with Intent: Improving Speech Recognition with Audio-to-Intent Front-End. CoRR abs/2105.07071 (2021) - [i16]Gokce Keskin, Minhua Wu, Brian John King, Sri Harish Mallidi, Yang Gao, Jasha Droppo, Ariya Rastrow, Roland Maas:
Do You Listen with One or Two Microphones? A Unified ASR Model for Single and Multi-Channel Audio. CoRR abs/2106.02750 (2021) - [i15]Iván Vallés-Pérez, Julian Roth, Grzegorz Beringer, Roberto Barra-Chicote, Jasha Droppo:
Improving multi-speaker TTS prosody variance with a residual encoder and normalizing flows. CoRR abs/2106.05762 (2021) - [i14]Rupak Vignesh Swaminathan, Brian John King, Grant P. Strimel, Jasha Droppo, Athanasios Mouchtaris:
CoDERT: Distilling Encoder Representations with Co-learning for Transducer-based Speech Recognition. CoRR abs/2106.07734 (2021) - [i13]Amin Fazel, Wei Yang, Yulan Liu, Roberto Barra-Chicote, Yixiong Meng, Roland Maas, Jasha Droppo:
SynthASR: Unlocking Synthetic Data for Speech Recognition. CoRR abs/2106.07803 (2021) - [i12]Jasha Droppo, Oguz Elibol:
Scaling Laws for Acoustic Models. CoRR abs/2106.09488 (2021) - [i11]I-Fan Chen, Brian King, Jasha Droppo:
Investigation of Training Label Error Impact on RNN-T. CoRR abs/2112.00350 (2021) - 2020
- [c60]Jinxi Guo, Gautam Tiwari, Jasha Droppo, Maarten Van Segbroeck, Che-Wei Huang, Andreas Stolcke, Roland Maas:
Efficient Minimum Word Error Rate Training of RNN-Transducer for End-to-End Speech Recognition. INTERSPEECH 2020: 2807-2811 - [i10]Jinxi Guo, Gautam Tiwari, Jasha Droppo, Maarten Van Segbroeck, Che-Wei Huang, Andreas Stolcke, Roland Maas:
Efficient minimum word error rate training of RNN-Transducer for end-to-end speech recognition. CoRR abs/2007.13802 (2020) - [i9]Daniel Korzekwa, Roberto Barra-Chicote, Szymon Zaporowski, Grzegorz Beringer, Jaime Lorenzo-Trueba, Alicja Serafinowicz, Jasha Droppo, Thomas Drugman, Bozena Kostek:
Detection of Lexical Stress Errors in Non-native (L2) English with Data Augmentation and Attention. CoRR abs/2012.14788 (2020)
2010 – 2019
- 2019
- [c59]Xiong Xiao, Zhuo Chen, Takuya Yoshioka, Hakan Erdogan, Changliang Liu, Dimitrios Dimitriadis, Jasha Droppo, Yifan Gong:
Single-channel Speech Extraction Using Speaker Inventory and Attention Network. ICASSP 2019: 86-90 - 2018
- [j12]Zhehuai Chen, Jasha Droppo, Jinyu Li, Wayne Xiong:
Progressive Joint Modeling in Unsupervised Single-Channel Overlapped Speech Recognition. IEEE ACM Trans. Audio Speech Lang. Process. 26(1): 184-196 (2018) - [c58]Zhehuai Chen, Jasha Droppo:
Sequence Modeling in Unsupervised Single-Channel Overlapped Speech Recognition. ICASSP 2018: 4809-4813 - [c57]Wayne Xiong, Lingfeng Wu, Fil Alleva, Jasha Droppo, Xuedong Huang, Andreas Stolcke:
The Microsoft 2017 Conversational Speech Recognition System. ICASSP 2018: 5934-5938 - 2017
- [j11]Wayne Xiong, Jasha Droppo, Xuedong Huang, Frank Seide, Michael L. Seltzer, Andreas Stolcke, Dong Yu, Geoffrey Zweig:
Toward Human Parity in Conversational Speech Recognition. IEEE ACM Trans. Audio Speech Lang. Process. 25(12): 2410-2423 (2017) - [c56]Jinyu Li, Guoli Ye, Rui Zhao, Jasha Droppo, Yifan Gong:
Acoustic-to-word model without OOV. ASRU 2017: 111-117 - [c55]Geoffrey Zweig, Chengzhu Yu, Jasha Droppo, Andreas Stolcke:
Advances in all-neural speech recognition. ICASSP 2017: 4805-4809 - [c54]Wayne Xiong, Jasha Droppo, Xuedong Huang, Frank Seide, Mike Seltzer, Andreas Stolcke, Dong Yu, Geoffrey Zweig:
The microsoft 2016 conversational speech recognition system. ICASSP 2017: 5255-5259 - [c53]Andreas Stolcke, Jasha Droppo:
Comparing Human and Machine Errors in Conversational Speech Transcription. INTERSPEECH 2017: 137-141 - [i8]Zhehuai Chen, Jasha Droppo, Jinyu Li, Wayne Xiong:
Progressive Joint Modeling in Unsupervised Single-channel Overlapped Speech Recognition. CoRR abs/1707.07048 (2017) - [i7]Wayne Xiong, Lingfeng Wu, Fil Alleva, Jasha Droppo, Xuedong Huang, Andreas Stolcke:
The Microsoft 2017 Conversational Speech Recognition System. CoRR abs/1708.06073 (2017) - [i6]Andreas Stolcke, Jasha Droppo:
Comparing Human and Machine Errors in Conversational Speech Transcription. CoRR abs/1708.08615 (2017) - [i5]Jinyu Li, Guoli Ye, Rui Zhao, Jasha Droppo, Yifan Gong:
Acoustic-To-Word Model Without OOV. CoRR abs/1711.10136 (2017) - 2016
- [c52]Pegah Ghahremani, Jasha Droppo, Michael L. Seltzer:
Linearly augmented deep neural network. ICASSP 2016: 5085-5089 - [c51]Charith Mendis, Jasha Droppo, Saeed Maleki, Madanlal Musuvathi, Todd Mytkowicz, Geoffrey Zweig:
Parallelizing WFST speech decoders. ICASSP 2016: 5325-5329 - [c50]Tianxing He, Jasha Droppo:
Exploiting LSTM structure in deep neural networks for speech recognition. ICASSP 2016: 5445-5449 - [c49]Pegah Ghahremani, Jasha Droppo:
Self-stabilized deep neural network. ICASSP 2016: 5450-5454 - [c48]Dong Yu, Wayne Xiong, Jasha Droppo, Andreas Stolcke, Guoli Ye, Jinyu Li, Geoffrey Zweig:
Deep Convolutional Neural Networks with Layer-Wise Context Expansion and Attention. INTERSPEECH 2016: 17-21 - [c47]Tianxing He, Yu Zhang, Jasha Droppo, Kai Yu:
On training bi-directional neural network language model with noise contrastive estimation. ISCSLP 2016: 1-5 - [i4]Tianxing He, Yu Zhang, Jasha Droppo, Kai Yu:
On Training Bi-directional Neural Network Language Model with Noise Contrastive Estimation. CoRR abs/1602.06064 (2016) - [i3]Wayne Xiong, Jasha Droppo, Xuedong Huang, Frank Seide, Mike Seltzer, Andreas Stolcke, Dong Yu, Geoffrey Zweig:
The Microsoft 2016 Conversational Speech Recognition System. CoRR abs/1609.03528 (2016) - [i2]Geoffrey Zweig, Chengzhu Yu, Jasha Droppo, Andreas Stolcke:
Advances in All-Neural Speech Recognition. CoRR abs/1609.05935 (2016) - [i1]Wayne Xiong, Jasha Droppo, Xuedong Huang, Frank Seide, Mike Seltzer, Andreas Stolcke, Dong Yu, Geoffrey Zweig:
Achieving Human Parity in Conversational Speech Recognition. CoRR abs/1610.05256 (2016) - 2015
- [j10]Chao Weng, Dong Yu, Michael L. Seltzer, Jasha Droppo:
Deep Neural Networks for Single-Channel Multi-Talker Speech Recognition. IEEE ACM Trans. Audio Speech Lang. Process. 23(10): 1670-1679 (2015) - [c46]Abdel-rahman Mohamed, Frank Seide, Dong Yu, Jasha Droppo, Andreas Stolcke, Geoffrey Zweig, Gerald Penn:
Deep bi-directional recurrent networks over spectral windows. ASRU 2015: 78-83 - [c45]Yu Zhang, Dong Yu, Michael L. Seltzer, Jasha Droppo:
Speech recognition with prediction-adaptation-correction recurrent neural networks. ICASSP 2015: 5004-5008 - [c44]Ritwik Giri, Michael L. Seltzer, Jasha Droppo, Dong Yu:
Improving speech recognition in reverberation using a room-aware deep neural network and multi-task learning. ICASSP 2015: 5014-5018 - 2014
- [c43]Frank Seide, Hao Fu, Jasha Droppo, Gang Li, Dong Yu:
On parallelizability of stochastic gradient descent for speech DNNS. ICASSP 2014: 235-239 - [c42]Nicolas Boulanger-Lewandowski, Jasha Droppo, Mike Seltzer, Dong Yu:
Phone sequence modeling with recurrent neural networks. ICASSP 2014: 5417-5421 - [c41]Chao Weng, Dong Yu, Michael L. Seltzer, Jasha Droppo:
Single-channel mixed speech recognition using deep neural networks. ICASSP 2014: 5632-5636 - [c40]Frank Seide, Hao Fu, Jasha Droppo, Gang Li, Dong Yu:
1-bit stochastic gradient descent and its application to data-parallel distributed training of speech DNNs. INTERSPEECH 2014: 1058-1062 - [c39]Dong Yu, Adam Eversole, Michael L. Seltzer, Kaisheng Yao, Brian Guenter, Oleksii Kuchaiev, Frank Seide, Huaming Wang, Jasha Droppo, Zhiheng Huang, Geoffrey Zweig, Christopher J. Rossbach, Jon Currey:
An introduction to computational networks and the computational network toolkit (invited talk). INTERSPEECH 2014 - 2013
- [c38]Michael L. Seltzer, Jasha Droppo:
Multi-task learning in deep neural networks for improved phoneme recognition. ICASSP 2013: 6965-6969 - 2012
- [c37]Rohit Prabhavalkar, Jasha Droppo:
A chunk-based phonetic score for mobile voice search. ICASSP 2012: 4729-4732 - [p1]Jasha Droppo:
Feature Compensation. Techniques for Noise Robustness in Automatic Speech Recognition 2012: 229-250 - 2011
- [c36]Brian Hutchinson, Jasha Droppo:
Learning non-parametric models of pronunciation. ICASSP 2011: 4904-4907 - [c35]Xing Fan, Michael L. Seltzer, Jasha Droppo, Henrique S. Malvar, Alex Acero:
Joint encoding of the waveform and speech recognition features using a transform codec. ICASSP 2011: 5148-5151 - [c34]Yun-Cheng Ju, Jasha Droppo:
Automatically Optimizing Utterance Classification Performance without Human in the Loop. INTERSPEECH 2011: 721-724 - 2010
- [j9]Ozlem Kalinli, Michael L. Seltzer, Jasha Droppo, Alex Acero:
Noise Adaptive Training for Robust Automatic Speech Recognition. IEEE Trans. Speech Audio Process. 18(8): 1889-1901 (2010) - [c33]Jasha Droppo, Alex Acero:
Context dependent phonetic string edit distance for automatic speech recognition. ICASSP 2010: 4358-4361 - [c32]Xiaoqiang Xiao, Jasha Droppo, Alex Acero:
Information retrieval methods for automatic speech recognition. ICASSP 2010: 5550-5553 - [c31]Geoffrey Zweig, Patrick Nguyen, Jasha Droppo, Alex Acero:
Continuous speech recognition with a TF-IDF acoustic model. INTERSPEECH 2010: 2854-2857 - [c30]Yun-Cheng Ju, Jasha Droppo:
Spontaneous Mandarin speech understanding using Utterance Classification: A case study. ISCSLP 2010: 256-260
2000 – 2009
- 2009
- [c29]Jasha Droppo, Alex Acero:
Experimenting with a global decision tree for state clustering in automatic speech recognition systems. ICASSP 2009: 4437-4440 - 2008
- [j8]Dong Yu, Li Deng, Jasha Droppo, Jian Wu, Yifan Gong, Alex Acero:
Robust Speech Recognition Using a Cepstral Minimum-Mean-Square-Error-Motivated Noise Suppressor. IEEE Trans. Speech Audio Process. 16(5): 1061-1070 (2008) - [c28]Ivan Tashev, Jasha Droppo, Michael L. Seltzer, Alex Acero:
Robust design of wideband loudspeaker arrays. ICASSP 2008: 381-384 - [c27]Dong Yu, Li Deng, Jasha Droppo, Jian Wu, Yifan Gong, Alex Acero:
A minimum-mean-square-error noise reduction algorithm on Mel-frequency cepstra for robust speech recognition. ICASSP 2008: 4041-4044 - [c26]Luis Buera, Jasha Droppo, Alex Acero:
Speech enhancement using a pitch predictive model. ICASSP 2008: 4885-4888 - [c25]Jasha Droppo, Michael L. Seltzer, Alex Acero, Yu-Hsiang Bosco Chiu:
Towards a non-parametric acoustic model: an acoustic decision tree for observation probability calculation. INTERSPEECH 2008: 289-292 - 2007
- [c24]Christopher White, Jasha Droppo, Alex Acero, Julian Odell:
Maximum Entropy Confidence Estimation for Speech Recognition. ICASSP (4) 2007: 809-812 - [c23]Jasha Droppo, Alex Acero:
A fine pitch model for speech. INTERSPEECH 2007: 2757-2760 - 2006
- [c22]Jasha Droppo, Alex Acero:
Joint Discriminative Front End and Back End Training for Improved Speech Recognition Accuracy. ICASSP (1) 2006: 281-284 - 2005
- [j7]Li Deng, Jian Wu, Jasha Droppo, Alex Acero:
Analysis and comparison of two speech feature extraction/compensation algorithms. IEEE Signal Process. Lett. 12(6): 477-480 (2005) - [j6]Li Deng, Jasha Droppo, Alex Acero:
Dynamic compensation of HMM variances using the feature enhancement uncertainty computed from a parametric model of speech distortion. IEEE Trans. Speech Audio Process. 13(3): 412-421 (2005) - [c21]Zicheng Liu, Amar Subramanya, Zhengyou Zhang, Jasha Droppo, Alex Acero:
Leakage Model and Teeth Clack Removal for Air- and Bone-Conductive Integrated Microphones. ICASSP (1) 2005: 1093-1096 - [c20]Jasha Droppo, Alex Acero:
Maximum mutual information SPLICE transform for seen and unseen conditions. INTERSPEECH 2005: 989-992 - [c19]Michael L. Seltzer, Alex Acero, Jasha Droppo:
Robust bandwidth extension of noise-corrupted narrowband speech. INTERSPEECH 2005: 1509-1512 - [c18]Amarnag Subramanya, Zhengyou Zhang, Zicheng Liu, Jasha Droppo, Alex Acero:
A graphical model for multi-sensory speech processing in air-and-bone conductive microphones. INTERSPEECH 2005: 2361-2364 - 2004
- [j5]Li Deng, Jasha Droppo, Alex Acero:
Enhancement of log Mel power spectra of speech using a phase-sensitive model of the acoustic environment and sequential estimation of the corrupting noise. IEEE Trans. Speech Audio Process. 12(2): 133-143 (2004) - [j4]Li Deng, Jasha Droppo, Alex Acero:
Estimating cepstrum of speech under the presence of noise using a joint prior of static and dynamic features. IEEE Trans. Speech Audio Process. 12(3): 218-233 (2004) - [j3]Li Deng, Ye-Yi Wang, Kuansan Wang, Alex Acero, Hsiao-Wuen Hon, Jasha Droppo, Constantinos Boulis, Milind Mahajan, Xuedong Huang:
Speech and Language Processing for Multimodal Human-Computer Interaction. J. VLSI Signal Process. 36(2-3): 161-187 (2004) - [c17]Zhengyou Zhang, Zicheng Liu, Mike Sinclair, Alex Acero, Li Deng, Jasha Droppo, Xuedong Huang, Yanli Zheng:
Multi-sensory microphones for robust speech detection, enhancement and recognition. ICASSP (3) 2004: 781-784 - [c16]Jasha Droppo, Alex Acero:
Noise robust speech recognition with a switching linear dynamic model. ICASSP (1) 2004: 953-956 - [c15]Zicheng Liu, Zhengyou Zhang, Alejandro Acero, Jasha Droppo, Xuedong Huang:
Direct filtering for air- and bone-conductive microphones. MMSP 2004: 363-366 - 2003
- [j2]Li Deng, Jasha Droppo, Alex Acero:
Recursive estimation of nonstationary noise using iterative stochastic approximation for robust speech recognition. IEEE Trans. Speech Audio Process. 11(6): 568-580 (2003) - [c14]Li Deng, Jasha Droppo, Alex Acero:
Incremental Bayes learning with prior evolution for tracking nonstationary noise statistics from noisy speech data. ICASSP (1) 2003: 672-675 - [c13]Jasha Droppo, Li Deng, Alex Acero:
A comparison of three non-linear observation models for noisy speech features. INTERSPEECH 2003: 681-684 - [c12]Michael L. Seltzer, Jasha Droppo, Alex Acero:
A harmonic-model-based front end for robust speech recognition. INTERSPEECH 2003: 1277-1280 - 2002
- [j1]Li Deng, Kuansan Wang, Alex Acero, Hsiao-Wuen Hon, Jasha Droppo, Constantinos Boulis, Ye-Yi Wang, Derek Jacoby, Milind Mahajan, Ciprian Chelba, Xuedong Huang:
Distributed speech processing in miPad's multimodal user interface. IEEE Trans. Speech Audio Process. 10(8): 605-619 (2002) - [c11]Li Deng, Alex Acero, Ye-Yi Wang, Kuansan Wang, Hsiao-Wuen Hon, Jasha Droppo, Milind Mahajan, Xuedong Huang:
A speech-centric perspective for human-computer interface. IEEE Workshop on Multimedia Signal Processing 2002: 263-267 - [c10]Jasha Droppo, Alex Acero, Li Deng:
Uncertainty decoding with SPLICE for noise robust speech recognition. ICASSP 2002: 57-60 - [c9]Li Deng, Jasha Droppo, Alex Acero:
A Bayesian approach to speech feature enhancement using the dynamic cepstral prior. ICASSP 2002: 829-832 - [c8]Jasha Droppo, Li Deng, Alex Acero:
Evaluation of SPLICE on the Aurora 2 and 3 tasks. INTERSPEECH 2002: 29-32 - [c7]Jasha Droppo, Alex Acero, Li Deng:
Noise from corrupted speech log mel-spectral energies. INTERSPEECH 2002: 1569-1572 - [c6]Li Deng, Jasha Droppo, Alex Acero:
Sequential MAP noise estimation and a phase-sensitive model of the acoustic environment. INTERSPEECH 2002: 1813-1816 - [c5]Li Deng, Jasha Droppo, Alex Acero:
Exploiting variances in robust feature extraction based on a parametric model of speech distortion. INTERSPEECH 2002: 2449-2452 - 2001
- [c4]Xuedong Huang, Alex Acero, Ciprian Chelba, Li Deng, Jasha Droppo, Doug Duchene, Joshua Goodman, Hsiao-Wuen Hon, Derek Jacoby, Li Jiang, Ricky Loynd, Milind Mahajan, Peter Mau, Scott Meredith, Salman Mughal, Salvado Neto, Mike Plumpe, Kuansan Steury, Gina Venolia, Kuansan Wang, Ye-Yi Wang:
MiPad: a multimodal interaction prototype. ICASSP 2001: 9-12 - [c3]Jasha Droppo, Alex Acero, Li Deng:
Efficient on-line acoustic environment estimation for FCDCN in a continuous speech recognition system. ICASSP 2001: 209-212 - [c2]Li Deng, Alex Acero, Li Jiang, Jasha Droppo, Xuedong Huang:
High-performance robust speech recognition using stereo training data. ICASSP 2001: 301-304 - [c1]Jasha Droppo, Li Deng, Alex Acero:
Evaluation of the SPLICE algorithm on the Aurora2 database. INTERSPEECH 2001: 217-220
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-08-05 21:22 CEST by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint