Urmăriți
Shubham Toshniwal
Shubham Toshniwal
Senior Research Scientist, NVIDIA
Adresă de e-mail confirmată pe nvidia.com - Pagina de pornire
Titlu
Citat de
Citat de
Anul
Beyond the imitation game: Quantifying and extrapolating the capabilities of language models
A Srivastava, A Rastogi, A Rao, AAM Shoeb, A Abid, A Fisch, AR Brown, ...
TMLR, 2023
11862023
Multilingual speech recognition with a single end-to-end model
S Toshniwal, TN Sainath, RJ Weiss, B Li, P Moreno, E Weinstein, K Rao
ICASSP 2018, 2018
2982018
Lingvo: a modular and scalable framework for sequence-to-sequence modeling
J Shen, P Nguyen, Y Wu, Z Chen, MX Chen, Y Jia, A Kannan, T Sainath, ...
arXiv preprint arXiv:1902.08295, 2019
2142019
A comparison of techniques for language model integration in encoder-decoder speech recognition
S Toshniwal, A Kannan, CC Chiu, Y Wu, TN Sainath, K Livescu
SLT 2018, 2018
1942018
Multitask learning with low-level auxiliary tasks for encoder-decoder based speech recognition
S Toshniwal, H Tang, L Lu, K Livescu
Interspeech 2017, 2017
1322017
Pre-Trained Text Embeddings for Enhanced Text-to-Speech Synthesis
T Hayashi, S Watanabe, T Toda, K Takeda, S Toshniwal, K Livescu
Interspeech 2019, 2019
912019
Parsing speech: a neural approach to integrating lexical and acoustic-prosodic information
T Tran, S Toshniwal, M Bansal, K Gimpel, K Livescu, M Ostendorf
NAACL 2018, 2017
85*2017
Learning to Ignore: Long Document Coreference with Bounded Memory Neural Networks
S Toshniwal, S Wiseman, A Ettinger, K Livescu, K Gimpel
EMNLP 2020, 2020
572020
Generating natural language dialog using a questions corpus
J Ajmera, AK Gupta, S Joshi, S Toshniwal
US Patent 10,049,152, 2018
572018
Jointly learning to align and convert graphemes to phonemes with neural attention models
S Toshniwal, K Livescu
SLT 2016, 2016
552016
Hierarchical multitask learning for ctc-based speech recognition
K Krishna, S Toshniwal, K Livescu
arXiv preprint arXiv:1807.06234, 2018
532018
Chess as a Testbed for Language Model State Tracking
S Toshniwal, S Wiseman, K Livescu, K Gimpel
AAAI 2022 36 (10), 11385-11393, 2022
46*2022
On Generalization in Coreference Resolution
S Toshniwal, P Xia, S Wiseman, K Livescu, K Gimpel
CRAC@EMNLP 2021, 2021
432021
A Cross-Task Analysis of Text Span Representations
S Toshniwal, H Shi, B Shi, L Gao, K Livescu, K Gimpel
RepL4NLP 2020, 2020
432020
Openmathinstruct-1: A 1.8 million math instruction tuning dataset
S Toshniwal, I Moshkov, S Narenthiran, D Gitman, F Jia, I Gitman
NeurIPS Datasets and Benchmark, 2024
422024
Nemotron-4 340B Technical Report
B Adler, N Agarwal, A Aithal, DH Anh, P Bhattacharya, A Brundyn, ...
arXiv preprint arXiv:2406.11704, 2024
392024
Adapting pretrained text-to-text models for long text sequences
W Xiong, A Gupta, S Toshniwal, Y Mehdad, W Yih
Findings of EMNLP 2023, 2023
262023
Learning to reason and memorize with self-notes
J Lanchantin, S Toshniwal, J Weston, S Sukhbaatar
NeurIPS 2023, 2023
232023
VibRein: an engaging and assistive mobile learning companion for students with intellectual disabilities
S Toshniwal, P Dey, N Rajput, S Srivastava
Proceedings of the annual meeting of the Australian special interest group …, 2015
162015
Code Pretraining Improves Entity Tracking Abilities of Language Models
N Kim, S Schuster, S Toshniwal
arXiv preprint arXiv:2405.21068, 2024
82024
Sistemul nu poate realiza operația în acest moment. Încercați din nou mai târziu.
Articole 1–20