Follow
Thomas Wolf
Thomas Wolf
Co-founder at HuggingFace
Verified email at polytechnique.edu - Homepage
Title
Cited by
Cited by
Year
Transformers: State-of-the-art natural language processing
T Wolf, L Debut, V Sanh, J Chaumond, C Delangue, A Moi, P Cistac, ...
Proceedings of the 2020 conference on empirical methods in natural language …, 2020
11322*2020
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
V Sanh, L Debut, J Chaumond, T Wolf
arXiv preprint arXiv:1910.01108, 2019
5448*2019
Multitask prompted training enables zero-shot task generalization
V Sanh, A Webson, C Raffel, SH Bach, L Sutawika, Z Alyafeai, A Chaffin, ...
arXiv preprint arXiv:2110.08207, 2021
8402021
Bloom: A 176b-parameter open-access multilingual language model
BS Workshop, TL Scao, A Fan, C Akiki, E Pavlick, S Ilić, D Hesslow, ...
arXiv preprint arXiv:2211.05100, 2022
7132022
Transfer learning in natural language processing
S Ruder, ME Peters, S Swayamdipta, T Wolf
Proceedings of the 2019 conference of the North American chapter of the …, 2019
6052019
Transfertransfo: A transfer learning approach for neural network based conversational agents
T Wolf, V Sanh, J Chaumond, C Delangue
arXiv preprint arXiv:1901.08149, 2019
4622019
Datasets: A community library for natural language processing
Q Lhoest, AV del Moral, Y Jernite, A Thakur, P von Platen, S Patil, ...
arXiv preprint arXiv:2109.02846, 2021
350*2021
Two-dimensional superconductivity at a Mott insulator/band insulator interface LaTiO3/SrTiO3
J Biscaras, N Bergeal, A Kushwaha, T Wolf, A Rastogi, RC Budhani, ...
Nature communications 1 (1), 89, 2010
3292010
Movement pruning: Adaptive sparsity by fine-tuning
V Sanh, T Wolf, A Rush
Advances in Neural Information Processing Systems 33, 20378-20389, 2020
2882020
A hierarchical multi-task approach for learning embeddings from semantic tasks
V Sanh, T Wolf, S Ruder
Proceedings of the AAAI Conference on Artificial Intelligence 33 (01), 6949-6956, 2019
2482019
Natural language processing with transformers
L Tunstall, L Von Werra, T Wolf
" O'Reilly Media, Inc.", 2022
1592022
StarCoder: may the source be with you!
R Li, LB Allal, Y Zi, N Muennighoff, D Kocetkov, C Mou, M Marone, C Akiki, ...
arXiv preprint arXiv:2305.06161, 2023
127*2023
Diffusers: State-of-the-art diffusion models
P von Platen, S Patil, A Lozhkov, P Cuenca, N Lambert, K Rasul, ...
1132022
Large-scale transfer learning for natural language generation
S Golovanov, R Kurbanov, S Nikolenko, K Truskovskyi, A Tselousov, ...
Proceedings of the 57th Annual Meeting of the Association for Computational …, 2019
932019
Learning from others' mistakes: Avoiding dataset biases without modeling them
V Sanh, T Wolf, Y Belinkov, AM Rush
arXiv preprint arXiv:2012.01300, 2020
772020
Huggingface’s transformers: State-of-the-art natural language processing. arXiv
T Wolf, L Debut, V Sanh, J Chaumond, C Delangue, A Moi, P Cistac, ...
arXiv preprint arXiv:1910.03771, 2019
572019
Strong field-matching effects in superconducting YBaCuO films with vortex energy landscapes engineered via masked ion irradiation
I Swiecicki, C Ulysse, T Wolf, R Bernard, N Bergeal, J Briatico, G Faini, ...
Physical Review B 85 (22), 224502, 2012
572012
Vimpac: Video pre-training via masked token prediction and contrastive learning
H Tan, J Lei, T Wolf, M Bansal
arXiv preprint arXiv:2106.11250, 2021
522021
& Rush, AM (2019). Huggingface's transformers: State-of-the-art natural language processing
T Wolf, L Debut, V Sanh, J Chaumond, C Delangue, A Moi
arXiv preprint arXiv:1910.03771, 1910
521910
The stack: 3 tb of permissively licensed source code
D Kocetkov, R Li, LB Allal, J Li, C Mou, CM Ferrandis, Y Jernite, M Mitchell, ...
arXiv preprint arXiv:2211.15533, 2022
452022
The system can't perform the operation now. Try again later.
Articles 1–20