Transformers: State-of-the-Art Natural Language Processing T Wolf arXiv preprint arXiv:1910.03771, 2020 | 6042 | 2020 |
Huggingface's transformers: State-of-the-art natural language processing T Wolf arXiv preprint arXiv:1910.03771, 2019 | 3565 | 2019 |
Bloom: A 176b-parameter open-access multilingual language model T Le Scao, A Fan, C Akiki, E Pavlick, S Ilić, D Hesslow, R Castagné, ... | 1748 | 2023 |
XLS-R: Self-supervised cross-lingual speech representation learning at scale A Babu, C Wang, A Tjandra, K Lakhotia, Q Xu, N Goyal, K Singh, ... arXiv preprint arXiv:2111.09296, 2021 | 720 | 2021 |
Diffusers: State-of-the-art diffusion models P Von Platen, S Patil, A Lozhkov, P Cuenca, N Lambert, K Rasul, ... | 443 | 2022 |
Datasets: A community library for natural language processing Q Lhoest, AV Del Moral, Y Jernite, A Thakur, P Von Platen, S Patil, ... arXiv preprint arXiv:2109.02846, 2021 | 301 | 2021 |
Lysandre Debut, Stas Bekman, Pierric Cistac, Thibault Goehringer, Victor Mustar, François Lagunas, Alexander Rush, and Thomas Wolf. 2021. Datasets: A community library for … Q Lhoest, AV Del Moral, Y Jernite, A Thakur, P Von Platen, S Patil, ... Proceedings of the 2021 Conference on Empirical Methods in Natural Language …, 2021 | 265 | 2021 |
Lcm-lora: A universal stable-diffusion acceleration module S Luo, Y Tan, S Patil, D Gu, P von Platen, A Passos, L Huang, J Li, ... arXiv preprint arXiv:2311.05556, 2023 | 113 | 2023 |
Huggingface’s transformers: State-of-the-art natural language processing. arXiv T Wolf, L Debut, V Sanh, J Chaumond, C Delangue, A Moi, P Cistac, ... arXiv preprint arXiv:1910.03771, 2019 | 92 | 2019 |
Distil-whisper: Robust knowledge distillation via large-scale pseudo labelling S Gandhi, P von Platen, AM Rush arXiv preprint arXiv:2311.00430, 2023 | 46 | 2023 |
How to generate text: using different decoding methods for language generation with Transformers P Von Platen Access mode: https://huggingface. co/blog/how-to-generate, 2020 | 43 | 2020 |
Esb: A benchmark for multi-domain end-to-end speech recognition S Gandhi, P Von Platen, AM Rush arXiv preprint arXiv:2210.13352, 2022 | 28 | 2022 |
Pixtral 12B P Agrawal, S Antoniak, EB Hanna, B Bout, D Chaplot, J Chudnovsky, ... arXiv preprint arXiv:2410.07073, 2024 | 25 | 2024 |
Multi-span acoustic modelling using raw waveform signals P von Platen, C Zhang, P Woodland arXiv preprint arXiv:1906.11047, 2019 | 23 | 2019 |
Xtreme-s: Evaluating cross-lingual speech representations A Conneau, A Bapna, Y Zhang, M Ma, P von Platen, A Lozhkov, C Cherry, ... arXiv preprint arXiv:2203.10752, 2022 | 22 | 2022 |
Progressive knowledge distillation of stable diffusion xl using layer level loss Y Gupta, VV Jaddipal, H Prabhala, S Paul, P Von Platen arXiv preprint arXiv:2401.02677, 2024 | 14 | 2024 |
amused: An open muse reproduction S Patil, W Berman, R Rombach, P von Platen arXiv preprint arXiv:2401.01808, 2024 | 10 | 2024 |
Huggingface’s transformers: State-of-the-art natural language processing. arXiv preprint arXiv: 191003771 T Wolf, L Debut, V Sanh, J Chaumond, C Delangue, A Moi, P Cistac, ... | 10 | 2019 |
Multi-task Siamese neural network for improving replay attack detection P von Platen, F Tao, G Tur arXiv preprint arXiv:2002.07629, 2020 | 8 | 2020 |
BLOOM: A 176B-Parameter Open-Access Multilingual Language Model V Danchev, V Nikoulina, V Laippala, V Lepercq, V Prabhu, Z Alyafeai, ... | | 2023 |