Clem Delangue 🤗 (@clemdelangue) 's Twitter Profile
Clem Delangue 🤗

@clemdelangue

Co-founder & CEO at Hugging Face 🤗. We teach computers to understand human language.

ID: 1049442839346724864

linkhttp://www.huggingface.co calendar_today08-10-2018 23:34:05

68 Tweet

1,1K Followers

5 Following

Clem Delangue 🤗 (@clemdelangue) 's Twitter Profile Photo

Best Long Paper #naacl2019 BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Jacob Devlin, Ming-Wei Chang, Kenton Lee and Kristina Toutanova #NLProc

Best Long Paper #naacl2019
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.
Jacob Devlin, Ming-Wei Chang, Kenton Lee and Kristina Toutanova #NLProc
Julien Chaumond (@julien_c) 's Twitter Profile Photo

🔥 Thrilled to release our Swift Core ML implementation of BERT for question answering.🔥🔥 Transformers models now also live on the edge. 📱📲 You now CAN do state-of-the-art NLP on mobile devices! github.com/huggingface/sw… Built w/ Lysandre and Thomas Wolf at Hugging Face

Thomas Wolf (@thom_wolf) 's Twitter Profile Photo

New release of Transformers repo is shaping up & I'm very excited! Gifts for all: -SOTA Lovers: new XLNet & XLM archi + 6 new Bert/GPT trained chkpt -Research Lovers: unified model API, attention/hidden-state outputs to swap/study models -Speed Lovers: Torchscript & head pruning!

New release of Transformers repo is shaping up & I'm very excited!
Gifts for all:
-SOTA Lovers: new XLNet & XLM archi + 6 new Bert/GPT trained chkpt
-Research Lovers: unified model API, attention/hidden-state outputs to swap/study models
-Speed Lovers: Torchscript & head pruning!
Thomas Wolf (@thom_wolf) 's Twitter Profile Photo

A question I get from time to time is how to convert a pretrained TensorFlow model in PyTorch easily and reliably. We're starting to be quite familiar with the process so I've written a short blog post summarizing our workflow and some lessons learned 👇 medium.com/huggingface/fr…

Hugging Face (@huggingface) 's Twitter Profile Photo

💃PyTorch-Transformers 1.1.0 is live💃 It includes RoBERTa, the transformer model from @facebookai, current state-of-the-art on the SuperGLUE leaderboard! Thanks to Myle Ott Julien Chaumond Lysandre and all the 100+ contributors!

💃PyTorch-Transformers 1.1.0 is live💃

It includes RoBERTa, the transformer model from @facebookai, current state-of-the-art on the SuperGLUE leaderboard! Thanks to <a href="/myleott/">Myle Ott</a> <a href="/julien_c/">Julien Chaumond</a> <a href="/LysandreJik/">Lysandre</a> and all the 100+ contributors!
Kyosuke Nishida (@kyoun) 's Twitter Profile Photo

DistilBERT (huggingface) BERT baseから蒸留にて6層に小型化(40%減)。推論は60%高速化、精度はGLUEで95%程度保持。8個の16GB V100 GPUで3.5日ぐらいで学習。hidden sizeは768のままで、層数の方が高速化には効果があるとのこと。github github.com/huggingface/py… blog medium.com/huggingface/di…

Julien Chaumond (@julien_c) 's Twitter Profile Photo

GPT-2 on device is blazing fast on iPhone 11 ⚡️ Core ML 3 is officially out so we can do state-of-the-art text generation on mobile (117M parameters running ~3 times par second on the neural engine!) We put together a small video benchmark ⬇️

Timothy Liu (@timothy_lkh_) 's Twitter Profile Photo

Happy to have a small PR accepted to the HuggingFace Transformer library demonstrating substantial mixed precision speed-up with @NVIDIA Tensor Core #GPU even at small batch size in the demo script github.com/huggingface/tr…

Thomas Wolf (@thom_wolf) 's Twitter Profile Photo

The @SustaiNLP2020 workshop at #EMNLP2020 will try to remove a little bit of SOTA addiction from NLP research 😉 We'll promote sensible trade-offs between performances & models that are - computationally more efficient - conceptually simpler ... [1/2] x.com/DoingJobs/stat…

chansung (@algo_diver) 's Twitter Profile Photo

Some more results. Now I made it fully supported by all kinds of model and vocabs. Good experience to use Hugging Face with Slack. And it looks pretty smart

Some more results. Now I made it fully supported by all kinds of model and vocabs. 

Good experience to use <a href="/huggingface/">Hugging Face</a> with <a href="/SlackHQ/">Slack</a>. And it looks pretty smart
Thomas Wolf (@thom_wolf) 's Twitter Profile Photo

Interesting work (and a nice large and clean dataset as well, looking forward to see it released): "Compressive Transformers for Long-Range Sequence Modelling" by Jack W. Rae, Anna Potapenko, Siddhant M. Jayakumar, Timothy P. Lillicrap (at DeepMind) Paper: arxiv.org/abs/1911.05507

Interesting work (and a nice large and clean dataset as well, looking forward to see it released):
"Compressive Transformers for Long-Range Sequence Modelling"
by Jack W. Rae, Anna Potapenko, Siddhant M. Jayakumar, Timothy P. Lillicrap (at DeepMind)
Paper: arxiv.org/abs/1911.05507
Soumith Chintala (@soumithchintala) 's Twitter Profile Photo

The first full paper on @pytorch after 3 years of development. It describes our goals, design principles, technical details uptil v0.4 Catch the poster at #NeurIPS2019 Authored by Adam Paszke , Sam Gross et. al. arxiv.org/abs/1912.01703