fairseq(@fairseq) 's Twitter Profileg
fairseq

@fairseq

Sequence modeling toolkit for @PyTorch

ID:1257733655503437829

linkhttps://github.com/pytorch/fairseq/ calendar_today05-05-2020 18:07:55

12 Tweets

1,6K Followers

11 Following

Follow People
Mikel Artetxe(@artetxem) 's Twitter Profile Photo

We are releasing a family of dense and MoE language models with up to 13B and 1.1T parameters. We find that MoEs are more efficient, but the gap narrows at scale and varies greatly across domains and tasks.

Paper: arxiv.org/abs/2112.10684

Models & code: github.com/pytorch/fairse…

We are releasing a family of dense and MoE language models with up to 13B and 1.1T parameters. We find that MoEs are more efficient, but the gap narrows at scale and varies greatly across domains and tasks. Paper: arxiv.org/abs/2112.10684 Models & code: github.com/pytorch/fairse…
account_circle
fairseq(@fairseq) 's Twitter Profile Photo

Mixture of experts training in fairseq is now 40% faster thanks to Microsoft's Tutel library!
Blog: microsoft.com/en-us/research…
Fairseq code: github.com/pytorch/fairse…
Tutel code: github.com/microsoft/tutel

account_circle
AI at Meta(@AIatMeta) 's Twitter Profile Photo

We’re introducing GSLM, the first language model that breaks free completely of the dependence on text for training. This “textless NLP” approach learns to generate expressive speech using only raw audio recordings as input. Learn more and get the code:
ai.facebook.com/blog/textless-…

We’re introducing GSLM, the first language model that breaks free completely of the dependence on text for training. This “textless NLP” approach learns to generate expressive speech using only raw audio recordings as input. Learn more and get the code: ai.facebook.com/blog/textless-…
account_circle
fairseq(@fairseq) 's Twitter Profile Photo

fairseq now supports CPU offloading and full parameter+optimizer state sharding via fairscale's FullyShardedDataParallel module. See our tutorial to train a 13B parameter LM on 1 GPU: fb.me/fairseq_fsdp

account_circle
fairseq(@fairseq) 's Twitter Profile Photo

We just released 0.10.0, which is our last significant release before 1.0.0 when we will migrate to Hydra. Changelog: github.com/pytorch/fairse…

account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

Fairseq includes support for sequence to sequence learning for speech and audio recognition tasks, faster exploration and prototyping of new research ideas while offering a clear path to production. bit.ly/2WfP85X

account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

fairseq now supports the training of gated convolutional language models (arxiv.org/abs/1612.08083). It can train a Google Billion Word language model on 128 GPUs in less than a day.

account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

FairSeq Toolkit - Major Update
- Distributed Training
- Transformer models (big Transformer on WMT Eng-German in < 5 hours on DGX-1)
- Fast Inference: translations @ 92 sent/sec for big Transformer
- Story Generation
Read more at Michael Auli's post: facebook.com/photo.php?fbid…

FairSeq Toolkit - Major Update - Distributed Training - Transformer models (big Transformer on WMT Eng-German in < 5 hours on DGX-1) - Fast Inference: translations @ 92 sent/sec for big Transformer - Story Generation Read more at Michael Auli's post: facebook.com/photo.php?fbid…
account_circle
Yann LeCun(@ylecun) 's Twitter Profile Photo

Fairseq, now in PyTorch!
The open-source convolutional sequence-to-sequence engine from FAIR is now available in... fb.me/1gCPauX6V

account_circle