Thomas Chaton (@chaton_thomas) 's Twitter Profile
Thomas Chaton

@chaton_thomas

Research Enginering Manager at @PyTorchLightnin | @gridai_

ID: 1262371033480400899

calendar_today18-05-2020 13:14:50

91 Tweet

112 Followers

22 Following

TuringPost (@theturingpost) 's Twitter Profile Photo

Amazing PyTorch-lightning tutorials with code by Phillip Lippe. Tutorials cover many deep learning topics like transformers, energy-based models, GNNs, and more! We recommend the tutorials for everyone starting their DL journey: pytorch-lightning.readthedocs.io/en/latest/

Amazing PyTorch-lightning tutorials with code by <a href="/phillip_lippe/">Phillip Lippe</a>.

Tutorials cover many deep learning topics like transformers, energy-based models, GNNs, and more!

We recommend the tutorials for everyone starting their DL journey: pytorch-lightning.readthedocs.io/en/latest/
Marc Skov Madsen, PhD, CFA® (@marcskovmadsen) 's Twitter Profile Photo

I've created a template repo for easily using LIGHTNING.AI and Panel to build, scale and deploy powerful #machinelearning #dataapp in #python Check it out github.com/marcskovmadsen… #TheToolsYouKnowAndLove ❤️ ⚡New Handle: @LightningAI ⚡ PyTorch #dataviz #DataScience

Marc Skov Madsen, PhD, CFA® (@marcskovmadsen) 's Twitter Profile Photo

Would you like to be able to add a Panel #dataapp to your ⚡New Handle: @LightningAI ⚡ App? Check out my Live App: …xzp3npc9cr2dmthh.litng-ai-03.litng.ai/view/Home Repo: github.com/marcskovmadsen… PR to Lightning Docs: github.com/Lightning-AI/l… PyTorch #python #machinelearning #deeplearning

Would you like to be able to add a <a href="/Panel_org/">Panel</a> #dataapp to your <a href="/PyTorchLightnin/">⚡New Handle: @LightningAI ⚡</a> App?

Check out my 

Live App: …xzp3npc9cr2dmthh.litng-ai-03.litng.ai/view/Home
Repo: github.com/marcskovmadsen…
PR to Lightning Docs: github.com/Lightning-AI/l…

<a href="/PyTorch/">PyTorch</a> #python #machinelearning #deeplearning
Lightning AI ⚡️ (@lightningai) 's Twitter Profile Photo

With just a few days of work, you too can build an intelligent app that is: 📈 Scalable 💪 Performant & distributed 🎨 Fully customizable Learn how 👉👉👉 bit.ly/3fYUf6n #Diffusion #AI #ML #BuildWithLightning

William Falcon ⚡️ (@_willfalcon) 's Twitter Profile Photo

What if you could launch new products and AI startups in only a few days with Lightning AI ⚡️ and our new apps framework! This Lightning app for stable diffusion took 2 weeks to build with 2 engs. Shout out to Emad and Stability AI for providing awesome opensource models

Lightning AI ⚡️ (@lightningai) 's Twitter Profile Photo

Introducing Boltus: The God of AI ⚡ Binge 🍿 all four episodes in this 🧵! Let Boltus teach you how to deploy diffusion models at scale 👉👉 bit.ly/3YvdWE2 (1/4)

Lightning AI ⚡️ (@lightningai) 's Twitter Profile Photo

Progress update!🦙🔥🤓 Lit-LLaMA now implements the LLaMA-Adapter method for efficient fine-tuning 🔧⚡️ The core idea can be implemented in about 11 lines of code🤯 (see screenshot) Link to repo👉 github.com/Lightning-AI/l… Link to Adapter paper👉arxiv.org/abs/2303.16199

Progress update!🦙🔥🤓

Lit-LLaMA now implements the LLaMA-Adapter method for efficient fine-tuning 🔧⚡️

The core idea can be implemented in about 11 lines of code🤯 (see screenshot)

Link to repo👉 github.com/Lightning-AI/l…
Link to Adapter paper👉arxiv.org/abs/2303.16199
Tri Dao (@tri_dao) 's Twitter Profile Photo

Announcing FlashAttention-2! We released FlashAttention a year ago, making attn 2-4 faster and is now widely used in most LLM libraries. Recently I’ve been working on the next version: 2x faster than v1, 5-9x vs standard attn, reaching 225 TFLOPs/s training speed on A100. 1/

Announcing FlashAttention-2! We released FlashAttention a year ago, making attn 2-4 faster and is now widely used in most LLM libraries. Recently I’ve been working on the next version: 2x faster than v1, 5-9x vs standard attn, reaching 225 TFLOPs/s training speed on A100. 1/
Thomas Chaton (@chaton_thomas) 's Twitter Profile Photo

Learn how to scrape web data to finetune LLMs using Lightning AI Studio. Everything is included, the code, the data, the Python dependencies, etc.. You can start in seconds instead of hours. lightning.ai/lightning-ai/s…

Thomas Chaton (@chaton_thomas) 's Twitter Profile Photo

Prepare a 1 trillion token dataset to train LLMs from scratch in under 4 hours instead of days with Lightning AI ⚡️ Studio! Everything is included, the final datasets, the code, dependencies, etc... Get started in seconds as no setup is needed. lightning.ai/lightning-ai/s…

Linus (@thesephist) 's Twitter Profile Photo

A while ago I complained here about persistent storage in Google Colab. Have been using Lightning AI ⚡️ Studios for a while now for: - Full VSCode (incl. GH Copilot) - Persisted files shared across notebooks - Multi-GPU/node (!!) It's been great. Feels like a remote ML workstation

A while ago I complained here about persistent storage in Google Colab.

Have been using <a href="/LightningAI/">Lightning AI ⚡️</a> Studios for a while now for:
- Full VSCode (incl. GH Copilot)
- Persisted files shared across notebooks
- Multi-GPU/node (!!)

It's been great. Feels like a remote ML workstation
DeepSeek (@deepseek_ai) 's Twitter Profile Photo

🚀 Introducing NSA: A Hardware-Aligned and Natively Trainable Sparse Attention mechanism for ultra-fast long-context training & inference! Core components of NSA: • Dynamic hierarchical sparse strategy • Coarse-grained token compression • Fine-grained token selection 💡 With

🚀 Introducing NSA: A Hardware-Aligned and Natively Trainable Sparse Attention mechanism for ultra-fast long-context training &amp; inference!

Core components of NSA:
• Dynamic hierarchical sparse strategy
• Coarse-grained token compression
• Fine-grained token selection

💡 With
NVIDIA AI Developer (@nvidiaaidev) 's Twitter Profile Photo

Introducing DeepSeek-R1 optimizations for Blackwell, delivering 25x more revenue at 20x lower cost per token, compared with NVIDIA H100 just four weeks ago. Fueled by TensorRT DeepSeek optimizations for our Blackwell architecture, including FP4 performance with state-of-the-art

Introducing DeepSeek-R1 optimizations for Blackwell, delivering 25x more revenue at 20x lower cost per token, compared with NVIDIA H100 just four weeks ago.

Fueled by TensorRT DeepSeek optimizations for our Blackwell architecture, including FP4 performance with state-of-the-art
Dan Biderman (@dan_biderman) 's Twitter Profile Photo

How can we use small LLMs to shift more AI workloads onto our laptops and phones? In our paper and open-source code, we pair on-device LLMs (ollama) with frontier LLMs in the cloud (@openai, @together), to solve token-intensive workloads on your 💻 at 17.5% of the cloud cost