Patrick Devaney (@patrickbdevaney) 's Twitter Profile
Patrick Devaney

@patrickbdevaney

ID: 1595527208998690819

calendar_today23-11-2022 21:18:48

108 Tweet

82 Takipçi

175 Takip Edilen

Saurabh Kumar (@drummatick) 's Twitter Profile Photo

Literally a beast of a book. Emphasizes heavily on code and modern deep learning architectures Important concepts are highlighted so it’s easier to understand and focus.

Literally a beast of a book.
Emphasizes heavily on code and modern deep learning architectures 

Important concepts are highlighted so it’s easier to understand and focus.
v0 (@v0) 's Twitter Profile Photo

v0 can now: • Create and run full-stack Next.js and React applications • Create multiple files in one generation • Link and deploy to Vercel projects • Use Vercel project environment variables

𝚐𝔪𝟾𝚡𝚡𝟾 (@gm8xx8) 's Twitter Profile Photo

REDUCIO! Generating 1024×1024 Video within 16 Seconds using Extremely Compressed Motion Latents code: github.com/microsoft/Redu… paper: arxiv.org/abs/2411.13552

Mohit Mishra (@chessman786) 's Twitter Profile Photo

The Algorithm Design Manual - Practical approach - Real-world examples - Problem-solving strategies - Good book for someone trying to understand algorithms - It will require some understanding of any language. - Resources: github.com/mohitmishra786…

The Algorithm Design Manual
- Practical approach
- Real-world examples
- Problem-solving strategies
- Good book for someone trying to understand algorithms 
- It will require some understanding of any language.
- Resources: github.com/mohitmishra786…
Maxime Labonne (@maximelabonne) 's Twitter Profile Photo

📈 The State of Generative AI in the Enterprise Interesting report from Menlo Ventures that shows the evolution of Gen AI in companies from 2023 to 2024: • Uses cases: Code generation, chatbots, search, data, and meeting summarization are the top generative AI use cases in

📈 The State of Generative AI in the Enterprise

Interesting report from Menlo Ventures that shows the evolution of Gen AI in companies from 2023 to 2024:

• Uses cases: Code generation, chatbots, search, data, and meeting summarization are the top generative AI use cases in
Min Choi (@minchoi) 's Twitter Profile Photo

Less than 48 hours ago, DeepSeek AI from China just dropped their AI reasoning model. And it's on par with OpenAI o1-preview. Major shift. 10 examples (and how to try):

Less than 48 hours ago, DeepSeek AI from China just dropped their AI reasoning model.

And it's on par with OpenAI o1-preview.  Major shift.

10 examples (and how to try):
Andrew Ng (@andrewyng) 's Twitter Profile Photo

A small number of people are posting text online that’s intended for direct consumption not by humans, but by LLMs (large language models). I find this a fascinating trend, particularly when writers are incentivized to help LLM providers better serve their users! People who post

Eric Ciarla (hiring) (@ericciarla) 's Twitter Profile Photo

Introducing llms.txt Generator ✨ You can now concatenate any website into a single text file that can be fed into any LLM. We crawl the whole website with Firecrawl and extract data with gpt-4o-mini. Create your own llms.txt at llmstxt.firecrawl.dev!

Unsloth AI (@unslothai) 's Twitter Profile Photo

You can finetune Llama-3.2-Vision-11B for free on Colab now! Unsloth finetunes VLMs 2x faster, with 50% less VRAM, 6x longer context - with no accuracy loss. Documentation: docs.unsloth.ai GitHub: github.com/unslothai/unsl… Finetuning Colab: colab.research.google.com/drive/1j0N4XTY…

Daniel Han (@danielhanchen) 's Twitter Profile Photo

Vision finetuning is finally in🦥Unsloth AI! It took a while, but Llama 3.2 Vision, Pixtral, Qwen2 VL & all Llava variants now work! 1. QLoRA / LoRA is 1.3x to 2x faster for each 2. 30-70% less VRAM usage 3. 3 examples - Radiography, LaTeX, Q&A Extra stuff: 1. Pixtral chat

Vision finetuning is finally in🦥<a href="/UnslothAI/">Unsloth AI</a>! It took a while, but Llama 3.2 Vision, Pixtral, Qwen2 VL &amp; all Llava variants now work!

1. QLoRA / LoRA is 1.3x to 2x faster for each
2. 30-70% less VRAM usage
3. 3 examples - Radiography, LaTeX, Q&amp;A

Extra stuff:
1. Pixtral chat
Yaroslav Bulatov (@yaroslavvb) 's Twitter Profile Photo

Anyone who thinks you need 100k GPUs to make progress should watch Hannaneh Hajishirzi COLM keynote. Molmo appeared to beat Llama 3.2 in quality with same release day, all open-science on a 1k GPU cluster youtube.com/watch?v=qMTzor…

swarms (@swarms_corp) 's Twitter Profile Photo

Introducing an all-new suite of tools built on swarms - the production-grade framework for autonomous agent swarms ⎆ Documentation Intelligence ⎆ Cross-language Compilation ⎆ Multi-agent Architecture ⎆ Financial Enterprise Solutions Here's what our lead developer

Introducing an all-new suite of tools built on swarms - the production-grade framework for autonomous agent swarms

⎆ Documentation Intelligence
⎆ Cross-language Compilation
⎆ Multi-agent Architecture
⎆ Financial Enterprise Solutions

Here's what our lead developer
DailyPapers (@huggingpapers) 's Twitter Profile Photo

Distilling LLM Agents! 🧪 New work shows how to transfer the reasoning & task-solving power of large language model agents into smaller, more efficient models by cloning their tool-using behavior with retrieval and code!

Distilling LLM Agents! 🧪 New work shows how to transfer the reasoning &amp; task-solving power of large language model agents into smaller, more efficient models by cloning their tool-using behavior with retrieval and code!
Naksh Jain (@nakshsonigara) 's Twitter Profile Photo

Fractal, an Indian AI company, dropped Fathom-R1-14B open-source reasoning model that achieves performance comparable to o4-mini on math benchmarks within a 16K context window, trained for just $499. Built on top of DeepSeek-R1-Distill-Qwen-14B, It beats o3-mini-low.

Fractal, an Indian AI company, dropped Fathom-R1-14B open-source reasoning model that achieves performance comparable to o4-mini on math benchmarks within a 16K context window, trained for just $499.

Built on top of DeepSeek-R1-Distill-Qwen-14B, It beats o3-mini-low.
Miami AI Hub (@miamiaihub) 's Twitter Profile Photo

🚨 Speaker Alert! 🚨 We’re kicking off Le Robot Hackathon Miami (June 14-15) with an amazing panel featuring clem 🤗, Co-Founder & CEO of Hugging Face. Clem turned open-source AI into a global movement—now he’s jetting to the 305 to talk robotics, community, and why the

🚨 Speaker Alert! 🚨
We’re kicking off Le Robot Hackathon Miami (June 14-15) with an amazing panel featuring <a href="/ClementDelangue/">clem 🤗</a>, Co-Founder &amp; CEO of <a href="/huggingface/">Hugging Face</a>. Clem turned open-source AI into a global movement—now he’s jetting to the 305 to talk robotics, community, and why the
Mustafa Shukor (@mustafashukor1) 's Twitter Profile Photo

The Worldwide LeRobot hackathon is in 2 weeks, and we have been cooking something for you… Introducing SmolVLA, a Vision-Language-Action model with light-weight architecture, pretrained on community datasets, with an asynchronous inference stack, to control robots🧵

The Worldwide <a href="/LeRobotHF/">LeRobot</a>  hackathon is in 2 weeks, and we have been cooking something for you… 
Introducing SmolVLA, a Vision-Language-Action model with light-weight architecture, pretrained on community datasets, with an asynchronous inference stack, to control robots🧵
merve (@mervenoyann) 's Twitter Profile Photo

H Company released Holo-1: 3B and 7B GUI Action Vision Language Models for various web and computer agent tasks 🤗 Holo-1 has Apache 2.0 license and Hugging Face transformers support 🔥 more details in their blog post (next ⤵️)

H Company released Holo-1: 3B and 7B GUI Action Vision Language Models for various web and computer agent tasks 🤗

Holo-1 has Apache 2.0 license and <a href="/huggingface/">Hugging Face</a>  transformers support 🔥
more details in their blog post (next ⤵️)
Sam Rodriques (@sgrodriques) 's Twitter Profile Photo

Today we are releasing ether0, our first scientific reasoning model. We trained Mistral 24B with RL on several molecular design tasks in chemistry. Remarkably, we found that LLMs can learn some scientific tasks more much data-efficiently than specialized models trained from