m_ric (@aymericroucher) 's Twitter Profile
m_ric

@aymericroucher

Building Agents at Hugging Face 🤗

ID: 1454785961158746115

calendar_today31-10-2021 12:24:16

465 Tweet

5,5K Takipçi

217 Takip Edilen

不鍊金丹不坐禪 (@zzwz) 's Twitter Profile Photo

m_ric Andrew Ng Post: My little homework project (DeepWebSearch AgentKit App. Build with 🤗 Hugging Face smolagents framework) github.com/lwyBZss8924d/D…

m_ric (@aymericroucher) 's Twitter Profile Photo

Who said Transformers couldn't be good at forecasting? Datadog's new open model tops forecasting benchmarks! 💥 And boy did they cook. They followed the playbook to build the best model: 1. The best benchmark They release a new benchmark named BOOM, based on observability

Who said Transformers couldn't be good at forecasting? Datadog's new open model tops forecasting benchmarks! 💥

And boy did they cook. They followed the playbook to build the best model:

1. The best benchmark

They release a new benchmark named BOOM, based on observability
m_ric (@aymericroucher) 's Twitter Profile Photo

TIL: When distilling reasoning capability from a teacher LLM to a smaller LLM, you should use Agent traces instead of CoT traces. Advantages are: 1. Increased generalization Intuitively, this is because your agent can encounter more "surprising" results by interacting with its

TIL: When distilling reasoning capability from a teacher LLM to a smaller LLM, you should use Agent traces instead of CoT traces.

Advantages are:
1. Increased generalization
Intuitively, this is because your agent can encounter more "surprising" results by interacting with its
Allan (@niemerg) 's Twitter Profile Photo

Huge fan of Claude Code—so I built a python version using smolagents! Introducing SmolCC 🤖📟🛠️ An open source coding agent with Claude Code style tools (bash, grep, edit…✨) that can be easily customized.

Huge fan of Claude Code—so I built a python version using smolagents!

Introducing SmolCC 🤖📟🛠️

An open source coding agent with Claude Code style tools (bash, grep, edit…✨)  that can be easily customized.
m_ric (@aymericroucher) 's Twitter Profile Photo

If you didn't yet, you should read the technical report for SmolVLA, published yesterday by the Hugging Face robotics team! ➡️Amongst other ideas, it introduces "Async inference" to boost their robot actions. Robots have a problem: performing the actions takes time (Unlike

If you didn't yet, you should read the technical report for SmolVLA, published yesterday by the <a href="/huggingface/">Hugging Face</a> robotics team!
➡️Amongst other ideas, it introduces "Async inference" to boost their robot actions.

Robots have a problem: performing the actions takes time (Unlike
m_ric (@aymericroucher) 's Twitter Profile Photo

Qwen silently dropped the new standard for embeddings on the Hub! - 0.6B, 4B and 8B versions (probably would use only the 0.6B) - 32k context length 📏 - 100 languages 🌍 - SOTA on MTEB, but like real SOTA, with 10 points margin on the second bests 🤯

m_ric (@aymericroucher) 's Twitter Profile Photo

English Wikipedia is 29.4B characters, with ~1.5 bits of information per character. So that means, at 3.6 bits per parameter, a 12B model could memorize all english Wikipedia, it seems wild!

m_ric (@aymericroucher) 's Twitter Profile Photo

Almost didn't take time to celebrate : Smolagents just wooshed past 20k stars on Github! ⭐️ Congrats to the team, especially Albert Villanova who spends lots of time improving the library. And thanks a lot to our community members for pushing is forward! 👏 It's still time to

Almost didn't take time to celebrate : Smolagents just wooshed past 20k stars on Github! ⭐️

Congrats to the team, especially <a href="/avillanovamoral/">Albert Villanova</a>  who spends lots of time improving the library. And thanks a lot to our community members for pushing is forward! 👏

It's still time to