Sylvain Kalache (@sylvainkalache) 's Twitter Profile
Sylvain Kalache

@sylvainkalache

Leading the AI Lab and developer relation @RootlyHQ

Former @HolbertonSchool co-founder, @LinkedIn & @Slideshare SRE

ID: 3153611

linkhttps://www.sylvainkalache.com/ calendar_today01-04-2007 20:31:51

12,12K Tweet

13,13K Followers

1,1K Following

Rootly (@rootlyhq) 's Twitter Profile Photo

“AI stands for Angry Intern,” says Aaron Erickson (NVIDIA). It’s fast, ambitious, and constantly pushing to ship (whether it’s ready or not). Platform engineers are on the front line as production requests surge, even from outside technical teams. The question: how do you keep

Flo Crivello (@altimor) 's Twitter Profile Photo

Announcing a new chapter for AI agents: Lindy 3.0. Our vision has always been the "Al employee": as capable as humans (can do anything on a computer) and as easy to use (just ask). 3.0 takes 3 giant steps in this direction, with Agent Builder, Autopilot, and Team Collaboration.

Rootly (@rootlyhq) 's Twitter Profile Photo

Gemini CLI has surpassed 60K stars on GitHub in under two months. Joana Carrasqueira, Head of DevRel at Google DeepMind, shares why developers are adding it to their terminals nonstop.

Rootly (@rootlyhq) 's Twitter Profile Photo

The Rootly SRE-oriented LLM benchmark was featured at ICML Conference & ACL 2025, and you can now run it with a single command through Groq Inc OpenBench. Groq open sourced OpenBench so the community could evaluate models in standardized, reproducible ways. Today it supports 18

Aarush Sah (@aarushsah_) 's Twitter Profile Photo

6/ 🚒 Rootly GMCQ: Real-world SRE Benchmark Thanks to Rootly, we’ve added Rootly GMCQ to OpenBench. This benchmark is specifically designed to test real-world SRE tasks, such as incident triage, log analysis, and outage mitigation. This is the first-open-sourcing of an eval

The New Stack (@thenewstack) 's Twitter Profile Photo

By automating complex tasks, platform teams can significantly accelerate their engineering team maturity. By Sylvain Kalache thenewstack.io/ai-first-platf…

Rootly (@rootlyhq) 's Twitter Profile Photo

As AI code assistants become core to the development lifecycle, Robert Zuber (CTO at CircleCI) asks: should we still measure “good code” the same way? Or should machine-written code follow a different standard than human-written code? In this episode of Humans of Reliability,

Rootly (@rootlyhq) 's Twitter Profile Photo

Join us next week at GitHub HQ in San Francisco to explore where AI meets reliability, guided by engineering leaders shaping the future – Rama Akkiraju (NVIDIA, VP of AI/ML for IT) – Renaud Gaubert (OpenAI, Member of Technical Staff) – Colin McGrath (Baseten, VP of

Rootly (@rootlyhq) 's Twitter Profile Photo

Agentic AI is moving fast. The energy feels a lot like the early cloud or even the early web. Everyone is racing to ship and stake a claim in the agent-driven future. At our recent San Francisco panel with @AarushSah_h (Head of Evals at Groq Inc), Gregg Mojica (Co-Founder & CEO

Rootly (@rootlyhq) 's Twitter Profile Photo

LLMs can easily become the least reliable part of your stack. Ertan Dogrultan, Director of Engineering at Replit ⠕, breaks down the challenges platform engineers face when delivering a seamless user experience, from juggling multiple model providers to ensuring committed

Tiarnán de Burca (@nycdubliner) 's Twitter Profile Photo

Sure, but where am I going to find some mundane problems! Great talk by @Sylvainkalache on AI's impact on Code, and rates of failure.

Sure, but where am I going to find some mundane problems! 

Great talk by @Sylvainkalache on AI's impact on Code, and rates of failure.
Rootly (@rootlyhq) 's Twitter Profile Photo

While Sonnet-4.5 remains a popular choice among developers, our benchmarks show it underperforms GPT-5 on SRE-related tasks when both are run with default parameters. However, using the Not Diamond prompt adaptation platform, Sonnet-4.5 achieved up to a 2x performance