Kaizhao Liang (@kyleliang5) 's Twitter Profile
Kaizhao Liang

@kyleliang5

@UTCompSci @IllinoisCDS

ID: 1070931144512823296

linkhttps://kyleliang919.github.io calendar_today07-12-2018 06:40:56

3,3K Tweet

598 Takipçi

67 Takip Edilen

Dynamics Lab (@dynamicslab_ai) 's Twitter Profile Photo

💥💥BANG! Experience the future of gaming with our real-time world model for video games!🕹️🕹️ Not just PLAY—but CREATE! Introducing Mirage, the world’s first AI-native UGC game engine. Now featuring real-time playable demos of two games: 🏙️ GTA-style urban chaos 🏎️ Forza

clem 🤗 (@clementdelangue) 's Twitter Profile Photo

Every $ invested in open-source generates $2,000 of value. Might be orders of magnitude more for open-source AI! x.com/clementdelangu…

Kaizhao Liang (@kyleliang5) 's Twitter Profile Photo

More compute = more intelligence Very cool work scaling up the compute in attention itself arxiv.org/pdf/2306.02896 arxiv.org/html/2507.0275…

More compute = more intelligence
Very cool work scaling up the compute in attention itself  
arxiv.org/pdf/2306.02896
arxiv.org/html/2507.0275…
elie (@eliebakouch) 's Twitter Profile Photo

Super excited to share SmolLM3, a new strong 3B model. SmolLM3 is fully open, we share the recipe, the dataset, the training codebase and much more! > Train on 11T token on 384 H100 for 220k GPU hours > Support long context up to 128k thanks to NoPE and intra document masking >

Super excited to share SmolLM3, a new strong 3B model.

SmolLM3 is fully open, we share the recipe, the dataset, the training codebase and much more!

> Train on 11T token on 384 H100 for 220k GPU hours
> Support long context up to 128k thanks to NoPE and intra document masking
>
Maxime Labonne (@maximelabonne) 's Twitter Profile Photo

Liquid AI open-sources a new generation of edge LLMs! 🥳 I'm so happy to contribute to the open-source community with this release on Hugging Face! LFM2 is a new architecture that combines best-in-class inference speed and quality into 350M, 700M, and 1.2B models.

Liquid AI open-sources a new generation of edge LLMs! 🥳

I'm so happy to contribute to the open-source community with this release on <a href="/huggingface/">Hugging Face</a>! 

LFM2 is a new architecture that combines best-in-class inference speed and quality into 350M, 700M, and 1.2B models.
Tri Dao (@tri_dao) 's Twitter Profile Photo

I really like Phil Tillet's framing of different tools having different tradeoffs in productivity and performance: torch compile, triton, CUDA, PTX. It's still early but CuTe-DSL and similar Python-based DSL might bend this curve. And soon we can probably get LLMs to generate

I really like Phil Tillet's framing of different tools having different tradeoffs in productivity and performance: torch compile, triton, CUDA, PTX. It's still early but CuTe-DSL and similar Python-based DSL might bend this curve. And soon we can probably get LLMs to generate
Kaizhao Liang (@kyleliang5) 's Twitter Profile Photo

RL scaling gives quantum bubble sort vibe. Spawning parallel universes of all possible order of a string and then destroy the universes in which the string is unsorted( verifiable reward)

RL scaling gives quantum bubble sort vibe. Spawning parallel universes of all possible order of a string and then destroy the universes in which the string is unsorted( verifiable reward)
Kimi.ai (@kimi_moonshot) 's Twitter Profile Photo

🚀 Hello, Kimi K2! Open-Source Agentic Model! 🔹 1T total / 32B active MoE model 🔹 SOTA on SWE Bench Verified, Tau2 & AceBench among open models 🔹Strong in coding and agentic tasks 🐤 Multimodal & thought-mode not supported for now With Kimi K2, advanced agentic intelligence

🚀 Hello, Kimi K2!  Open-Source Agentic Model!
🔹 1T total / 32B active MoE model
🔹 SOTA on SWE Bench Verified, Tau2 &amp; AceBench among open models
🔹Strong in coding and agentic tasks
🐤 Multimodal &amp; thought-mode not supported for now

With Kimi K2, advanced agentic intelligence
Cartesia (@cartesia_ai) 's Twitter Profile Photo

We're excited to announce a new research release from the Cartesia team, as part of a long-term collaboration to advance deep learning architectures. We've always believed that model architectures remain a fundamental bottleneck in building truly intelligent systems. H-Nets are

Tri Dao (@tri_dao) 's Twitter Profile Photo

I played w it for 1h. Went through my usual prompts (math derivations, floating point optimizations, …). It’s a good model, feels comparable to the best frontier models

Danny Limanseta (@dannylimanseta) 's Twitter Profile Photo

I took Grok 4 for a spin this weekend to build this game prototype. I used SuperGrok Chat to generate the initial game prototype and then brought it over to Cursor to continue coding with Grok 4 MAX. Grok 4 in Cursor is like a no-nonsense agent. Doesn't speak much, but

JingyuanLiu (@jingyuanliu123) 's Twitter Profile Photo

can not agree more! simple translation: open source requires higher tech standard and forced the developers to deliver better models more aligned with AGI for those who do not know, moonshot used to be closed sourced and embraced open source just half year ago.

can not agree more! simple translation: open source requires higher tech standard and forced the developers to deliver better models more aligned with AGI

for those who do not know, moonshot used to be closed sourced and embraced open source just half year ago.
Elon Musk (@elonmusk) 's Twitter Profile Photo

We are creating a multi-agent AI software company @xAI, where @Grok spawns hundreds of specialized coding and image/video generation/understanding agents all working together and then emulates humans interacting with the software in virtual machines until the result is excellent.

Artificial Analysis (@artificialanlys) 's Twitter Profile Photo

While Moonshot AI’s Kimi k2 is the leading open weights non-reasoning model in the Artificial Analysis Intelligence Index, it outputs ~3x more tokens than other non-reasoning models, blurring the lines between reasoning & non-reasoning Kimi k2 is the largest major open weights

While Moonshot AI’s Kimi k2 is the leading open weights non-reasoning model in the Artificial Analysis Intelligence Index, it outputs ~3x more tokens than other non-reasoning models, blurring the lines between reasoning &amp; non-reasoning

Kimi k2 is the largest major open weights