Mingfei Li (@mingfei_x) 's Twitter Profile
Mingfei Li

@mingfei_x

ID: 1682652688662802432

linkhttp://mingfei.io calendar_today22-07-2023 07:24:15

109 Tweet

54 Takipçi

128 Takip Edilen

rez0 (@rez0__) 's Twitter Profile Photo

The three areas of need are: 1. Agent Authentication & Authorization 2. Prompt Injection-related attacks 3. Secure Agent Architecture For #1, Agent authentication can likely be put into existing stuff like okta etc. Authorization is the really hard part and will need to be

Jim Fan (@drjimfan) 's Twitter Profile Photo

We are living in a timeline where a non-US company is keeping the original mission of OpenAI alive - truly open, frontier research that empowers all. It makes no sense. The most entertaining outcome is the most likely. DeepSeek-R1 not only open-sources a barrage of models but

We are living in a timeline where a non-US company is keeping the original mission of OpenAI alive - truly open, frontier research that empowers all. It makes no sense. The most entertaining outcome is the most likely.

DeepSeek-R1 not only open-sources a barrage of models but
Andrej Karpathy (@karpathy) 's Twitter Profile Photo

For friends of open source: imo the highest leverage thing you can do is help construct a high diversity of RL environments that help elicit LLM cognitive strategies. To build a gym of sorts. This is a highly parallelizable task, which favors a large community of collaborators.

Aran Komatsuzaki (@arankomatsuzaki) 's Twitter Profile Photo

Stanford presents: s1: Simple test-time scaling - Seeks the simplest approach to achieve test-time scaling and strong reasoning performance - Exceeds o1-preview on competition math questions by up to 27% (MATH and AIME24) - Model, data, and code are open-source

Stanford presents: 

s1: Simple test-time scaling

- Seeks the simplest approach to achieve test-time scaling and strong reasoning performance
- Exceeds o1-preview on competition math questions by up to 27% (MATH and AIME24)
- Model, data, and code are open-source
xlr8harder (@xlr8harder) 's Twitter Profile Photo

A funny thing about the deepseek-v3 training cost everyone is freaking out about is that they reported comparable training efficiency in the deepseek-v2 paper in May 2024. 172.8K hours * 14.8T tokens = 2.557M hours vs 2.788M hours

A funny thing about the deepseek-v3 training cost everyone is freaking out about is that they reported comparable training efficiency in the deepseek-v2 paper in May 2024. 

172.8K hours * 14.8T tokens = 2.557M hours vs 2.788M hours
Stefano Ermon (@stefanoermon) 's Twitter Profile Photo

Excited to share that I’ve been working on scaling up diffusion language models at Inception. A new generation of LLMs with unprecedented capabilities is coming!

Neel Nanda (@neelnanda5) 's Twitter Profile Photo

The best way to judge a forecaster is their track record. In 2021 Daniel Kokotajlo predicted o1-style models. I think we should all be very interested in the new predictions he's making in 2025! I've read it and highly recommend - it's thought provoking and stressfully plausible

The best way to judge a forecaster is their track record. In 2021 Daniel Kokotajlo predicted o1-style models. I think we should all be very interested in the new predictions he's making in 2025!

I've read it and highly recommend - it's thought provoking and stressfully plausible
Andrej Karpathy (@karpathy) 's Twitter Profile Photo

I attended a vibe coding hackathon recently and used the chance to build a web app (with auth, payments, deploy, etc.). I tinker but I am not a web dev by background, so besides the app, I was very interested in what it's like to vibe code a full web app today. As such, I wrote

I attended a vibe coding hackathon recently and used the chance to build a web app (with auth, payments, deploy, etc.). I tinker but I am not a web dev by background, so besides the app, I was very interested in what it's like to vibe code a full web app today. As such, I wrote
Percy Liang (@percyliang) 's Twitter Profile Photo

Wrapped up Stanford CS336 (Language Models from Scratch), taught with an amazing team Tatsunori Hashimoto Marcel Rød Neil Band Rohith Kuditipudi. Researchers are becoming detached from the technical details of how LMs work. In CS336, we try to fix that by having students build everything:

Danielle Baskin (@djbaskin) 's Twitter Profile Photo

SF's flyering game is incredible right now. There's 1000s of flyers around the city that are completely unrelated and all completely sincere

SF's flyering game is incredible right now. There's 1000s of flyers around the city that are completely unrelated and all completely sincere
Floor Eijkelboom (@feijkelboom) 's Twitter Profile Photo

Flow Matching (FM) is one of the hottest ideas in generative AI - and it’s everywhere at #ICML2025. But what is it? And why is it so elegant? 🤔 This thread is an animated, intuitive intro into (Variational) Flow Matching - no dense math required. Let's dive in! 🧵👇

Kevin Lu (@_kevinlu) 's Twitter Profile Photo

Why you should stop working on RL research and instead work on product // The technology that unlocked the big scaling shift in AI is the internet, not transformers I think it's well known that data is the most important thing in AI, and also that researchers choose not to work

Why you should stop working on RL research and instead work on product //
The technology that unlocked the big scaling shift in AI is the internet, not transformers

I think it's well known that data is the most important thing in AI, and also that researchers choose not to work
Latent.Space (@latentspacepod) 's Twitter Profile Photo

🆕 Everything you should know about Context Engineering youtube.com/watch?v=_IlTcW… Works like Chroma's Context Rot research and Drew Breunig's Context Fails show a lot of issues with naive long context usage: - Context Poisoning - Context Distraction - Context Confusion - Context

Khurram Javed (@khurramjaved_96) 's Twitter Profile Photo

This is a big deal. It is the first large-scale demonstration of the advantage of real-time reinforcement learning. The recipe is scalable and requires no intervention in principle; the model can adapt forever as long as it is being used. There is no way to achieve similar

Nathan Lambert (@natolambert) 's Twitter Profile Photo

The first research on the fundamentals of character training -- i.e. applying modern post training techniques to ingrain specific character traits into models. All models, datasets, code etc released. Really excited about this project! Sharan was a joy to work with.

The first research on the fundamentals of character training -- i.e. applying modern post training techniques to ingrain specific character traits into models.

All models, datasets, code etc released.
Really excited about this project! Sharan was a joy to work with.
Neel Nanda (@neelnanda5) 's Twitter Profile Photo

It was great to help with this interactive tutorial on SAEs, what they can be used for, and how they work. Fantastic work by the team!