Linjie (Lindsey) Li (@linjiefun) 's Twitter Profile
Linjie (Lindsey) Li

@linjiefun

researching @Microsoft, @UW, contributing to openai.com/dall-e-3

ID: 761527442

calendar_today16-08-2012 13:11:24

145 Tweet

2,2K Takipçi

390 Takip Edilen

Xueyan Zou (@xyz2maureen) 's Twitter Profile Photo

As an AI researcher, are you interested in tracking trends from CV/NLP/ML to robotics—even Nature/Science. Our paper “Real Deep Research for AI, Robotics & Beyond” automates survey generation and trend/topic discovery across fields 🔥Explore RDR at realdeepresearch.github.io

Huaxiu Yao✈️ICLR 2025🇸🇬 (@huaxiuyaoml) 's Twitter Profile Photo

🚨 #EMNLP2025 Oral: AI Can’t Actually “See” Videos - New Benchmark Exposes the Truth LVLMs aren’t thinking with video - they’re performing it. What looks like understanding is just visual theater. Introducing GLIMPSE - a benchmark revealing how today’s models fail when

DailyPapers (@huggingpapers) 's Twitter Profile Photo

Microsoft Research unveils VCode A new multimodal coding benchmark that turns images into expressive SVG code for advanced visual reasoning. It's time to bridge the gap between pixel-based vision and symbolic understanding!

Microsoft Research unveils VCode

A new multimodal coding benchmark that turns images into expressive SVG code for advanced visual reasoning. It's time to bridge the gap between pixel-based vision and symbolic understanding!
Kevin Lin (@kevinqhlin) 's Twitter Profile Photo

Thanks AK sharing our work! 🚀 Glad to introduce our newest work — VCode! 🎨 VCode: A Multimodal Coding Benchmark with SVG as Symbolic Visual Representation For decades, RGB pixels have been the default medium for representing images. But in the agentic era, how can we

Kevin Lin (@kevinqhlin) 's Twitter Profile Photo

🤗Excited to open-source GroundCUA! 🚀A large-scale, human-annotated dataset for precise UI grounding to advance Computer-Use Agents. - 3.56M+ high-quality human annotations - 56K screenshots - 87 desktop applications - all datasets and models are available project page:

🤗Excited to open-source GroundCUA!

🚀A large-scale, human-annotated dataset for precise UI grounding to advance Computer-Use Agents.

- 3.56M+ high-quality human annotations
- 56K screenshots
- 87 desktop applications
- all datasets and models are available

project page:
Qineng Wang (@qineng_wang) 's Twitter Profile Photo

Most VLM benchmarks watch the world; few ask how actions *change* it from a robot's eye. Embodied cognition tells us that intelligence isn't just watching – it's enacted through interaction. 👉We introduce ENACT: A benchmark that tests if VLMs can track the evolution of a

Kevin Lin (@kevinqhlin) 's Twitter Profile Photo

🤔Gemini3-Pro can generate fantasy, human-appealing UIs, but have you ever wondered whether these UIs are actually good for agents? 🚀 Excited to share our latest work: Computer-Use Agents as Judges for Generative User Interfaces 📖 TL;DR: Can agents provide feedback to help

Kevin Lin (@kevinqhlin) 's Twitter Profile Photo

Are your UIs ready for agents? Good summary! 👉Check out AUI -- Agent-friendly User Interface. 📄arXiv: arxiv.org/abs/2511.15567 🌐Website: showlab.github.io/AUI/ 💻GitHub: github.com/showlab/AUI/ 🤗Dailypaper: huggingface.co/papers/2511.15… 🐙Gradio demo: huggingface.co/spaces/showlab…

Xueyan Zou (@xyz2maureen) 's Twitter Profile Photo

I will join Tsinghua University, College of AI, as an Assistant Professor in the coming month. I am actively looking for 2026 spring interns and future PhDs (ping me if you are in #NeurIPS). It has been an incredible journey of 10 years since I attended an activity organized by

I will join Tsinghua University, College of AI, as an Assistant Professor in the coming month. I am actively looking for 2026 spring interns and future PhDs (ping me if you are in #NeurIPS).

It has been an incredible journey of 10 years since I attended an activity organized by
Zihan Wang - on RAGEN (@wzihanw) 's Twitter Profile Photo

VAGEN poster 𝐭𝐨𝐦𝐨𝐫𝐫𝐨𝐰 at #NeurIPS! 🎮🧠 - 🕚 11am–2pm Wed - 📍 Exhibit Hall C,D,E #5502 We had much fun exploring: • How 𝐰𝐨𝐫𝐥𝐝 𝐦𝐨𝐝𝐞𝐥𝐢𝐧𝐠 helps VLM RL agents learn better policies • 𝐌𝐮𝐥𝐭𝐢-𝐭𝐮𝐫𝐧 𝐏𝐏𝐎 credit assignment via 𝐭𝐰𝐨-𝐥𝐞𝐯𝐞𝐥

VAGEN poster 𝐭𝐨𝐦𝐨𝐫𝐫𝐨𝐰 at #NeurIPS! 🎮🧠

- 🕚 11am–2pm Wed
- 📍 Exhibit Hall C,D,E #5502

We had much fun exploring:
• How 𝐰𝐨𝐫𝐥𝐝 𝐦𝐨𝐝𝐞𝐥𝐢𝐧𝐠 helps VLM RL agents learn better policies
• 𝐌𝐮𝐥𝐭𝐢-𝐭𝐮𝐫𝐧 𝐏𝐏𝐎 credit assignment via 𝐭𝐰𝐨-𝐥𝐞𝐯𝐞𝐥
Jinpeng Wang (@awinyimgprocess) 's Twitter Profile Photo

Thanks AK for sharing our work. We speed up Qwen-image and Flux Inference speed from 50 steps to less than 10steps. 1 sample is enough for speeding up diffusion model in specific domain. Paper Link: arxiv.org/abs/2512.02899 Code Link: github.com/CSU-JPG/Glance

Manling Li (@manlingli_) 's Twitter Profile Photo

VAGEN poster at #NeurIPS: ⏲️11am-2pm Wed 📍Exhibit Hall C,D,E #5502 We look forward to discussing with you about: 1. MDP → POMDP 2. World modeling in agent internal belief 3. What is a good representation in agent internal belief for visual states? 4. How to use World

Liwei Jiang (@liweijianglw) 's Twitter Profile Photo

Super happy to receive the Best Paper Award at #NeurIPS2025 for our Artificial Hivemind paper!! (Really enjoyed giving oral talk at NeurIPS as well!)

Super happy to receive the Best Paper Award at #NeurIPS2025 for our Artificial Hivemind paper!! (Really enjoyed giving oral talk at NeurIPS as well!)
jack morris (@jxmnop) 's Twitter Profile Photo

Wondering how to attend an ML conference the right way? ahead of NeurIPS 2025 (30k attendees!) here are ten pro tips: 1. Your main goals: (i) meet people (ii) regain excitement about work (iii) learn things – in that order. 2. Make a list of papers you like

Wondering how to attend an ML conference the right way?

ahead of NeurIPS 2025 (30k attendees!) here are ten pro tips:  

1. Your main goals:
    (i) meet people
    (ii) regain excitement about work
    (iii) learn things
    – in that order. 
2. Make a list of papers you like
Alisa Liu (@alisawuffles) 's Twitter Profile Photo

Presenting Broken Tokens at the 4:30pm poster session today with Brian Zheng! We'll demystify how LMs can understand brand new tokenizations ([␣, c, a, t] instead of ␣cat) entirely at test-time 😱

Jiawei Gu (@kuvvius) 's Twitter Profile Photo

🔥 Big news! Our ThinkMorph will be discussed at MAR 2025 @ NeurIPS! Two keynotes you don't want to miss: 1⃣ Lindsey Li Linjie (Lindsey) Li : Pictures think harder than words - Evaluating and Building Visual Thinking in Multimodal Models (9:45 AM, Dec 7) 2⃣Yu Cheng Yu Cheng :

🔥 Big news! Our ThinkMorph will be discussed at MAR 2025 @ NeurIPS!

Two keynotes you don't want to miss:
1⃣ Lindsey Li <a href="/LINJIEFUN/">Linjie (Lindsey) Li</a> : Pictures think harder than words - Evaluating and Building Visual Thinking in Multimodal Models (9:45 AM, Dec 7) 
2⃣Yu Cheng <a href="/YuCheng348997/">Yu Cheng</a> :
LMSYS Org (@lmsysorg) 's Twitter Profile Photo

How long have you been "planning to understand" how modern LLM inference works? We just gave you a readable version of SGLang you can finish over the weekend. Introducing mini-SGLang ⚡ We distilled SGLang from 300K into 5,000 lines. Kept the core design, cut the complexity.

How long have you been "planning to understand" how modern LLM inference works?

We just gave you a readable version of SGLang you can finish over the weekend.

Introducing mini-SGLang ⚡

We distilled SGLang from 300K into 5,000 lines. Kept the core design, cut the complexity.