Muthu Kumar Chandrasekaran, PhD (@muthukumarc87) 's Twitter Profile
Muthu Kumar Chandrasekaran, PhD

@muthukumarc87

Scientist #NLProc #PhD #AI #ML | Social Injustice in Education | Tweets don't represent my employer

ID: 53433085

linkhttp://linkedin.com/in/muthukumarc87 calendar_today03-07-2009 16:45:10

1,1K Tweet

264 Followers

506 Following

Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

The paper shows a few reinforcement learning tweaks let small LLM agents use tools better and beat larger ones. A 4B model matches or exceeds 32B agents on hard math, science, and code tasks. Old training stitched fake tool traces together, which taught clumsy timing for tool

The paper shows a few reinforcement learning tweaks let small LLM agents use tools better and beat larger ones. 

A 4B model matches or exceeds 32B agents on hard math, science, and code tasks.

Old training stitched fake tool traces together, which taught clumsy timing for tool
Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

This paper speeds up diffusion LLM decoding by updating the stored keys and values cache only when and where needed. Reported gains reach up to 45.1x in the longest cases. Most existing decoders recompute queries, keys, and values for every token and layer at every step, which

This paper speeds up diffusion LLM decoding by updating the stored keys and values cache only when and where needed.

Reported gains reach up to 45.1x in the longest cases.

Most existing decoders recompute queries, keys, and values for every token and layer at every step, which
Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

👨‍🔧 Github: RAG-Anything: All-in-One RAG Framework 7.6k Stars ⭐️ All-in-One Multimodal Document Processing RAG system built on LightRAG. You can query documents containing interleaved text, visual diagrams, structured tables, and mathematical formulations through one interface.

👨‍🔧 Github: RAG-Anything: All-in-One RAG Framework

7.6k Stars ⭐️

All-in-One Multimodal Document Processing RAG system built on LightRAG.

You can query documents containing interleaved text, visual diagrams, structured tables, and mathematical formulations through one interface.
Data Science Dojo (@datasciencedojo) 's Twitter Profile Photo

🚨 Finally, A Scientific Definition of AGI (and It’s Not What You Think) 🚨 For years, “Artificial General Intelligence” has been the most misused and mystified term in AI. Now, a team of leading researchers, including Dan Hendrycks, Yoshua Bengio, Dawn Song, Gary Marcus, and

🚨 Finally, A Scientific Definition of AGI (and It’s Not What You Think) 🚨
For years, “Artificial General Intelligence” has been the most misused and mystified term in AI.

Now, a team of leading researchers, including Dan Hendrycks, Yoshua Bengio, Dawn Song, Gary Marcus, and
Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

Great paper on AI's recursive self-improvement. Builds a single loop that lets a search agent teach itself. One part writes new tasks, one part tries to solve them, and one part judges the answers. A 3-role loop can keep improving a search agent without human labels. The

Great paper on AI's recursive self-improvement.

Builds a single loop that lets a search agent teach itself. 

One part writes new tasks, one part tries to solve them, and one part judges the answers.

A 3-role loop can keep improving a search agent without human labels.

The
Adina Yakup (@adinayakup) 's Twitter Profile Photo

DeepSeek-OCR is out 🔥 huggingface.co/deepseek-ai/De… ✨High-accuracy OCR - MIT license ✨Fast GPU inference (FlashAttention 2, BF16) ✨Docs > Markdown ✨Works with transformers

Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

The paper shows small models can reason better by retrieving step by step instructions during inference. Gives a simple recipe that turns reasoning into reusable text the model can fetch and follow They build a library of 2 part guides by clustering training questions. Each

The paper shows small models can reason better by retrieving step by step instructions during inference.

Gives a simple recipe that turns reasoning into reusable text the model can fetch and follow

They build a library of 2 part guides by clustering training questions.

Each
NVIDIA AI Developer (@nvidiaaidev) 's Twitter Profile Photo

✨ From prototype to production, the new NVIDIA DGX Spark puts GB10 Superchip performance in your hands. Compact. 4 TB ready. Game-changing speed to run your LLMs locally. Available here ➡️ marketplace.nvidia.com/en-us/develope… #SparkSomethingBig ✨

✨ From prototype to production, the new NVIDIA DGX Spark puts GB10 Superchip performance in your hands.

Compact. 4 TB ready. Game-changing speed to run your LLMs locally.

Available here ➡️ marketplace.nvidia.com/en-us/develope…

 #SparkSomethingBig ✨
Emily Xiao (@xiaoemily41333) 's Twitter Profile Photo

Can we train LLMs to be good prompt engineers? 🚀We propose Prompt-MII: Meta-Learning Instruction Induction for LLMs Our models out-perform strong baselines like ICL and GEPA with 13x fewer tokens. 🧵

Can we train LLMs to be good prompt engineers?

🚀We propose Prompt-MII: Meta-Learning Instruction Induction for LLMs

Our models out-perform strong baselines like ICL and GEPA with 13x fewer tokens. 🧵
Jessy Lin (@realjessylin) 's Twitter Profile Photo

🧠 How can we equip LLMs with memory that allows them to continually learn new things? In our new paper with AI at Meta, we show how sparsely finetuning memory layers enables targeted updates for continual learning, w/ minimal interference with existing knowledge. While full

🧠 How can we equip LLMs with memory that allows them to continually learn new things?

In our new paper with <a href="/AIatMeta/">AI at Meta</a>, we show how sparsely finetuning memory layers enables targeted updates for continual learning, w/ minimal interference with existing knowledge.

While full
Guilherme Penedo (@gui_penedo) 's Twitter Profile Photo

New dataset release: 🌐FineWiki This is an updated and better extracted version of Wikipedia, covering 325+ languages. Unlike the old dataset from 2023, we kept all the math content, tables, properly rendered templates, and extracted key facts. Examples and highlights below.

New dataset release: 🌐FineWiki

This is an updated and better extracted version of Wikipedia, covering 325+ languages.

Unlike the old dataset from 2023, we kept all the math content, tables, properly rendered templates, and extracted key facts.

Examples and highlights below.
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

We used DeepSeek OCR to extract every dataset from tables/charts across 500k+ AI arXiv papers for $1000 🚀 See which benchmarks are trending and discover datasets you didn't know existed Doing the same task with Mistral OCR would've cost $7500 👀

elvis (@omarsar0) 's Twitter Profile Photo

Fundamentals of Building Autonomous LLM Agents Great overview of LLM-based agents. Great if you are just getting started with AI agents. This covers the basics good.

Fundamentals of Building Autonomous LLM Agents

Great overview of LLM-based agents.

Great if you are just getting started with AI agents.

This covers the basics good.
Sharon Y. Li (@sharonyixuanli) 's Twitter Profile Photo

Deception is one of the most concerning behaviors that advanced AI systems can display. If you are not concerned yet, this paper might change your view. We built a multi-agent framework to study: 👉 How deceptive behaviors can emerge and evolve in LLM agents during realistic

Deception is one of the most concerning behaviors that advanced AI systems can display. If you are not concerned yet, this paper might change your view.

We built a multi-agent framework to study:
👉 How deceptive behaviors can emerge and evolve in LLM agents during realistic
Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

This paper explains how to build LLM agents that can perceive, reason, remember, and act autonomously. Humans finish 72.36% of OSWorld tasks while top agents reach 42.9%, so there is a big gap. Paper gives one clear recipe and maps common failure points. Workflows run a fixed

This paper explains how to build LLM agents that can perceive, reason, remember, and act autonomously.

Humans finish 72.36% of OSWorld tasks while top agents reach 42.9%, so there is a big gap.

Paper gives one clear recipe and maps common failure points.

Workflows run a fixed
Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

New AMD paper shows a simple way to add reasoning to vision language models cheaply. It reaches results close to heavy methods using 4K examples and about 2 hours of fine tuning. Most multimodal models read images, yet they struggle to connect steps and get the final answer.

New AMD paper shows a simple way to add reasoning to vision language models cheaply. 

It reaches results close to heavy methods using 4K examples and about 2 hours of fine tuning.

Most multimodal models read images, yet they struggle to connect steps and get the final answer.
Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

The paper shows most LLM agents break from early mistakes and offers a way to catch and fix them. Targeted debugging raises task success by up to 26%. Agents run through memory, reflection, planning, and action, and a wrong move early tends to cascade. The authors define a

The paper shows most LLM agents break from early mistakes and offers a way to catch and fix them.

Targeted debugging raises task success by up to 26%.

Agents run through memory, reflection, planning, and action, and a wrong move early tends to cascade.

The authors define a
DailyPapers (@huggingpapers) 's Twitter Profile Photo

NVIDIA just released Audio Flamingo 3 on Hugging Face! This fully open, state-of-the-art Large Audio-Language Model excels at understanding & reasoning across speech, sounds, and music, setting new benchmarks on 20+ tasks. huggingface.co/nvidia/audio-f…

Yueqi Song (@yueqi_song) 's Twitter Profile Photo

We just built and released the largest dataset for supervised fine-tuning of agentic LMs, 1.27M trajectories (~36B tokens)! Up until now, large-scale SFT for agents is rare - not for lack of data, but because of fragmentation across heterogeneous formats, tools, and interfaces.