Abhinav Chinta (@abhinavchinta10) 's Twitter Profile
Abhinav Chinta

@abhinavchinta10

NLP Research @ UIUC | @convai_uiuc

ID: 1704543418867654657

linkhttp://abhinavchinta.com calendar_today20-09-2023 17:10:19

65 Tweet

78 Takipçi

413 Takip Edilen

Sumuk (@sumukx) 's Twitter Profile Photo

Feel GPT's style is not quite what you like? Wish that it could understand your preferences implicitly? Our latest work from ConvAI@UIUC, Unsupervised Human Preference Learning (now accepted to the EMNLP 2024 main conference), offers a novel, decoupled solution to the alignment

Sumuk (@sumukx) 's Twitter Profile Photo

happening today at riverfront hall from 10:30 - 12:00! do drop by - we’ve done something quite clever and think you’ll be super excited!

Abhinav Chinta (@abhinavchinta10) 's Twitter Profile Photo

With a bit of prompting gemini 2.0 flash experimental was finally able to generate this. Tried with every other model and nothing even came close. “A horse riding an astronaut.”

With a bit of prompting gemini 2.0 flash experimental was finally able to generate this. Tried with every other model and nothing even came close.

“A horse riding an astronaut.”
Sumuk (@sumukx) 's Twitter Profile Photo

we're launching 🤗 yourbench today, an open source tool for custom benchmarking and synthetic data generation from ANY of your documents. it's a big step towards improving how model evaluations work early access link in replies! (1/8)

we're launching 🤗 yourbench today, an open source tool for custom benchmarking and synthetic data generation from ANY of your documents. it's a big step towards improving how model evaluations work

early access link in replies!

(1/8)
Sagnik Mukherjee (@saagnikkk) 's Twitter Profile Photo

🚀Our ICML 2025 paper introduces "Premise-Augmented Reasoning Chains" - a structured approach to induce explicit dependencies in reasoning chains. By revealing the dependencies within chains, we significantly improve how LLM reasoning can be verified. 🧵[1/n]

🚀Our ICML 2025 paper introduces "Premise-Augmented Reasoning Chains" - a structured approach to induce explicit dependencies in reasoning chains. 

By revealing the dependencies within chains, we significantly improve how LLM reasoning can be verified.

🧵[1/n]
Sagnik Mukherjee (@saagnikkk) 's Twitter Profile Photo

🚨 Paper Alert: “RL Finetunes Small Subnetworks in Large Language Models” From DeepSeek V3 Base to DeepSeek R1 Zero, a whopping 86% of parameters were NOT updated during RL training 😮😮 And this isn’t a one-off. The pattern holds across RL algorithms and models. 🧵A Deep Dive

🚨 Paper Alert: “RL Finetunes Small Subnetworks in Large Language Models”

From DeepSeek V3 Base to DeepSeek R1 Zero, a whopping 86% of parameters were NOT updated during RL training 😮😮
And this isn’t a one-off. The pattern holds across RL algorithms and models.
🧵A Deep Dive
Sagnik Mukherjee (@saagnikkk) 's Twitter Profile Photo

🚀 Headed to #ICML2025 in Vancouver (July 13-19) ! We will present our paper in the poster session at East Exhibition Hall on Tuesday (15th) at 4:30 PM PDT. Happy to chat regarding reasoning, post-training and anything LLMs in general !

🚀 Headed to #ICML2025 in Vancouver (July 13-19) !
We will present our paper in the poster session at East Exhibition Hall on Tuesday (15th) at 4:30 PM PDT. 

Happy to chat regarding reasoning, post-training and anything LLMs in general !
Abhinav Chinta (@abhinavchinta10) 's Twitter Profile Photo

🚨Presenting our ICML paper “Premise Augmented Reasoning Chains” at 4:30PM PT today in the East Exhibition Hall A-B (E-2410). Come check out our work on improving error detection in COT using DAGs! Link to paper: abhinavchinta.com/parc/ #icml25

Sumuk (@sumukx) 's Twitter Profile Photo

100% agree with will. building my first 3090 cluster with Abhinav Chinta was such a great learning experiment in sourcing cheap components from shenzhen, dealing with riser retiming issues, hacking power supplies together etc. just buy the service “advice” is toxic.

Anjiang Wei (@anjiangw) 's Twitter Profile Photo

We introduce SuperCoder, the first work to successfully apply LLMs as superoptimizers for assembly code 🚀. Our RL-trained model achieves 95% correctness ✅ and 1.46× speedup ⚡over gcc -O3. 📄arxiv.org/pdf/2505.11480 💻github.com/Anjiang-Wei/Su… #LLM #Compilers #Code #Optimization

We introduce SuperCoder, the first work to successfully apply LLMs as superoptimizers for assembly code 🚀. Our RL-trained model achieves 95% correctness ✅ and 1.46× speedup ⚡over gcc -O3.
📄arxiv.org/pdf/2505.11480
💻github.com/Anjiang-Wei/Su…
#LLM #Compilers #Code #Optimization