Scott Niekum (@scottniekum) 's Twitter Profile
Scott Niekum

@scottniekum

Associate professor at UMass Amherst CICS. AIignment, safety, reinforcement learning, imitation learning, and robotics.

ID: 1091815542334337024

linkhttps://people.cs.umass.edu/~sniekum/ calendar_today02-02-2019 21:48:05

572 Tweet

3,3K Takipçi

369 Takip Edilen

Greg Durrett (@gregd_nlp) 's Twitter Profile Photo

This project started with us annoyed at papers evaluating CoT "reasoning" with only GSM8k & MATH. We didn't expect to find such strong evidence that these are the only type of problem where CoT helps! Credit to Juan Diego Rodríguez (he/him) & Kyle Mahowald for driving the rigorous meta-analysis!

Harshit Sikchi (@harshit_sikchi) 's Twitter Profile Photo

Our cross-university(s) collaborative work on "Scaling laws for Reward Model Overoptimization in Direct Alignment Algorithms" is accepted at NeurIPS Conference!

David Krueger (@davidskrueger) 's Twitter Profile Photo

"Predicting Future Actions of Reinforcement Learning Agents" - Chung et al. We introduce the problem of predicting RL agents' behavior, which could have important safety implications. We find that RL agents that perform explicit (or implicit) planning can be more predictable.

Marlos C. Machado (@marloscmachado) 's Twitter Profile Photo

For those interested, the keynotes of the RL_Conference 2024 are now available online: youtube.com/@RL-conference… Unfortunately, Doina Precup's talk was not recorded, but we have: Andy Barto, Emma Brunskill, Finale Doshi-Velez, Sergey Levine, David Silver, and Peter Stone.

Eugene Vinitsky 🍒🦋 (@eugenevinitsky) 's Twitter Profile Photo

In our new paper, we find that LLMs can efficiently do RLHF in-context! Our method, in-context preference learning (ICPL), iterates LLMs writing reward functions, training agents, and putting preferences into context. We see a 30x boost in query efficiency over baseline RLHF!

In our new paper, we find that LLMs can efficiently do RLHF in-context! 
Our method, in-context preference learning (ICPL), iterates LLMs writing reward functions, training agents, and putting preferences into context. We see a 30x boost in query efficiency over baseline RLHF!
Zizhao Wang (@duke_zzwang) 's Twitter Profile Photo

In multi-object env, why do most Unsupervised Skill Discovery methods fail to learn complex skills like tool use? Because they simply maximize state coverage. Introducing our solution SkiLD: Skill Discovery Guided by Factor Interactions (NeurIPS24) wangzizhao.github.io/SkiLD/

Meghan E. Huber (@meghanehuber) 's Twitter Profile Photo

Come join our team at UMass Robotics!! We are hiring at the Associate/Full level for a joint appointment in engineering and computer science. Feel free to reach out if you have any questions. RTs appreciated :) careers.umass.edu/amherst/en-us/…

RLDM (@rldmdublin2025) 's Twitter Profile Photo

Save the date! RLDM 2025, The Multi-disciplinary Conference on Reinforcement Learning and Decision Making, is only around the corner. Visit our website to keep an eye on our submission deadlines👀 rldm.org

Save the date! 

RLDM 2025, The Multi-disciplinary Conference on Reinforcement Learning and Decision Making, is only around the corner. Visit our website to keep an eye on our submission deadlines👀

rldm.org
brendan o'connor (@brendan642) 's Twitter Profile Photo

We're hiring new #nlproc faculty this year! Asst or Assoc Professors in NLP at UMass CICS -- careers.umass.edu/amherst/en-us/…

RL_Conference (@rl_conference) 's Twitter Profile Photo

The call for papers for RLC is now up! Abstract deadline of 2/14, submission deadline of 2/21! Please help us spread the word. rl-conference.cc/callforpapers.…

Scott Niekum (@scottniekum) 's Twitter Profile Photo

I'm quite excited about this and still a bit shocked that it works as well as it does. Imitation via distribution matching has always felt like a clunky, brittle way to teach agents. Language + zero-shot RL is natural and scales well, due to the unsupervised nature of RL Zero.

Greg Durrett (@gregd_nlp) 's Twitter Profile Photo

Huge congrats to Prasann Singhal for being one of the 8 CRA Outstanding Undergraduate Researcher Award winners! It has been an absolute privilege to work with Prasann during his time at UT. (And he's applying for PhD programs this year...hint hint...) Prasann's work... 🧵

Huge congrats to <a href="/prasann_singhal/">Prasann Singhal</a> for being one of the 8 CRA Outstanding Undergraduate Researcher Award winners! It has been an absolute privilege to work with Prasann during his time at UT. (And he's applying for PhD programs this year...hint hint...)

Prasann's work... 🧵
Gokul Swamy (@g_k_swamy) 's Twitter Profile Photo

1.5 yrs ago, we set out to answer a seemingly simple question: what are we *actually* getting out of RL in fine-tuning? I'm thrilled to share a pearl we found on the deepest dive of my PhD: the value of RL in RLHF seems to come from *generation-verification gaps*. Get ready to🤿!

1.5 yrs ago, we set out to answer a seemingly simple question: what are we *actually* getting out of RL in fine-tuning? I'm thrilled to share a pearl we found on the deepest dive of my PhD: the value of RL in RLHF seems to come from *generation-verification gaps*. Get ready to🤿!
Scott Niekum (@scottniekum) 's Twitter Profile Photo

I'm extremely proud of the work that Harshit has done and looking forward to seeing what he does next. Congratulations, Harshit!

RL_Conference (@rl_conference) 's Twitter Profile Photo

Reminder that early registration for RLC closes on the 30th! Please register early to save yourself some money and help us get the word out.

Harshit Sikchi (@harshit_sikchi) 's Twitter Profile Photo

Behavioral Foundation Models (BFMs) trained with RL are secretly more powerful than we think. BFM’s directly output a policy believed to be near-optimal given any reward function. Our new work shows that they can actually do much better: