Mansi Sakarvadia
@mansi__s
Bluesky: @mansisakarvadia.bsky.social
Computer Science/Machine Learning Ph.D. Student @UChicago & @globus. @doecsgf Computational Science Graduate Fellow.
ID: 3434463047
https://msakarvadia.github.io/ 21-08-2015 17:45:19
44 Tweet
84 Followers
163 Following
Interested in understanding how #LLM s work, why they often fail to reason, and how to improve performance? One tool to boost multi-hop reasoning is with targeted memory injections. This improves desired token probability by up to 424%! 🎥 Watch the talk by Mansi Sakarvadia now:
Mansi Sakarvadia presented her Master's thesis on "Memory Injections: Correcting Multi-Hop Reasoning Failures during Inference in Transformer-Based Language Models". Watch the recording here: youtube.com/watch?v=4EE9DI…
Jordan Pettyjohn, Nathaniel Hudson, Mansi Sakarvadia, Aswathy Ajith, and Kyle Chard just published new work demonstrating detoxification strategies on Language Model outputs at BlackboxNLP! ""Mind Your Manners: Detoxifying Language Models via Attention Head Intervention" Congrats All!
Language models can memorize sensitive data! 🔒 Our new research by the team (Mansi Sakarvadia, Nathaniel Hudson, and others) with TinyMem shows unlearning methods like BalancedSubnet effectively mitigate memorization while keeping performance high. #AI #Privacy mansisak.com/memorization/
Excited to share our latest work: "SoK: On Finding Common Ground in Loss Landscapes Using Deep Model Merging Techniques"! 🧠 arxiv.org/abs/2410.12927 By Arham Khan, Todd Nief, Nathaniel Hudson, Mansi Sakarvadia, Daniel Grzenda, Aswathy Ajith, Jordan Pettyjohn, Kyle Chard, and Ian Foster.