Lars Quaedvlieg (@lars_quaedvlieg) 's Twitter Profile
Lars Quaedvlieg

@lars_quaedvlieg

Research Assistant @ CLAIRE | MSc Data Science @ EPFL ๐Ÿ‡จ๐Ÿ‡ญ | Interested in reasoning with foundation models and sequential decision-making ๐Ÿง 

ID: 1694031255262617600

linkhttps://lars-quaedvlieg.github.io/ calendar_today22-08-2023 16:58:25

19 Tweet

43 Followers

163 Following

Grigoris Chrysos (@grigoris_c) 's Twitter Profile Photo

๐Ÿ—’๏ธ#NeurIPS paper on tackling NP-hard problem(s) is out! How can self-training meet dynamic programming? Paper: arxiv.org/abs/2310.18672 1/2

Skander Moalla (@skandermoalla) 's Twitter Profile Photo

Have you ever been left puzzled by your PPO agent collapsing out of nowhere? ๐Ÿ“ˆ๐Ÿคฏ๐Ÿ“‰ Weโ€™ve all been there... We can help you with a hint: monitor your representations!๐Ÿ’ก ๐Ÿš€ We show that PPO suffers from degrading representations and that this breaks its trust region ๐Ÿ’”

Have you ever been left puzzled by your PPO agent collapsing out of nowhere? ๐Ÿ“ˆ๐Ÿคฏ๐Ÿ“‰
Weโ€™ve all been there... We can help you with a hint: monitor your representations!๐Ÿ’ก
๐Ÿš€ We show that PPO suffers from degrading representations and that this breaks its trust region ๐Ÿ’”
Caglar Gulcehre (@caglarml) 's Twitter Profile Photo

Submissions to our workshop at ICML on the next "Next Generation of Sequence Modelling" are open now ๐ŸŽ‰๐ŸŽ‡. If you have interesting ideas on improving research related to sequence models, consider submitting them to our workshop: sites.google.com/view/ngsmworksโ€ฆ

Justin Deschenaux (@jdeschena) 's Twitter Profile Photo

How is Stable Diffusion or Dall.E able to synthesize novel images? Does this arise from the architecture? Is it the training data? ๐Ÿค” Our new paper in ย #ICML2024ย ๐Ÿš€ ICML Conference focuses on the interpolation aspect of generalization๐Ÿงต 1/8

How is Stable Diffusion or Dall.E able to synthesize novel images? Does this arise from the architecture? Is it the training data? ๐Ÿค” Our new paper in ย #ICML2024ย ๐Ÿš€ <a href="/icmlconf/">ICML Conference</a> focuses on the interpolation aspect of generalization๐Ÿงต 1/8
Justin Deschenaux (@jdeschena) 's Twitter Profile Photo

๐ŸŒŸย Excited to share our latest work on making diffusion language models (DLMs) faster than autoregressive (AR) models! โšก Itโ€™s been great to work on this with Caglar Gulcehre ๐Ÿ˜Ž Lately, DLMs are gaining traction as a promising alternative to autoregressive sequence modeling ๐Ÿ‘€ 1/14 ๐Ÿงต

Xiuying Wei@Neurips (Wed11am East #2010) (@xiuyingwei966) 's Twitter Profile Photo

Want to make the big FFN more efficient? Check out our work tomorrow! ๐Ÿ“… Wednesday, 11 am โ€“ 2 pm PST ๐Ÿ“ East Exhibit Hall Aโ€“C #2010 Stop by and explore how structured matrices can make LLM training more efficient. ๐Ÿš€ Caglar Gulcehre Skander Moalla

Want to make the big FFN more efficient?
Check out our work tomorrow!
๐Ÿ“… Wednesday, 11 am โ€“ 2 pm PST
๐Ÿ“ East Exhibit Hall Aโ€“C #2010
Stop by and explore how structured matrices can make LLM training more efficient. ๐Ÿš€
<a href="/caglarml/">Caglar Gulcehre</a> <a href="/SkanderMoalla/">Skander Moalla</a>
Lars Quaedvlieg (@lars_quaedvlieg) 's Twitter Profile Photo

Proud to have contributed to EvoTune โ€” a new method that combines evolutionary search and RL fine-tuning to accelerate algorithm discovery with LLMs! ๐Ÿš€ Check out claire-labo.github.io/EvoTune for more info and cool visualizations :)