Jonas Pfeiffer (@pfeiffjo) 's Twitter Profile
Jonas Pfeiffer

@pfeiffjo

Research Scientist @GoogleDeepMind | @AdapterHub | previously @nyuniversity @TUDarmstadt @UKPLab @MetaAI @spotify | pfeiffer.ai | (he/him)

ID: 795620317532143616

calendar_today07-11-2016 13:34:00

648 Tweet

3,3K Followers

686 Following

Marktechpost AI Research News ⚡ (@marktechpost) 's Twitter Profile Photo

Google DeepMind Introduces Differentiable Cache Augmentation: A Coprocessor-Enhanced Approach to Boost LLM Reasoning and Efficiency Researchers from Google DeepMind have introduced a method called Differentiable Cache Augmentation. This technique uses a trained coprocessor to

Google DeepMind Introduces Differentiable Cache Augmentation: A Coprocessor-Enhanced Approach to Boost LLM Reasoning and Efficiency

Researchers from Google DeepMind have introduced a method called Differentiable Cache Augmentation. This technique uses a trained coprocessor to
Raj Dabre (@prajdabre1) 's Twitter Profile Photo

Paper #2: Language and Task Arithmetic with Parameter-Efficient Layers for Zero-Shot Summarization Link: aclanthology.org/2024.mrl-1.7/ Ever wondered how we can do LLM weight arithmetic to enable models to handle tasks in languages in zero shot style? The authors have a solution.

Paper #2: Language and Task Arithmetic with Parameter-Efficient Layers for Zero-Shot Summarization
Link: aclanthology.org/2024.mrl-1.7/

Ever wondered how we can do LLM weight arithmetic to enable models to handle tasks in languages in zero shot style? The authors have a solution.
Raj Dabre (@prajdabre1) 's Twitter Profile Photo

I've missed out several technical details so please read the paper, it contains some interesting information nuggets. Overall another cool work by: Alexandra Chronopoulou, Jonas Pfeiffer, Joshua Maynez, Xinyi Wang (Cindy), Sebastian Ruder, and Priyanka Agrawal Thanks for the cool paper!

Douwe Kiela (@douwekiela) 's Twitter Profile Photo

I’m really sad that my dear friend Felix Hill is no longer with us. He had many friends and colleagues all over the world - to try to ensure we reach them, his family have asked to share this webpage for the celebration of his life: pp.events/felix

I’m really sad that my dear friend <a href="/FelixHill84/">Felix Hill</a> is no longer with us. He had many friends and colleagues all over the world - to try to ensure we reach them, his family have asked to share this webpage for the celebration of his life: pp.events/felix
Leshem Choshen C U @ ICLR 🤖🤗 (@lchoshen) 's Twitter Profile Photo

Thanks to our invited speakers Jonas Pfeiffer Alisa Liu who delivered inspiring talks on Modular Deep Learning and decoding-time experts for language model adaptation. A heartfelt thank you to our sponsors Hugging Face Sakana AI Arcee.ai ! Making the competition possible.

Arthur Douillard (@ar_douillard) 's Twitter Profile Photo

Workshop alert 🚨 We'll host in ICLR 2025 a workshop on modularity, encompassing collaborative + decentralized + continual learning. Those topics are on the critical path to building better AIs. Interested? submit a paper and join us in Singapore! sites.google.com/corp/view/mcdc…

Workshop alert 🚨

We'll host in ICLR 2025 a workshop on modularity, encompassing collaborative + decentralized + continual learning.

Those topics are on the critical path to building better AIs.

Interested? submit a paper and join us in Singapore!

sites.google.com/corp/view/mcdc…
AdapterHub (@adapterhub) 's Twitter Profile Photo

🎁 A new update of the Adapters library is out! Check out all the novelties, changes & fixes here: github.com/adapter-hub/ad…

Aishwarya Kamath (@ashkamath20) 's Twitter Profile Photo

Super excited to announce what I’ve been working on for the past few months 💃 GEMMA 3 is out today! It supports 140+ languages, has a context length of 128k tokens and the best part? It’s natively multimodal! 📸

Super excited to announce what I’ve been working on for the past few months 💃

GEMMA 3 is out today! It supports 140+ languages, has a context length of 128k tokens and the best part? It’s natively multimodal! 📸
JB Alayrac (@jalayrac) 's Twitter Profile Photo

Congratulations to the whole Gemma team for the launch and especially Aishwarya Kamath who did an amazing job pushing the MM capability of the model 🚀. Give a try to the model 🔥

Aishwarya Kamath (@ashkamath20) 's Twitter Profile Photo

Put together a small demo with some fun examples of how you can use Gemma3’s new vision capability with multilinguality and reasoning!

Jonas Pfeiffer (@pfeiffjo) 's Twitter Profile Photo

I am hiring a Student Researcher for our Modularity team at the Google DeepMind office in Zurich🇨🇭 Please fill out the interest form if you would like to work with us! The role would start mid/end 2025 and would be in-person in Zurich with 80-100% at GDM forms.gle/N94ViTmKHCCAcv…

Alexandre Ramé (@ramealexandre) 's Twitter Profile Photo

Hiring two student researchers for Gemma post-training team at Google DeepMind Paris! First topic is about diversity in RL for LLMs (merging, generalization, exploration & creativity), second is about distillation (with Nino Vieillard). Ideal if you're finishing PhD. DMs open!

Leon Engländer (@leonenglaender) 's Twitter Profile Photo

Thrilled about our new Adapters release!🎉I had a blast working on this version, especially contributing to the new plugin interface (like adding ModernBERT) and helping with the VeRA adapter method. Have a look at the full thread for all the awesome updates from our team 👇