nfiedel (@nfiedel) 's Twitter Profile
nfiedel

@nfiedel

ID: 14693322

calendar_today07-05-2008 23:17:48

314 Tweet

314 Takipçi

212 Takip Edilen

Monica Dinculescu (@notwaldorf) 's Twitter Profile Photo

We've all been feeling pretty isolated lately so I made a cute little app to try to help! Here's Make a Song Together, so that you and your friends can..… make a song together!!! 👩‍🎤 g.co/magenta/make-a… 📝 magenta.tensorflow.org/make-a-song-to…

Jascha Sohl-Dickstein (@jaschasd) 's Twitter Profile Photo

CALL FOR TASKS CAPTURING LIMITATIONS OF LARGE LANGUAGE MODELS We are soliciting contributions of tasks to a *collaborative* benchmark designed to measure and extrapolate the capabilities and limitations of large language models. Submit tasks at github.com/google/BIG-Ben… #BIGbench

CALL FOR TASKS CAPTURING LIMITATIONS OF LARGE LANGUAGE MODELS

We are soliciting contributions of tasks to a *collaborative* benchmark designed to measure and extrapolate the capabilities and limitations of large language models. Submit tasks at github.com/google/BIG-Ben… 
#BIGbench
Brian Lester (@blester125) 's Twitter Profile Photo

My first Google AI residency project was accepted to EMNLP 2025 #EMNLP2021! Prompt Tuning can condition a frozen T5 XXL model to perform new tasks while only adding 0.003% more parameters and no performance loss. Camera Ready 📸: arxiv.org/abs/2104.08691 Quick Thread 🧵(1/7)

My first <a href="/GoogleAI/">Google AI</a> residency project was accepted to <a href="/emnlpmeeting/">EMNLP 2025</a> #EMNLP2021!

Prompt Tuning can condition a frozen T5 XXL model to perform new tasks while only adding 0.003% more parameters and no performance loss.

Camera Ready 📸: arxiv.org/abs/2104.08691

Quick Thread 🧵(1/7)
Adam Roberts (@ada_rob) 's Twitter Profile Photo

For a year, the T5 team has collab'd with FLAX and JAX to build a successor to our research library, using it to train models at many scales 📈 on TPU... ...and now you can too! T5X is still in rapid development, but you can use it or find inspiration at goo.gle/t5x!

nfiedel (@nfiedel) 's Twitter Profile Photo

Am so proud of the team’s exceptional research & engineering over the past year+! We are excited to share PaLM 🌴 with the world! The paper is at: goo.gle/palm-paper

nfiedel (@nfiedel) 's Twitter Profile Photo

Tool Augmented Language Models. abs: arxiv.org/abs/2205.12255 Smaller tool augmented models outperform larger non-augmented models in two domains (thus far), and on out-of-distribution examples. Great collaboration w/AaronParisi and @YaoZhaoAI!

Jascha Sohl-Dickstein (@jaschasd) 's Twitter Profile Photo

After 2 years of work by 442 contributors across 132 institutions, I am thrilled to announce that the github.com/google/BIG-ben… paper is now live: arxiv.org/abs/2206.04615. BIG-bench consists of 204 diverse tasks to measure and extrapolate the capabilities of large language models.

After 2 years of work by 442 contributors across 132 institutions, I am thrilled to announce that the github.com/google/BIG-ben… paper is now live: arxiv.org/abs/2206.04615. BIG-bench consists of 204 diverse tasks to measure and extrapolate the capabilities of large language models.
Zoubin Ghahramani (@zoubinghahrama1) 's Twitter Profile Photo

What happens when you combine the best of language models with robots that operate in the real world? Take a look at our new work from Google AI and Everyday Robots!

Jason Baldridge (@jasonbaldridge) 's Twitter Profile Photo

Exciting news: #Parti and #Imagen teamed up to create a hybrid system with Parti creating 256x256 images which then recieve Imagen super resolution to produce 1024x1024 pixels! See the diagram below for how it works. See thread for more info and new images with this system!

Exciting news: #Parti and #Imagen teamed up to create a hybrid system with Parti creating 256x256 images which then recieve Imagen super resolution to produce 1024x1024 pixels! See the diagram below for how it works.

See thread for more info and new images with this system!
Aleksandra Faust (@aleksandrafaust) 's Twitter Profile Photo

Common HTML understanding tasks can be done without custom NN architecture design and with orders of magnitude less data by fine-tuning LLMs. Bidirectional attention appears to be crucial, and context windows remain the bottleneck.

Jeff Dean (@jeffdean) 's Twitter Profile Photo

Bard is now available in the US and UK, w/more countries to come. It’s great to see early Google AI work reflected in it—advances in sequence learning, large neural nets, Transformers, responsible AI techniques, dialog systems & more. You can try it at bard.google.com

nfiedel (@nfiedel) 's Twitter Profile Photo

Building Gemma together with an exceptional team has been a delight, and now we're thrilled to share it with the world. A huge congrats to the entire team! Special thanks to Kathleen & Alek, Tris Warkentin, Armand Joulin, Clément – you are all amazing :)

Clément (@clmt) 's Twitter Profile Photo

Gemma is expanding.... we just announced CodeGemma, a version of Gemma tuned for code generation. And bonus... Gemma is now bumped to v1.1, addressing lots of feedback we got. Congrats Gemma team for one more amazing release! developers.googleblog.com/2024/04/gemma-…

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

We’re introducing new additions to Gemma: our family of open models built with the same technology as Gemini. 🔘 PaliGemma: a powerful open vision-language model 🔘 Gemma 2: coming soon in various sizes, including 27 billion parameters → dpmd.ai/3QKEteK #GoogleIO

We’re introducing new additions to Gemma: our family of open models built with the same technology as Gemini.

🔘 PaliGemma: a powerful open vision-language model
🔘 Gemma 2: coming soon in various sizes, including 27 billion parameters

→ dpmd.ai/3QKEteK #GoogleIO
Clément (@clmt) 's Twitter Profile Photo

Gemma 2 is out! As with our first model, we're super focused on creating models at useful, practical sizes, so that they can be easily deployable... all the while being amazing in quality. We upgraded our 9B so that it's truly awesome and best in class across many benchmarks.

Gemma 2 is out!

As with our first model, we're super focused on creating models at useful, practical sizes, so that they can be easily deployable... all the while being amazing in quality. 

We upgraded our 9B so that it's truly awesome and best in class across many benchmarks.
Demis Hassabis (@demishassabis) 's Twitter Profile Photo

Gemma 2 is available to researchers & developers. At 27B it delivers best-in-class performance for its size and is competitive even to models over twice its size! Proud to continue our tradition of thoughtfully bringing cutting-edge research to the open models ecosystem.

Armand Joulin (@armandjoulin) 's Twitter Profile Photo

Are small models still undertrained? We are releasing a 2B model that beats GPT-3.5. The crazy part is that it was distill on only 2T tokens from a small model. Distillation is the future of LLMs with the growing availability of large and efficient open models!

Are small models still undertrained? 
We are releasing a 2B model that beats GPT-3.5. The crazy part is that it was distill on only 2T tokens from a small model. 
Distillation is the future of LLMs with the growing availability of large and efficient open models!