Eric Frank @ SIGGRAPH (@ihavesweaters) 's Twitter Profile
Eric Frank @ SIGGRAPH

@ihavesweaters

research scientist, formerly founding team @ uber ai labs & @ml_collective & geometric intelligence, ex toy designer @KiteandRocket šŸŒˆāœ”ļø

ID: 520153866

calendar_today10-03-2012 06:33:17

1,1K Tweet

1,1K Followers

2,2K Following

AK (@_akhaliq) 's Twitter Profile Photo

Google presents LUMIERE A Space-Time Diffusion Model for Video Generation paper page: huggingface.co/papers/2401.12ā€¦ Demonstrate state-of-the-art text-to-video generation results, and show that our design easily facilitates a wide range of content creation tasks and video editing

Alex Albert (@alexalbert__) 's Twitter Profile Photo

Fun story from our internal testing on Claude 3 Opus. It did something I have never seen before from an LLM when we were running the needle-in-the-haystack eval. For background, this tests a modelā€™s recall ability by inserting a target sentence (the "needle") into a corpus of

Fun story from our internal testing on Claude 3 Opus. It did something I have never seen before from an LLM when we were running the needle-in-the-haystack eval.

For background, this tests a modelā€™s recall ability by inserting a target sentence (the "needle") into a corpus of
Jeff Clune (@jeffclune) 's Twitter Profile Photo

Fascinating! MAML to make LLMs NOT learn fast. If true/robust, it could really help with AI safety. Also, great name! "Results verify that fine-tuning SOPHON-protected models incurs an overhead comparable to or even greater than training from scratch." arxiv.org/pdf/2404.12699

Ben Kellman (@bkell1123) 's Twitter Profile Photo

So why does it matter that glycans are genetically encoded? Glycans can be >50% of proteinā€™s weight, cover >50% of protein surface, and vital in immunology, metastasis, and autoimmune pathogenesis. Yet, many biologists ignored them because they are hard to measure and control

Shreyas Kapur (@shreyaskapur) 's Twitter Profile Photo

My first PhD paper!šŸŽ‰We learn *diffusion* models for code generation that learn to directly *edit* syntax trees of programs. The result is a system that can incrementally write code, see the execution output, and debug it. šŸ§µ1/n

Aran Komatsuzaki (@arankomatsuzaki) 's Twitter Profile Photo

Google presents Open-Endedness is Essential for Artificial Superhuman Intelligence - Argues that the ingredients are now in place to achieve openendedness in AI systems - Claims that such open-endedness is an essential property of any ASI arxiv.org/abs/2406.04268

Google presents Open-Endedness is Essential for Artificial Superhuman Intelligence

- Argues that the ingredients are now in place to achieve openendedness in AI systems
- Claims that such open-endedness is an essential property of any ASI

arxiv.org/abs/2406.04268
Eric Frank @ SIGGRAPH (@ihavesweaters) 's Twitter Profile Photo

Reminds me of small world LSTMs from OpenAI in 2017 openai.com/index/block-spā€¦ The brain is modular, our models should be too

Kenneth Stanley (@kenneth0stanley) 's Twitter Profile Photo

ā€œFoundation modelā€ is a poor metaphor for LLMs. A 3-year old is a foundation model. A 3-year old has a foundation of skills and basic concepts upon which almost anything can be built. An LLM is a multiple-personality mega-vac - it literally sucks up every concept and

Maxence Faldor (@maxencefaldor) 's Twitter Profile Photo

Quality-Diversity + Lenia is a match made in heaven. ā¤ļø With Antoine Cully, we are excited to introduce Leniabreeder, a framework designed to automate the discovery of diverse artificial life. šŸ§¬ šŸŒ Website: leniabreeder.github.io šŸ“„ Paper: arxiv.org/abs/2406.04235 šŸ§µ 1/11

Owain Evans (@owainevans_uk) 's Twitter Profile Photo

New paper, surprising result: We finetune an LLM on just (x,y) pairs from an unknown function f. Remarkably, the LLM can: a) Define f in code b) Invert f c) Compose f ā€”without in-context examples or chain-of-thought. So reasoning occurs non-transparently in weights/activations!

New paper, surprising result:
We finetune an LLM on just (x,y) pairs from an unknown function f. Remarkably, the LLM can:
a) Define f in code
b) Invert f
c) Compose f
ā€”without in-context examples or chain-of-thought.
So reasoning occurs non-transparently in weights/activations!
Orion Reed (@orionreedone) 's Twitter Profile Photo

Here's a WIP browser extension which lets you pull apart any website, create views and transformations with LLMs, add new UI, pull in context from knowledgebases and APIs, and mix pieces of multiple websites together.

samim (@samim) 's Twitter Profile Photo

New Experiment: SerendipityLM - Interactive evolutionary exploration of generative design spaces with large language models. Check out the extensive experiment report, incl. many generative art examples & link to open source repo: samim.io/studio/work/seā€¦ Find out more in this

Eric Frank @ SIGGRAPH (@ihavesweaters) 's Twitter Profile Photo

"Illusion knitting produces knit planes that, when viewed head-on, appear to be meaningless and shapeless gray rectangles... walk around the mounted image, however, and magic happens: an image emerges from the noise." uwplse.org/2024/03/18/Illā€¦

"Illusion knitting produces knit planes that, when viewed head-on, appear to be meaningless and shapeless gray rectangles... walk around the mounted image, however, and magic happens: an image emerges from the noise."
uwplse.org/2024/03/18/Illā€¦
Machine Learning Street Talk (@mlstreettalk) 's Twitter Profile Photo

This is Kenneth Stanley and Tim RocktƤschel on the importance of open-endedness in AI research. Edward Hughes, TimR and colleagues at Google DeepMind have just written "Open-Endedness is Essential for Artificial Superhuman Intelligence"

Yaroslav Bulatov (@yaroslavvb) 's Twitter Profile Photo

How did ML get fixated on the idea of using the same step size for every example? This is bad if your data has fat tails. For Cauchy-distributed data, SGD fails to converge for every setting of step-size and batch-size colab.research.google.com/drive/1KQQGJvJā€¦

AI21 Labs (@ai21labs) 's Twitter Profile Photo

We released the #Jamba 1.5 open model family: - 256K #contextwindow - Up to 2.5X faster on #longcontext in its size class - Native support for structured JSON output, function calling, digesting doc objects & generating citations twtr.to/giIEE #AI #LLM #AI21Jamba

We released the #Jamba 1.5 open model family:

- 256K #contextwindow 
- Up to 2.5X faster on #longcontext in its size class
- Native support for structured JSON output, function calling, digesting doc objects & generating citations 

twtr.to/giIEE

 #AI #LLM #AI21Jamba
Hugh Zhang (@hughbzhang) 's Twitter Profile Photo

Enabling LLMs to reason more deeply at inference time via search is one of the most exciting directions in AI right now. We introduce PlanSearch, a novel method for code generation that searches over high-level "plans" in natural language as a means of encouraging diversity.

Enabling LLMs to reason more deeply at inference time via search is one of the most exciting directions in AI right now. We introduce PlanSearch, a novel method for code generation that searches over high-level "plans" in natural language as a means of encouraging diversity.