Andrew Gordon Wilson(@andrewgwils) 's Twitter Profileg
Andrew Gordon Wilson

@andrewgwils

Machine Learning Professor

ID:2800204849

linkhttps://cims.nyu.edu/~andrewgw calendar_today09-09-2014 16:14:15

2,0K Tweets

26,5K Followers

717 Following

Andrew Gordon Wilson(@andrewgwils) 's Twitter Profile Photo

A distinguishing ambition of ML research: 'We want AI agents that can discover like we can, not which contain what we have discovered. Building in our discoveries only makes it harder to see how the discovering process can be done.' From Rich Sutton's 'Bitter Lesson'.

account_circle
Sanae Lotfi(@LotfiSanae) 's Twitter Profile Photo

Thanks a lot Cohere For AI for the invitation. I had a great time presenting our work on generalization bounds for LLMs and the questions were insightful.

The process of preparing a talk can be very rewarding 🥳 Thanks Micah Goldblum for helping me craft the story of this talk.

account_circle
Amazon Science(@AmazonScience) 's Twitter Profile Photo

Amazon researchers have released an open-source family of pretrained models for time series forecasting. Chronos is built on a language model architecture and trained with billions of tokenized time series observations to provide accurate zero-shot forecasts for new time series.

account_circle
Bernie Wang(@LoVVgE) 's Twitter Profile Photo

Are all these time-series-specific model design necessary for deep forecasters / foundation forecasting models? In Chronos, we claim no novelty in time series modeling, but that's exactly the point. 'Everything should be made as simple as possible, but no simpler.'

account_circle
Abdul Fatir(@solitarypenman) 's Twitter Profile Photo

🚀 Excited to share that we released Chronos today. Chronos is a framework for building pretrained time series models based on language model architectures.

Simple idea: quantize time series into tokens and feed them into 🤗 Hugging Face models.

🧵

🚀 Excited to share that we released Chronos today. Chronos is a framework for building pretrained time series models based on language model architectures. Simple idea: quantize time series into tokens and feed them into 🤗 @huggingface models. 🧵
account_circle
Lorenzo Stella(@lostella) 's Twitter Profile Photo

Great work, great Team: today we released Chronos, a family of pretrained, LLM-based time series models. Simple core idea: quantize series to get tokens, feed it into 🤗 Hugging Face models, done.

📝 arxiv.org/abs/2403.07815
💻 github.com/amazon-science…
🤗 huggingface.co/amazon/chronos…

Great work, great Team: today we released Chronos, a family of pretrained, LLM-based time series models. Simple core idea: quantize series to get tokens, feed it into 🤗 @huggingface models, done. 📝 arxiv.org/abs/2403.07815 💻 github.com/amazon-science… 🤗 huggingface.co/amazon/chronos…
account_circle
Andrew Gordon Wilson(@andrewgwils) 's Twitter Profile Photo

I really enjoyed my visit at BU! Such a vibrant group of students and faculty. Really great questions, and interactive discussions. And everyone was so welcoming and friendly.

account_circle
Andrew Gordon Wilson(@andrewgwils) 's Twitter Profile Photo

I'm excited to be speaking tomorrow at Boston University, as part of their distinguished speaker series. My talk will be on prescriptive foundations for building autonomous intelligent systems. Talk details: bu.edu/hic/air-distin…

I'm excited to be speaking tomorrow at Boston University, as part of their distinguished speaker series. My talk will be on prescriptive foundations for building autonomous intelligent systems. Talk details: bu.edu/hic/air-distin…
account_circle
Micah Goldblum(@micahgoldblum) 's Twitter Profile Photo

Do LLMs simply memorize and parrot their pretraining data or do they learn patterns that generalize? Let’s put this to the test! We compute the first generalization guarantees for LLMs.

w/ Sanae Lotfi, Marc Finzi, Yilun Kuang, Tim G. J. Rudner, Andrew Gordon Wilson

arxiv.org/abs/2312.17173

1/9

account_circle
Marc Finzi(@m_finzi) 's Twitter Profile Photo

In this work we construct the first nonvacuous generalization bounds for LLMs, helping to explain why these models generalize.
w/ Sanae Lotfi, Yilun Kuang, Tim G. J. Rudner Micah Goldblum, Andrew Gordon Wilson

arxiv.org/abs/2312.17173

A 🧵on how we make these bounds
1/9

account_circle
Andrew Gordon Wilson(@andrewgwils) 's Twitter Profile Photo

We introduce non-vacuous generalization bounds for large language models. LLMs are hardly parroting their training data! They achieve significant compression.

account_circle
Andrew Gordon Wilson(@andrewgwils) 's Twitter Profile Photo

It wasn’t that long ago that these generated faces were considered a spectacular breakthrough. How far we’ve come.

It wasn’t that long ago that these generated faces were considered a spectacular breakthrough. How far we’ve come.
account_circle
Andrew Gordon Wilson(@andrewgwils) 's Twitter Profile Photo

I'm glad to see losslandscape.com is still going strong. Javier Ideami has beautiful visualizations. The geometric properties of neural network training objectives, such as mode connectivity, make deep learning truly distinct.

I'm glad to see losslandscape.com is still going strong. @ideami has beautiful visualizations. The geometric properties of neural network training objectives, such as mode connectivity, make deep learning truly distinct.
account_circle
Ofir Press(@OfirPress) 's Twitter Profile Photo

5 tips for finding research ideas:
>Focus on finding a problem, not a solution
>Iterate quickly through multiple ideas, don't over-analyze just one idea
>Write a paper that many people would want to read
>Make your problems/solutions/code/writing simple
📝 ofir.io/Tips-for-Findi…

5 tips for finding research ideas: >Focus on finding a problem, not a solution >Iterate quickly through multiple ideas, don't over-analyze just one idea >Write a paper that many people would want to read >Make your problems/solutions/code/writing simple 📝 ofir.io/Tips-for-Findi…
account_circle
Andrew Gordon Wilson(@andrewgwils) 's Twitter Profile Photo

Pre-training LLMs on text completion surprisingly helps a lot with generating stable materials! CrystalLLM can also learn relevant symmetries (e.g. rotation), and provide compelling text-conditional generation!

account_circle