Sanyam Bhutani(@bhutanisanyam1) 's Twitter Profileg
Sanyam Bhutani

@bhutanisanyam1

๐Ÿ‘จโ€๐Ÿ’ป Sr Data Scientist @h2oai | Previously: @weights_biases ๐ŸŽ™ Podcast Host @ctdsshow ๐Ÿ‘จโ€๐ŸŽ“ International Fellow @fastdotai ๐ŸŽฒ Grandmaster @Kaggle

ID:784597005825871876

linkhttps://www.youtube.com/c/chaitimedatascience calendar_today08-10-2016 03:31:18

7,9K Tweets

34,5K Followers

994 Following

Jeremy Howard(@jeremyphoward) 's Twitter Profile Photo

Today at Answer.AI we've got something new for you: FSDP/QDoRA. We've tested it with AI at Meta Llama3 and the results blow away anything we've seen before.

I believe that this combination is likely to create better task-specific models than anything else at any cost. ๐Ÿงต

Today at @answerdotai we've got something new for you: FSDP/QDoRA. We've tested it with @AIatMeta Llama3 and the results blow away anything we've seen before. I believe that this combination is likely to create better task-specific models than anything else at any cost. ๐Ÿงต
account_circle
Andrej Karpathy(@karpathy) 's Twitter Profile Photo

Congrats to AI at Meta on Llama 3 release!! ๐ŸŽ‰
ai.meta.com/blog/meta-llamโ€ฆ
Notes:

Releasing 8B and 70B (both base and finetuned) models, strong-performing in their model class (but we'll see when the rankings come in @ lmsys.org :))
400B is still training, but already encroachingโ€ฆ

account_circle
Hamel Husain(@HamelHusain) 's Twitter Profile Photo

RAG vs. Fine-Tuning debates drive me nuts. They are not fungible replacements for each other. Dan Becker and I are teaching a free lighting course on the topic

maven.com/p/787a2f/rag-aโ€ฆ

account_circle
Logan Kilpatrick(@OfficialLoganK) 's Twitter Profile Photo

I read over 100 pitch decks in the last 24 hours for AI companies, most people need to 10x the level of their ambition.

account_circle
Logan Kilpatrick(@OfficialLoganK) 's Twitter Profile Photo

A few weeks ago, I had an incredible conversation with Jeremy Howard

We chatted all things AnswerAI (his new startup), FastAI, and more. I hope you enjoy this as much as I did ๐Ÿซถ

youtu.be/OujUZnXf4J0?siโ€ฆ

account_circle
Hamel Husain(@HamelHusain) 's Twitter Profile Photo

There are a growing number of voices expressing disillusionment with fine-tuning.

I'm curious about the sentiment more generally. (I am withholding sharing my opinion rn).

Tweets below are from Emmanuel Ameisen anton Ethan Mollick

There are a growing number of voices expressing disillusionment with fine-tuning. I'm curious about the sentiment more generally. (I am withholding sharing my opinion rn). Tweets below are from @mlpowered @abacaj @emollick
account_circle
Sebastian Raschka(@rasbt) 's Twitter Profile Photo

I've been coding on GitHub quite consistently for about 12 years, but I honestly never expected to find myself up there ๐Ÿ˜Š

I've been coding on GitHub quite consistently for about 12 years, but I honestly never expected to find myself up there ๐Ÿ˜Š
account_circle
Sanyam Bhutani(@bhutanisanyam1) 's Twitter Profile Photo

The incredible Radek Osmulski ๐Ÿ‡บ๐Ÿ‡ฆ was kind enough to interview me ๐Ÿ™

We talked about:

- My journey into programming
- How fastai changed my life
- Why I started my podcast series
- My approach to goal setting
- Reading LLM papers

Thanks again Radek!

m.youtube.com/watch?v=JrwO7fโ€ฆ

The incredible @radekosmulski was kind enough to interview me ๐Ÿ™ We talked about: - My journey into programming - How fastai changed my life - Why I started my podcast series - My approach to goal setting - Reading LLM papers Thanks again Radek! m.youtube.com/watch?v=JrwO7fโ€ฆ
account_circle
Kurian Benoy ๐Ÿ’ป(@kurianbenoy2) 's Twitter Profile Photo

You need to put 2x the quantity as your goal, where x is the quantity you can achieve easily in a week.

Eg: you should aim daily 2 hours to read paper, if you can comfortably read 1 hour now.

- Sanyam Bhutani said this during an interview with Radek.

youtu.be/JrwO7f2C__U?siโ€ฆ

account_circle
Chinmaya Meher(@ckmeher_) 's Twitter Profile Photo

After watching the video
youtu.be/haRCu4hoI2A?siโ€ฆ

I have started reading this book.
I will be posting some ideas from this book.

Thanks Sanyam Bhutani and Radek Osmulski ๐Ÿ‡บ๐Ÿ‡ฆ

After watching the video youtu.be/haRCu4hoI2A?siโ€ฆ I have started reading this book. I will be posting some ideas from this book. Thanks @bhutanisanyam1 and @radekosmulski
account_circle
Sanyam Bhutani(@bhutanisanyam1) 's Twitter Profile Photo

Last week, Radek Osmulski ๐Ÿ‡บ๐Ÿ‡ฆ and I got to hangout with the CEO of Answer.ai ๐Ÿ™

Jeremy Howard has always been very kind to us with his time

Also, he owns a large beach in Brisbane so that makes it an awesome trip ๐Ÿ˜

Last week, @radekosmulski and I got to hangout with the CEO of Answer.ai ๐Ÿ™ @jeremyphoward has always been very kind to us with his time Also, he owns a large beach in Brisbane so that makes it an awesome trip ๐Ÿ˜
account_circle
Sanyam Bhutani(@bhutanisanyam1) 's Twitter Profile Photo

A perfect intro to open source LLMs! ๐Ÿ™

The course by Amit Sangani is now my top recommendation for getting started with Large Language Models:

- Just enough theory for a whole picture

- Teaches prompting, special tokens and conversational agents

- Perfectly abstracts theโ€ฆ

A perfect intro to open source LLMs! ๐Ÿ™ The course by @asangani7 is now my top recommendation for getting started with Large Language Models: - Just enough theory for a whole picture - Teaches prompting, special tokens and conversational agents - Perfectly abstracts theโ€ฆ
account_circle
Andrej Karpathy(@karpathy) 's Twitter Profile Photo

New (2h13m ๐Ÿ˜…) lecture: 'Let's build the GPT Tokenizer'

Tokenizers are a completely separate stage of the LLM pipeline: they have their own training set, training algorithm (Byte Pair Encoding), and after training implement two functions: encode() from strings to tokens, andโ€ฆ

New (2h13m ๐Ÿ˜…) lecture: 'Let's build the GPT Tokenizer' Tokenizers are a completely separate stage of the LLM pipeline: they have their own training set, training algorithm (Byte Pair Encoding), and after training implement two functions: encode() from strings to tokens, andโ€ฆ
account_circle
Sebastiรกn Ramรญrez(@tiangolo) 's Twitter Profile Photo

uv venv creates a virtual environment

โœจ with a sensible default (.venv)
๐Ÿคฉ includes a .gitignore for it

chefkiss ๐Ÿ‘ฉโ€๐Ÿณ

uv pip compile and uv pip install are crazy fast ๐Ÿ˜ฎ

I have to manually check the files to confirm it actually did something, it's annoyingly fast ๐Ÿคฏ

account_circle
Sanyam Bhutani(@bhutanisanyam1) 's Twitter Profile Photo

LLM Fine-Tuning Benchmarks! ๐Ÿ™

Super excited to finally publish this report comparing different GPUs and precisions:

- First, why do it and what is it?

- There are MANY GPU benchmarks but few specific to Large Language Models

- I wanted to compare their behaviour fromโ€ฆ

LLM Fine-Tuning Benchmarks! ๐Ÿ™ Super excited to finally publish this report comparing different GPUs and precisions: - First, why do it and what is it? - There are MANY GPU benchmarks but few specific to Large Language Models - I wanted to compare their behaviour fromโ€ฆ
account_circle