Sarah Catanzaro(@sarahcat21) 's Twitter Profileg
Sarah Catanzaro

@sarahcat21

“All methods are sacred if they are internally necessary” (GP @amplifypartners, prev @canvasvc; Head of Data @Mattermark; @palantirtech; @c4ads)

ID:2539212208

linkhttp://starckmarcandthefarc.wordpress.com/ calendar_today10-05-2014 15:10:06

5,5K Tweets

12,4K Followers

1,1K Following

Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

I’m most excited about techniques to apply foundation models to low data regimes and/or strategies to generate pretraining/fine-tuning datasets in contexts where data is not abundant. The future is not just LLMs trained on public web data.

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

Stoked for the conversation to shift from vector databases to new file formats optimized for ML/generative AI and designed for today’s hardware/cloud environments:

vldb.org/pvldb/vol17/p1…

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

Reading the Mistborn series based on Ari Morcos rec. but there’s just one problem…

Why hasn’t anyone recognized that the mists present the ideal climate for cultivating Pinot Noir?

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

Fine-tuning is important. Fine-tuning is hard (to do well).

arxiv.org/abs/2405.05904

I hope papers like this get lots of eyeballs so we can be more proactive in controlling the model behaviors that might emerge from FT.

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

AI safety and the impact of regulation on innovation is an important topic but there seems to be such a strong positive correlation between size of ego and number of tweets on this topic (on both sides of the table)

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

It’s all about the data. It always has been.

Plus a phenomenal team and a great product! So proud of our portfolio company DatologyAI!

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

There is so much research/innovation focused on coding agents - and much of it is really cool. But is coding really the only/best approach to get to machine reasoning? Curious why we don’t see more focus on scientific reasoning, mathematical reasoning etc

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

Apparently, attention is not all you need.

You also need a clean diaper and a full belly.

Lessons learned after Camilla Isla Stubbs arrived on March 19, 2024.

Welcome world.

Apparently, attention is not all you need. You also need a clean diaper and a full belly. Lessons learned after Camilla Isla Stubbs arrived on March 19, 2024. Welcome world.
account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

Evaluating LLMs can be less pain, more automated, and more rigorous - but you still need to look at your outputs.

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

First we demanded more compute to train models; then we demanded more compute efficient models; now we want more data to train models, guess what happens next?

(In case you lost the plot, the answer is data efficient models)

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

It was fun to participate in the AI50 list as a judge; congrats to all the companies that are playing such an active role in bringing about a paradigm shift and radically disrupting how we work and live with generative AI.

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

You won’t generate a masterpiece by crafting a good prompt; but you may if you use AI as a tool to enable and execute creative expression.

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

Lots of interesting insights from the Replit ⠕ AI team, but perhaps the most interesting tidbit shows up in the first paragraph.

Will other companies pretrain models that are designed to operate in their specific applications/with their specific interfaces?

Lots of interesting insights from the @Replit AI team, but perhaps the most interesting tidbit shows up in the first paragraph. Will other companies pretrain models that are designed to operate in their specific applications/with their specific interfaces?
account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

Can’t keep a good man down

(Although we can debate if Josh is good, he’s certainly not good at retiring…)

account_circle
Sarah Catanzaro(@sarahcat21) 's Twitter Profile Photo

There’s discussion on data curation for LLMs and discussion on training agent foundation models but not enough talk on what datasets to use when training agents to learn to complete complex multi-step reasoning.

account_circle