Nicole Fitzgerald (@ninklefitz) 's Twitter Profile
Nicole Fitzgerald

@ninklefitz

on sabbatical? prev co-founder @alpacaml (acq); research @MicrosoftResearch, @MosaicML, @Mila_Quebec. Writing at fov.dev

ID: 897115119662006273

linkhttps://ninkle.github.io/ calendar_today14-08-2017 15:18:06

227 Tweet

1,1K Takipçi

589 Takip Edilen

Matthew Leavitt (@leavittron) 's Twitter Profile Photo

DatologyAI and I are at #ICML2025, and we have two best-in-class things at the conference: 1. Swag: fidget spinners 2. Jobs: We're looking for talented researchers and engineers (full-time & interns). We're especially interested in someone to lead research on data curation for

<a href="/datologyai/">DatologyAI</a> and I are at #ICML2025, and we have two best-in-class things at the conference:
1. Swag: fidget spinners
2. Jobs: We're looking for talented researchers and engineers (full-time &amp; interns). We're especially interested in someone to lead research on data curation for
Linus (@thesephist) 's Twitter Profile Photo

I've learned from many writer heroes of mine on Dialectic with Jackson Dahl. Humbled and excited to be able to share this wide ranging conversation about agency, technology, interfaces, and what I've been working on lately with the inimitable Jackson Dahl 💙

Jonah Kallenbach (@jonahkallenbach) 's Twitter Profile Photo

I just wrote my first blog post, “Everything is Training Data Now.” I noticed myself starting to doubt LLM performance on certain benchmarks due to data contamination / gaming, and wanted to see how real this is. Spoiler alert: there’s a lot of variance across models and labs.

I just wrote my first blog post, “Everything is Training Data Now.”

I noticed myself starting to doubt LLM performance on certain benchmarks due to data contamination / gaming, and wanted to see how real this is. Spoiler alert: there’s a lot of variance across models and labs.
Sid (@sid_srk) 's Twitter Profile Photo

Trying to raise one last batch of funds for season 1 of Toronto School of Foundation Modelling! We are nearing the end of the first season with many in the cohort already training their own models, discovering scaling laws, building tokenizers etc. I’m planning a few surprises

Kexin Huang (@kexinhuang5) 's Twitter Profile Photo

🙌Biomni-R0-32B, the first RL agent model trained end-to-end for biology research, is now open-weight on Hugging Face! To benchmark progress, we’re also releasing Biomni-Eval1 — 443 data points across 10 biomedical research tasks! 🔗 Data: huggingface.co/datasets/biomn… 🔗 Model:

alana goyal (@alanaagoyal) 's Twitter Profile Photo

certain words are so nuanced that their translation is an impossible task lexiconic is a digital exploration of those words basecase.vc/lexiconic

Sara Hooker (@sarahookr) 's Twitter Profile Photo

We are hiring a data adaptation research engineer. 🔥 This is a role I am really excited about. Our hiring says a lot about what we stand for, and how we want to build technology. So, here we go...

Nicole Fitzgerald (@ninklefitz) 's Twitter Profile Photo

I am specifically looking for a talented early career researcher (or engineer looking to break into research)! - Fast ramp-up to work on frontier capability evaluation at pre-training & post-training stages - Abundant technical mentorship and exposure to major model labs -

Nicole Fitzgerald (@ninklefitz) 's Twitter Profile Photo

Happily overwhelmed by the response to this! Addendum: am also looking for a very special type of mid-career researcher, more below: - Opportunity to stretch and grow as a leader, a team-builder, and a setter of research direction - Again, ample technical and personal

Cody Blakeney (@code_star) 's Twitter Profile Photo

I've got something new for everyone. My first substack article! Not the one I planned to do first, but a fun one! I have made a handy calculator base on the DeepSeek v1 coefficients for finding optimal LR and batch sizes for dense LLMs.

I've got something new for everyone. My first substack article! Not the one I planned to do first, but a fun one!

I have made a handy calculator base on the DeepSeek v1 coefficients for finding optimal LR and batch sizes for dense LLMs.