David Marx (@digthatdata) 's Twitter Profile
David Marx

@digthatdata

Generative AI MLE, FOSS toolmaker, innovation catalyst @CoreWeave + @AiEleuther. AI enhanced creativity, philosophy of mind/science/probability

ID: 2211601081

linkhttps://github.com/dmarx calendar_today24-11-2013 00:56:11

10,10K Tweet

4,4K Followers

1,1K Following

RebeccaWatson@bsky.social why not (@rebeccawatson) 's Twitter Profile Photo

Good morning! Yesterday I shot a 10-hour time lapse of my calathea and now my boyfriend is scared it’s going to try to kill us in the night

Richard Sutton (@richardssutton) 's Twitter Profile Photo

If you are looking to conduct research full-time on the foundations of AI, and • you have read the RL textbook and done the exercises, • you agree with the Alberta Plan for AI Research, • you already have a PhD, • you are open to spending some time in Edmonton, then the

Tom Silver (@tomssilver) 's Twitter Profile Photo

I am hiring multiple PhD students this year to launch the Princeton Robot Planning and Learning lab: tomsilver.github.io/hiring Thanks for helping me spread the word!

clem 🤗 (@clementdelangue) 's Twitter Profile Photo

Did you notice how most video AIs are commercial, closed-source & secret? It sucks and a big reason for the lack of open-source & transparency is that there are very few high-quality open video datasets. The Hugging Face team is changing this today by releasing FineVideo, a

Did you notice how most video AIs are commercial, closed-source & secret? It sucks and a big reason for the lack of open-source & transparency is that there are very few high-quality open video datasets. 

The <a href="/huggingface/">Hugging Face</a> team is changing this today by releasing FineVideo, a
Naomi Saphra (@nsaphra) 's Twitter Profile Photo

Trying to get faster finetuning? How about 𝟏𝟎𝐱 𝐟𝐚𝐬𝐭𝐞𝐫? Our new method, Fast Forward, will help you trim an order of magnitude off your already-efficient LoRA or DoRA finetuning by using … line search, the most basic optimization algorithm.

Trying to get faster finetuning? How about 𝟏𝟎𝐱 𝐟𝐚𝐬𝐭𝐞𝐫? Our new method, Fast Forward, will help you trim an order of magnitude off your already-efficient LoRA or DoRA finetuning by using … line search, the most basic optimization algorithm.
OpenAI (@openai) 's Twitter Profile Photo

We're releasing a preview of OpenAI o1—a new series of AI models designed to spend more time thinking before they respond. These models can reason through complex tasks and solve harder problems than previous models in science, coding, and math. openai.com/index/introduc…

David Marx (@digthatdata) 's Twitter Profile Photo

How am I only just learning about this (article is 3mos old). This is a crime against humanity. Trump is basically guilty of green lighting biological warfare. reuters.com/investigates/s…

Vitaliy Chiley (@vitaliychiley) 's Twitter Profile Photo

It's actually WILD that OAI just dropped a plot where inference compute is log scale and the entire ML community is hyped If you were worried about global warming before... gg earth, it's been a real one :pour-one-out:

David Marx (@digthatdata) 's Twitter Profile Photo

Apparently, someone leaked the onboarding handbook for Mr. Beast's production team. I don't actually care for Mr. Beast videos, but this was legit a super interesting and insightful read. Strong recommend, especially if you are a youtube creator.

Martin Nebelong (@martinnebelong) 's Twitter Profile Photo

Remember how we used to play as kids, how reality and creativity would mix and we'd create new worlds and stories out of cardboard boxes and kitchen utensils.. how nature around us could turn from a bush in our garden into a dense jungle on an alien planet? As adults, a lot of

Daniel Han (@danielhanchen) 's Twitter Profile Photo

A transformer's depth affects its reasoning capabilities, whilst model size affects its knowledge capacity High recommend Zeyuan Allen-Zhu's video on reasoning in transformers. Experiments show wider nets don't affect reasoning but more depth helps. Video: youtu.be/bpp6Dz8N2zY

A transformer's depth affects its reasoning capabilities, whilst model size affects its knowledge capacity

High recommend <a href="/ZeyuanAllenZhu/">Zeyuan Allen-Zhu</a>'s video on reasoning in transformers. Experiments show wider nets don't affect reasoning but more depth helps. Video: youtu.be/bpp6Dz8N2zY
Liliang Ren (@liliang_ren) 's Twitter Profile Photo

Microsoft Research Deep Learning Group and Microsoft GenAI are hiring self-motivated part-time research interns working on long sequence modeling. We have hundreds of H100/A100 dedicated to this project. Please send CV to [email protected] and [email protected].