David Liu (@davidwnliu) 's Twitter Profile
David Liu

@davidwnliu

PhD Machine Learning & Computational Neuroscience @Cambridge_Eng | BA/MSci Computational and Theoretical Physics @DeptofPhysics

dev @thedavindicode

ID: 1446743206624903168

linkhttp://davindicode.github.io calendar_today09-10-2021 07:44:41

52 Tweet

106 Followers

131 Following

Guillaume Bellec (@bellecguill) 's Twitter Profile Photo

Pre-print: machine learning for neuroscience We build interpretable biological network reconstructions from electrode recordings with ML and optimal transport. Towards models of mechanisms driving behavior, we focus on single-trial neural activity and trial variability 1/6

Pre-print: machine learning for neuroscience

We build interpretable biological network reconstructions from electrode recordings with ML and optimal transport.

Towards models of mechanisms driving behavior, we focus on single-trial neural activity and trial variability 1/6
Laura Ruis (@lauraruis) 's Twitter Profile Photo

How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledge🦜? In our new preprint, we look at the pretraining data and find evidence against this: Procedural knowledge in pretraining drives LLM reasoning ⚙️🔢 🧵⬇️

How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledge🦜? In our new preprint, we look at the pretraining data and find evidence against this:

Procedural knowledge in pretraining drives LLM reasoning ⚙️🔢

🧵⬇️
Wayne Soo (@soowmwayne) 's Twitter Profile Photo

Continuous-time RNNs are used in neuroscience to model neural dynamics. CNNs are used in vision neuroscience for image processing. So what's the right architecture to model the biological visual system? We propose a hybrid. (#NeurIPS2024 spotlight!) openreview.net/forum?id=ZZ94a…

Scott Linderman (@scott_linderman) 's Twitter Profile Photo

I'm excited to share our #NeurIPS2024 paper, "Modeling Latent Neural Dynamics with Gaussian Process Switching Linear Dynamical Systems" 🧠✨ We introduce the gpSLDS, a new model for interpretable analysis of latent neural dynamics! 🧵 1/10

Darshan 🦖 (@darshan) 's Twitter Profile Photo

The most misunderstood condition: Brain fog. It's not just fatigue. It's not just stress. Here's what's really happening inside your body:

The most misunderstood condition:

Brain fog.

It's not just fatigue. It's not just stress.

Here's what's really happening inside your body:
James Campbell (@jam3scampbell) 's Twitter Profile Photo

The Road to AGI along with Emiliano (who's awesome, go follow), I built an interactive timeline of everything in AI the past few years we're living through the most exciting time in history and this site hopes to document it! go visit: ai-timeline dot org (link below)

David D. Baek (@dbaek__) 's Twitter Profile Photo

1/9 🚨 New Paper Alert: Cross-Entropy Loss is NOT What You Need! 🚨 We introduce harmonic loss as alternative to the standard CE loss for training neural networks and LLMs! Harmonic loss achieves 🛠️significantly better interpretability, ⚡faster convergence, and ⏳less grokking!

Aleksander Madry (@aleks_madry) 's Twitter Profile Photo

Do current LLMs perform simple tasks (e.g., grade school math) reliably? We know they don't (is 9.9 larger than 9.11?), but why? Turns out that, for one reason, benchmarks are too noisy to pinpoint such lingering failures. w/ Josh Vendrow Eddie Vendrow Sara Beery 1/5

Do current LLMs perform simple tasks (e.g., grade school math) reliably?

We know they don't (is 9.9 larger than 9.11?), but why?

Turns out that, for one reason, benchmarks are too noisy to pinpoint such lingering failures.

w/ <a href="/josh_vendrow/">Josh Vendrow</a> <a href="/EdwardVendrow/">Eddie Vendrow</a> <a href="/sarameghanbeery/">Sara Beery</a>
1/5
Brian S. Kim (@itchdoctor) 's Twitter Profile Photo

Cancer neuroimmunology is real. Nociceptive neurons promote gastric tumour progression via a CGRP–RAMP1 axis | Nature nature.com/articles/s4158…

Miles Cranmer (@milescranmer) 's Twitter Profile Photo

Why 'I don’t know' is the true test for AGI—it’s a strictly harder problem than text generation! This magnificent 62-page paper (arxiv.org/abs/2408.02357) formally proves AGI hallucinations are inevitable, with 50 pages (!!) of supplementary proofs.

Why 'I don’t know' is the true test for AGI—it’s a strictly harder problem than text generation!

This magnificent 62-page paper (arxiv.org/abs/2408.02357) formally proves AGI hallucinations are inevitable, with 50 pages (!!) of supplementary proofs.
David Duvenaud (@davidduvenaud) 's Twitter Profile Photo

LLMs have complex joint beliefs about all sorts of quantities. And my postdoc James Requeima visualized them! In this thread we show LLM predictive distributions conditioned on data and free-form text. LLMs pick up on all kinds of subtle and unusual structure: 🧵

Akira Yoshiyama ⁂ (@yoshiyama_akira) 's Twitter Profile Photo

Happy to announce we outperformed OpenAI o1 with a 7B model :) We released two self-improvement methods for verifiable domains in our preliminary paper -->

Happy to announce we outperformed <a href="/OpenAI/">OpenAI</a> o1 with a 7B model :)

We released two self-improvement methods for verifiable domains in our preliminary paper --&gt;
Bindu Reddy (@bindureddy) 's Twitter Profile Photo

Mercury Is The First Diffusion LLM! AI simply groks the patterns of the universe. Diffusion LLMs literally manifest the LLM response and are so next generation This is Mercury! The world’s first diffusion LLM

MatthewBerman (@matthewberman) 's Twitter Profile Photo

We knew very little about how LLMs actually work...until now. Anthropic just dropped the most insane research paper, detailing some of the ways AI "thinks." And it's completely different than we thought. Here are their wild findings: 🧵

We knew very little about how LLMs actually work...until now.

<a href="/AnthropicAI/">Anthropic</a> just dropped the most insane research paper, detailing some of the ways AI "thinks."

And it's completely different than we thought.

Here are their wild findings: 🧵
AI at Meta (@aiatmeta) 's Twitter Profile Photo

Today is the start of a new era of natively multimodal AI innovation. Today, we’re introducing the first Llama 4 models: Llama 4 Scout and Llama 4 Maverick — our most advanced models yet and the best in their class for multimodality. Llama 4 Scout • 17B-active-parameter model

Today is the start of a new era of natively multimodal AI innovation.

Today, we’re introducing the first Llama 4 models: Llama 4 Scout and Llama 4 Maverick —  our most advanced models yet and the best in their class for multimodality.

Llama 4 Scout
• 17B-active-parameter model
AI at Meta (@aiatmeta) 's Twitter Profile Photo

Introducing DINOv3: a state-of-the-art computer vision model trained with self-supervised learning (SSL) that produces powerful, high-resolution image features. For the first time, a single frozen vision backbone outperforms specialized solutions on multiple long-standing dense

Andrej Karpathy (@karpathy) 's Twitter Profile Photo

I quite like the new DeepSeek-OCR paper. It's a good OCR model (maybe a bit worse than dots), and yes data collection etc., but anyway it doesn't matter. The more interesting part for me (esp as a computer vision at heart who is temporarily masquerading as a natural language

Spencer Baggins (@bigaiguy) 's Twitter Profile Photo

🚨 MIT just humiliated every major AI lab and nobody’s talking about it. They built a new benchmark called WorldTest to see if AI actually understands the world… and the results are brutal. Even the biggest models Claude, Gemini 2.5 Pro, OpenAI o3 got crushed by humans.

🚨 MIT just humiliated every major AI lab and nobody’s talking about it.

They built a new benchmark called WorldTest to see if AI actually understands the world… and the results are brutal.

Even the biggest models Claude, Gemini 2.5 Pro, OpenAI o3 got crushed by humans.