Bing Yan (@bingyan4science) 's Twitter Profile
Bing Yan

@bingyan4science

AI+Chemistry PhD student @nyuniversity | Prev PhD of Chemistry @mit

ID: 905020625579905025

linkhttp://bingyan.me calendar_today05-09-2017 10:51:46

13 Tweet

76 Takipçi

185 Takip Edilen

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Ever wondered how nondeterministic GPT-4 is even with greedy decoding (T=0)? I built a website that asks GPT-4 to draw a unicorn every hour and tracks if the results stay consistent over time (spoiler alert: they don't! 🦄). Explore the findings: openaiwatch.com

Ever wondered how nondeterministic GPT-4 is even with greedy decoding (T=0)? I built a website that asks GPT-4 to draw a unicorn every hour and tracks if the results stay consistent over time (spoiler alert: they don't! 🦄).

Explore the findings:
openaiwatch.com
Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Meet the Chalk-Diagrams plugin for ChatGPT! Based on Dan Oneață & Sasha Rush's lib, it lets you create vector graphics with language instructions. Try it with ChatGPT plugins: 1️⃣ Plugin store 2️⃣ Install unverified plugin 3️⃣ chalk-diagrams.com Test it out: "draw a pizza" 🍕

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Excited to share that I'm joining Waterloo's Cheriton School of Computer Science as an Assistant Professor and Vector Institute as a Faculty Affiliate in Fall '24. Before that, I'm doing a postdoc at Ai2 with Yejin Choi. Immensely grateful to my PhD advisors Sasha Rush and Stuart Shieber (@[email protected]). This journey wouldn't have

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Can LMs solve reasoning tasks without showing their work? "Implicit Chain of Thought Reasoning via Knowledge Distillation" teaches LMs to reason internally to solve tasks like 5×5 multiplication. Here's how we bypass human-like step-by-step reasoning bit.ly/implicitCoT 1/6

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

I am hiring NLP/ML PhD students at UWaterloo, home to 5 NLP professors! Apply by Dec 1 Strong consideration will be given to those who can tackle the below challenge: Can we use LM's hidden states to reason multiple problems simultaneously? ​​Retweets/shares appreciated🥰

I am hiring NLP/ML PhD students at UWaterloo, home to 5 NLP professors! Apply by Dec 1

Strong consideration will be given to those who can tackle the below challenge: Can we use LM's hidden states to reason multiple problems simultaneously?

​​Retweets/shares appreciated🥰
Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

What do people use ChatGPT for? We built WildVis, an interactive tool to visualize the embeddings of million-scale chat datasets like WildChat. Work done with Wenting Zhao Jack Hessel Sean Ren Claire Cardie Yejin Choi 📝huggingface.co/papers/2409.03… 🔗wildvisualizer.com/embeddings/eng… 1/7

FAIR Chemistry (@opencatalyst) 's Twitter Profile Photo

Introducing Meta’s Open Materials 2024 (OMat24) Dataset and Models! All under permissive open licenses for commercial and non-commercial use! Paper: arxiv.org/abs/2410.12771 Dataset: huggingface.co/datasets/fairc… Models: huggingface.co/fairchem/OMAT24 🧵1/x

Introducing Meta’s Open Materials 2024 (OMat24) Dataset and Models! All under permissive open licenses for commercial and non-commercial use!

Paper: arxiv.org/abs/2410.12771
Dataset: huggingface.co/datasets/fairc…
Models: huggingface.co/fairchem/OMAT24

🧵1/x
Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

How many reasoning tokens does OpenAI o1 use? It turns out they are almost always multiples of 64 (99+% of the time in 100K collected turns)🤔Could it be that the model only uses multiples of 64 tokens to think? Or maybe OpenAI rounds the token count in the returned usage? 1/4

How many reasoning tokens does OpenAI o1 use? It turns out they are almost always multiples of 64 (99+% of the time in 100K collected turns)🤔Could it be that the model only uses multiples of 64 tokens to think? Or maybe OpenAI rounds the token count in the returned usage? 1/4
NYU Center for Data Science (@nyudatascience) 's Twitter Profile Photo

Yuntian Deng from University of Waterloo presented at CILVR on internalizing reasoning in language models. His team's finetuning approach enabled GPT-2 Small to solve 20x20 multiplication with 99.5% accuracy, while standard training couldn't go beyond 4x4. Demo up at huggingface.co/spaces/yuntian…

<a href="/YuntianDeng/">Yuntian Deng</a> from <a href="/UWaterloo/">University of Waterloo</a> presented at CILVR on internalizing reasoning in language models. His team's finetuning approach enabled GPT-2 Small to solve 20x20 multiplication with 99.5% accuracy, while standard training couldn't go beyond 4x4.

Demo up at huggingface.co/spaces/yuntian…
Aaron Havens (@aaronjhavens) 's Twitter Profile Photo

New paper out with FAIR(+FAIR-Chemistry): Adjoint Sampling: Highly Scalable Diffusion Samplers via Adjoint Matching We present a scalable method for sampling from unnormalized densities beyond classical force fields. 📄: arxiv.org/abs/2504.11713

Ricky T. Q. Chen (@rickytqchen) 's Twitter Profile Photo

We've open sourced Adjoint Sampling! It's part of a bundled release showcasing FAIR's research and open source commitment to AI for science. github.com/facebookresear… x.com/AIatMeta/statu…

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Can we build an operating system entirely powered by neural networks? Introducing NeuralOS: towards a generative OS that directly predicts screen images from user inputs. Try it live: neural-os.com Paper: huggingface.co/papers/2507.08… Inspired by Andrej Karpathy's vision. 1/5

Bing Yan (@bingyan4science) 's Twitter Profile Photo

Our new paper Digital Discovery is out💡Kyunghyun Cho Angelica Chen Do LLMs capture intrinsic chemistry, beyond string representations? Across SMILES & IUPAC: ❌ SOTA models: very low consistency (<1%) ⚖️ 1-to-1 mapped finetune ↑ consistency, not accuracy Read: pubs.rsc.org/en/Content/Art…

Our new paper <a href="/digital_rsc/">Digital Discovery</a> is out💡<a href="/kchonyc/">Kyunghyun Cho</a> <a href="/_angie_chen/">Angelica Chen</a>
Do LLMs capture intrinsic chemistry, beyond string representations? Across SMILES &amp; IUPAC:
❌ SOTA models: very low consistency (&lt;1%)
⚖️ 1-to-1 mapped finetune ↑ consistency, not accuracy
Read: pubs.rsc.org/en/Content/Art…
Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

🚀New dataset release: WildChat-4.8M 4.8M real user-ChatGPT conversations collected from our public chatbots: - 122K from reasoning models (o1-preview, o1-mini): represent real uses in the wild and very costly to collect - 2.5M from GPT-4o 🔗 hf.co/datasets/allen… (1/4)