Bing Yan (@bingyan4science) 's Twitter Profile
Bing Yan

@bingyan4science

AI+Chemistry PhD student @nyuniversity | Prev PhD of Chemistry @mit

ID: 905020625579905025

linkhttp://bingyan.me calendar_today05-09-2017 10:51:46

13 Tweet

76 Followers

185 Following

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Ever wondered how nondeterministic GPT-4 is even with greedy decoding (T=0)? I built a website that asks GPT-4 to draw a unicorn every hour and tracks if the results stay consistent over time (spoiler alert: they don't! šŸ¦„). Explore the findings: openaiwatch.com

Ever wondered how nondeterministic GPT-4 is even with greedy decoding (T=0)? I built a website that asks GPT-4 to draw a unicorn every hour and tracks if the results stay consistent over time (spoiler alert: they don't! šŸ¦„).

Explore the findings:
openaiwatch.com
Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Meet the Chalk-Diagrams plugin for ChatGPT! Based on Dan Oneață & Sasha Rush's lib, it lets you create vector graphics with language instructions. Try it with ChatGPT plugins: 1ļøāƒ£ Plugin store 2ļøāƒ£ Install unverified plugin 3ļøāƒ£ chalk-diagrams.com Test it out: "draw a pizza" šŸ•

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Excited to share that I'm joining Waterloo's Cheriton School of Computer Science as an Assistant Professor and Vector Institute as a Faculty Affiliate in Fall '24. Before that, I'm doing a postdoc at Ai2 with Yejin Choi. Immensely grateful to my PhD advisors Sasha Rush and Stuart Shieber (@[email protected]). This journey wouldn't have

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Can LMs solve reasoning tasks without showing their work? "Implicit Chain of Thought Reasoning via Knowledge Distillation" teaches LMs to reason internally to solve tasks like 5Ɨ5 multiplication. Here's how we bypass human-like step-by-step reasoning bit.ly/implicitCoT 1/6

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

I am hiring NLP/ML PhD students at UWaterloo, home to 5 NLP professors! Apply by Dec 1 Strong consideration will be given to those who can tackle the below challenge: Can we use LM's hidden states to reason multiple problems simultaneously? ​​Retweets/shares appreciated🄰

I am hiring NLP/ML PhD students at UWaterloo, home to 5 NLP professors! Apply by Dec 1

Strong consideration will be given to those who can tackle the below challenge: Can we use LM's hidden states to reason multiple problems simultaneously?

​​Retweets/shares appreciated🄰
Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

What do people use ChatGPT for? We built WildVis, an interactive tool to visualize the embeddings of million-scale chat datasets like WildChat. Work done with Wenting Zhao Jack Hessel Sean Ren Claire Cardie Yejin Choi šŸ“huggingface.co/papers/2409.03… šŸ”—wildvisualizer.com/embeddings/eng… 1/7

FAIR Chemistry (@opencatalyst) 's Twitter Profile Photo

Introducing Meta’s Open Materials 2024 (OMat24) Dataset and Models! All under permissive open licenses for commercial and non-commercial use! Paper: arxiv.org/abs/2410.12771 Dataset: huggingface.co/datasets/fairc… Models: huggingface.co/fairchem/OMAT24 🧵1/x

Introducing Meta’s Open Materials 2024 (OMat24) Dataset and Models! All under permissive open licenses for commercial and non-commercial use!

Paper: arxiv.org/abs/2410.12771
Dataset: huggingface.co/datasets/fairc…
Models: huggingface.co/fairchem/OMAT24

🧵1/x
Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

How many reasoning tokens does OpenAI o1 use? It turns out they are almost always multiples of 64 (99+% of the time in 100K collected turns)šŸ¤”Could it be that the model only uses multiples of 64 tokens to think? Or maybe OpenAI rounds the token count in the returned usage? 1/4

How many reasoning tokens does OpenAI o1 use? It turns out they are almost always multiples of 64 (99+% of the time in 100K collected turns)šŸ¤”Could it be that the model only uses multiples of 64 tokens to think? Or maybe OpenAI rounds the token count in the returned usage? 1/4
NYU Center for Data Science (@nyudatascience) 's Twitter Profile Photo

Yuntian Deng from University of Waterloo presented at CILVR on internalizing reasoning in language models. His team's finetuning approach enabled GPT-2 Small to solve 20x20 multiplication with 99.5% accuracy, while standard training couldn't go beyond 4x4. Demo up at huggingface.co/spaces/yuntian…

<a href="/YuntianDeng/">Yuntian Deng</a> from <a href="/UWaterloo/">University of Waterloo</a> presented at CILVR on internalizing reasoning in language models. His team's finetuning approach enabled GPT-2 Small to solve 20x20 multiplication with 99.5% accuracy, while standard training couldn't go beyond 4x4.

Demo up at huggingface.co/spaces/yuntian…
Aaron Havens (@aaronjhavens) 's Twitter Profile Photo

New paper out with FAIR(+FAIR-Chemistry): Adjoint Sampling: Highly Scalable Diffusion Samplers via Adjoint Matching We present a scalable method for sampling from unnormalized densities beyond classical force fields. šŸ“„: arxiv.org/abs/2504.11713

Ricky T. Q. Chen (@rickytqchen) 's Twitter Profile Photo

We've open sourced Adjoint Sampling! It's part of a bundled release showcasing FAIR's research and open source commitment to AI for science. github.com/facebookresear… x.com/AIatMeta/statu…

Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

Can we build an operating system entirely powered by neural networks? Introducing NeuralOS: towards a generative OS that directly predicts screen images from user inputs. Try it live: neural-os.com Paper: huggingface.co/papers/2507.08… Inspired by Andrej Karpathy's vision. 1/5

Bing Yan (@bingyan4science) 's Twitter Profile Photo

Our new paper Digital Discovery is outšŸ’”Kyunghyun Cho Angelica Chen Do LLMs capture intrinsic chemistry, beyond string representations? Across SMILES & IUPAC: āŒ SOTA models: very low consistency (<1%) āš–ļø 1-to-1 mapped finetune ↑ consistency, not accuracy Read: pubs.rsc.org/en/Content/Art…

Our new paper <a href="/digital_rsc/">Digital Discovery</a> is outšŸ’”<a href="/kchonyc/">Kyunghyun Cho</a> <a href="/_angie_chen/">Angelica Chen</a>
Do LLMs capture intrinsic chemistry, beyond string representations? Across SMILES &amp; IUPAC:
āŒ SOTA models: very low consistency (&lt;1%)
āš–ļø 1-to-1 mapped finetune ↑ consistency, not accuracy
Read: pubs.rsc.org/en/Content/Art…
Yuntian Deng (@yuntiandeng) 's Twitter Profile Photo

šŸš€New dataset release: WildChat-4.8M 4.8M real user-ChatGPT conversations collected from our public chatbots: - 122K from reasoning models (o1-preview, o1-mini): represent real uses in the wild and very costly to collect - 2.5M from GPT-4o šŸ”— hf.co/datasets/allen… (1/4)