Shenbin Qian (@shenbinqian) 's Twitter Profile
Shenbin Qian

@shenbinqian

PhD Student at University of Surrey in Natural Language Processing

ID: 2856348493

calendar_today15-10-2014 04:43:58

32 Tweet

30 Followers

101 Following

AK (@_akhaliq) 's Twitter Profile Photo

Toolformer: Language Models Can Teach Themselves to Use Tools introduce Toolformer, a model trained to decide which APIs to call, when to call them, what arguments to pass, and how to best incorporate the results into future token prediction abs: arxiv.org/abs/2302.04761

Toolformer: Language Models Can Teach Themselves to Use Tools 

introduce Toolformer, a model trained to decide which APIs to call, when to call them, what arguments to pass, and how to best incorporate the results into future token prediction 

abs: arxiv.org/abs/2302.04761
Chen Lu (@_chen_lu_) 's Twitter Profile Photo

I wrote a UNet diffusion model in pure CUDA: github.com/clu0/unet.cu This project was inspired by Andrej Karpathy 's llm.c (github.com/karpathy/llm.c). I also learnt a lot about CUDA kernels from Simon Boehm 's Matmul blog (siboehm.com/articles/22/CU…). (1/3)

Sanchit Gandhi (@sanchitgandhi99) 's Twitter Profile Photo

Local Gemma: a package for running LARGE language models with SMALL amounts of memory 🤏 The Gemma-2 27b model typically requires 70GB GPU memory. With Local Gemma, this is just 5GB 🤯 Try it for yourself in 2-steps: pip install local-gemma local-gemma --preset memory_extreme

Local Gemma: a package for running LARGE language models with SMALL amounts of memory 🤏

The Gemma-2 27b model typically requires 70GB GPU memory. With Local Gemma, this is just 5GB 🤯

Try it for yourself in 2-steps:

pip install local-gemma
local-gemma --preset memory_extreme
Tanishq Mathew Abraham, Ph.D. (@iscienceluvr) 's Twitter Profile Photo

AI Agents That Matter abs: arxiv.org/abs/2407.01502 Performs a careful analysis of existing benchmarks, analyzing across additional axes like cost, proposes new baselines 1. AI agent evaluations must be cost-controlled 2. Jointly optimizing accuracy and cost can yield better

AI Agents That Matter

abs: arxiv.org/abs/2407.01502

Performs a careful analysis of existing benchmarks, analyzing across additional axes like cost, proposes new baselines

1. AI agent evaluations must be cost-controlled
2. Jointly optimizing accuracy and cost can yield better
Unsloth AI (@unslothai) 's Twitter Profile Photo

We made a step-by-step tutorial on how to finetune Llama-3 with Google Colab & deploy it to @Ollama Tutorial: docs.unsloth.ai/tutorials/how-… Colab notebook: colab.research.google.com/drive/1WZDi7AP… Blog post & video coming soon. 🦥

We made a step-by-step tutorial on how to finetune Llama-3 with Google Colab & deploy it to @Ollama

Tutorial: docs.unsloth.ai/tutorials/how-…

Colab notebook: colab.research.google.com/drive/1WZDi7AP…

Blog post & video coming soon. 🦥
Zhaorun Chen @ICLR2025 (@zrchen_aisafety) 's Twitter Profile Photo

We know LLM agents 🤖 are powerful and popular these days, but can they be subverted to act as killer agents 😈 just like in Westworld?😱 Sadly, the answer is YES! 😱😱 🔥🔥 We reveal the vulnerability and potential threats of generic LLM agents in our new work AgentPoison:

We know LLM agents 🤖 are powerful and popular these days, but can they be subverted to act as killer agents 😈 just like in Westworld?😱

Sadly, the answer is YES! 😱😱

🔥🔥 We reveal the vulnerability and potential threats of generic LLM agents in our new work AgentPoison:
Andrej Karpathy (@karpathy) 's Twitter Profile Photo

To help explain the weirdness of LLM Tokenization I thought it could be amusing to translate every token to a unique emoji. This is a lot closer to truth - each token is basically its own little hieroglyph and the LLM has to learn (from scratch) what it all means based on

To help explain the weirdness of LLM Tokenization I thought it could be amusing to translate every token to a unique emoji. This is a lot closer to truth - each token is basically its own little hieroglyph and the LLM has to learn (from scratch) what it all means based on
PyTorch (@pytorch) 's Twitter Profile Photo

Introducing FlexAttention: a new API that lets you implement diverse attention variants in just a few lines of idiomatic PyTorch code. 🔥 Check out the blog post for more details: hubs.la/Q02KsKNR0

Introducing FlexAttention: a new API that lets you implement diverse attention variants in just a few lines of idiomatic PyTorch code. 🔥

Check out the blog post for more details: hubs.la/Q02KsKNR0
AI at Meta (@aiatmeta) 's Twitter Profile Photo

Using structured weight pruning and knowledge distillation, the NVIDIA AI research team refined Llama 3.1 8B into a new Llama-3.1-Minitron 4B. They're releasing the new models on Hugging Face and shared a deep dive on how they did it ➡️ go.fb.me/b2h2c8

Using structured weight pruning and knowledge distillation, the <a href="/NVIDIAAI/">NVIDIA AI</a> research team refined Llama 3.1 8B into a new Llama-3.1-Minitron 4B. 

They're releasing the new models on <a href="/huggingface/">Hugging Face</a> and shared a deep dive on how they did it ➡️ go.fb.me/b2h2c8
Shenbin Qian (@shenbinqian) 's Twitter Profile Photo

Thank you Sui Sui HE スイ for the invitation! I'm looking forward to sharing my research on MT evaluation at the upcoming seminar. Hope to see many of you there for an engaging discussion!

CTS Surrey (@cts_surrey) 's Twitter Profile Photo

Due to its popularity, we are extending our 'ASR for interpreting' course applications! Are you a company interested in #ASR & #AI for #interpreting? Fill in our survey to share your experience & join a 4-day online course on #bespoke interpreting ASR. 🤩surreyfbel.qualtrics.com/jfe/form/SV_8u…

Due to its popularity, we are extending our 'ASR for interpreting' course applications! Are you a company interested in #ASR &amp; #AI for #interpreting? Fill in our survey to share your experience &amp; join a 4-day online course on #bespoke interpreting ASR.

🤩surreyfbel.qualtrics.com/jfe/form/SV_8u…
Cheng Han Chiang (姜成翰) (@dcml0714) 's Twitter Profile Photo

🚀 New Paper Alert! 🚀 Want better LLM-as-a-Judge? TRACT: 🧠 CoT + Regression-Aware Fine-tuning (RAFT) = Better numerical predictions! 📊 arxiv.org/abs/2503.04381 🧵👇 A thread on TRACT:

🚀 New Paper Alert! 🚀
Want better LLM-as-a-Judge?
TRACT: 🧠 CoT + Regression-Aware Fine-tuning (RAFT) = Better numerical predictions! 📊
arxiv.org/abs/2503.04381
🧵👇 A thread on TRACT: