Dhruv Diddi (@dhruvdiddi) 's Twitter Profile
Dhruv Diddi

@dhruvdiddi

Powering Hardware Aware Inference | SLMs | Computer Vision | XGBoost | Formerly at Google, Turo #OwnYourAI #GetSoloTech #PhysicalAI

ID: 1110234349465538560

linkhttps://www.linkedin.com/in/dhruvdiddi/ calendar_today25-03-2019 17:37:50

335 Tweet

77 Followers

220 Following

Ji-Ha (@ji_ha_kim) 's Twitter Profile Photo

I stumbled upon a highly underrated paper It studies the dimensionality self-attention geometry via algebraic methods, focusing on a simplified version without softmax but also a short section with it at the end, they have some conjectures if anyone will can prove it

I stumbled upon a highly underrated paper
It studies the dimensionality self-attention geometry via algebraic methods, focusing on a simplified version without softmax but also a short section with it at the end, they have some conjectures if anyone will can prove it
Yohei Nishitsuji (@yoheinishitsuji) 's Twitter Profile Photo

float i,e,R,s;vec3 q,p,d=vec3(FC.xy/r-vec2(.6,.5),.7);for(q.zx--;i++<99.;){o.rgb+=hsv(.1,.2,min(e*s,.7-e)/35.);s=1.;p=q+=d*e*R*.1;p=vec3(log2(R=length(p))-t,exp(1.-p.z/R),atan(p.y,p.x)+cos(t)*.2);for(e=--p.y;s<3e2;s+=s)e+=sin(dot(sin(p.zxy*s)-.5,1.-cos(p.yxz*s)))/s;}#つぶやきGLSL

BlinkDL (@blinkdl_ai) 's Twitter Profile Photo

RWKV7-G1 "GooseOne" first release: reasoning @ 0.1b params, pure RNN (attention-free), fully multilingual. Demo & weights on RWKV.com 🪿 Larger G1 training in progress.

RWKV7-G1 "GooseOne" first release: reasoning @ 0.1b params, pure RNN (attention-free), fully multilingual. Demo &amp; weights on RWKV.com 🪿 Larger G1 training in progress.
Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

Qwen2.5-Math-7B-Instruct can scale to o1 level accuracy in only 32 rollouts. This paper's methods has a 4–16x better scaling rate over our deterministic search counterparts. Current inference-time scaling often relies on imperfect reward models that cause “reward hacking.”

Qwen2.5-Math-7B-Instruct can scale to o1 level accuracy in only 32 rollouts.

This paper's methods has a 4–16x better scaling rate over our deterministic search counterparts.

Current inference-time scaling often relies on imperfect reward models that cause “reward hacking.”
Charles Guillemet (@p3b7_) 's Twitter Profile Photo

At Ledger, you might know that we have the Ledger Donjon, our dedicated team constantly conducting open security research. We recently worked with Trezor, revealing that their Trezor Safe 3 was susceptible to physical supply chain attacks. Here's a thread on our findings:🧵

At <a href="/Ledger/">Ledger</a>, you might know that we have the <a href="/DonjonLedger/">Ledger Donjon</a>, our dedicated team constantly conducting open security research. 

We recently worked with Trezor, revealing that their Trezor Safe 3 was susceptible to physical supply chain attacks. Here's a thread on our findings:🧵
Andi Marafioti (@andimarafioti) 's Twitter Profile Photo

🚀We just dropped SmolDocling: a 256M open-source vision LM for complete document OCR!📄✨ It's lightning fast, process a page in 0.35 sec on consumer GPU using < 500MB VRAM⚡ SOTA in document conversion, beating every competing model we tested up to 27x larger🤯 But how? 🧶⬇️

🚀We just dropped SmolDocling: a 256M open-source vision LM for complete document OCR!📄✨
It's lightning fast, process a page in 0.35 sec on consumer GPU using &lt; 500MB VRAM⚡
SOTA in document conversion, beating every competing model we tested up to 27x larger🤯
But how? 🧶⬇️
Maxime Labonne (@maximelabonne) 's Twitter Profile Photo

⚡ AutoQuant I updated AutoQuant to make the GGUF versions of Gemma 3 abliterated. It implements imatrix and can split the model into multiple files. The GGUF code is based on gguf-my-repo, maintained by Xuan-Son Nguyen and Vaibhav (VB) Srivastav It also supports GPTQ, ExLlamaV2, AWQ, and HQQ!

⚡ AutoQuant

I updated AutoQuant to make the GGUF versions of Gemma 3 abliterated.

It implements imatrix and can split the model into multiple files.

The GGUF code is based on gguf-my-repo, maintained by <a href="/ngxson/">Xuan-Son Nguyen</a> and <a href="/reach_vb/">Vaibhav (VB) Srivastav</a> 

It also supports GPTQ, ExLlamaV2, AWQ, and HQQ!
Yohei Nishitsuji (@yoheinishitsuji) 's Twitter Profile Photo

float i,e,R,s;vec3 q,p,d=vec3(FC.xy/r,1);for(q.zy--;i++<229.;){e+=i/6e4;o.rgb+=hsv(q.z-.75,R,e*R*i/30.);s=2.;p=q+=d*e*R*R*.15;p=vec3(log(R=length(p-.2))-t*.25,exp(mod(-p.z,s)/R),p+sin(t)*.07);for(e=--p.y;s<4e3;s+=s)e+=-abs(dot(sin(p*s),max(p,exp(atan(.4,e))))/s*.1);}#つぶやきGLSL