Aditya Kane (@adityakane1) 's Twitter Profile
Aditya Kane

@adityakane1

Issuing one CTA at a time.

ID: 1244310176225787907

linkhttp://adityakane2001.github.io calendar_today29-03-2020 17:07:29

569 Tweet

181 Followers

374 Following

Humphrey Shi (@humphrey_shi) 's Twitter Profile Photo

We are releasing a major NATTEN upgrade that brings you new Hopper & Blackwell sparse attention kernels, both capable of realizing Theoretical Max Speedup: 90% sparsity -> 10X speedup. Thanks to the great efforts by Ali Hassani & @NVIDIA cutlass team! natten.org

We are releasing a major NATTEN upgrade that brings you new Hopper & Blackwell sparse attention kernels, both capable of realizing Theoretical Max Speedup: 
90% sparsity -> 10X speedup.

Thanks to the great efforts by <a href="/AliHassaniJr/">Ali Hassani</a> &amp; @NVIDIA cutlass team!

natten.org
Sharon Goldman (@sharongoldman) 's Twitter Profile Photo

Exclusive: Ex-Meta AI leaders debut an agent that scours the web for you in a push to ultimately give users their own digital ‘chief of staff’ Devi Parikh Abhishek Das @DhruvBatraDB fortune.com/2025/06/10/exc…

Devi Parikh (@deviparikh) 's Twitter Profile Photo

It's here! Introducing Scouts by Yutori. Scouts is like having a team of agents monitoring the web for information that matters to you. We're letting more users in everyday. Join the waitlist!

Dhruv Batra (@dhruvbatradb) 's Twitter Profile Photo

Scouts by Yutori. AI agents that monitor the web for things you care about. So you can focus on the meaningful things in life and experience a bit more yutori.

Aditya Kane (@adityakane1) 's Twitter Profile Photo

Scouts from Yutori keep tabs on things you care about so you never miss an important update! An update like AMD securing a big win in the AI chip market :)

Scouts from <a href="/yutori_ai/">Yutori</a>  keep tabs on things you care about so you never miss an important update!

An update like AMD securing a big win in the AI chip market :)
Stas Bekman (@stasbekman) 's Twitter Profile Photo

Oh wow, the newly released nccl finally started to use fp32 accumulation for reduction ops with half precision inputs! This is so important! Thank you NCCL team! github.com/NVIDIA/nccl/co… I'd imagine we will see this version in pytorch>=2.8 I hope, unless you build your own nccl.

Oh wow, the newly released nccl finally started to use fp32 accumulation for reduction ops with half precision inputs! This is so important! Thank you NCCL team!

github.com/NVIDIA/nccl/co…

I'd imagine we will see this version in pytorch&gt;=2.8 I hope, unless you build your own nccl.
Ali Hassani (@alihassanijr) 's Twitter Profile Photo

Cosmos-Predict2 meets NATTEN. We just released variants of Cosmos-Predict2 where we replace most self attentions with neighborhood attention, bringing up to 2.6X end-to-end speedup, with minimal effect on quality! github.com/nvidia-cosmos/… (1/5)

Humphrey Shi (@humphrey_shi) 's Twitter Profile Photo

Sparse Attention is now pushing World Foundation Models to the Speed of Light! Attention powers modern AI (Transformers, ViTs, DiTs), and Sparse Attention is the next frontier. Neighborhood Attention (NA) is the first multidimensional sparse attention infrastructure that: -

Ali Hassani (@alihassanijr) 's Twitter Profile Photo

Watch my talk about NATTEN on GPU MODE this Saturday at 3PM ET / noon PT. I'll go over all the exciting new features we shipped very recently, especially our Hopper and Blackwell FNA kernels, now speeding up video / world models by up to 2.6X e2e! youtube.com/watch?v=mF_H_J

Humphrey Shi (@humphrey_shi) 's Twitter Profile Photo

Check out Ali's talk tomorrow on GPU MODE if you breathe GPUs! This is our 2nd GPU MODE talk — last time we unveiled Distributed GEMM: a CUTLASS-based Tensor Parallelism implementation that helps push NVL-based AI systems to the next level - transforming a network of GPUs into

Check out Ali's talk tomorrow on <a href="/GPU_MODE/">GPU MODE</a> if you breathe GPUs! 
This is our 2nd GPU MODE talk — last time we unveiled Distributed GEMM: a CUTLASS-based Tensor Parallelism implementation that helps push NVL-based AI systems to the next level - transforming a network of GPUs into
FFmpeg (@ffmpeg) 's Twitter Profile Photo

Not sure why but we have lots of new followers! FFmpeg makes extensive use of hand-written assembly code for huge (10-50x) speed increases so we are providing assembly lessons to teach a new generation of assembly language programmers. Learn more here: github.com/FFmpeg/asm-les…