Andreas Grivas (@andreasgrv) 's Twitter Profile
Andreas Grivas

@andreasgrv

Interested in Bottlenecks in Neural Networks; Unargmaxable Outputs.

Postdoc in ML/NLP at the University of Edinburgh.

ID: 3468362176

linkhttps://grv.unargmaxable.ai/ calendar_today28-08-2015 12:21:01

544 Tweet

505 Followers

658 Following

antonio vergari - hiring PhD students (@tetraduzione) 's Twitter Profile Photo

Less than two weeks to submit your papers on: 📈 #lowrank adapters and #factorizations 🧊 #tensor networks 🔌 probabilistic #circuits 🎓 #theory of factorizations to the first workshop on connecting them in #AI #ML at AAAI please share! 🔁 👇👇👇 april-tools.github.io/colorai/

antonio vergari - hiring PhD students (@tetraduzione) 's Twitter Profile Photo

If in the ELLIS application portal you select one of these topics: 🎲Bayesian & Probabilistic Learning 🦾Robust & Trustworthy ML 🧠Symbolic Machine Learning 🥽Unsupervised Learning consider listing me as a supervisor and check what we do here: april-tools.github.io

Tom Hosking (@tomhosking) 's Twitter Profile Photo

I'll be presenting this paper as a poster on Weds 1600-1730! Come and chat about combining discrete hierarchies with LLMs for opinion summarisation 🥳

Verna Dankers (@vernadankers) 's Twitter Profile Photo

My "looking for a postdoc" stickers and I are in Miami for #EMNLP2024! 🤩 Do you have/know of a postdoc for summer/autumn '25 related to interpretability, figlang and/or memorisation (vs generalisation)? Reach out! Looking forward to #GenBench2024 on Saturday and the many many...

My "looking for a postdoc" stickers and I are in Miami for #EMNLP2024! 🤩 Do you have/know of a postdoc for summer/autumn '25 related to interpretability, figlang and/or memorisation (vs generalisation)? Reach out! Looking forward to #GenBench2024 on Saturday and the many many...
Piotr Nawrot (@p_nawrot) 's Twitter Profile Photo

🥳 We release educational implementations of recent plug-and-play Sparse Attention methods 🥳 While plug-and-play Sparse Attention offers a powerful way to improve LLM inference efficiency, most implementations rely on complex Triton or custom CUDA kernels to match dense

🥳 We release educational implementations of recent plug-and-play Sparse Attention methods 🥳

While plug-and-play Sparse Attention offers a powerful way to improve LLM inference efficiency, most implementations rely on complex Triton or custom CUDA kernels to match dense
Edoardo Ponti (@pontiedoardo) 's Twitter Profile Photo

Another nano.* gem from my amazing student Piotr Nawrot: a repo on SOTA sparse attention methods (both for prefilling and generation) It will also feature in the NeurIPS Conference tutorial "Dynamic Sparsity in Machine Learning" I will give with Andre Martins!

Edoardo Ponti (@pontiedoardo) 's Twitter Profile Photo

Is sparsity the key to conditional computation, interpretability, long context/generation, and more in foundation models? Find out at my #NeurIPS2024 tutorial on Dynamic Sparsity in Machine Learning with Andre Martins! Followed by a panel with Sara Hooker and Alessandro Sordoni 🧵

Andre Martins (@andre_t_martins) 's Twitter Profile Photo

Come to our #NeurIPS2024 tutorial on Tuesday Dec 10 9:30 AM to find the answers to these questions (or to ask new questions!) 🔬

Come to our #NeurIPS2024 tutorial on Tuesday Dec 10 9:30 AM to find the answers to these questions (or to ask new questions!) 🔬
Andreas Grivas (@andreasgrv) 's Twitter Profile Photo

📕⬇️ My thesis on 🚫unargmaxable outputs is online! Check it out if you want to learn more about how output layers constrain what neural networks can and cannot predict 👉 era.ed.ac.uk/handle/1842/42…

Edoardo Ponti (@pontiedoardo) 's Twitter Profile Photo

I have been elected an ELLIS Scholar! On a related note: it's more urgent than ever for ELLIS to take a leading role in advancing an AI plan for Europe and developing foundation models autonomously. DeepSeek-R1 showed us it's possible. ellis.eu/fellows-schola…

NeSy 2025 (@nesyconf) 's Twitter Profile Photo

Don't forget to submit to NeSy 2025🐲! We have two deadlines this year, so no excuse to submit (but be careful to be on time with US visa's!)

Edoardo Ponti (@pontiedoardo) 's Twitter Profile Photo

I have a scholarship for a PhD on efficient memory and tokenization in LLMs at EdinburghNLP! Eligibility: UK home fee status Starting date: from July 2025 onwards. informatics.ed.ac.uk/study-with-us/… Please contact me if you're interested!

Piotr Nawrot (@p_nawrot) 's Twitter Profile Photo

Sparse attention is one of the most promising strategies to unlock long-context processing and long generation reasoning in LLMs. We performed the most comprehensive study on training-free sparse attention to date. Here is what we found:

Sparse attention is one of the most promising strategies to unlock long-context processing and long generation reasoning in LLMs.

We performed the most comprehensive study on training-free sparse attention to date.

Here is what we found:
Piotr Nawrot (@p_nawrot) 's Twitter Profile Photo

We built sparse-frontier — a clean abstraction that lets you focus on your custom sparse attention implementation while automatically inheriting vLLM’s optimizations and model support. As a PhD student, I've learned that sometimes the bottleneck in research isn't ideas — it's

We built sparse-frontier — a clean abstraction that lets you focus on your custom sparse attention implementation while automatically inheriting vLLM’s optimizations and model support.

As a PhD student, I've learned that sometimes the bottleneck in research isn't ideas — it's