Gedas Bertasius (@gberta227) 's Twitter Profile
Gedas Bertasius

@gberta227

Assistant Professor at @unccs, previously a postdoc at @facebookai, PhD from @Penn, a basketball enthusiast.

ID: 1276571011152900096

linkhttps://www.gedasbertasius.com/ calendar_today26-06-2020 17:41:10

462 Tweet

1,1K Followers

952 Following

Mohaiminul (Emon) Islam (@mmiemon) 's Twitter Profile Photo

🚀 On the job market! Final-year PhD @ UNC Chapel Hill working on computer vision, video understanding, multimodal LLMs & AI agents. 2x Research Scientist Intern Meta 🔍 Seeking Research Scientist/Engineer roles! 🔗 md-mohaiminul.github.io 📧 mmiemon [at] cs [dot] unc [dot] edu

Ziyang Wang (@ziyangw00) 's Twitter Profile Photo

🚨Introducing Video-RTS: Resource-Efficient RL for Video Reasoning with Adaptive Video TTS! While RL-based video reasoning with LLMs has advanced, the reliance on large-scale SFT with extensive video data and long CoT annotations remains a major bottleneck. Video-RTS tackles

🚨Introducing Video-RTS: Resource-Efficient RL for Video Reasoning with Adaptive Video TTS! 

While RL-based video reasoning with LLMs has advanced, the reliance on large-scale SFT with extensive video data and long CoT annotations remains a major bottleneck.

Video-RTS tackles
Mohaiminul (Emon) Islam (@mmiemon) 's Twitter Profile Photo

Checkout our new paper: Video-RTS 🎥 A data-efficient RL method for complex video reasoning tasks. 🔹 Pure RL w/ output-based rewards. 🔹 Novel sparse-to-dense Test-Time Scaling (TTS) to expand input frames via self-consistency. 💥 96.4% less training data! More in the thread👇

Ziyang Wang (@ziyangw00) 's Twitter Profile Photo

🎉Our Video-RTS paper has been accepted at #EMNLP2025 Main!! We propose a novel video reasoning approach that combines data-efficient reinforcement learning (GRPO) with video-adaptive test-time scaling, improving reasoning performance while maintaining efficiency on multiple

Junlin (Hans) Han (@han_junlin) 's Twitter Profile Photo

Excited to share our new work: “Learning to See Before Seeing”! 🧠➡️👀 We investigate an interesting phenomeno: how do LLMs, trained only on text, learn about the visual world? Project page: junlinhan.github.io/projects/lsbs/

Excited to share our new work: “Learning to See Before Seeing”! 🧠➡️👀 We investigate an interesting phenomeno: how do LLMs, trained only on text, learn about the visual world? 
Project page:  junlinhan.github.io/projects/lsbs/
Zaid Khan (@codezakh) 's Twitter Profile Photo

How can an agent reverse engineer the underlying laws of an unknown, hostile & stochastic environment in “one life”, without millions of steps + human-provided goals / rewards? In our work, we: 1️⃣ infer an executable symbolic world model (a probabilistic program capturing

Gedas Bertasius (@gberta227) 's Twitter Profile Photo

Can AI models teach you to shoot like Steph Curry? 🏀 Come to my talk on Challenges in Expert-Level Skill Analysis at 4:30 pm in Room 318-A tomorrow (Sunday) to find out! sauafg-workshop.github.io #ICCV2025

Homanga Bharadhwaj (@mangahomanga) 's Twitter Profile Photo

I'll be joining the faculty Johns Hopkins University late next year as a tenure-track assistant professor in JHU Computer Science Looking for PhD students to join me tackling fun problems in robot manipulation, learning from human data, understanding+predicting physical interactions, and beyond!

I'll be joining the faculty <a href="/JohnsHopkins/">Johns Hopkins University</a> late next year as a tenure-track assistant professor in <a href="/JHUCompSci/">JHU Computer Science</a> 

Looking for PhD students to join me tackling fun problems in robot manipulation, learning from human data, understanding+predicting physical interactions, and beyond!
Zaid Khan (@codezakh) 's Twitter Profile Photo

🥳 Honored and grateful to be awarded an NDSEG Fellowship in Computer Science! 💫🇺🇸 Big thanks to my advisor Mohit Bansal for his guidance, and shoutout to my lab mates at UNC AI, collaborators, internship advisors, and mentors for their support 🤗 Excited to continue

Jaehong Yoon (on the faculty job market) (@jaeh0ng_yoon) 's Twitter Profile Photo

🎉 Excited to share that 5/5 of my papers (3 main, 2 findings) have been accepted at #EMNLP2025, in video/multimodal reasoning, instructional video editing, and efficient LLM adaptation & reasoning! 🚨 I’m recruiting Ph.D. students to join the Multimodal AI Group at NTU College

🎉 Excited to share that 5/5 of my papers (3 main, 2 findings) have been accepted at #EMNLP2025, in video/multimodal reasoning, instructional video editing, and efficient LLM adaptation &amp; reasoning!

🚨 I’m recruiting Ph.D. students to join the Multimodal AI Group at NTU College
Mohit Bansal (@mohitban47) 's Twitter Profile Photo

🚨 Check out our awesome students/postdocs' papers at #EMNLP2025 and say hi to them 👋! Also, I will give a keynote (virtually) on "Attributable, Conflict-Robust, and Multimodal Summarization with Multi-Source Retrieval" at the NewSumm workshop. -- Jaehong (in-person) finished

🚨 Check out our awesome students/postdocs' papers at #EMNLP2025 and say hi to them 👋! 

Also, I will give a keynote (virtually) on "Attributable, Conflict-Robust, and Multimodal Summarization with Multi-Source Retrieval" at the NewSumm workshop.

-- Jaehong (in-person) finished
Yiyang Zhou (@aiyiyangz) 's Twitter Profile Photo

🚨 BREAKING: AI Can't Actually See Videos. New benchmark shows mainstream LVLMs barely hit 60% accuracy—while humans reach 94.82%. This isn’t a glitch—it’s a fundamental failure in video understanding. LVLMs are doing visual theater, not real comprehension.

🚨 BREAKING: AI Can't Actually See Videos.
New benchmark shows mainstream LVLMs barely hit 60% accuracy—while humans reach 94.82%.
This isn’t a glitch—it’s a fundamental failure in video understanding. LVLMs are doing visual theater, not real comprehension.
Tanveer Hannan (@hannan_tanveer) 's Twitter Profile Photo

Our latest paper, DocSLM, developed during my internship at Microsoft, is now on arXiv: arxiv.org/abs/2511.11313. It is an efficient & compact Vision-Language Model to process long & complex documents while operating on resource-constrained edge devices like mobiles & laptops.

Eivinas Butkus (@eivinasbutkus) 's Twitter Profile Photo

1/ Can causal models and causal inference engines emerge through next-token prediction? Judea Pearl and others (Matej Zečević) have argued no. We present behavioral and mechanistic evidence that this is possible. #neurips2025 #NeurIPS

Tyler Zhu (@tyleryzhu) 's Twitter Profile Photo

Today seems to be a fitting day for Google DeepMind news, so I'm excited to announce our new preprint! Prior work suggests that text & img repr's are converging, albeit weakly. We found these same models actually have strong alignment; the inputs were too impoverished to see it!

Today seems to be a fitting day for <a href="/GoogleDeepMind/">Google DeepMind</a> news, so I'm excited to announce our new preprint!

Prior work suggests that text &amp; img repr's are converging, albeit weakly. We found these same models actually have strong alignment; the inputs were too impoverished to see it!