Jaemin Cho (on faculty job market) (@jmin__cho) 's Twitter Profile
Jaemin Cho (on faculty job market)

@jmin__cho

On faculty job market! PhD candidate @UNCCS | @Bloomberg PhD Fellow | Prev: @GoogleAI @MSFTResearch @AdobeResearch @Allen_AI | 🦋: jmincho.bsky.social

ID: 243126428

linkhttps://j-min.io calendar_today26-01-2011 10:49:24

987 Tweet

1,1K Followers

1,1K Following

Chong Zeng (@iam_ncj) 's Twitter Profile Photo

What if a Transformer could render? Not text → image. But mesh → image — with global illumination. No rasterizers. No ray-tracers. Just a Transformer without per-scene training. RenderFormer does exactly that. #SIGGRAPH2025 🔗microsoft.github.io/renderformer

What if a Transformer could render?
Not text → image.
But mesh → image — with global illumination.

No rasterizers. No ray-tracers. Just a Transformer without per-scene training.

RenderFormer does exactly that.

#SIGGRAPH2025 
🔗microsoft.github.io/renderformer
David Bau (@davidbau) 's Twitter Profile Photo

Dear MAGA friends, I have been worrying about STEM in the US a lot, because right now the Senate is writing new laws that cut 75% of the STEM budget in the US. Sorry for the long post, but the issue is really important, and I want to share what I know about it. The entire

Joykirat (@joykiratsingh) 's Twitter Profile Photo

I’m thrilled to share that I’ll be joining the University of North Carolina at Chapel Hill for my CS PhD this fall!! 🎓💙 UNC-Chapel Hill I’ll be working with the amazing Mohit Bansal at UNC NLP. Grateful to everyone who’s supported me, excited for this new chapter! 🚀

Minghao Wu (@wuminghao_nlp) 's Twitter Profile Photo

Excited to share that I’ll be joining UNC Computer Science and UNC NLP as a Postdoctoral Research Associate, working with the incredible Mohit Bansal! Can’t wait to collaborate with the amazing students and faculty there! 🎉 A huge thank you to my supervisor Reza Haffari, my colleagues at

Excited to share that I’ll be joining <a href="/unccs/">UNC Computer Science</a> and <a href="/uncnlp/">UNC NLP</a> as a Postdoctoral Research Associate, working with the incredible <a href="/mohitban47/">Mohit Bansal</a>! Can’t wait to collaborate with the amazing students and faculty there! 🎉

A huge thank you to my supervisor Reza Haffari, my colleagues at
Daeun Lee (@danadaeun) 's Twitter Profile Photo

Excited to share Video-Skill-CoT🎬🛠️– a new framework for domain-adaptive video reasoning with skill-aware Chain-of-Thought (CoT) supervision! ⚡️Key Highlights: ➡️ Automatically extracts domain-specific reasoning skills from questions and organizes them into a unified taxonomy,

Jaemin Cho (on faculty job market) (@jmin__cho) 's Twitter Profile Photo

Introducing Video-Skill-CoT 📽️ , a new framework for domain-adaptive video understanding with skill-specific chain-of-thought reasoning! ✅ Automatically discovers reasoning skills from video data ✅ Trains skill-specific expert modules with skill-specific CoT rationales ✅

Jaehong Yoon (on the faculty job market) (@jaeh0ng_yoon) 's Twitter Profile Photo

🚨 New Release: Video-Skill-CoT! Domain-Adaptive, Skill-Based Video Reasoning💡 ✅ Automatically extracts domain-specific reasoning skills ✅ Generates tailored, skill-based CoT rationales ✅ Trains with skill-specific experts for stronger domain adaptation 🚀 Outperforms

Zun Wang (@zunwang919) 's Twitter Profile Photo

🚨Check my amazing labmate's latest work 🎬 Video-Skill-CoT 🛠️, a powerful and elegant framework for domain-adaptive video reasoning with skill-aware CoT 🧠✨, achieving strong results across multiple tasks! 📊🔥

Elias Stengel-Eskin (on the faculty job market) (@eliaseskin) 's Twitter Profile Photo

🚨 CLATTER treats entailment as a reasoning process, guiding models to follow concrete steps (decomposition, attribution/entailment, and aggregation). CLATTER improves hallucination detection via NLI, with gains on ClaimVerify, LFQA, and TofuEval especially on long-reasoning

Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

This paper proposes VIDEO-SKILL-COT to improve domain adaptation using skill-aware Chain-of-Thought supervision and expert learning modules. Methods 🔧: → The framework automatically constructs skill-based Chain-of-Thought annotations by extracting skills from questions,

This paper proposes VIDEO-SKILL-COT to improve domain adaptation using skill-aware Chain-of-Thought supervision and expert learning modules.

Methods 🔧:

→ The framework automatically constructs skill-based Chain-of-Thought annotations by extracting skills from questions,
David Wan (@meetdavidwan) 's Twitter Profile Photo

Excited to share our new work, CLaMR! 🚀 We tackle multimodal content retrieval by jointly considering video, speech, OCR, and metadata. CLaMR learns to dynamically pick the right modality for your query, boosting retrieval by 25 nDCG@10 over single modality retrieval! 🧐

Excited to share our new work, CLaMR! 🚀

We tackle multimodal content retrieval by jointly considering video, speech, OCR, and metadata. CLaMR learns to dynamically pick the right modality for your query, boosting retrieval by 25 nDCG@10 over single modality retrieval!

🧐
Elias Stengel-Eskin (on the faculty job market) (@eliaseskin) 's Twitter Profile Photo

Excited to announce CLaMR, our new retriever for multimodal documents! Strong performance improvements (+25 nDGC@10) compared to both multimodal and unimodal retrieval baselines. 🤝 CLaMR jointly encodes multiple modalities and selects the most relevant ones for each query. 🏋️‍♂️

Han Wang (@hanwang98) 's Twitter Profile Photo

How can a multimodal retriever accurately retrieve docs from massive online video content that spans multiple modalities? We introduce CLaMR, a contextualized late-interaction retriever that jointly encodes all modalities and dynamically selects those containing the relevant

Jaemin Cho (on faculty job market) (@jmin__cho) 's Twitter Profile Photo

Introducing CLaMR -- a late-interaction retriever for complex multimodal video content! 📽️📚 ➡️ Jointly encodes frames, speech, on-screen text, and metadata to answer diverse queries grounded across modalities ➡️ Trained with a new dataset we introduce, MultiVENT 2.0++, a

Ziyang Wang (@ziyangw00) 's Twitter Profile Photo

Excited to present VideoTree🌲 at #CVPR2025 Fri at 10:30AM! VideoTree improves long-video QA via smart sampling: -Query-adaptive: finds the parts of the video relevant to the query -Coarse-to-fine structure: structured hierarchically to sample granularly from relevant segments

Omar Khattab (@lateinteraction) 's Twitter Profile Photo

Wow I missed this extra fancy ColBERT model. > A late-interaction retriever which jointly encodes/contextualizes information from many modalities, allowing for fine-grained matching between the query and implicitly finding the most relevant modality.

Mohit Bansal (@mohitban47) 's Twitter Profile Photo

Welcome Jaewoo to the MURGe-Lab + UNC NLP + UNC Computer Science family & the beautiful Chapel Hill + Research Triangle area! 🎉 Looking forward to the exciting research and fun together in your PhD journey 💙

hyunji amy lee (@hyunji_amy_lee) 's Twitter Profile Photo

🚨 Want models to better utilize and ground on the provided knowledge? We introduce Context-INformed Grounding Supervision (CINGS)! Training LLM with CINGS significantly boosts grounding abilities in both text and vision-language models compared to standard instruction tuning.

🚨 Want models to better utilize and ground on the provided knowledge? We introduce Context-INformed Grounding Supervision (CINGS)! Training LLM with CINGS significantly boosts grounding abilities in both text and vision-language models compared to standard instruction tuning.