Jaemin Cho (on faculty job market) (@jmin__cho) 's Twitter Profile
Jaemin Cho (on faculty job market)

@jmin__cho

On faculty job market! PhD candidate @UNCCS | @Bloomberg PhD Fellow | Prev: @GoogleAI @MSFTResearch @AdobeResearch @Allen_AI | ๐Ÿฆ‹: jmincho.bsky.social

ID: 243126428

linkhttps://j-min.io calendar_today26-01-2011 10:49:24

987 Tweet

1,1K Followers

1,1K Following

Chong Zeng (@iam_ncj) 's Twitter Profile Photo

What if a Transformer could render? Not text โ†’ image. But mesh โ†’ image โ€” with global illumination. No rasterizers. No ray-tracers. Just a Transformer without per-scene training. RenderFormer does exactly that. #SIGGRAPH2025 ๐Ÿ”—microsoft.github.io/renderformer

What if a Transformer could render?
Not text โ†’ image.
But mesh โ†’ image โ€” with global illumination.

No rasterizers. No ray-tracers. Just a Transformer without per-scene training.

RenderFormer does exactly that.

#SIGGRAPH2025 
๐Ÿ”—microsoft.github.io/renderformer
David Bau (@davidbau) 's Twitter Profile Photo

Dear MAGA friends, I have been worrying about STEM in the US a lot, because right now the Senate is writing new laws that cut 75% of the STEM budget in the US. Sorry for the long post, but the issue is really important, and I want to share what I know about it. The entire

Joykirat (@joykiratsingh) 's Twitter Profile Photo

Iโ€™m thrilled to share that Iโ€™ll be joining the University of North Carolina at Chapel Hill for my CS PhD this fall!! ๐ŸŽ“๐Ÿ’™ UNC-Chapel Hill Iโ€™ll be working with the amazing Mohit Bansal at UNC NLP. Grateful to everyone whoโ€™s supported me, excited for this new chapter! ๐Ÿš€

Minghao Wu (@wuminghao_nlp) 's Twitter Profile Photo

Excited to share that Iโ€™ll be joining UNC Computer Science and UNC NLP as a Postdoctoral Research Associate, working with the incredible Mohit Bansal! Canโ€™t wait to collaborate with the amazing students and faculty there! ๐ŸŽ‰ A huge thank you to my supervisor Reza Haffari, my colleagues at

Excited to share that Iโ€™ll be joining <a href="/unccs/">UNC Computer Science</a> and <a href="/uncnlp/">UNC NLP</a> as a Postdoctoral Research Associate, working with the incredible <a href="/mohitban47/">Mohit Bansal</a>! Canโ€™t wait to collaborate with the amazing students and faculty there! ๐ŸŽ‰

A huge thank you to my supervisor Reza Haffari, my colleagues at
Daeun Lee (@danadaeun) 's Twitter Profile Photo

Excited to share Video-Skill-CoT๐ŸŽฌ๐Ÿ› ๏ธโ€“ a new framework for domain-adaptive video reasoning with skill-aware Chain-of-Thought (CoT) supervision! โšก๏ธKey Highlights: โžก๏ธ Automatically extracts domain-specific reasoning skills from questions and organizes them into a unified taxonomy,

Jaemin Cho (on faculty job market) (@jmin__cho) 's Twitter Profile Photo

Introducing Video-Skill-CoT ๐Ÿ“ฝ๏ธ , a new framework for domain-adaptive video understanding with skill-specific chain-of-thought reasoning! โœ… Automatically discovers reasoning skills from video data โœ… Trains skill-specific expert modules with skill-specific CoT rationales โœ…

Jaehong Yoon (on the faculty job market) (@jaeh0ng_yoon) 's Twitter Profile Photo

๐Ÿšจ New Release: Video-Skill-CoT! Domain-Adaptive, Skill-Based Video Reasoning๐Ÿ’ก โœ… Automatically extracts domain-specific reasoning skills โœ… Generates tailored, skill-based CoT rationales โœ… Trains with skill-specific experts for stronger domain adaptation ๐Ÿš€ Outperforms

Zun Wang (@zunwang919) 's Twitter Profile Photo

๐ŸšจCheck my amazing labmate's latest work ๐ŸŽฌ Video-Skill-CoT ๐Ÿ› ๏ธ, a powerful and elegant framework for domain-adaptive video reasoning with skill-aware CoT ๐Ÿง โœจ, achieving strong results across multiple tasks! ๐Ÿ“Š๐Ÿ”ฅ

Elias Stengel-Eskin (on the faculty job market) (@eliaseskin) 's Twitter Profile Photo

๐Ÿšจ CLATTER treats entailment as a reasoning process, guiding models to follow concrete steps (decomposition, attribution/entailment, and aggregation). CLATTER improves hallucination detection via NLI, with gains on ClaimVerify, LFQA, and TofuEval especially on long-reasoning

Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

This paper proposes VIDEO-SKILL-COT to improve domain adaptation using skill-aware Chain-of-Thought supervision and expert learning modules. Methods ๐Ÿ”ง: โ†’ The framework automatically constructs skill-based Chain-of-Thought annotations by extracting skills from questions,

This paper proposes VIDEO-SKILL-COT to improve domain adaptation using skill-aware Chain-of-Thought supervision and expert learning modules.

Methods ๐Ÿ”ง:

โ†’ The framework automatically constructs skill-based Chain-of-Thought annotations by extracting skills from questions,
David Wan (@meetdavidwan) 's Twitter Profile Photo

Excited to share our new work, CLaMR! ๐Ÿš€ We tackle multimodal content retrieval by jointly considering video, speech, OCR, and metadata. CLaMR learns to dynamically pick the right modality for your query, boosting retrieval by 25 nDCG@10 over single modality retrieval! ๐Ÿง

Excited to share our new work, CLaMR! ๐Ÿš€

We tackle multimodal content retrieval by jointly considering video, speech, OCR, and metadata. CLaMR learns to dynamically pick the right modality for your query, boosting retrieval by 25 nDCG@10 over single modality retrieval!

๐Ÿง
Elias Stengel-Eskin (on the faculty job market) (@eliaseskin) 's Twitter Profile Photo

Excited to announce CLaMR, our new retriever for multimodal documents! Strong performance improvements (+25 nDGC@10) compared to both multimodal and unimodal retrieval baselines. ๐Ÿค CLaMR jointly encodes multiple modalities and selects the most relevant ones for each query. ๐Ÿ‹๏ธโ€โ™‚๏ธ

Han Wang (@hanwang98) 's Twitter Profile Photo

How can a multimodal retriever accurately retrieve docs from massive online video content that spans multiple modalities? We introduce CLaMR, a contextualized late-interaction retriever that jointly encodes all modalities and dynamically selects those containing the relevant

Jaemin Cho (on faculty job market) (@jmin__cho) 's Twitter Profile Photo

Introducing CLaMR -- a late-interaction retriever for complex multimodal video content! ๐Ÿ“ฝ๏ธ๐Ÿ“š โžก๏ธ Jointly encodes frames, speech, on-screen text, and metadata to answer diverse queries grounded across modalities โžก๏ธ Trained with a new dataset we introduce, MultiVENT 2.0++, a

Ziyang Wang (@ziyangw00) 's Twitter Profile Photo

Excited to present VideoTree๐ŸŒฒ at #CVPR2025 Fri at 10:30AM! VideoTree improves long-video QA via smart sampling: -Query-adaptive: finds the parts of the video relevant to the query -Coarse-to-fine structure: structured hierarchically to sample granularly from relevant segments

Omar Khattab (@lateinteraction) 's Twitter Profile Photo

Wow I missed this extra fancy ColBERT model. > A late-interaction retriever which jointly encodes/contextualizes information from many modalities, allowing for fine-grained matching between the query and implicitly finding the most relevant modality.

Mohit Bansal (@mohitban47) 's Twitter Profile Photo

Welcome Jaewoo to the MURGe-Lab + UNC NLP + UNC Computer Science family & the beautiful Chapel Hill + Research Triangle area! ๐ŸŽ‰ Looking forward to the exciting research and fun together in your PhD journey ๐Ÿ’™

hyunji amy lee (@hyunji_amy_lee) 's Twitter Profile Photo

๐Ÿšจ Want models to better utilize and ground on the provided knowledge? We introduce Context-INformed Grounding Supervision (CINGS)! Training LLM with CINGS significantly boosts grounding abilities in both text and vision-language models compared to standard instruction tuning.

๐Ÿšจ Want models to better utilize and ground on the provided knowledge? We introduce Context-INformed Grounding Supervision (CINGS)! Training LLM with CINGS significantly boosts grounding abilities in both text and vision-language models compared to standard instruction tuning.