Siddharth Joshi (@sjoshi804) 's Twitter Profile
Siddharth Joshi

@sjoshi804

ML PhD at @UCLA under @baharanm | Data Curation for Efficient & Robust SSL | Prev @MSFTResearch, @Cisco Research, @Microsoft

ID: 892839778390917120

linkhttp://sjoshi804.github.io calendar_today02-08-2017 20:09:26

105 Tweet

1,1K Takipรงi

2,2K Takip Edilen

Yihe Deng (@yihe__deng) 's Twitter Profile Photo

๐Ÿ“ข New paper alert! Introducing STIC (Self-Training on Image Comprehension) that enhances the understanding and reasoning capabilities of LVLMs through self-generated data ๐ŸŒŸ ๐Ÿ“„ Read the paper: arxiv.org/abs/2405.19716 ๐Ÿ”— Project page: stic-lvlm.github.io ๐Ÿ’ป GitHub Repo:

๐Ÿ“ข New paper alert!

Introducing STIC (Self-Training on Image Comprehension) that enhances the understanding and reasoning capabilities of LVLMs through self-generated data ๐ŸŒŸ

๐Ÿ“„ Read the paper: arxiv.org/abs/2405.19716
๐Ÿ”— Project page: stic-lvlm.github.io
๐Ÿ’ป GitHub Repo:
Siddharth Joshi (@sjoshi804) 's Twitter Profile Photo

Starting my internship with Microsoft Research AI Frontiers Team in Redmond today! Excited for the next 3 months here. Hmu if you're around in the greater Seattle area, would love to chat about data-efficient and robust learning!

Starting my internship with <a href="/MSFTResearch/">Microsoft Research</a> AI Frontiers Team in Redmond today!

Excited for the next 3 months here.  Hmu if you're around in the greater Seattle area, would love to chat about data-efficient and robust learning!
Siddharth Joshi (@sjoshi804) 's Twitter Profile Photo

๐Ÿš€ Exciting News! ๐Ÿš€ Join Baharan Mirzasoleiman and me for a 2-hour tutorial on Data-Efficient Learning! Learn the principles behind data curation: the secret sauce powering todayโ€™s AI revolution! โšก๏ธ See you at 1pm on Monday CEST in Hall A8! ๐Ÿ™Œ ๐Ÿ”— More details: sjoshi804.github.io/data-efficientโ€ฆ

ROHAN WADHAWAN (@rohanwadhawan7) 's Twitter Profile Photo

๐Ÿš€ Exciting news! ConTextual (con-textual.github.io) is headed to Vienna for #ICML2024! ๐ŸŽ‰ ๐Ÿ“Š Leaderboard updates: GPT4o-mini is now 2nd, just 1% behind GPT4o๐Ÿฅณ Claude-3.5-Sonnet takes 3rd, outperforming Claude-3-Opus by 19% ๐Ÿ˜ฒ

Besmira Nushi ๐Ÿ’™๐Ÿ’› (@besanushi) 's Twitter Profile Photo

Excited to announce the release of Eureka, an open-source framework for evaluating and understanding large foundation models! ๐ŸŒŸ Eureka offers: ๐Ÿ”In-depth analysis of 12 cutting-edge models ๐Ÿง  Multimodal & language capability testing beyond single-score reporting and rankings ๐Ÿ“ˆ

Excited to announce the release of Eureka, an open-source framework for evaluating and understanding large foundation models! ๐ŸŒŸ

Eureka offers: ๐Ÿ”In-depth analysis of 12 cutting-edge models ๐Ÿง  Multimodal &amp; language capability testing beyond single-score reporting and rankings ๐Ÿ“ˆ
Natasha Butt (@natashaeve4) 's Twitter Profile Photo

Introducing BenchAgents: a framework for automated benchmark creation, using multiple LLM agents that interact with each other and with developers to generate diverse, high-quality, and challenging benchmarks w/ Varun Chandrasekaran Neel Joshi Besmira Nushi ๐Ÿ’™๐Ÿ’› Vidhisha Balachandran Microsoft Research ๐Ÿงต1/8

Introducing BenchAgents: a framework for automated benchmark creation, using multiple LLM agents that interact with each other and with developers to generate diverse, high-quality, and challenging benchmarks w/ <a href="/VarunChandrase3/">Varun Chandrasekaran</a> <a href="/neelsj/">Neel Joshi</a> <a href="/besanushi/">Besmira Nushi ๐Ÿ’™๐Ÿ’›</a> <a href="/vidhisha_b/">Vidhisha Balachandran</a> <a href="/MSFTResearch/">Microsoft Research</a> ๐Ÿงต1/8
Yu Yang (@yuyang_i) 's Twitter Profile Photo

1/ I'll be at #NeurIPS2024 presenting our work SmallToLarge (S2L): Data-efficient Fine-tuning of LLMs! ๐Ÿš€ Whatโ€™s S2L? Itโ€™s a scalable data selection method that trains a small proxy model to guide fine-tuning for larger models, reducing costs while preserving performance. ๐Ÿ‘‡

1/ I'll be at #NeurIPS2024 presenting our work SmallToLarge (S2L): Data-efficient Fine-tuning of LLMs! ๐Ÿš€

Whatโ€™s S2L? Itโ€™s a scalable data selection method that trains a small proxy model to guide fine-tuning for larger models, reducing costs while preserving performance. ๐Ÿ‘‡