Xin Wang (@xinw_ai) 's Twitter Profile
Xin Wang

@xinw_ai

Researcher @OpenAI | ex Microsoft Research, Apple AI/ML | @Berkeley_EECS PhD

ID: 1656799478

linkhttps://xinw.ai/ calendar_today09-08-2013 03:31:12

245 Tweet

4,4K Takipçi

1,1K Takip Edilen

Kaichun Mo (@kaichunmo) 's Twitter Profile Photo

Presenting STOW at #CoRL2023 our latest work learning to segment and track previously unseen objects for robot stowing and fetching on cluttered shelves in warehouses. Check out our poster at Poster Sec. 6 on Thur 😄

Besmira Nushi 💙💛 (@besanushi) 's Twitter Profile Photo

Our internship job post on Evaluating & Understanding Foundation Models is out. Sharing here a list of open challenges our team is excited to explore together with future research interns. Application link: jobs.careers.microsoft.com/global/en/job/… Vibhav Vineet Neel Joshi @hmd_palangi Ece Kamar

Our internship job post on Evaluating & Understanding Foundation Models is out. Sharing here a list of open challenges our team is excited to explore together with future research interns. Application link: jobs.careers.microsoft.com/global/en/job/…

<a href="/VibhavVineet/">Vibhav Vineet</a> Neel Joshi @hmd_palangi <a href="/ecekamar/">Ece Kamar</a>
Ilija Radosavovic (@ir413) 's Twitter Profile Photo

we have trained a humanoid transformer with large-scale reinforcement learning in simulation and deployed it to the real world zero-shot

Microsoft Research (@msftresearch) 's Twitter Profile Photo

Today, we share our teams’ latest contributions, Phi-2 and promptbase. Phi-2 outperforms other existing small language models, yet it’s small enough to run on a laptop or mobile device. msft.it/6040ipYH6

AK (@_akhaliq) 's Twitter Profile Photo

Microsoft releases Phi-2 on Hugging Face model: huggingface.co/microsoft/phi-2 a 2.7 billion-parameter language model that demonstrates outstanding reasoning and language understanding capabilities, showcasing state-of-the-art performance among base language models with less than 13

Microsoft releases Phi-2 on <a href="/huggingface/">Hugging Face</a> 

model: huggingface.co/microsoft/phi-2

a 2.7 billion-parameter language model that demonstrates outstanding reasoning and language understanding capabilities, showcasing state-of-the-art performance among base language models with less than 13
Sebastien Bubeck (@sebastienbubeck) 's Twitter Profile Photo

Check out this short video for a brief discussion of the phi series with Microsoft CTO Kevin Scott , including why "textbooks" in "Textbooks Are All You Need" might not be exactly what you have in mind. youtu.be/O-DjHgZt-Uk?si…

Sebastien Bubeck (@sebastienbubeck) 's Twitter Profile Photo

We're so pumped to see phi-2 at the top of trending models on Hugging Face ! It's sibling phi-1.5 has already half a million downloads. Can't wait to see the mechanistic interpretability works that will come out of this & their impact on all the important LLM research questions!

We're so pumped to see phi-2 at the top of trending models on <a href="/huggingface/">Hugging Face</a> ! It's sibling phi-1.5 has already half a million downloads. Can't wait to see the mechanistic interpretability works that will come out of this &amp; their impact on all the important LLM research questions!
Baifeng (@baifeng_shi) 's Twitter Profile Photo

Are larger vision models always necessary? We find scaling on **image scales** (e.g., 224->448->672) is usually better than scaling on model size (e.g., Base->Large->Giant). With one line of code, improve any vision model for Multimodal LLMs or various vision and robotic tasks!

Are larger vision models always necessary?

We find scaling on **image scales** (e.g., 224-&gt;448-&gt;672) is usually better than scaling on model size (e.g., Base-&gt;Large-&gt;Giant).

With one line of code, improve any vision model for Multimodal LLMs or various vision and robotic tasks!
Xin Wang (@xinw_ai) 's Twitter Profile Photo

We are releasing phi-3 mini today! Finally, we have an open-sourced SLM (3.8B) at GPT-3.5 level! Checkout the models and technical report at here huggingface.co/microsoft/Phi-… 🥳

Baifeng (@baifeng_shi) 's Twitter Profile Photo

S2 is officially integrated into NVIIDA VILA! Checkpoint for VILA-3b with S2 is released. More checkpoints on the way! S2 enables any vision model to perceive higher resolution with as few as **one line of code**. Try it out here: github.com/bfshi/scaling_…

Sebastien Bubeck (@sebastienbubeck) 's Twitter Profile Photo

Amazing work on these new benchmarks, keep them coming!!! And notice our little phi-3-mini (3.8B) ahead of 34B models :-). Quite curious to see where phi-3-medium (14B) lands!

Amazing work on these new benchmarks, keep them coming!!! And notice our little phi-3-mini (3.8B) ahead of 34B models :-). Quite curious to see where phi-3-medium (14B) lands!
Sebastien Bubeck (@sebastienbubeck) 's Twitter Profile Photo

Updated phi-3 tech report with final numbers for 7B/14B and a new section on phi-3-V (e.g., MMMU at 40.4, in the ballpark of Claude 3-haiku and Gemini-1.0 pro) : arxiv.org/abs/2404.14219

Updated phi-3 tech report with final numbers for 7B/14B and a new section on phi-3-V (e.g., MMMU at 40.4, in the ballpark of Claude 3-haiku and Gemini-1.0 pro) : arxiv.org/abs/2404.14219
Xin Wang (@xinw_ai) 's Twitter Profile Photo

Great work Shishir Patil Tianjun Zhang It still feels like yesterday when we kicked out the project. Great to see the work continue to influence the function calling space 😉

Sebastien Bubeck (@sebastienbubeck) 's Twitter Profile Photo

Surprise #NeurIPS2024 drop for y'all: phi-4 available open weights and with amazing results!!! Tl;dr: phi-4 is in Llama 3.3-70B category (win some lose some) with 5x fewer parameters, and notably outperforms on pure reasoning like GPQA (56%) and MATH (80%).

Surprise #NeurIPS2024 drop for y'all: phi-4 available open weights and with amazing results!!!

Tl;dr: phi-4 is in Llama 3.3-70B category (win some lose some) with 5x fewer parameters, and notably outperforms on pure reasoning like GPQA (56%) and MATH (80%).
Hongyu Ren (@ren_hongyu) 's Twitter Profile Photo

The models are high again. We bring to you, o3 & o4-mini, our absolute best text & VISUAL reasoning models that truly manage to use any tools to solve hard tasks: canvas, browser, python, memory, ... & IMAGEGEN. The secret trick is to talk to the models in images🤫

The models are high again. We bring to you, o3 &amp; o4-mini, our absolute best text &amp; VISUAL reasoning models that truly manage to use any tools to solve hard tasks: canvas, browser, python, memory, ... &amp; IMAGEGEN. 

The secret trick is to talk to the models in images🤫