Tao Li (@tao__li) 's Twitter Profile
Tao Li

@tao__li

Research Engineer @GoogleDeepMind | Formerly @GoogleAI | PhD @UtahNLP @UUtah | Intern @allen_ai Aristo, @Amazon A9, @PhilipsNA.

ID: 1135286288204959744

linkhttps://www.cs.utah.edu/~tli/ calendar_today02-06-2019 20:45:18

46 Tweet

175 Takipçi

461 Takip Edilen

Vivek Gupta (@keviv9) 's Twitter Profile Photo

Excited to share our #ACL2023NLP (#NLProc) paper on "Information Synchronization Across Multilingual Semi-Structured Tables"! 📚🔍 We explore the problem of Information Synchronization across multilingual tables and construct a large-scale dataset InfoSync. #NLPResearch -1/n

Excited to share our #ACL2023NLP (#NLProc) paper on "Information Synchronization Across Multilingual Semi-Structured Tables"! 📚🔍

We explore the problem of Information Synchronization across multilingual tables and construct a large-scale dataset InfoSync. #NLPResearch  -1/n
Vivek Gupta (@keviv9) 's Twitter Profile Photo

Excited to share our #ACL2023NLP (#NLProc) work on "Evaluating Inter-Bilingual Semantic Parsing for Indian Languages" appearing in @5thnlp4convai workshop. We propose IE-SEMPARSE, an Inter-bilingual Seq2seq Semantic parsing dataset for 11 Indian languages. - 1/n

Excited to share our #ACL2023NLP (#NLProc) work on "Evaluating Inter-Bilingual Semantic Parsing for Indian Languages" appearing in @5thnlp4convai  workshop. We propose IE-SEMPARSE, an Inter-bilingual Seq2seq Semantic parsing dataset for 11 Indian languages. - 1/n
Maitrey Mehta (@my_tray) 's Twitter Profile Photo

I’ll be presenting our work “Verifying Annotation Agreement without Multiple Experts: A Case Study with Gujarati SNACS” as a virtual poster on Tuesday at #ACL2023 and in-person at LAW on Thursday. Joint work with Vivek Srikumar Paper: tinyurl.com/3989kvnp (1/3)

AK (@_akhaliq) 's Twitter Profile Photo

A Zero-Shot Language Agent for Computer Control with Structured Reflection paper page: huggingface.co/papers/2310.08… Large language models (LLMs) have shown increasing capacity at planning and executing a high-level goal in a live computer environment (e.g. MiniWoB++). To perform a

A Zero-Shot Language Agent for Computer Control with Structured Reflection

paper page: huggingface.co/papers/2310.08…

Large language models (LLMs) have shown increasing capacity at planning and executing a high-level goal in a live computer environment (e.g. MiniWoB++). To perform a
Maitrey Mehta (@my_tray) 's Twitter Profile Photo

New preprint 🚨 "Do LLM predictors provide structurally consistent outputs in the zero- and few-shot regime?" Our new work "Promptly Predicting Structures: The Return of Inference" shows that they do not, and we show how to fix it. (1/n) 🧵

New preprint 🚨

"Do LLM predictors provide structurally consistent outputs in the zero- and few-shot regime?"

Our new work "Promptly Predicting Structures: The Return of Inference" shows that they do not, and we show how to fix it.

(1/n) 🧵
Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Introducing SIMA: the first generalist AI agent to follow natural-language instructions in a broad range of 3D virtual environments and video games. 🕹️ It can complete tasks similar to a human, and outperforms an agent trained in just one setting. 🧵 dpmd.ai/3TiYV7d

Haoyu Wang (@haoyu_wang_97) 's Twitter Profile Photo

Excited to introduce our new paper BLINK! It’s a new benchmark for MLLMs, focusing on visual perception capabilities. We show that there’s still a gap between SOTA MLLMs and human performance in 14 tasks that can be solved by humans within a blink~

fly51fly (@fly51fly) 's Twitter Profile Photo

[AI] Devil's Advocate: Anticipatory Reflection for LLM Agents H Wang, T Li, Z Deng, D Roth, Y Li [Google DeepMind] (2024) arxiv.org/abs/2405.16334 - Proposes a novel approach that integrates introspection into LLM agents to enhance their consistency and adaptability in solving

[AI] Devil's Advocate: Anticipatory Reflection for LLM Agents
H Wang, T Li, Z Deng, D Roth, Y Li [Google DeepMind] (2024)
arxiv.org/abs/2405.16334

- Proposes a novel approach that integrates introspection into LLM agents to enhance their consistency and adaptability in solving
Haoyu Wang (@haoyu_wang_97) 's Twitter Profile Photo

Multiple Reflections NOT helping much? Tired of changing plans and NOT seeing utmost effort in their execution? Introducing Devil’s Advocate 😈: Equipping LLM Agents with *Anticipatory* Reflection before action execution #LLM #Agent #AI #ML arxiv.org/pdf/2405.16334…

Multiple Reflections NOT helping much? Tired of changing plans and NOT seeing utmost effort in their execution?

Introducing Devil’s Advocate 😈: Equipping LLM Agents with *Anticipatory* Reflection before action execution 

#LLM #Agent #AI #ML 

arxiv.org/pdf/2405.16334…
Tao Li (@tao__li) 's Twitter Profile Photo

Reflection doesn't have to be post-hoc. We show that agent can benefit from "anticipatory" failures. An extra bonus of doing so is that reflective trials can now run in parallel.

Haoyu Wang (@haoyu_wang_97) 's Twitter Profile Photo

NeurIPS’24 authors: if you are desk rejected due to a missing checklist, add your email to this appeal letter! docs.google.com/document/d/16_… Getting desk rejected after filling out the checklist form carefully in OpenReview… This does not make any sense!

Google AI (@googleai) 's Twitter Profile Photo

Congratulations to the authors of the “Rich Human Feedback for Text-to-Image Generation” paper, which received the #CVPR2024 Best Paper Award. Check out the paper at: arxiv.org/pdf/2312.10240

Congratulations to the authors of the “Rich Human Feedback for Text-to-Image Generation” paper, which received the #CVPR2024 Best Paper Award. Check out the paper at: arxiv.org/pdf/2312.10240
Zhichao Xu Brutus (@zhichaoxu_ir) 's Twitter Profile Photo

Are compressed LLMs less toxic and biased against different demographic groups❓In this new📜, we study 4 pruning methods and 3 quantization methods and evaluate on 7 bias/toxicity benchmarks. arxiv.org/abs/2407.04965 (Un)surprising answer is: they are not less toxic/biased

Qingyao Ai (@qingyaoai) 's Twitter Profile Photo

Thrilled to know that our paper on Scaling Laws for Dense Retrieval has won the #SIGIR2024 Best Paper Award! 🏆Our study reveals a power-law scaling of dense retrieval models, which can help optimize training and resource allocation. Huge thanks and congrats to all collaborators!

Thrilled to know that our paper on Scaling Laws for Dense Retrieval has won the #SIGIR2024 Best Paper Award! 🏆Our study reveals a power-law scaling of dense retrieval models, which can help optimize training and resource allocation. Huge thanks and congrats to all collaborators!
Jeff Dean (@jeffdean) 's Twitter Profile Photo

Got a picture that isn't quite right? Try our native image generation in Gemini Flash 2.0. "Can you remove the stuff on the couch?". "Can you make the curtains light green?" "Can you put a unicorn horn on the person in the green pants?" Editing in human language, not image

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Think you know Gemini? 🤔 Think again. Meet Gemini 2.5: our most intelligent model 💡 The first release is Pro Experimental, which is state-of-the-art across many benchmarks - meaning it can handle complex problems and give more accurate responses. Try it now →