JIE GAO (@jerrygaodextrys) 's Twitter Profile
JIE GAO

@jerrygaodextrys

Researcher in NLP/text analysis, semantic/content technology, misinformation/disinformation; retweets are bookmarks for myself; Husband, father; reasonable cook

ID: 38875070

linkhttps://jerrygaolondon.github.io/ calendar_today09-05-2009 15:54:30

916 Tweet

212 Takipçi

977 Takip Edilen

Google AI (@googleai) 's Twitter Profile Photo

Today on the blog we introduce a notion of sufficient context to examine retrieval augmented generation (RAG) systems, developing a method to classify instances, analyzing failures of RAG systems & proposing a way to reduce hallucinations. Read more →goo.gle/43gp3Vk

Today on the blog we introduce a notion of sufficient context to examine retrieval augmented generation (RAG) systems, developing a method to classify instances, analyzing failures of RAG systems & proposing a way to reduce hallucinations. Read more →goo.gle/43gp3Vk
Omar Khattab (@lateinteraction) 's Twitter Profile Photo

Google folks continues to do awesome late interaction work. Compared to vanilla ColBERT, a version of this new “CRISP achieves an 11x reduction in the number of vectors—with only a 3.6% quality loss”.

JIE GAO (@jerrygaodextrys) 's Twitter Profile Photo

Beyond simple Q/A pairs or triplet based data—it creates complex synthetic data for end to end RAG components, covering both single & multi-hop queries with varying logic and cross-docs & cross-sents "clues"

tomaarsen (@tomaarsen) 's Twitter Profile Photo

Wow, ColBERT/Late Interaction/Multi-vector search models just keep on winning right now. State-of-the-art on several domains on the BRIGHT benchmark (reasoning-intensive retrieval) with just 149M parameters, while outperforming 8B ones. Trained in 2 hours. Wild!

tomaarsen (@tomaarsen) 's Twitter Profile Photo

Relabeling datasets for Information Retrieval improves NDCG@10 of both embedding models & cross-encoder rerankers. This was already the prevalent belief, but now it's been confirmed. Great job Nandan Thakur, Crystina Zhang, Xueguang Ma & Jimmy Lin

Relabeling datasets for Information Retrieval improves NDCG@10 of both embedding models & cross-encoder rerankers. This was already the prevalent belief, but now it's been confirmed. 

Great job <a href="/beirmug/">Nandan Thakur</a>, Crystina Zhang, <a href="/xueguang_ma/">Xueguang Ma</a> &amp; <a href="/lintool/">Jimmy Lin</a>
Charlie Marsh (@charliermarsh) 's Twitter Profile Photo

You can set `UV_TORCH_BACKEND=auto` and uv will automatically install the right CUDA-enabled PyTorch for your machine, zero configuration

You can set `UV_TORCH_BACKEND=auto` and uv will automatically install the right CUDA-enabled PyTorch for your machine, zero configuration
JIE GAO (@jerrygaodextrys) 's Twitter Profile Photo

Based on simple yet effective semantic selection criterion: 1. Negatives closer to the query than positives; 2. Yet far enough from the positive to avoid noise; Use clustering and dimensionality reduction to do negative sampling at scale.

Sumit (@_reachsumit) 's Twitter Profile Photo

Towards Better Instruction Following Retrieval Models Yuchen Zhuang et al. introduce a large-scale training corpus with over 38,000 instruction-query-passage triplets for enhancing retrieval models in instruction-following IR 📝arxiv.org/abs/2505.21439 👨🏽‍💻huggingface.co/datasets/InF-I…

elvis (@omarsar0) 's Twitter Profile Photo

New Lens on RAG Systems RAG systems are more brittle than you think, even when provided sufficient context. Great work from Google and collaborators. Good tips for devs included. Here are my notes:

New Lens on RAG Systems

RAG systems are more brittle than you think, even when provided sufficient context.

Great work from Google and collaborators.

Good tips for devs included.

Here are my notes:
Ravid Shwartz Ziv (@ziv_ravid) 's Twitter Profile Photo

You know all those arguments that LLMs think like humans? Turns out it's not true. 🧠 In our paper "From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning" we test it by checking if LLMs form concepts the same way humans do Yann LeCun Chen Shani Dan Jurafsky

You know all those arguments that LLMs think like humans? Turns out it's not true.

🧠 In our paper  "From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning" we test it by checking if LLMs form concepts the same way humans do  <a href="/ylecun/">Yann LeCun</a> <a href="/ChenShani2/">Chen Shani</a>  <a href="/jurafsky/">Dan Jurafsky</a>
Eugene Vinitsky 🍒🦋 (@eugenevinitsky) 's Twitter Profile Photo

There is no AI research program in the US without Chinese and Indian students. If you think otherwise, it’s because you’re not a researcher

Omar Khattab (@lateinteraction) 's Twitter Profile Photo

🤩 in some cases up to 554% speedup for ColBERT models against PLAID, which is already a ridiculously fast engine for late interaction

Qwen (@alibaba_qwen) 's Twitter Profile Photo

🚀 Proud to introduce the Qwen3-Embedding and Qwen3-Reranker Series – setting new standards in multilingual text embedding and relevance ranking! ✨ Highlights: ✅ Available in 0.6B / 4B / 8B versions ✅ Supports 119 languages ✅ State-of-the-Art performance on MMTEB , MTEB ,

🚀 Proud to introduce the Qwen3-Embedding and Qwen3-Reranker Series – setting new standards in multilingual text embedding and relevance ranking!

✨ Highlights:
✅ Available in 0.6B / 4B / 8B versions
✅ Supports 119 languages
✅ State-of-the-Art performance on MMTEB , MTEB ,
tomaarsen (@tomaarsen) 's Twitter Profile Photo

Qwen is continuing their habit of state-of-the-art releases with 3 extraordinarily strong embedding models and 3 powerful reranker models, focusing on multilingual text retrieval and more. Details in 🧵

Qwen is continuing their habit of state-of-the-art releases with 3 extraordinarily strong embedding models and 3 powerful reranker models, focusing on multilingual text retrieval and more. 

Details in 🧵
Xueguang Ma (@xueguang_ma) 's Twitter Profile Photo

Very strong embedding model!!! If anyone is interested in further fine-tuning Qwen3-embed with custom data. Here is the command with Tevatron. github.com/texttron/tevat…

Very strong embedding model!!!

If anyone is interested in further fine-tuning Qwen3-embed with custom data. Here is the command with Tevatron.  github.com/texttron/tevat…
Zhijing Jin✈️ ICLR Singapore (@zhijingjin) 's Twitter Profile Photo

Really excited about our recent large collaboration work on NLP for Social Good. The work stems from our discussions at the NLP for Positive Impact Workshop (EMNLP 2024) Workshop at #EMNLP2024 EMNLP 2025. Thanks to all our awesome collaborators, workshop attendees and all supporters!

Really excited about our recent large collaboration work on NLP for Social Good. The work stems from our discussions at the <a href="/NLP4PosImpact/">NLP for Positive Impact Workshop (EMNLP 2024)</a> Workshop at #EMNLP2024 <a href="/emnlpmeeting/">EMNLP 2025</a>. Thanks to all our awesome collaborators, workshop attendees and all supporters!
Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Extract – a system built by the UK government, using our Gemini foundational model – will help council planners make faster decisions. 🚀 Using multimodal reasoning, it turns complex planning documents – even handwritten notes and blurry maps – into digital data in just 40s.

Michael Moor (@michael_d_moor) 's Twitter Profile Photo

Excited to announce MIRIAD — a large-scale dataset of 5,821,948 medical question-answer pairs, each rephrased from passages in the medical literature. Great collab with Qinyue Zheng, Salman Abdullah, Sam Rawal, MD, Cyril Zakka, MD, Sophie Ostmeier, Maximilian Purk, Eduardo Reis, Eric Topol &