vishwajeet kumar (@vishwajeet_86) 's Twitter Profile
vishwajeet kumar

@vishwajeet_86

Research Scientist at IBM Research AI | Interested in Natural Language Processing and Machine Learning | IIT Bombay

ID: 155120050

calendar_today13-06-2010 05:02:51

2,2K Tweet

119 Followers

1,1K Following

Tom Yeh (@proftomyeh) 's Twitter Profile Photo

I just edited my lecture - Beginner's Guide to RAG - and posted to YouTube. I gave this lecture last May. Do you like it? If so, I will edit and post more lectures like this, whenever I have some free time. Link is in the comment below. ๐Ÿ‘‡

I just edited my lecture - Beginner's Guide to RAG - and posted to YouTube. I gave this lecture last May. Do you like it? If so, I will edit and post more lectures like this, whenever I have some free time. Link is in the comment below. ๐Ÿ‘‡
Fangyu Lei (@fangyu_lei) 's Twitter Profile Photo

Wow, congratulations ๐ŸŽ‰! A team achieved a performance of 24.68% on Spider 2.0-Snow. Are there any better methods out there? ๐Ÿง spider2-sql.github.io

Wow, congratulations ๐ŸŽ‰! A team achieved a performance of 24.68% on Spider 2.0-Snow. Are there any better methods out there? ๐Ÿง
spider2-sql.github.io
Paul Couvert (@itspaulai) 's Twitter Profile Photo

Wow Mistral has released its new model tailor-made for AI code assistants Codestral 25.01 (that's its name) is debuting at #1 on the LMsys copilot arena leaderboard ๐Ÿ”ฅ You can already use it for free in Continue (100% open-source) for VS Code

Jay Alammar (@jayalammar) 's Twitter Profile Photo

Alphaxiv is an awesome way to discuss ML papers -- often with the authors themselves. Here's an intro and demo by Raj Palleti at #neurips2024 .

Philipp Schmid (@_philschmid) 's Twitter Profile Photo

Combine RAG with reasoning models like o1. Search-o1 modifies the RAG to work with reasoning models like OpenAI o1 or QwQ o,utperforming traditional RAG systems by integrating retrieved information into the Chain of Thought! ๐Ÿ‘€ Implementation 1๏ธโƒฃ Choose a Reasoning LLM, e.g.

Combine RAG with reasoning models like o1. Search-o1 modifies the RAG to work with reasoning models like OpenAI o1 or QwQ o,utperforming traditional RAG systems by integrating retrieved information into the Chain of Thought! ๐Ÿ‘€

Implementation
1๏ธโƒฃ Choose a Reasoning LLM, e.g.
Sean Welleck (@wellecks) 's Twitter Profile Photo

Excited to teach Advanced NLP at CMU this semester! Slides are on the course page as the course proceeds: cmu-l3.github.io/anlp-spring202โ€ฆ Lectures will be uploaded to Youtube: youtube.com/playlist?list=โ€ฆ

elvis (@omarsar0) 's Twitter Profile Photo

Foundations of LLMs This amazing new LLM book just dropped on arXiv. 200+ pages! It covers areas such as pre-training, prompting, and alignment methods. It looks like a great intro to LLMs for devs and researchers.

Foundations of LLMs

This amazing new LLM book just dropped on arXiv. 

200+ pages!

It covers areas such as pre-training, prompting, and alignment methods. 

It looks like a great intro to LLMs for devs and researchers.
DAIR.AI (@dair_ai) 's Twitter Profile Photo

7). Enhancing RAG - systematically explores the factors and methods that improve RAG systems such as retrieval strategies, query expansion, contrastive in-context learning, prompt design, and chunking. x.com/omarsar0/statuโ€ฆ

Tom Yeh (@proftomyeh) 's Twitter Profile Photo

I taught Lesson 1 - Agent yesterday. I am so glad to receive so many submissions from people all over the world! Here are some of their beautiful drawings! Join the course ๐Ÿ‘‰ byhand.ai/p/introductionโ€ฆ

I taught Lesson 1 - Agent yesterday. I am so glad to receive so many submissions from people all over the world! Here are some of their beautiful drawings! Join the course ๐Ÿ‘‰ byhand.ai/p/introductionโ€ฆ
elvis (@omarsar0) 's Twitter Profile Photo

NEW: Google DeepMind just introduced Gemma 3 Gemma 3 looks like a strong open long-context and multimodal model. Gemma 3 is a lightweight open model family (1Bโ€“27B parameters) that integrates vision understanding, multilingual coverage, and extended context windows (up to 128K

NEW: Google DeepMind just introduced Gemma 3

Gemma 3 looks like a strong open long-context and multimodal model.

Gemma 3 is a lightweight open model family (1Bโ€“27B parameters) that integrates vision understanding, multilingual coverage, and extended context windows (up to 128K
LangChain (@langchainai) 's Twitter Profile Photo

Fully local multi-agent systems with LangGraph With the release of OpenAI agent SDK, there's high interest in multi-agent systems. We review Swarm and Supervisor based multi-agent systems and run both locally w/ ollama + LangGraph. ๐Ÿ“ฝ๏ธ: youtu.be/4oC1ZKa9-Hs

Fully local multi-agent systems with LangGraph

With the release of OpenAI agent SDK, there's high interest in multi-agent systems.

We review Swarm and Supervisor based multi-agent systems and run both locally w/ <a href="/ollama/">ollama</a> + LangGraph.

๐Ÿ“ฝ๏ธ:
youtu.be/4oC1ZKa9-Hs
elvis (@omarsar0) 's Twitter Profile Photo

Universal RAG RAG is dead, they said. Then you see papers like this and it gives you a better understanding of the opportunities and challenges ahead. Lots of great ideas in this paper. I've summarized a few below:

Universal RAG

RAG is dead, they said.

Then you see papers like this and it gives you a better understanding of the opportunities and challenges ahead.

Lots of great ideas in this paper. I've summarized a few below:
Taiwei Shi (@taiwei_shi) 's Twitter Profile Photo

Want to ๐œ๐ฎ๐ญ ๐‘๐…๐“ ๐ญ๐ซ๐š๐ข๐ง๐ข๐ง๐  ๐ญ๐ข๐ฆ๐ž ๐›๐ฒ ๐ฎ๐ฉ ๐ญ๐จ ๐Ÿร— and boost performance? ๐Ÿš€ Meet ๐‘จ๐’…๐’‚๐‘น๐‘ญ๐‘ป โ€” a lightweight, plug-and-play curriculum learning method you can drop into any mainstream RFT algorithms (PPO, GRPO, REINFORCE). Less compute. Better results. ๐Ÿงต 1/n

Want to ๐œ๐ฎ๐ญ ๐‘๐…๐“ ๐ญ๐ซ๐š๐ข๐ง๐ข๐ง๐  ๐ญ๐ข๐ฆ๐ž ๐›๐ฒ ๐ฎ๐ฉ ๐ญ๐จ ๐Ÿร— and boost performance? ๐Ÿš€

Meet ๐‘จ๐’…๐’‚๐‘น๐‘ญ๐‘ป โ€” a lightweight, plug-and-play curriculum learning method you can drop into any mainstream RFT algorithms (PPO, GRPO, REINFORCE).

Less compute. Better results. ๐Ÿงต 1/n
NeurIPS Conference (@neuripsconf) 's Twitter Profile Photo

๐Ÿ“ข A reminder that the NeurIPS deadline for full paper submission is May 15th (Anywhere on Earth, AOE). We look forward to receiving your work, good luck to all submitting authors!

Haruki Sakajo (@sjh4i) 's Twitter Profile Photo

Our paper, Dictionaries to the Rescue: Cross-Lingual Vocabulary Transfer for Low-Resource Languages Using Bilingual Dictionaries, has been accepted to #ACL2025NLP Findings! Thanks to the co-authors, Yusuke Ide , Justin, yusuke_sakai , Yingtao Tian , Hidetaka Kamigaito , tarowatanabe !

Shivalika Singh (@singhshiviii) 's Twitter Profile Photo

Super thrilled to share GMMLU is accepted to #ACL2025 main conference ๐ŸŽ‰ It was also recently recognised by Stanford HAI as one of the significant AI releases of 2024 ๐Ÿš€ I had a blast collaborating on this closely with Beyza ErmiลŸ and all our collaborators! Huge congrats!๐Ÿ’™

Kevin Patrick Murphy (@sirbayes) 's Twitter Profile Photo

I am pleased to announce a new version of my RL tutorial. Major update to the LLM chapter (eg DPO, GRPO, thinking), minor updates to the MARL and MBRL chapters and various sections (eg offline RL, DPG, etc). Enjoy! arxiv.org/abs/2412.05265

I am pleased to announce a new version of my RL tutorial. Major update to the LLM chapter (eg DPO, GRPO, thinking), minor updates to the MARL and MBRL chapters and various sections (eg offline RL, DPG, etc). Enjoy!
arxiv.org/abs/2412.05265
Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

A 24-trillion-token web dataset with document-level metadata just dropped on Hugging Face License: apache-2.0 ESSENTIAL-WEB v1.0 collects 24 trillion tokens from Common Crawl. Each document is labeled with a 12-field taxonomy covering topic, page type, complexity, and quality

A 24-trillion-token web dataset with document-level metadata just dropped on <a href="/huggingface/">Hugging Face</a> 

License: apache-2.0

ESSENTIAL-WEB v1.0 collects 24 trillion tokens from Common Crawl. Each document is labeled with a 12-field taxonomy covering topic, page type, complexity, and quality