V. Antonopoulos (@v_antonop) 's Twitter Profile
V. Antonopoulos

@v_antonop

Co-founder 11tensors.com, a pioneering AI startup at the forefront of knowledge extraction from diverse sources. LLM and Deep Learning model training.

ID: 84534940

linkhttps://11tensors.com calendar_today23-10-2009 08:04:04

384 Tweet

262 Takipçi

766 Takip Edilen

AK (@_akhaliq) 's Twitter Profile Photo

LLaMA-Factory github: github.com/hiyouga/LLaMA-… Easy-to-use LLM fine-tuning framework (LLaMA, BLOOM, Mistral, Baichuan, Qwen, ChatGLM)

Haihao Shen (@haihaoshen) 's Twitter Profile Photo

🤗Intel Extension for Transformers supports Mixtral-8-7B with 8-bit and 4-bit inference optimizations on Intel platforms! Start from CPUs🚀 🙌Don't hesitate to give a try. Sample code below👇 🎯Project: github.com/intel/intel-ex… #iamintel #intelai Intel Hugging Face

🤗Intel Extension for Transformers supports Mixtral-8-7B with 8-bit and 4-bit inference optimizations on Intel platforms! Start from CPUs🚀
🙌Don't hesitate to give a try. Sample code below👇 
🎯Project: github.com/intel/intel-ex…
#iamintel #intelai <a href="/intel/">Intel</a> <a href="/huggingface/">Hugging Face</a>
Haihao Shen (@haihaoshen) 's Twitter Profile Photo

🙌A nice blog from 11tensors V. Antonopoulos that gives the hands-on practice to build AI Agent: Intel NeuralChat-7B LLM performs very well and can replace Open AI Function Calling to build your AI Agent! 🎯medium.com/11tensors/conn… #iamintel #intelai @intelai Huma Abidi Wei Li

Haihao Shen (@haihaoshen) 's Twitter Profile Photo

📢Just change the model name, you can run LLMs blazingly fast on your PC using Intel Extension for Transformers powered by SOTA low-bit quantization! 🎯Code: github.com/intel/intel-ex…, supporting Mistral, Llama2, Mixtral-MOE, Phi2, Solar, most recent LLMs. 🤗github.com/intel/intel-ex…

Argilla (@argilla_io) 's Twitter Profile Photo

🚀 Open-source AI strikes again! Announcing Notux 8x7B, a fine-tune of Mixtral Instruct with high-quality chat data and DPO. Notux now the top ranked MoE on the Open LLM leaderboard. huggingface.co/argilla/notux-…

🚀 Open-source AI strikes again! Announcing Notux 8x7B, a fine-tune of Mixtral Instruct with high-quality chat data and DPO.

Notux now the top ranked MoE on the Open LLM leaderboard.

huggingface.co/argilla/notux-…
Yam Peleg (@yampeleg) 's Twitter Profile Photo

Mixtral-8x7B on free Google Colab just dropped! TL;DR: - Novel offloading trick & mixed quantization. - Managed to run Mixtral-8x7B on free-tier Google Colab. Notebook: github.com/dvmazur/mixtra… Code: github.com/dvmazur/mixtra… Paper: arxiv.org/pdf/2312.17238…

Mixtral-8x7B on free Google Colab just dropped!

TL;DR:
- Novel offloading trick &amp; mixed quantization.
- Managed to run Mixtral-8x7B on free-tier Google Colab. 

Notebook: github.com/dvmazur/mixtra…
Code: github.com/dvmazur/mixtra…
Paper: arxiv.org/pdf/2312.17238…
Maxime Labonne (@maximelabonne) 's Twitter Profile Photo

Fine-tune a Mistral-7b model using Direct Preference Optimization (DPO). Just published a tutorial on Towards Data Science about using DPO to enhance the performance of SFT models. Funnily enough, I created NeuralHermes-2.5 for this article. towardsdatascience.com/fine-tune-a-mi…

elvis (@omarsar0) 's Twitter Profile Photo

LLMs for Generative Information Extraction One of the most common uses of LLMs is to analyze and extract information. It's also probably one of the easiest and most effective ways to use LLMs if done right. Here is a new survey paper on methods, trends, domains, insights, and

LLMs for Generative Information Extraction

One of the most common uses of LLMs is to analyze and extract information. It's also probably one of the easiest and most effective ways to use LLMs if done right.

Here is a new survey paper on methods, trends, domains, insights, and
Sophia Yang, Ph.D. (@sophiamyang) 's Twitter Profile Photo

NEW PAPER: Mixtral of Experts 🧡arxiv.org/abs/2401.04088 Check out how Mixtral 8x7B works! We are also hosting an office hour on Mistral AI Discord tomorrow. Join us if you have any questions about Mistral models or la plateforme: discord.gg/mistralai

NEW PAPER: Mixtral of Experts 🧡arxiv.org/abs/2401.04088

Check out how Mixtral 8x7B works! 

We are also hosting an office hour on <a href="/MistralAI/">Mistral AI</a> Discord tomorrow. Join us if you have any questions about Mistral models or la plateforme: discord.gg/mistralai
11tensors (@11tensors) 's Twitter Profile Photo

A biLSTM architecture for the creation of semantic links in the scientific publications of the European Space Agency medium.com/11tensors/a-bi…

lmsys.org (@lmsysorg) 's Twitter Profile Photo

[Arena] Exciting update! Mistral Medium has gathered 6000+ votes and is showing remarkable performance, reaching the level of Claude. Congrats Mistral AI! We have also revamped our leaderboard with more Arena stats (votes, CI). Let us know any thoughts :) Leaderboard

[Arena] Exciting update! Mistral Medium has gathered 6000+ votes and is showing remarkable performance, reaching the level of Claude. Congrats <a href="/MistralAI/">Mistral AI</a>!

We have also revamped our leaderboard with more Arena stats (votes, CI). Let us know any thoughts :)

Leaderboard
Maxime Labonne (@maximelabonne) 's Twitter Profile Photo

🐶 NeuralBeagle14-7B It's the best-performing 7B parameter model on the Open LLM Leaderboard. Remarkably, it also ranks as the 10th best-performing model overall on the Open LLM Leaderboard. In just 7B parameters! Merge + DPO = profit

🐶 NeuralBeagle14-7B

It's the best-performing 7B parameter model on the Open LLM Leaderboard.

Remarkably, it also ranks as the 10th best-performing model overall on the Open LLM Leaderboard. In just 7B parameters!

Merge + DPO = profit
Arthur Mensch (@arthurmensch) 's Twitter Profile Photo

An over-enthusiastic employee of one of our early access customers leaked a quantised (and watermarked) version of an old model we trained and distributed quite openly. To quickly start working with a few selected customers, we retrained this model from Llama 2 the minute we got

Arthur Mensch (@arthurmensch) 's Twitter Profile Photo

We’re announcing a new optimised model today! Mistral Large has top-tier reasoning capacities, is multi-lingual by design, has native function calling capacities and a 32k model. The pre-trained model has 81.2% accuracy on MMLU. Learn more on mistral.ai/news/mistral-l…. Mistral

Vaibhav (VB) Srivastav (@reach_vb) 's Twitter Profile Photo

Introducing Command R Plus ⚡ > Beats claude-3, mistral-large, gpt-4 turbo. > 104 Billion parameters. > Built with multi-step tool use and RAG. > Supports 10 languages. > Context length of 128K. > Trained with grounded generation capabilities - citations and responses based on

Introducing Command R Plus ⚡

&gt; Beats claude-3, mistral-large, gpt-4 turbo.
&gt; 104 Billion parameters.
&gt; Built with multi-step tool use and RAG.
&gt; Supports 10 languages.
&gt; Context length of 128K.
&gt; Trained with grounded generation capabilities - citations and responses based on
João Moura (@joaomdmoura) 's Twitter Profile Photo

🚀crewAI 0.27.0 is out!! Super reliable and high quality results Really appreciate RTs ❤ 🧠 Memory long, short and entity 🤚🏼 Native Human Input 🌐 Local models RAG Tools 🔍 Finer Cache Control ⛓️ Improved Cross Agent Collab 📝 Improved tool usage 🔒 Security improvements More!

V. Antonopoulos (@v_antonop) 's Twitter Profile Photo

An evaluation of various models as generators in a RAG application handling Greek text. 11tensors new blogpost medium.com/11tensors/are-… #Llama3 #Mixtral #RAG #LLMs