πŸ‘‹ Jan (@jandotai) 's Twitter Profile
πŸ‘‹ Jan

@jandotai

Run AI models locally. Jan is an open source ChatGPT-alternative that runs 100% offline. Built by @menloresearch.

Community: discord.gg/TE5wMUa7b6

ID: 1714161384265854976

linkhttps://jan.ai/ calendar_today17-10-2023 06:08:27

827 Tweet

4,4K Followers

1,1K Following

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

Someone got DeepSeek-R1-0528-Qwen3-8B running locally on a Samsung Galaxy Tab S10 Ultra (MediaTek Dimensity 9300+). Runs via MNN Chat using OpenCL backend with 6 threads. Phones, tablets, 8B models are going places. via AaronFeng47 on r/LocalLLaMA

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

We get this questions several times a day: what's the difference between Jan and LM Studio? Jan's open-source, built in public, and a bit easier to use. But the main difference is in the roadmap. You'll start to notice that with the next few updates.

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

NVIDIA just released Llama-Nemotron-Nano-VL-8B-V1, an 8B vision model that reads dense documents, charts, and video frames. It's #1 on OCRBench V2 (English), with layout and OCR fused end-to-end. huggingface.co/nvidia/Llama-3…

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

Shisa V2 405B is the strongest Japanese LLM to date. It's fine-tuned from Llama 3.1 405B and outperforming GPT-4 and Turbo in both Japanese & English. It goes toe-to-toe with GPT-4o and DeepSeek-V3 on Japanese MT-Bench. huggingface.co/shisa-ai/shisa…

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

Pandas 3.0 is moving closer to PyArrow. It will require PyArrow, with pyarrow.string as the default for string data. Early tests show: 2.2GB CSV -> 379MB Parquet, load time drops from 56s to 9s. thenewstack.io/python-pandas-…

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

You don't need a 32B model to fix grammar. You just need one that's actually trained to do it. GRMR-V3 (1B–4.3B) is a set of models that correct grammar reliably without touching meaning. huggingface.co/collections/qi…

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

Jan's beta drops soon, with the new design and a few features we think you'll like. Join us on Discord to try it early, first announcement goes there. πŸ€™ discord.gg/Exe46xPMbK

Jan's beta drops soon, with the new design and a few features we think you'll like. Join us on Discord to try it early, first announcement goes there.

πŸ€™ discord.gg/Exe46xPMbK
πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

A guy spent $60 making LLMs play Town of Salem 100 times to test how well they lie, strategize, and persuade. - GPT-4.1 and DeepSeek crushed it - Claude and Grok bombed - Qwen-D (yes, the distilled one) outplayed Claude Wild stuff. github.com/summersonnn/To…

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

Qwen team just released a 0.6B embedding model that beats 7B+ models on multilingual tasks. It supports 32k context & custom dims. huggingface.co/Qwen/Qwen3-Emb…

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

You can run LLMs 2x faster using 26% less memory. Sparse Transformers fuse MLP ops and reuse weights with smart caching. MLP time drops from 30ms to 6ms. github.com/NimbleEdge/spa…

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

OpenThinker3 is out. It's a 7B model that beats every open 7B/8B model on math, code, and science - including RL-trained ones. To run it locally, click Use this model on Hugging Face and select πŸ‘‹ Jan. huggingface.co/bartowski/open…

πŸ‘‹ Jan (@jandotai) 's Twitter Profile Photo

Hey to the new folks πŸ‘‹ Quick heads-up: Jan's part of a bigger update, soon you can try out MCPs + build your own assistants. To test the beta version, join our Discord - link's in the profile.

Hey to the new folks πŸ‘‹

Quick heads-up: Jan's part of a bigger update, soon you can try out MCPs + build your own assistants.

To test the beta version, join our Discord - link's in the profile.