matt (@mattjcly) 's Twitter Profile
matt

@mattjcly

Systems/ML Software Engineer @ LM Studio

ID: 1760464029300981760

linkhttps://www.linkedin.com/in/matthewjrclayton/ calendar_today22-02-2024 00:38:38

23 Tweet

42 Takipçi

53 Takip Edilen

LM Studio (@lmstudio) 's Twitter Profile Photo

Happy 2025! Introducing LM Studio 0.3.6 🚀 - New vision models: Qwen2VL and QVQ (both GGUF + MLX) 🤩 - Function Calling API (in beta) 🧰 - New installer on Windows: choose drive (finally 😮‍💨) - In-app updates are much smaller & have a progress bar! 🟩🟩⬜️⬜️ - Update your

LM Studio (@lmstudio) 's Twitter Profile Photo

DeepSeek R1 Distilled models are now available in LM Studio! 1.5B, 7B, 8B, 13B, 14B, 32B, and 70B variants. Please update to LM Studio 0.3.7 first. Let's go! 🚀

JetBrains (@jetbrains) 's Twitter Profile Photo

✨You can now connect AI chat to locally hosted models via LM Studio in JetBrains AI Assistant! Configure the connection in AI Assistant’s settings to gain full control of your AI experience. Check out this and other model updates here: jb.gg/rowjbz

✨You can now connect AI chat to locally hosted models via <a href="/lmstudio/">LM Studio</a> in JetBrains AI Assistant!
Configure the connection in AI Assistant’s settings to gain full control of your AI experience.
Check out this and other model updates here: jb.gg/rowjbz
LM Studio (@lmstudio) 's Twitter Profile Photo

LM Studio 0.3.10 is here with 🔮 Speculative Decoding! This provides inferencing speedups, in some cases 2x or more, with no degradation in quality. - Works for both GGUF/llama.cpp and MLX models! - Easily experiment with different draft models - Visualize accepted draft token

yags (@yagilb) 's Twitter Profile Photo

We just released LM Studio 0.3.10 with Speculative Decoding support! It's a technique that pairs together a large model, and a more efficient (smaller) "draft" model that can generate tokens fast (draft generation). The larger model then verifies/rejects the draft tokens. If

LM Studio (@lmstudio) 's Twitter Profile Photo

🚢 LM Studio 0.3.14 is out now with new powerful controls for Multi-GPU setups! > Enable/disable specific GPUs > Choose which GPUs to fill first, second, etc. > Limit offload to Dedicated GPU memory > Bug fixes and more! See it in action in 🧵👇

🚢 LM Studio 0.3.14 is out now with new powerful controls for Multi-GPU setups!

&gt; Enable/disable specific GPUs
&gt; Choose which GPUs to fill first, second, etc.
&gt; Limit offload to Dedicated GPU memory
&gt; Bug fixes and more!

See it in action in 🧵👇
LM Studio (@lmstudio) 's Twitter Profile Photo

✨LM Studio 0.3.15 is out now. - Support for NVIDIA RTX 50-series (CUDA 12.8) - GLM-4 enabled in llama.cpp and MLX - New system prompt editor UI 📝 - tool_choice param in OpenAI-like API ... and more 👾 Release notes lmstudio.ai/blog/lmstudio-…

✨LM Studio 0.3.15 is out now.

- Support for NVIDIA RTX 50-series (CUDA 12.8)
- GLM-4 enabled in llama.cpp and MLX
- New system prompt editor UI 📝
- tool_choice param in OpenAI-like API
... and more 👾

Release notes lmstudio.ai/blog/lmstudio-…
yags (@yagilb) 's Twitter Profile Photo

.matt fixed an mlx-engine bug that caused undesired prompt reprocessing. The difference in performance is huge, and the fix is pretty interesting. Check out the PR to see how this issue related to repetition penalty: github.com/lmstudio-ai/ml…

LM Studio (@lmstudio) 's Twitter Profile Photo

Qwen3 is available on LM Studio in GGUF and MLX! Sizes: 0.6B, 1.7B, 4B, 8B, 14B, 30B MoE, 32B, and 235B MoE Happy Qwen day to all who celebrate! 🎉

LM Studio (@lmstudio) 's Twitter Profile Photo

Engine update: LM Studio llama.cpp/1.29.0 - Qwen2.5VL now supported in GGUF (h/t Xuan-Son Nguyen) - Support for Nomic AI's new MoE embedding model 🚂 Cmd/Ctrl + Shift + R → Update

Engine update: LM Studio llama.cpp/1.29.0

- Qwen2.5VL now supported in GGUF (h/t <a href="/ngxson/">Xuan-Son Nguyen</a>) 
- Support for <a href="/nomic_ai/">Nomic AI</a>'s new MoE embedding model

🚂 Cmd/Ctrl + Shift + R → Update