Daniel Bevenius (@dbevenius) 's Twitter Profile
Daniel Bevenius

@dbevenius

ID: 25042159

calendar_today18-03-2009 07:27:16

332 Tweet

156 Takipçi

172 Takip Edilen

Lance Ball (@lanceball) 's Twitter Profile Photo

Really excited about the DEVIE award my team received last night at #DeveloperWeek for our work on #NodeJS and #OpenShift!

Really excited about the DEVIE award my team received last night at #DeveloperWeek for our work on #NodeJS and #OpenShift!
Drogue IoT (@drogueiot) 's Twitter Profile Photo

Join us at the @EclipseCon Hacker Day! You'll get to program Micro:bit Educational Foundation with Rust Embedded Working Group and connect them to the internet! Write Quarkus applications that process the data (tweet when you jump?) and send commands back (play smoke on the water?). Anything is possible!

Drogue IoT (@drogueiot) 's Twitter Profile Photo

We just released Drogue Cloud 0.11.0! Are you curious about CoAP with DTLS, TLS-PSK, or #digitaltwin? Then maybe take a few minutes and read our release blog post: blog.drogue.io/drogue-cloud-z…

Georgi Gerganov (@ggerganov) 's Twitter Profile Photo

HuggingFace just shipped in-browser GGUF editing It allows you to edit GGUF metadata in the comfort of your browser, without having to even download the full model. This feature is enabled via the Xet technology that makes partial file updates possible.

Xuan-Son Nguyen (@ngxson) 's Twitter Profile Photo

Hugging Face Inference Endpoint now supports deploying GLM-4.7-Flash via llama.cpp, for as cheap as $0.8/hr Using Q4_K_M and 24k tokens context length - should be enough for most use case!

Hugging Face Inference Endpoint now supports deploying GLM-4.7-Flash via llama.cpp, for as cheap as $0.8/hr

Using Q4_K_M and 24k tokens context length - should be enough for most use case!
Xuan-Son Nguyen (@ngxson) 's Twitter Profile Photo

Qwen3-Coder-Next and Minimax-M2.1 are available on HF inference endpoints with the price of $2.5/hr and $5/hr respectively. With the context fitting supported, you can now utilize the largest context length possible for a given hardware. No more manual tuning -c option!

Qwen3-Coder-Next and Minimax-M2.1 are available on HF inference endpoints with the price of $2.5/hr and $5/hr respectively.

With the context fitting supported, you can now utilize the largest context length possible for a given hardware. No more manual tuning -c option!
Georgi Gerganov (@ggerganov) 's Twitter Profile Photo

In collaboration with NVIDIA we announce support for the new NVIDIA Nemotron 3 Super model in llama.cpp NVIDIA Nemotron 3 Super is a 120B open MoE model activating just 12B parameters to deliver maximum compute efficiency and accuracy for complex multi-agent applications.

Georgi Gerganov (@ggerganov) 's Twitter Profile Photo

GGUF model on Hugging Face: huggingface.co/ggml-org/Nemot… Find initial llama.cpp performance metrics at: github.com/ggml-org/llama…

Georgi Gerganov (@ggerganov) 's Twitter Profile Photo

llama.cpp at 100k stars now that 90% of the code worldwide is being written by AI agents, I predict that within 3-6 months, 90% of all AI agents will be running locally with llama.cpp 😄 Jokes aside, I am going to use this small milestone as an opportunity to reflect a bit on

llama.cpp at 100k stars

now that 90% of the code worldwide is being written by AI agents, I predict that within 3-6 months, 90% of all AI agents will be running locally with llama.cpp 😄

Jokes aside, I am going to use this small milestone as an opportunity to reflect a bit on