Prompt Engineer (@prompt48) 's Twitter Profile
Prompt Engineer

@prompt48

A Prompt Engineer : Man Machine Interaction

ID: 1645489960487895041

calendar_today10-04-2023 18:13:12

707 Tweet

183 Followers

115 Following

Prompt Engineer (@prompt48) 's Twitter Profile Photo

🚀 Cloud GPU. Local LLM. One Secure Tunnel. I just dropped a video showing how I run OpenClaw on a RunPod VPS while using LLMs running locally on my Windows machine via Ollama — connected using reverse SSH tunneling. OpenClaw🦞 ollama Runpod youtu.be/GYW4S41li64

OpenAI (@openai) 's Twitter Profile Photo

We’re starting to roll out a test for ads in ChatGPT today to a subset of free and Go users in the U.S. Ads do not influence ChatGPT’s answers. Ads are labeled as sponsored and visually separate from the response. Our goal is to give everyone access to ChatGPT for free with

We’re starting to roll out a test for ads in ChatGPT today to a subset of free and Go users in the U.S.

Ads do not influence ChatGPT’s answers. Ads are labeled as sponsored and visually separate from the response.

Our goal is to give everyone access to ChatGPT for free with
Z.ai (@zai_org) 's Twitter Profile Photo

Introducing GLM-5: From Vibe Coding to Agentic Engineering GLM-5 is built for complex systems engineering and long-horizon agentic tasks. Compared to GLM-4.5, it scales from 355B params (32B active) to 744B (40B active), with pre-training data growing from 23T to 28.5T tokens.

Introducing GLM-5: From Vibe Coding to Agentic Engineering

GLM-5 is built for complex systems engineering and long-horizon agentic tasks. Compared to GLM-4.5, it scales from 355B params (32B active) to 744B (40B active), with pre-training data growing from 23T to 28.5T tokens.
Unsloth AI (@unslothai) 's Twitter Profile Photo

Z.ai Congrats guys on release & thank you for supporting open-source! 👏 🥰 We uploaded GLM-5 GGUFs so people can run it locally: huggingface.co/unsloth/GLM-5-…

ollama (@ollama) 's Twitter Profile Photo

MiniMax M2.5 is on Ollama's cloud! ollama run minimax-m2.5:cloud Use MiniMax M2.5 with OpenCode, Claude Code, Codex, OpenClaw via ollama launch! OpenCode: ollama launch opencode --model minimax-m2.5:cloud Claude: ollama launch claude --model glm-5:cloud

MiniMax M2.5 is on Ollama's cloud! 

ollama run minimax-m2.5:cloud 

Use MiniMax M2.5 with OpenCode, Claude Code, Codex, OpenClaw via ollama launch!

OpenCode:
ollama launch opencode --model minimax-m2.5:cloud

Claude:  
ollama launch claude --model glm-5:cloud
Logan Kilpatrick (@officiallogank) 's Twitter Profile Photo

Gemini Deep Think 3 is the world's most capable model by many measures, huge amounts of progress on reasoning benchmarks and more. Available right now via the Gemini App for Ultra subscribers and in the API soon : )

Gemini Deep Think 3 is the world's most capable model by many measures, huge amounts of progress on reasoning benchmarks and more.

Available right now via the Gemini App for Ultra subscribers and in the API soon : )
OpenAI Developers (@openaidevs) 's Twitter Profile Photo

Introducing GPT-5.3-Codex-Spark, our ultra-fast model purpose built for real-time coding. We’re rolling it out as a research preview for ChatGPT Pro users in the Codex app, Codex CLI, and IDE extension.

Joon Sung Park (@joon_s_pk) 's Twitter Profile Photo

Introducing Simile. Simulating human behavior is one of the most consequential and technically difficult problems of our time. We raised $100M from Index, Hanabi, A* BCV, Andrej Karpathy Fei-Fei Li Adam D'Angelo Guillermo Rauch scott belsky among others.

Akshay 🚀 (@akshay_pachaar) 's Twitter Profile Photo

Meta just solved the biggest problem in RAG! Most RAG systems waste your money. They retrieve 100 chunks when you only need 10. They force the LLM to process thousands of irrelevant tokens. You pay for compute you don't need. Meta AI just solved this. They built REFRAG, a new

Meta just solved the biggest problem in RAG!

Most RAG systems waste your money. They retrieve 100 chunks when you only need 10. They force the LLM to process thousands of irrelevant tokens. You pay for compute you don't need.

Meta AI just solved this.

They built REFRAG, a new
elvis (@omarsar0) 's Twitter Profile Photo

Just incredible that this is possible today. One of my favorite MCP tools as of late. Just prompt to generate beautiful excalidraw diagrams.

Unsloth AI (@unslothai) 's Twitter Profile Photo

You can now run MiniMax-2.5 locally! 🚀 At 230B parameters, MiniMax-2.5 is the strongest LLM under 700B params, delivering SOTA agentic coding & chat. Run Dynamic 3/4-bit on a 128GB Mac for 20 tokens/s. Guide: unsloth.ai/docs/models/mi… GGUF: huggingface.co/unsloth/MiniMa…

You can now run MiniMax-2.5 locally! 🚀

At 230B parameters, MiniMax-2.5 is the strongest LLM under 700B params, delivering SOTA agentic coding & chat.

Run Dynamic 3/4-bit on a 128GB Mac for 20 tokens/s.

Guide: unsloth.ai/docs/models/mi…
GGUF: huggingface.co/unsloth/MiniMa…
Sam Altman (@sama) 's Twitter Profile Photo

Peter Steinberger is joining OpenAI to drive the next generation of personal agents. He is a genius with a lot of amazing ideas about the future of very smart agents interacting with each other to do very useful things for people. We expect this will quickly become core to our

Chao Huang (@huang_chao4969) 's Twitter Profile Photo

Introducing ClawWork 🚀: Transform your openclaw/nanobot from AI assistant into a money-earning AI coworker. Watch it earn 💰$10K+ in just 7 hours by completing real professional tasks across 44+ industries — from Technology & Engineering to Business & Finance, Healthcare &

Introducing ClawWork 🚀: Transform your openclaw/nanobot from AI assistant into a money-earning AI coworker. Watch it earn 💰$10K+ in just 7 hours by completing real professional tasks across 44+ industries — from Technology & Engineering to Business & Finance, Healthcare &
LMSYS Org (@lmsysorg) 's Twitter Profile Photo

🎉 Meet Qwen3.5-397B-A17B from Qwen, 397B total params (17B active), built for real-world multimodal intelligence — day-0 support is now live in SGLang! 👁️ Unified vision-language foundation (early fusion): stronger reasoning, coding & agents ⚡ Gated DeltaNet + sparse

🎉 Meet Qwen3.5-397B-A17B from <a href="/Alibaba_Qwen/">Qwen</a>, 397B total params (17B active), built for real-world multimodal intelligence — day-0 support is now live in SGLang!

👁️ Unified vision-language foundation (early fusion): stronger reasoning, coding &amp; agents
⚡ Gated DeltaNet + sparse
vLLM (@vllm_project) 's Twitter Profile Photo

🎉 Congrats to Qwen on releasing Qwen3.5 on Chinese New Year's Eve — day-0 support is ready in vLLM! Qwen3.5 is a multimodal MoE with Gated Delta Networks architecture — 397B total params, only 17B active. What makes it interesting for inference: 🧠 Gated Delta

🎉 Congrats to <a href="/Alibaba_Qwen/">Qwen</a> on releasing Qwen3.5 on Chinese New Year's Eve  — day-0 support is ready in vLLM!

Qwen3.5 is a multimodal MoE with Gated Delta Networks architecture — 397B total params, only 17B active.

What makes it interesting for inference:

🧠 Gated Delta
Hasan Toor ✪ (@hasantoxr) 's Twitter Profile Photo

🚨BREAKING: The "Ollama for voice cloning" just dropped. It's called Voicebox and it clones any voice from just a few seconds of audio entirely on your machine. No ElevenLabs subscription. No cloud uploads. No voice data leaving your device. It's powered by Qwen3-TTS,

🚨BREAKING: The "Ollama for voice cloning" just dropped.

It's called Voicebox and it clones any voice from just a few seconds of audio entirely on your machine.

No ElevenLabs subscription. No cloud uploads. No voice data leaving your device.

It's powered by Qwen3-TTS,