Inference (@inference_net) 's Twitter Profile
Inference

@inference_net

AI inference APIs at 90% lower cost. Switch in two minutes. Sign up today and start saving.

ID: 1849902577808297984

linkhttps://inference.net?join=inference&utm_source=inferencexbio calendar_today25-10-2024 19:55:59

20 Tweet

782 Takipçi

4 Takip Edilen

Sam Hogan 🇺🇸 (@0xsamhogan) 's Twitter Profile Photo

The cost of LLM tokens has dropped more than 50% over the last 12 months. Groq Inc was leading the pack for a while, but just isn't competitive any more Llama 3.1 70B tokens on inference.net are now 63% cheaper than Groq Inc and almost 90% cheaper than Replicate

The cost of LLM tokens has dropped more than 50% over the last 12 months.

<a href="/GroqInc/">Groq Inc</a> was leading the pack for a while, but just isn't competitive any more

Llama 3.1 70B tokens on inference.net are now 63% cheaper than <a href="/GroqInc/">Groq Inc</a> and almost 90% cheaper than <a href="/replicate/">Replicate</a>
Sam Hogan 🇺🇸 (@0xsamhogan) 's Twitter Profile Photo

This log scale graph from OpenRouter clearly shows the cost of LLM inference has been decreasing exponentially since the release of GPT-3 in late 2022 New inference providers like Inference will continue this trend into the foreseeable future 👇

This log scale graph from <a href="/OpenRouterAI/">OpenRouter</a> clearly shows the cost of LLM inference has been decreasing exponentially since the release of GPT-3 in late 2022

New inference providers like <a href="/inference_net/">Inference</a> will continue this trend into the foreseeable future 👇
Sam Hogan 🇺🇸 (@0xsamhogan) 's Twitter Profile Photo

If you're working on synthetic data generation and need cheap batch inference, hit my DMs. We're cooking up something specifically for this on inference .net inference.net

Sam Hogan 🇺🇸 (@0xsamhogan) 's Twitter Profile Photo

The inference .net grants program is live! Apply and get $10k for AI projects that need LLM inference from top models: - R1 and V3 from DeepSeek - Llama 70b and 405b from AI at Meta - High rate-limits, batch support, and more If you're building open source, read below  👇

The inference .net grants program is live!

Apply and get $10k for AI projects that need LLM inference from top models:

- R1 and V3 from <a href="/deepseek_ai/">DeepSeek</a>
- Llama 70b and 405b from <a href="/AIatMeta/">AI at Meta</a>
- High rate-limits, batch support, and more

If you're building open source, read below  👇
Ibrahim Ahmed (@atbeme) 's Twitter Profile Photo

We're excited to be one of the first inference APIs to host Gemma 3, starting with the 27b variant Inference ! It has incredible vision capabilities, rivaling models that are 2-6x its size in quality Give it a try and let us know what you think!

We're excited to be one of the first inference APIs to host Gemma 3, starting with the 27b variant <a href="/inference_net/">Inference</a> !

It has incredible vision capabilities, rivaling models that are 2-6x its size in quality

Give it a try and let us know what you think!
Keywords AI (YC W24) (@keywordsai) 's Twitter Profile Photo

We worked with OpenAI to build a native integration for the OpenAI Agents SDK. Today, we are very excited to launch as a tracing processor. With just a few lines of code, you can trace all your agent workflows and debug them much faster. Check out this quick demo to get

Sam Hogan 🇺🇸 (@0xsamhogan) 's Twitter Profile Photo

Inference .net is hiring a Developer Relations Lead in San Francisco! If you: - Love working with AI to amazing build products - Have experience creating world-class educational content for developers - Enjoy solving developer problems and fostering community growth We want to

Ibrahim Ahmed (@atbeme) 's Twitter Profile Photo

So Groq Inc is cool, but it’s no longer magic With basic out of the box optimizations from SGLang we’re achieving over 250 tokens per second in llama 3.1 8b at full precision on a 4090 (consumer gaming card) Inference And we think we can go even higher (1k)

Keywords AI (YC W24) (@keywordsai) 's Twitter Profile Photo

Excited to announce our integration with Inference - a new LLM provider hosting faster and more cost-effective open-source models! Inference[dot]net can reduce your LLM costs by up to 90%. Models like Llama 3.3 70B, DeepSeek V3, and DeepSeek R1 are significantly faster than

Excited to announce our integration with <a href="/inference_net/">Inference</a> - a new LLM provider hosting faster and more cost-effective open-source models!

Inference[dot]net can reduce your LLM costs by up to 90%. Models like Llama 3.3 70B, DeepSeek V3, and DeepSeek R1 are significantly faster than