Artificial Analysis
@artificialanlys
Independent analysis of AI models and hosting providers - choose the best model and API provider for your use-case
ID: 1743487864934162432
http://artificialanalysis.ai/ 06-01-2024 04:21:21
423 Tweet
10,10K Followers
426 Following
We know that Jamba 1.5 models are the fastest, but the question is - how fast? Artificial Analysis tested our models to find out 😎The image below shows the throughput for various models (with prompt length = 10K tokens). Jamba 1.5 models are a whole lot faster – and that speed
More details in this thread from Artificial Analysis. x.com/ArtificialAnly…
I've been playing with SambaNova Systems's API serving fast Llama 3.1 405B tokens. Really cool to see leading model running at speed. Congrats to Samba Nova for hitting a 114 tokens/sec speed record (and also thanks Kunle Olukotun for getting me an API key!) sambanova.ai/blog/speed-rec…
Verified by Artificial Analysis, Cerebras Inference achieves 1,850 tokens/sec on Llama 3.1 8B and 450 tokens/sec on Llama 3.1 70B! By dramatically reducing processing time, we're enabling more complex AI workflows and enhancing real-time LLM intelligence. This includes a new class