Josh Lipe-Melton (@joshlipemelton) 's Twitter Profile
Josh Lipe-Melton

@joshlipemelton

ID: 422825765

calendar_today27-11-2011 18:52:24

59 Tweet

79 Takipçi

275 Takip Edilen

Madhav Singhal (@madhavsinghal_) 's Twitter Profile Photo

At Replit, we extensively use Spark on Databricks for all our training data work. We run highly customized transformations on code like parsing, deduping, PII redaction, code filtering, tokenization, and more, written with low-level Spark primitives. High quality data is key.

Josh Lipe-Melton (@joshlipemelton) 's Twitter Profile Photo

POC with big proprietary models, deploy to production with smaller specialized models. Replit a great example of smaller specialized model offering improvements over larger proprietary ones at a better price

Abhi Venigalla (@ml_hardware) 's Twitter Profile Photo

Back in June we . showed that our LLM Foundry training stack runs seamlessly on AMD MI250 GPUs. Today, I'm happy to share that we've scaled up to 128xMI250, with great multi-node performance!

Back in June we <a href="/MosaicML/">.</a> showed that our LLM Foundry training stack runs seamlessly on <a href="/AMD/">AMD</a>  MI250 GPUs.

Today, I'm happy to share that we've scaled up to 128xMI250, with great multi-node performance!
Brian Armstrong (@brian_armstrong) 's Twitter Profile Photo

AI should be decentralized as much as possible. Open source is a great step toward this. Crypto should be decentralized. Self-custodial wallets and protocols help this. Permissionless access creates innovation instead of gatekeepers. No single entity to capture.

Naveen Rao (@naveengrao) 's Twitter Profile Photo

Meet our new AI, #DBRX DBRX is an advance in what language models can do per $. These economics will have profound impacts on how AI is used, and we've built this to democratize these capabilities! It's the best open model in the world. It closes the gap to closed models in a

Garry Tan (@garrytan) 's Twitter Profile Photo

Perplexity is actually just better than Google for clear well cited answers. The quality of results and speed to answer is significantly better.

Ali Ghodsi (@alighodsi) 's Twitter Profile Photo

Databricks to acquire Tabular (now part of Databricks), a data platform from the original creators of Apache Iceberg. Together, we will bring format compatibility to the lakehouse for Delta Lake and Apache Iceberg databricks.com/blog/databrick…

Databricks Mosaic Research (@dbrxmosaicai) 's Twitter Profile Photo

Lynx is a new hallucination detection model for #LLMs that is especially suited for real-world applications in industries like healthcare and fintech. PatronusAI trained Lynx on Databricks Mosaic AI using Composer, our open source PyTorch-based training library.

Matei Zaharia (@matei_zaharia) 's Twitter Profile Photo

Great post on how AI can be engineered "brick-by-brick" to get the best results, including through the design of compound AI systems. The teams that figure out how to get the right data pipelines and system design are the ones getting transformational results.

Josh Lipe-Melton (@joshlipemelton) 's Twitter Profile Photo

Per insider sources: Inter Miami will receive 37 minutes of extra time if they aren’t winning at the end of regulation tonight

Josh Lipe-Melton (@joshlipemelton) 's Twitter Profile Photo

o3 or its checkpoints have been available to OpenAI developers and PMs internally for a while now. With virtually unlimited PhD level AI developers and capitol, why haven’t they shipped more world class software? Their other product releases were less exciting than google’s, no?

Josh Lipe-Melton (@joshlipemelton) 's Twitter Profile Photo

OpenAI's high percentage of revenue from consumer apps could push them to commoditize their models via open source. Ironically, this would turn ChatGPT into a "thin model wrapper”