Matt Upson (@m_a_upson) 's Twitter Profile
Matt Upson

@m_a_upson

Founder @mantisnlp. @PublicDigitalHQ Network. Fellow @SoftwareSaved. Previously @gdsteam. Mostly #NLP and #MLOps.

ID: 1394290140

linkhttp://www.machinegurning.blog calendar_today01-05-2013 10:27:26

2,2K Tweet

1,1K Takipçi

1,1K Takip Edilen

Jean-Rémi King (@jeanremiking) 's Twitter Profile Photo

🔥Preprint out: `Toward a realistic model of speech processing in the brain with self-supervised learning’: arxiv.org/abs/2206.01685 by J. Millet*, Charlotte Caucheteux @ICML24* and our wonderful team: The 3 main results summarized below 👇

Nick Sorros (@nsorros) 's Twitter Profile Photo

📢 If you are interested in how to scale 🚀 multilabel classsification to thousands or millions of labels 🏷 I am giving a talk on Sunday in PyData London titled “Extreme multilabel classification in the biomedical NLP domain” 🗣 Also super excited to be back in London 🇬🇧

Briar Adams (@briaradams) 's Twitter Profile Photo

Nick Sorros shares his approach to the challenges of extreme multilabel classifier prediction spaces with custom #neuralnetwork #architecture & inspiration from #XLinear & #BertMesh PyData London #biomedical #NLP #bigdata #DataScience #pydatalondon2022 Wellcome @mantisnlp

Nick Sorros shares his approach to the challenges of extreme multilabel classifier prediction spaces with custom #neuralnetwork #architecture &amp; inspiration from #XLinear &amp; #BertMesh <a href="/pydatalondon/">PyData London</a> #biomedical #NLP #bigdata #DataScience #pydatalondon2022 <a href="/wellcometrust/">Wellcome</a> @mantisnlp
Jay Alammar (@jayalammar) 's Twitter Profile Photo

Great talk by Nick Sorros on Extreme Multilabel Classification: classification problems with thousands or millions of classes. - svm.sparsify() is useful - TFIDF vocabs can be made smaller with BPE/WordPiece - "Multilabel attention" is used in these models #PyDataLDN

Great talk by <a href="/nsorros/">Nick Sorros</a> on Extreme Multilabel Classification: classification problems with thousands or millions of classes.

- svm.sparsify() is useful
- TFIDF vocabs can be made smaller with BPE/WordPiece
- "Multilabel attention" is used in these models

#PyDataLDN
Matt Upson (@m_a_upson) 's Twitter Profile Photo

How did I miss this? rostrum.blog/2022/06/07/ass… Matt Dray is the arch troll of the #RStats world. I feel like there should be an award for this.

Rachael Tatman @rctatman@mastodon.rctatman.com (@rctatman) 's Twitter Profile Photo

I've got a ~surprise~ video for you today! I had a great conversation with Nick Sorros and Matt Upson from @mantisnlp about reproducibility, open source, data ethics & more. I hope you enjoy it as much as I did. 😁▶️ youtube.com/watch?v=uztK9g…

Kim Calders (@kimcalders) 's Twitter Profile Photo

After more than five years of work: our new publication on how to almost double woodland carbon overnight; or, when models work a bit too well. Full paper: bit.ly/3uXiG8D [1/10]

After more than five years of work: our new publication on how to almost double woodland carbon overnight; or, when models work a bit too well.

Full paper: bit.ly/3uXiG8D

[1/10]
Nick Sorros (@nsorros) 's Twitter Profile Photo

✨Here is quick example of a retrieval based #ChatGPT equivalent you can easily build using your own data and the excellent library from deepset, makers of Haystack

Jim Fan (@drjimfan) 's Twitter Profile Photo

The Adam optimizer is at the heart of modern AI. Researchers have been trying to dethrone Adam for years. How about we ask a machine to do a better job? Google AI uses evolution to discover a simpler & efficient algorithm with remarkable features. It’s just 8 lines of code: 🧵

The Adam optimizer is at the heart of modern AI. Researchers have been trying to dethrone Adam for years.

How about we ask a machine to do a better job? <a href="/GoogleAI/">Google AI</a> uses evolution to discover a simpler &amp; efficient algorithm with remarkable features.

It’s just 8 lines of code: 🧵
Nick Sorros (@nsorros) 's Twitter Profile Photo

We recently switched to using Hugging Face inference endpoints as it makes it much easier to deploy and manage the models for our clients. We also created a nice CLI tool called hugie 🐻 to make inference endpoints available from the command line 🚀

Nick Sorros (@nsorros) 's Twitter Profile Photo

Here is a nice CLI tool we created on top of the Hugging Face inference endpoints which allow you to deploy any transformer model from the command line or a CI/CD pipeline. Kudos mainly to Matt Upson 👏

Prakash (Ate-a-Pi) (@8teapi) 's Twitter Profile Photo

Vicious Self-Degradation > you Google > Quora spots query and id’s as frequent > Quora uses ChatGPT to generate answer > ChatGPT hallucinates > Google picks up Quora answer as highest probability correct answer > ChatGPT hallucination is now canonical Google answer

Vicious Self-Degradation 

&gt; you Google 
&gt; Quora spots query and id’s as frequent
&gt; Quora uses ChatGPT to generate answer 
&gt; ChatGPT hallucinates 
&gt; Google picks up Quora answer as highest probability correct answer 
&gt; ChatGPT hallucination is now canonical Google answer
Argilla (@argilla_io) 's Twitter Profile Photo

🔎 What if only a base model were needed for preference alignment? Welcome to ORPO, introduced in our latest blog with @MantisNLP. argilla.io/blog/mantisnlp…

Matt Upson (@m_a_upson) 's Twitter Profile Photo

🎉 Sieves v0.11.0 is out! Now featuring model distillation for classification tasks using SetFit/Model2Vec! Plus smart document caching to reduce compute waste.🔗 sieves.ai

🎉 Sieves v0.11.0 is out! Now featuring model distillation for classification tasks using SetFit/Model2Vec! Plus smart document caching to reduce compute waste.🔗 sieves.ai
Matt Upson (@m_a_upson) 's Twitter Profile Photo

> Self-styled prophets are claiming they have "awakened" chatbots and accessed the secrets of the universe through ChatGPT Sobering reading. Who saw this coming?rollingstone.com/culture/cultur…