slm tokens (@tulkenss) 's Twitter Profile
slm tokens

@tulkenss

🦑 Machine learning biped

ID: 3020046891

linkhttps://stephantul.github.io calendar_today05-02-2015 16:43:13

1,1K Tweet

245 Takipçi

554 Takip Edilen

slm tokens (@tulkenss) 's Twitter Profile Photo

This is interesting to me because static models are the fastest and worst possible you can probably make. There is a market for very fast+bad: there are some tasks you can only do if you are very very fast. This is what the value prop for static models is.

slm tokens (@tulkenss) 's Twitter Profile Photo

Obvious observation perhaps, but if your agent has trouble writing tests for your code, you've not yet hit the right abstraction level. Instead of letting the agent persevere, or asking it to refactor, just take a look. (this is a look at your data post)

slm tokens (@tulkenss) 's Twitter Profile Photo

I'm curious about batching in services. I've always done batch size 1 inference for requests (i.e., using FastAPI + a model). Inefficient, but easy. So it seems to me that actual latency for transformers on GPUs is much higher per instance than is typically claimed. Anyone?

slm tokens (@tulkenss) 's Twitter Profile Photo

The first good modernbert static model: huggingface.co/stephantulkens… Still need to push the scores, but I'm simply too stoked!