XMaster96 (@_xmaster96) 's Twitter Profile
XMaster96

@_xmaster96

Former Senior AI researcher @Aleph__Alpha
EVE Online player since 2013

ID: 1779555675405688832

calendar_today14-04-2024 17:01:58

58 Tweet

86 Followers

57 Following

XMaster96 (@_xmaster96) 's Twitter Profile Photo

Okay, but I have to say after I told o3-mini-highy what the actual problem was and how to solve it, its implementation was way nicer than the version from R1

XMaster96 (@_xmaster96) 's Twitter Profile Photo

I think that GPT-4.5 is a general and clear uplift in many categories where you can’t just generate a simple reward signal…. Which happens to be many areas a lot of tech folks just don't care about….

XMaster96 (@_xmaster96) 's Twitter Profile Photo

One thing that all of this Vibe coding SaaS bros don’t seem to get is that if they can generate an app / product with no programming knowledge, so can their potential customers.

DHH (@dhh) 's Twitter Profile Photo

Our S3 exit is going full steam ahead for a final departure this summer (when our 4-year contract expires!). Look at that beautiful Pure NVMe gear! 😍

Our S3 exit is going full steam ahead for a final departure this summer (when our 4-year contract expires!). Look at that beautiful Pure NVMe gear! 😍
XMaster96 (@_xmaster96) 's Twitter Profile Photo

I think the most confusing thing about VLLM in its current state is that they are apparently right doing a major re-factor, and so the code base still has lots of duplicate code

XMaster96 (@_xmaster96) 's Twitter Profile Photo

We’re hyped to be at the frontier of the next chapter in entertainment, pushing to build the models that will empower the greatest stories this world has ever seen.

XMaster96 (@_xmaster96) 's Twitter Profile Photo

One interesting thing we found out while working on improving existing TTS models, was that basically all of the open source audio encoders are god awful and are really holding back current open source TTS models. The best audio encoder we found was the Moshi one while SNAC was

XMaster96 (@_xmaster96) 's Twitter Profile Photo

Don't worry we are on it! Let me explain why it is hard... While transformers are generally amazing, they do not have length generalisation. This means if you want a model that outputs a long consistent text you need to train it specifically for that.

Don't worry we are on it! Let me explain why it is hard...

While transformers are generally amazing, they do not have length generalisation. This means if you want a model that outputs a long consistent text you need to train it specifically for that.