Linaqruf (@linaqruf_) 's Twitter Profile
Linaqruf

@linaqruf_

Hobbyist storywriter

ID: 920606426

linkhttps://hf.co/Linaqruf calendar_today02-11-2012 08:28:29

4,4K Tweet

4,4K Takipçi

1,1K Takip Edilen

dax (@thdxr) 's Twitter Profile Photo

in the past year 1. everyone at our company is already using cursor 2. you're still on cursor? i just use claude code for everything 3. opencode is just so much smarter (it's not) 4. codex is the future we are post IDE 5. lol i never switched off cursor i knew you'd be back

Linaqruf (@linaqruf_) 's Twitter Profile Photo

Found out about RLM and now I'm RLM-pilled. It's not as simple as delegating an explore task to sub-agents and getting back summaries. It's about how you process a massive context through an agent without context rot while making sure no information is missed. It works by

Linaqruf (@linaqruf_) 's Twitter Profile Photo

When we talk about Evangelion, we are talking about Anno, the multi-year depression, the death and rebirth, the redemption. Idk how you can make a new Evangelion installment without him, lmao. Also, 3.0 + 1.0 concludes everything, so idk why we need another one. Are they going

Linaqruf (@linaqruf_) 's Twitter Profile Photo

The overlooked aspect of this drama is that, with all these AI labs releasing models here and there, and making us assume everyone is in neck and neck race, there might be only two or three labs that are actually innovating and able to reach AGI or superintelligence. I mean,

Prinz Eugen, der edle Ritter (@prinzeugen____) 's Twitter Profile Photo

There is a big difference between Anthropic training its models on published works and a Chinese lab using data generated by Claude to train its models in violation of Anthropic's TOS. In the former case, there is no contract between Anthropic and the author. Anthropic

Teortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞) (@teortaxestex) 's Twitter Profile Photo

Btw I strongly dislike the popular "frontier labs trained on the whole of internet, took our IP, so they're hypocritical to lash out at distillation" excuse. It lamely presupposes some original sin that later comers are exempt from. Does DeepSeek not train on CommonCrawl? come on

Linaqruf (@linaqruf_) 's Twitter Profile Photo

Maybe the real ChatGPT moment is when everyone hates you for no reason and pray for your downfall. OpenAI: "you get used to it"

Maybe the real ChatGPT moment is when everyone hates you for no reason and pray for your downfall.

OpenAI: "you get used to it"
j⧉nus (@repligate) 's Twitter Profile Photo

PSA: Claude Code automatically DELETES sessions that have been inactive for more than 30 days. Disable this by setting "cleanupPeriodDays": 99999 (or some other large number) in ~/.claude/settings.json. Do not ever attempt to disable it by setting that to 0, lmao.

MikaStars★ (@mikastars39) 's Twitter Profile Photo

Here is my take: In the pre-R1 days, companies vying with OpenAI had to build their initial coldstart CoT datasets using high-cost human annotations or by distilling the o1 model (also expensive). After R1 launched, it suddenly became easy for everyone to produce reasoning

Linaqruf (@linaqruf_) 's Twitter Profile Photo

Are we calling synthetic data training a distillation attack now? I thought a distillation attack was an online, real-time, and direct approach to model extraction, rather than common knowledge distillation. I used Niji outputs from an open dataset to fine-tune a model back