Josh Z. (@jmziegler) 's Twitter Profile
Josh Z.

@jmziegler

Computationish linguist. Voice tech and skepticism. Believable proxy of human behavior.

ID: 20062251

calendar_today04-02-2009 15:43:53

493 Tweet

81 Followers

184 Following

Josh Z. (@jmziegler) 's Twitter Profile Photo

People in the replies saying things like "just make your GPT's prompt more secure" as if a) putting that burden on the creator is reasonable* b) that will work *remember, this whole paradigm means you "don't have to know how to code", so cybersecurity is a non-starter

Josh Z. (@jmziegler) 's Twitter Profile Photo

This is a well-written high-level intro to some of the nonsense swirling around the tech world in recent months and why "EA" and "e/acc" both deserve the most exaggerated of eye rolls. Worth a read (maybe on his blog instead of Twitter, due to formatting).

Josh Z. (@jmziegler) 's Twitter Profile Photo

When you have zero chance of actually fixing your service, just add an arbitrary rule for users. This is the most hilarious, blatant admission yet that LLMs are not "intelligent" and that "alignment" is smoke and mirrors.

Josh Z. (@jmziegler) 's Twitter Profile Photo

Maybe Tay's legacy is making MS actually care about training data; if so, totally worth it. Of *course* your model's going to be better behaved if you don't train it on reddit and 4chan. Once again, these are statistical models, not conscious beings. microsoft.com/en-us/research…

Josh Z. (@jmziegler) 's Twitter Profile Photo

OK, the detail is better than our old favorite Will Smith eating spaghetti, but...the nose. The cat's magical third front leg. The disembodied hand. Don't say it's there until it's there, folks.

Josh Z. (@jmziegler) 's Twitter Profile Photo

Beautiful. Looks like everyone can start suing for all those $1 cars they got dealership chatbots to promise them. arstechnica.com/tech-policy/20…

Josh Z. (@jmziegler) 's Twitter Profile Photo

Please tell me this and the next tweet in the thread are real (seems plausible, given Sydney's history). "AI alignment" is a joke, and the fact that they keep doubling down on it rather than fixing their foundations should be all we need to see the perverse incentives at play.

Josh Z. (@jmziegler) 's Twitter Profile Photo

This is your quarterly reminder that <latest AI model release> is only as amazing as the marketing says if all the answers it's supposedly giving are accurate, and that you can only verify said answers if you already know them or can get them from another trusted source.

Josh Z. (@jmziegler) 's Twitter Profile Photo

"Do LLMs exhibit theory of mind?" "Hm, I don't know...let's find out using some well-established theory of mind tests." This is where I stopped reading, as "well-established" implies a sizable written corpus...which would show up in LLM training data. nature.com/articles/s4156…

Josh Z. (@jmziegler) 's Twitter Profile Photo

This doesn't mean that the model is "simulating people". You do not have to simulate a person to invent a survey response loosely based on background information that you've seen countless times before. A language model models language (or more broadly, text), not people.

Josh Z. (@jmziegler) 's Twitter Profile Photo

This is a good breakdown/takedown. My main question is: did the authors have their idea raters rate the idea of seeing if LLMs could generate research ideas? How did it score? I think that would tell us something about the quality of the rest of their ratings.

Josh Z. (@jmziegler) 's Twitter Profile Photo

This is, in fact, not what the test "proves" (it doesn't prove anything other than that most of the people he surveyed aren't art critics/historians/artists), and the original post's conclusion is nonsense. 2/10, recommend only reading the one thoughtful critique it quotes.

Josh Z. (@jmziegler) 's Twitter Profile Photo

I'm loving the new X feature where I get logged out every time I get to Twitter from clicking a link to a tweet/xeet/whatever, even if it's a link in my browser history. Can't believe they just now got around to implementing it.