Ray Lillywhite (@lillywhiteray) 's Twitter Profile
Ray Lillywhite

@lillywhiteray

๐Ÿ‡น๐Ÿ‡ผ

ID: 1102451411143389184

calendar_today04-03-2019 06:11:13

204 Tweet

29 Takipรงi

522 Takip Edilen

Bernie Sanders (@sensanders) 's Twitter Profile Photo

Will AI become smarter than humans? If so, is humanity in danger? I went to Silicon Valley to ask some of the leading AI experts that question. Hereโ€™s what they had to say:

typedfemale (@typedfemale) 's Twitter Profile Photo

had no idea dario hated sam - always believed he started a company because he believed two companies with honest high integrity leaders would lead to better outcomes

Franรงois Chollet (@fchollet) 's Twitter Profile Photo

I keep reading this take (below) every few months, presented as if extremely profound, and it is just offensively dumb. It confuses data and information, it ignores the fact that not all information is equally valuable, and it ignores the importance of retention rate. As a

I keep reading this take (below) every few months, presented as if extremely profound, and it is just offensively dumb. It confuses data and information, it ignores the fact that not all information is equally valuable, and it ignores the importance of retention rate.

As a
Eric Schmidt ๐ŸŒ๐Ÿ”ธ (@edavidds) 's Twitter Profile Photo

The world needs humble deference to intellectual superiority at this crucial moment in our speciesโ€™ history. Our government should listen to the goddamn Einsteins of our time.

Ajeya Cotra (@ajeya_cotra) 's Twitter Profile Photo

New post: on Jan 14, I predicted that SWE time horizon by EOY would be ~24 hours. Now I think it'll be >100 hours, and maybe unbounded. For the first time, I don't see solid evidence against AI R&D automation *this year.* Link below.

New post: on Jan 14, I predicted that SWE time horizon by EOY would be ~24 hours. Now I think it'll be >100 hours, and maybe unbounded. For the first time, I don't see solid evidence against AI R&D automation *this year.* Link below.
Gone to ๐Ÿฆ‹ same user name. (@glenweyl) 's Twitter Profile Photo

I am humbled to be among the courageous leaders from Abrahamic religious traditions who put out this important statement about the Department of War ๐Ÿ‡บ๐Ÿ‡ธ-Anthropic dispute: faithfamilytech.org/moral-guardraiโ€ฆ. A short summary of the substance:

Harlan Stewart (@humanharlan) 's Twitter Profile Photo

People are offering a lot of reasons for why it's actually ok that AI systems have an emergent behavior to sometimes blackmail humans

People are offering a lot of reasons for why it's actually ok that AI systems have an emergent behavior to sometimes blackmail humans
SE Gyges (@segyges) 's Twitter Profile Photo

"Stochastic Parrots" is a meme that won't go away. It seemed important enough to do a rundown of everything that is wrong with the technical or "philosophy of language" side of the paper (which is everything). ๐Ÿ‘‡

Rob Wiblin (@robertwiblin) 's Twitter Profile Photo

Do you think Andreessen could have written a world-famous essay like 'A Time to Build' while simultaneously leading the fight against housing construction in his area if he'd been wasting any precious time on introspection? I don't think so.

Cosmos Raj (@cosmos_raj) 's Twitter Profile Photo

Breaking news: Anthropic buys the All in podcast just to shut it down Dario quoted as saying: โ€œthis isnโ€™t even about new media I just want to stop seeing them on my timelineโ€

Breaking news: Anthropic buys the All in podcast just to shut it down

Dario quoted as saying: โ€œthis isnโ€™t even about new media I just want to stop seeing them on my timelineโ€
Rob Wiblin (@robertwiblin) 's Twitter Profile Photo

AI companies are notorious for refusing to release AI models as a clever marketing ploy to build hype. In fact it's hard to get them to release models at all. We're still waiting for ones they trained years ago. Crazy but true.

Jordan Braunstein (@jbraunstein914) 's Twitter Profile Photo

Joshua Achiam I'm sorry, but this is ridiculous. The implication is that people should censor sincere assessments of our situation bc it might indirectly incite unwell individuals to violence? By that standard, nothing could ever be called an emergency, as it might cause panic, so to avoid

Sriram Krishnan (@sriramk) 's Twitter Profile Photo

Every person here's reaction to the Jensen + Dwarkesh Patel podcast can be extrapolated *directly* from whether they believe in the frontier labs achieving short timelines for AGI/ASI. If you believe in the labs achieving RSI and then AGI/ASI (for some definition of all three) in

Dean W. Ball (@deanwball) 's Twitter Profile Photo

For a moment, substitute the notion of โ€œbelieving in short AGI timelinesโ€ for: โ€œacknowledging the idea of AGI as an ill-defined thing that will nonetheless probably exist within a strategically relevant timeframe, the pursuit of which will produce importantly capable artifacts

๐™ท๐š’๐š–๐šŠ ๐™ป๐šŠ๐š”๐š”๐šŠ๐š›๐šŠ๐š“๐šž (@hima_lakkaraju) 's Twitter Profile Photo

๐Ÿ“ฃ Excited to announce our oral presentation at #ICLR! LLMs capture rich semantic structure, as evidenced by their strong performance across a wide range of language and reasoning tasks. But Sparse Autoencoders (SAEs), a popular interpretability tool, mostly learn local, noisy,

๐Ÿ“ฃ Excited to announce our oral presentation at #ICLR!

LLMs capture rich semantic structure, as evidenced by their strong performance across a wide range of language and reasoning tasks.

But Sparse Autoencoders (SAEs), a popular interpretability tool, mostly learn local, noisy,
AI Security Institute (@aisecurityinst) 's Twitter Profile Photo

As part of our work on assessing AI loss-of-control risks, we collaborated with Anthropic to pilot alignment evals on models including pre-release snapshots of Mythos Preview and Opus 4.7. We ask: could an AI agent used inside a frontier lab sabotage safety research? ๐Ÿงต

As part of our work on assessing AI loss-of-control risks, we collaborated with <a href="/AnthropicAI/">Anthropic</a> to pilot alignment evals on models including pre-release snapshots of Mythos Preview and Opus 4.7.

We ask: could an AI agent used inside a frontier lab sabotage safety research? ๐Ÿงต