Thorben Jansen (@learnteachaied) 's Twitter Profile
Thorben Jansen

@learnteachaied

ID: 1167056088421818369

calendar_today29-08-2019 12:49:14

90 Tweet

130 Takipçi

833 Takip Edilen

René Kizilcec (@whynotyet) 's Twitter Profile Photo

What cultural values do GPT-4o, 4, 3.5, 3 express? Using World Values Survey questions, we find GPT consistently aligns with English-speaking countries/Protestant Europe. We show that Cultural Prompting improves alignment. arxiv.org/abs/2311.14096 Yan Tao Olga Viberg @BakerEDMLab

What cultural values do GPT-4o, 4, 3.5, 3 express? Using World Values Survey questions, we find GPT consistently aligns with English-speaking countries/Protestant Europe. We show that Cultural Prompting improves alignment. arxiv.org/abs/2311.14096 <a href="/yan_ytyt/">Yan Tao</a> <a href="/OlgaOvi/">Olga Viberg</a> @BakerEDMLab
IPN (@ipn_kiel) 's Twitter Profile Photo

🚀Startschuss für das Projekt GENIUS am IPN, gefördert von der telekomstiftung Ziel: Mit #KI die Beurteilungs- und Feedbackprozesse in der #Schule verbessern und neue Maßstäbe setzen🌟📚🤖 Mehr Infos: leibniz-ipn.de #DigitaleBildung Copyright Foto: Timo Wilke

🚀Startschuss für das Projekt GENIUS am IPN, gefördert von  der <a href="/telekomstiftung/">telekomstiftung</a> 

Ziel: Mit #KI die Beurteilungs- und Feedbackprozesse in der #Schule verbessern und neue Maßstäbe setzen🌟📚🤖

Mehr Infos: leibniz-ipn.de

#DigitaleBildung 
Copyright Foto: Timo Wilke
FelloFish (@fellofishcom) 's Twitter Profile Photo

Neuer Blogbeitrag: Kann KI Lehrkräfte bei der Beurteilung von Schüler:leistungen unterstützen? Dr. Thorben Jansen Thorben Jansen vom IPN fasst die aktuelle Forschungslage zusammen und leitet daraus Implikationen für die Praxis ab. fiete.ai/blog/kuenstlic…

Ethan Mollick (@emollick) 's Twitter Profile Photo

Hate it when you ask o1-preview a hard question and it thinks for less than a second. You really feel that you failed to interest the AI in your problem.

Dan Hendrycks (@danhendrycks) 's Twitter Profile Photo

Have a question that is challenging for humans and AI? We (Center for AI Safety + Scale AI) are launching Humanity's Last Exam, a massive collaboration to create the world's toughest AI benchmark. Submit a hard question and become a co-author. Best questions get part of $500,000 in

Have a question that is challenging for humans and AI?

We (<a href="/ai_risks/">Center for AI Safety</a> + <a href="/scale_AI/">Scale AI</a>) are launching Humanity's Last Exam, a massive collaboration to create the world's toughest AI benchmark.
Submit a hard question and become a co-author.
Best questions get part of $500,000 in
James Clear (@jamesclear) 's Twitter Profile Photo

Two simple rules: 1. You get better at what you practice. 2. Everything is practice. Look around and you may be surprised by what people are “practicing" each day. If you consider each moment a repetition, what are most people training for all day long? Many people are

Ethan Mollick (@emollick) 's Twitter Profile Photo

I read a lot of social science papers on AI and my conclusion is that there are far too few people rigorously studying the implications (good & bad) of LLMs Computer science is producing a tide of good AI work. Economics, management, psych, & sociology etc. need to do the same.

Ethan Mollick (@emollick) 's Twitter Profile Photo

Our lack of good deep measures of human creativity, reasoning, empathy, etc. is really a problem in AI right now. A lot of tests that were "good enough" for human research (RAT for creativity, Seeing the Mind in The Eyes for empathy) are not robust enough for benchmarks for AI.

Dan Hendrycks (@danhendrycks) 's Twitter Profile Photo

We’re releasing Humanity’s Last Exam, a dataset with 3,000 questions developed with hundreds of subject matter experts to capture the human frontier of knowledge and reasoning. State-of-the-art AIs get <10% accuracy and are highly overconfident. pak.ai Scale.ai

We’re releasing Humanity’s Last Exam, a dataset with 3,000 questions developed with hundreds of subject matter experts to capture the human frontier of knowledge and reasoning.

State-of-the-art AIs get &lt;10% accuracy and are highly overconfident.
<a href="/ai_risk/">pak.ai</a> <a href="/scaleai/">Scale.ai</a>
BOLD (@bold_insights) 's Twitter Profile Photo

“Self-beliefs in childhood and adolescence can influence important life outcomes years later.” Building competencies, with adult support, can help children develop positive self-beliefs, say Jennifer Meyer & Thorben Jansen. Jennifer Meyer Thorben Jansen boldscience.org/how-do-childre…

Andrej Karpathy (@karpathy) 's Twitter Profile Photo

We have to take the LLMs to school. When you open any textbook, you'll see three major types of information: 1. Background information / exposition. The meat of the textbook that explains concepts. As you attend over it, your brain is training on that data. This is equivalent

We have to take the LLMs to school.

When you open any textbook, you'll see three major types of information:

1. Background information / exposition. The meat of the textbook that explains concepts. As you attend over it, your brain is training on that data. This is equivalent
Sam Altman (@sama) 's Twitter Profile Photo

we trained a new model that is good at creative writing (not sure yet how/when it will get released). this is the first time i have been really struck by something written by AI; it got the vibe of metafiction so right. PROMPT: Please write a metafictional literary short story

Ethan Mollick (@emollick) 's Twitter Profile Photo

I suspect that a lot of "AI training" in companies and schools has become obsolete in the last few months As models get larger, the prompting tricks that used to be useful are no longer good; reasoners don't play well with Chain-of-Thought; hallucination rates have dropped, etc.

Alex Albert (@alexalbert__) 's Twitter Profile Photo

Most people don't realize they can significantly influence what frontier LLMs improve at, it just requires some work. Publish a high-quality eval on a task where models currently struggle, and I guarantee future models will show substantial improvement on it.

Paul Graham (@paulg) 's Twitter Profile Photo

I just realized something most people are going to lose when (as they inevitably will) they start using AIs to write everything for them. They'll lose the knowledge of how writing is constructed.

François Chollet (@fchollet) 's Twitter Profile Photo

The most important skill for a researcher is not technical ability. It's taste. The ability to identify interesting and tractable problems, and recognize important ideas when they show up. This can't be taught directly. It's cultivated through curiosity and broad reading.