Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profileg
Aran Komatsuzaki

@arankomatsuzaki

@TeraflopAI

ID:794433401591693312

linkhttps://arankomatsuzaki.wordpress.com/about-me/ calendar_today04-11-2016 06:57:37

4,9K Tweets

95,7K Followers

78 Following

Follow People
Christian Holz(@cholz) 's Twitter Profile Photo

new research: Ultra Inertial Poser: Scalable full-body tracking in the wild using sparse sensing


No cameras—just 6 wearables (IMU+UWB) for our graph model to estimate poses

We are first to process raw IMU signals and need no proprietary sensors for 3D orientation

account_circle
Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profile Photo

Self-Play Preference Optimization for Language Model Alignment

SPPO serves as the RLHF counterpart of SPIN and outperforms iterative DPO, Snorkel AI, Self-Rewarding LM, GPT-4 0613 etc

arxiv.org/abs/2405.00675

Self-Play Preference Optimization for Language Model Alignment SPPO serves as the RLHF counterpart of SPIN and outperforms iterative DPO, Snorkel AI, Self-Rewarding LM, GPT-4 0613 etc arxiv.org/abs/2405.00675
account_circle
Nikita Drobyshev(@NikDrob23) 's Twitter Profile Photo

I am thrilled to announce that my latest paper has been accepted at CVPR 2024:
EMOPortraits: Emotion-enhanced Multimodal One-shot Head Avatars
🔗 Project Page: neeek2303.github.io/EMOPortraits/

account_circle
Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profile Photo

Scale AI presents A Careful Examination of LLM Performance on Grade School Arithmetic

- Evaluate existing LLMs on a new test set of GSM8K
- Observe accuracy drops of up to 13%, with models like Phi and Mistral showing evidence of systematic overfitting

arxiv.org/abs/2405.00332

Scale AI presents A Careful Examination of LLM Performance on Grade School Arithmetic - Evaluate existing LLMs on a new test set of GSM8K - Observe accuracy drops of up to 13%, with models like Phi and Mistral showing evidence of systematic overfitting arxiv.org/abs/2405.00332
account_circle
Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profile Photo

Btw this multiple token training is not a panacea. The performance gain depends on the target task.

It leads to no perf gain or slight degradation on some multiple choice questions.

It leads to minior improvement on summarization and little to no improvement on arithmetic…

Btw this multiple token training is not a panacea. The performance gain depends on the target task. It leads to no perf gain or slight degradation on some multiple choice questions. It leads to minior improvement on summarization and little to no improvement on arithmetic…
account_circle
Chujie Zheng @ ICLR 2024(@ChujieZheng) 's Twitter Profile Photo

✨New Paper Alert✨
Excited to introduce ExPO, an extremely simple method to boost LLMs' alignment with human preference, via weak-to-strong model extrapolation
👇

✨New Paper Alert✨ Excited to introduce ExPO, an extremely simple method to boost LLMs' alignment with human preference, via weak-to-strong model extrapolation 👇 #LLMs #MachineLearning #NLProc #ArtificialIntelligence #AI
account_circle
Ziming Liu(@ZimingLiu11) 's Twitter Profile Photo

Aran Komatsuzaki Thanks for sharing our work! In case anyone's interested in digging more, here's my tweet: twitter.com/ZimingLiu11/st…

account_circle
Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profile Photo

KAN: Kolmogorov–Arnold Networks

Proposes an alternative to MLP that outperforms in terms of accuracy and interpretability

arxiv.org/abs/2404.19756

KAN: Kolmogorov–Arnold Networks Proposes an alternative to MLP that outperforms in terms of accuracy and interpretability arxiv.org/abs/2404.19756
account_circle
Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profile Photo

Meta presents Iterative Reasoning Preference Optimization

Increasing accuracy for Llama-2-70B-Chat:
- 55.6% -> 81.6% on GSM8K
- 12.5% -> 20.8% on MATH
- 77.8% -> 86.7% on ARC-Challenge

arxiv.org/abs/2404.19733

Meta presents Iterative Reasoning Preference Optimization Increasing accuracy for Llama-2-70B-Chat: - 55.6% -> 81.6% on GSM8K - 12.5% -> 20.8% on MATH - 77.8% -> 86.7% on ARC-Challenge arxiv.org/abs/2404.19733
account_circle
Tanishq Mathew Abraham, Ph.D.(@iScienceLuvr) 's Twitter Profile Photo

gpt2-chatbot → Generative Pretrained Transformer 2 Chatbot

This is clearly a scaled up version of the Transformer 2 archtecture!! ↓

account_circle
Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profile Photo

Meta presents Better & Faster Large Language Models via Multi-token Prediction

- training language models to predict multiple future tokens at once results in higher sample efficiency
- up to 3x faster at inference

arxiv.org/abs/2404.19737

Meta presents Better & Faster Large Language Models via Multi-token Prediction - training language models to predict multiple future tokens at once results in higher sample efficiency - up to 3x faster at inference arxiv.org/abs/2404.19737
account_circle
Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profile Photo

MotionLCM: Real-time Controllable Motion Generation via Latent Consistency Model

proj: dai-wenxun.github.io/MotionLCM-page/
abs: arxiv.org/abs/2404.19759

MotionLCM: Real-time Controllable Motion Generation via Latent Consistency Model proj: dai-wenxun.github.io/MotionLCM-page/ abs: arxiv.org/abs/2404.19759
account_circle
TeraflopAI(@TeraflopAI) 's Twitter Profile Photo

Awesome to see Joseph Spisak, AI Product Director, Meta, mention our previous research, YaRN, on stage at the Weights & Biases Fully Connected conference. We have another very exciting long-context release coming soon.

account_circle
Weiyan Shi(@shi_weiyan) 's Twitter Profile Photo

🚨New Paper🚨
We propose
1⃣CultureBank🌎 dataset sourced from TikTok & Reddit
2⃣An extensible pipeline to build cultural knowledge bases
3⃣Evaluation of LLMs’ cultural awareness
4⃣Insights into culturally-aware LLMs

Project: culturebank.github.io
Data: shorturl.at/hrtwP

🚨New Paper🚨 We propose 1⃣CultureBank🌎 dataset sourced from TikTok & Reddit 2⃣An extensible pipeline to build cultural knowledge bases 3⃣Evaluation of LLMs’ cultural awareness 4⃣Insights into culturally-aware LLMs Project: culturebank.github.io Data: shorturl.at/hrtwP
account_circle
Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profile Photo

Apple presents OpenELM

- An efficient LM family with open-source training and inference framework
- Performs on par with OLMo while requiring 2x fewer pre-training tokens

repo: github.com/apple/corenet
hf: huggingface.co/apple/OpenELM
abs: arxiv.org/abs/2404.14619

Apple presents OpenELM - An efficient LM family with open-source training and inference framework - Performs on par with OLMo while requiring 2x fewer pre-training tokens repo: github.com/apple/corenet hf: huggingface.co/apple/OpenELM abs: arxiv.org/abs/2404.14619
account_circle
Aran Komatsuzaki(@arankomatsuzaki) 's Twitter Profile Photo

SnapKV: LLM Knows What You are Looking for Before Generation

- Automatically compresses KV caches
- Consistent decoding speed with a 3.6x increase in generation speed and an 8.2x enhancement in memory efficiency

repo: github.com/FasterDecoding…
abs: arxiv.org/abs/2404.14469

SnapKV: LLM Knows What You are Looking for Before Generation - Automatically compresses KV caches - Consistent decoding speed with a 3.6x increase in generation speed and an 8.2x enhancement in memory efficiency repo: github.com/FasterDecoding… abs: arxiv.org/abs/2404.14469
account_circle