Julen Etxaniz
@juletxara
PhD Student in Language Analysis and Processing at @upvehu @Hitz_zentroa @IxaTaldea. Working on Improving Language Models for Low-resource Languages.
ID: 813409458546216961
https://julenetxaniz.eus 26-12-2016 15:41:42
1,1K Tweet
286 Takipรงi
416 Takip Edilen
๐จWe uncover a new vulnerability- Pre-Fine-Tuning Weight Recovery With a few LoRA fine-tuned models we recover the pre-fine-tuning weights๐๏ธof SoTA models, undoing Stable Diffusion personalization training and Mistral alignment๐ Project: vision.huji.ac.il/spectral_detunโฆ ๐งต๐
โก๐ ๐๐ makes RL faster โ but at the cost of performance. We present ๐ ๐ฅ๐๐ฌ๐ก๐๐, the first ๐จ๐ฉ๐๐งโ๐ฌ๐จ๐ฎ๐ซ๐๐ & ๐ฐ๐จ๐ซ๐ค๐ข๐ง๐ ๐๐ ๐ซ๐๐๐ข๐ฉ๐ that applies ๐๐๐๐/๐ ๐๐ for rollout ๐ฐ๐ข๐ญ๐ก๐จ๐ฎ๐ญ ๐ฅ๐จ๐ฌ๐ข๐ง๐ ๐ฉ๐๐ซ๐๐จ๐ซ๐ฆ๐๐ง๐๐ compared to ๐๐ ๐๐! ๐ Blog:
Are you afraid of LLMs teaching people how to build bioweapons? Have you tried just... not teaching LLMs about bioweapons? @AIEleuther and AI Security Institute joined forces to see what would happen, pretraining three 6.9B models for 500B tokens and producing 15 total models to study
With fresh support of $75M from U.S. National Science Foundation and $77M from @NVIDIA, weโre set to scale our open model ecosystem, bolster the infrastructure behind it, and fastโtrack reproducible AI research to unlock the next wave of scientific discovery. ๐ก