Marcell Fekete (@v4rmer) 's Twitter Profile
Marcell Fekete

@v4rmer

Marcell Fekete, NLP specialist Check me out on vrmer.github.io

ID: 595847137

calendar_today31-05-2012 21:30:01

322 Tweet

87 Followers

275 Following

Johannes Bjerva (@johannesbjerva) 's Twitter Profile Photo

Interested in a ph.d. position in NLP, in the beautiful city of Copenhagen? We're hiring for a project on the topic of Explainability and Factuality in Language Modelling, at AAU Copenhagen. AAU TECH Department of Computer Science, Aalborg University #NLProc #LLM Apply via this link: stillinger.aau.dk/phd-stillinger…

Interested in a ph.d. position in NLP, in the beautiful city of Copenhagen?
We're hiring for a project on the topic of Explainability and Factuality in Language Modelling, at AAU Copenhagen. <a href="/aautech/">AAU TECH</a> <a href="/CompSciAAU/">Department of Computer Science, Aalborg University</a> #NLProc #LLM Apply via this link: stillinger.aau.dk/phd-stillinger…
Johannes Bjerva (@johannesbjerva) 's Twitter Profile Photo

Our latest typology #NLProc paper was accepted to eaclmeeting main, with Emi who visited us from Mila - Institut québécois d'IA @McGillU, and Esther. We derive continuous word order features from treebanks, better reflecting the variability of language: arxiv.org/abs/2402.01513 Department of Computer Science, Aalborg University

Edoardo Ponti (@pontiedoardo) 's Twitter Profile Photo

Can open-source LLMs execute *chains of instructions* in a single query? Not so well, we found. However, they can learn this ability by: - augmenting examples from public SFT mixtures with chains of instructions automatically - performing *sequential instruction tuning* on them.

Can open-source LLMs execute *chains of instructions* in a single query? Not so well, we found.

However, they can learn this ability by:
- augmenting examples from public SFT mixtures with chains of instructions automatically
- performing *sequential instruction tuning* on them.
Piotr Nawrot (@p_nawrot) 's Twitter Profile Photo

The memory in Transformers grows linearly with the sequence length at inference time. In SSMs it is constant, but often at the expense of performance. We introduce Dynamic Memory Compression (DMC) where we retrofit LLMs to compress their KV cache while preserving performance

The memory in Transformers grows linearly with the sequence length at inference time.

In SSMs it is constant, but often at the expense of performance.

We introduce Dynamic Memory Compression (DMC) where we retrofit LLMs to compress their KV cache while preserving performance
Irina Saparina (@irisaparina) 's Twitter Profile Photo

Next week I’ll be in Malta 🇲🇹 to present our work on Improving Generalization in Semantic Parsing by Increasing Natural Language Variation at #EACL2024! 1/3

Next week I’ll be in Malta 🇲🇹 to present our work on Improving Generalization in Semantic Parsing by Increasing Natural Language Variation at #EACL2024!

1/3
Benjamin Minixhofer (@bminixhofer) 's Twitter Profile Photo

Introducing Zero-Shot Tokenizer Transfer (ZeTT) ⚡ ZeTT frees language models from their tokenizer, allowing you to use any model with any tokenizer, with little or no extra training. Super excited to (finally!) share the first project of my PhD🧵

Introducing Zero-Shot Tokenizer Transfer (ZeTT) ⚡

ZeTT frees language models from their tokenizer, allowing you to use any model with any tokenizer, with little or no extra training.

Super excited to (finally!) share the first project of my PhD🧵
Edoardo Ponti (@pontiedoardo) 's Twitter Profile Photo

Today I am joining NVIDIA part-time as a visiting professor I could not imagine a better place to explore new efficient architectures for LLMs and diffusion I am looking forward to collaborating with so many talented researchers!

Iñigo Alonso (@alonsonlp) 's Twitter Profile Photo

Reimagining table representation! In our new #ACL2024NLP paper we introduce PixT3: a family of image-based Table-to-Text Generation models that scale better at generating text from large tables, outperforming traditional text-based baselines. arxiv.org/abs/2311.09808

Marcell Fekete (@v4rmer) 's Twitter Profile Photo

Completed my research stay at The University of Edinburgh, supported by Otto Mønsteds Fond! Investigated linguistic variation using multi-agent communication. Gained insights, networked with top researchers, and made new friends! Edoardo Ponti Johannes Bjerva AAU TECH Department of Computer Science, Aalborg University EdinburghNLP

Completed my research stay at <a href="/EdinburghUni/">The University of Edinburgh</a>, supported by Otto Mønsteds Fond! Investigated linguistic variation using multi-agent communication. Gained insights, networked with top researchers, and made new friends! <a href="/PontiEdoardo/">Edoardo Ponti</a> <a href="/johannesbjerva/">Johannes Bjerva</a> <a href="/aautech/">AAU TECH</a> <a href="/CompSciAAU/">Department of Computer Science, Aalborg University</a> <a href="/EdinburghNLP/">EdinburghNLP</a>
babyLM (@babylmchallenge) 's Twitter Profile Photo

BabyLM is looking for organizers to join our team! If you are a mid-stage graduate student, interested in (sample-efficient) language modeling or cognitive science, you could be a great fit! Find out more information, and fill out our interest form here: forms.gle/uCf62Nk6mmQhhx…

BabyLM is looking for organizers to join our team! If you are a mid-stage graduate student, interested in (sample-efficient)  language modeling or cognitive science, you could be a great fit!
Find out more information, and fill out our interest form here:
forms.gle/uCf62Nk6mmQhhx…
Nathaniel R. Robinson (@robinson_n8) 's Twitter Profile Photo

Got to present our work in progress on leveraging adapters for machine translation of Creole languages MRL #EMNLP2024 🚀 Stay tuned for more on Creole language MT! aclanthology.org/2024.mrl-1.17/ @v4mer @heather_nlp @prajdabre1 Johannes Bjerva

Got to present our work in progress on leveraging adapters for machine translation of Creole languages  <a href="/mrl2024_emnlp/">MRL</a> #EMNLP2024 🚀 Stay tuned for more on Creole language MT! 
aclanthology.org/2024.mrl-1.17/

@v4mer @heather_nlp @prajdabre1 <a href="/johannesbjerva/">Johannes Bjerva</a>
Nathan Godey (@nthngdy) 's Twitter Profile Photo

I'm now officially looking for a post-doc position, starting in Spring! I would be happy to pursue my work on training dynamics, interpretability, or even more specifically on LM representations and character-level models Feel free to reach out!