Zhuang Li (@zhuang_li_nlp) 's Twitter Profile
Zhuang Li

@zhuang_li_nlp

PhD, Research Fellow @MonashUni | NLP | Previously SDE @Microsoft | Contribution @BigCodeProject

ID: 2456739656

calendar_today21-04-2014 15:07:11

82 Tweet

167 Followers

279 Following

Zhuang Li (@zhuang_li_nlp) 's Twitter Profile Photo

It's quite odd that the models from my ACL work has been downloaded over 3,000 times on Hugging Face but has only been cited once. ๐Ÿฅฒ I might have to advertise the work more.

Zhuang Li (@zhuang_li_nlp) 's Twitter Profile Photo

After ACL2023, our 'FACTUAL' model for text scene graph parsing, a joint effort with Wuhan University and Adobe Research, exceeded 4,000 downloads on Hugging Face! ๐Ÿ˜‚ I just spent a weekend updating it for easier use. Check it out: [github.com/zhuang-li/Factโ€ฆ].

Raj Dabre (@prajdabre1) 's Twitter Profile Photo

I am extremely honoured and pleased to share our survey titled: Natural Language Processing for Dialects: A Survey. arxiv.org/abs/2401.05632 Ours is the first of its kind, comprehensive survey of NLP for dialects. Aditya Joshi Diptesh Kanojia Reza Haffari

Leshem Choshen ๐Ÿค–๐Ÿค— (@lchoshen) 's Twitter Profile Photo

How ICL ๐˜ฆ๐˜ฎ๐˜ฆ๐˜ณ๐˜จ๐˜ฆ๐˜ด from unsupervised data? ๐˜๐˜ต ๐˜ญ๐˜ฆ๐˜ข๐˜ณ๐˜ฏ๐˜ด ๐˜ง๐˜ณ๐˜ฐ๐˜ฎ parallel phrases After deleting parallel parts the ICL ability was reduced by 51% deleting random words - only 2% ๐Ÿงต Yanda Chen Chen Zhao zhou Yu He He Columbia NLP arxiv.org/abs/2402.12530

How ICL ๐˜ฆ๐˜ฎ๐˜ฆ๐˜ณ๐˜จ๐˜ฆ๐˜ด from unsupervised data?
๐˜๐˜ต ๐˜ญ๐˜ฆ๐˜ข๐˜ณ๐˜ฏ๐˜ด ๐˜ง๐˜ณ๐˜ฐ๐˜ฎ parallel phrases

After deleting parallel parts the ICL ability was reduced by 51% deleting random words - only 2%

๐Ÿงต
<a href="/yanda_chen_/">Yanda Chen</a> <a href="/henryzhao4321/">Chen Zhao</a> <a href="/Zhou_Yu_AI/">zhou Yu</a> <a href="/hhexiy/">He He</a> 
<a href="/columbianlp/">Columbia NLP</a>
arxiv.org/abs/2402.12530
BigCode (@bigcodeproject) 's Twitter Profile Photo

Introducing: StarCoder2 and The Stack v2 โญ๏ธ StarCoder2 is trained with a 16k token context and repo-level information for 4T+ tokens. All built on The Stack v2 - the largest code dataset with 900B+ tokens. All code, data and models are fully open! hf.co/bigcode/starcoโ€ฆ

Introducing: StarCoder2 and The Stack v2 โญ๏ธ

StarCoder2 is trained with a 16k token context and repo-level information for 4T+ tokens. All built on The Stack v2 - the largest code dataset with 900B+ tokens. 

All code, data and models are fully open!

hf.co/bigcode/starcoโ€ฆ
Aaditya Singh (@aaditya6284) 's Twitter Profile Photo

Long (code) files may not be as high quality as you thinkโ€ฆ Excited for our new work, "Brevity is the soul of wit: Pruning long files for code generation". We find that long code files are often low quality and show benefits from pruning such files for code gen. Read on ๐Ÿ”Žโฌ

Zhuang Li (@zhuang_li_nlp) 's Twitter Profile Photo

who knows what is the board line score for ACL rolling reviewโ€ฆ I thought it was 3 but now I am a little bit confused โ€ฆ

Zhuang Li (@zhuang_li_nlp) 's Twitter Profile Photo

I found 2 out of 4 ARR reviews were generated with Claude. I can even reproduce their reviews with my Claude accountโ€ฆ can I complain to the chairs. But their scores are actually good to me

Zhuang Li (@zhuang_li_nlp) 's Twitter Profile Photo

BTW. I found the LLM review tends to add a short title with a paragraph of long explanation. if anyone uses LLMs to review the paper, please at least modify the writing style. Otherwise it is too obviousโ€ฆ