Or Patashnik (@opatashnik) 's Twitter Profile
Or Patashnik

@opatashnik

PhD student @ Tel-Aviv University

ID: 1125521043932758017

linkhttp://orpatashnik.github.io calendar_today06-05-2019 22:01:42

266 Tweet

1,1K Takipçi

418 Takip Edilen

naveen manwani (@naveenmanwani17) 's Twitter Profile Photo

🚨 Paper Alert 🚨 ➡️Paper Title: Kontinuous Kontext: Continuous Strength Control for Instruction-based Image Editing 🌟Few pointers from the paper 🎯Instruction-based image editing offers a powerful and intuitive way to manipulate images through natural language. 🎯 Yet,

Shelly Golan (@shelly_golan1) 's Twitter Profile Photo

T2I models excel at realism, but true creativity means generating what doesn't exist yet. How do you prompt for something you can't describe? 🎨 We introduce VLM-Guided Adaptive Negative Prompting: inference time method that promotes creative image generation. 1/6

T2I models excel at realism, but true creativity means generating what doesn't exist yet. How do you prompt for something you can't describe? 🎨

We introduce VLM-Guided Adaptive Negative Prompting: inference time method that promotes creative image generation.

1/6
Or Patashnik (@opatashnik) 's Twitter Profile Photo

Generative models have made huge progress, but creativity still lags behind. We show that these models already have the capacity for it, we just need to guide them properly!

Zarloya Vinzot (@nirgoren) 's Twitter Profile Photo

The initial noise in diffusion models is surprisingly correlated with the final image. Our NoisePrints paper exploits this to provide a lightweight, distortion-free, cryptographically secure watermark for proving authorship of generated images & videos, requiring no model access.

The initial noise in diffusion models is surprisingly correlated with the final image.
Our NoisePrints paper exploits this to provide a lightweight, distortion-free, cryptographically secure watermark for proving authorship of generated images & videos, requiring no model access.
Rishubh Parihar (@rishubhparihar) 's Twitter Profile Photo

✨ I’ll be presenting our work on depth-aware image editing at #ICCV2025 in Hawaii 🌴 next week! 📅 Oct 22 | 📍 Exhibit Hall I | 🧩 Poster #82 🌍 Project: rishubhpar.github.io/DAEdit/ 🤝 Working on image generation or editing? I’d love to chat at ICCV! Vision and AI Lab, IISc

Kfir Aberman (@abermankfir) 's Twitter Profile Photo

Can’t wait to join this incredible lineup of speakers at the Personalization in Generative AI Workshop! See you all at #ICCV2025 next week 🙌✨

Nupur Kumari (@nupurkmr9) 's Twitter Profile Photo

🚀 New preprint! We present NP-Edit, a framework for training an image editing diffusion model without paired supervision. We use differentiable feedback from Vision-Language Models (VLMs) combined with distribution-matching loss (DMD) to learn editing directly. webpage:

Etai Sella (@etai_sella) 's Twitter Profile Photo

Hey #ICCV2025 goers, I’ll be presenting this cool point cloud editing work at 14:45-16:45 in exhibit hall 1 poster #1771. Come say hi!

Guy Yariv (@guy_yariv) 's Twitter Profile Photo

We present DyPE, a framework for ultra high resolution image generation. DyPE adjusts positional embeddings to evolve dynamically with the spectral progression of diffusion. This lets pre-trained DiTs create images with 16M+ pixels without retraining or extra inference cost. 🧵👇

We present DyPE, a framework for ultra high resolution image generation.
DyPE adjusts positional embeddings to evolve dynamically with the spectral progression of diffusion.
This lets pre-trained DiTs create images with 16M+ pixels without retraining or extra inference cost.
🧵👇
Mor Ventura (@mor_ventura95) 's Twitter Profile Photo

“What big teeth you have!” said Red.👩‍🦰 “All because my model suffers from semantic leakage,” said the Wolf.🐺 When Text-to-Image models blur boundaries, identities collapse. Meet 𝐃𝐞𝐋𝐞𝐚𝐤𝐞𝐫, a lightweight inference-time fix that mitigates semantic leakage! 👇

“What big teeth you have!” said Red.👩‍🦰
 “All because my model suffers from semantic leakage,” said the Wolf.🐺

When Text-to-Image models blur boundaries, identities collapse.
Meet 𝐃𝐞𝐋𝐞𝐚𝐤𝐞𝐫, a lightweight inference-time fix that mitigates semantic leakage!
👇
Shai Yehezkel (@yehezkelshai) 's Twitter Profile Photo

Visual Diffusion Models are Geometric Solvers We cast geometry as images: a plain diffusion model denoises into valid solutions. It is simple, general and effective. Shown on Inscribed Square, Steiner Tree, and Maximum Area Polygonization - all classic hard problems.

Visual Diffusion Models are Geometric Solvers

We cast geometry as images: a plain diffusion model denoises into valid solutions. It is simple, general and effective.
Shown on Inscribed Square, Steiner Tree, and Maximum Area Polygonization - all classic hard problems.
Ron Mokady (@mokadyron) 's Twitter Profile Photo

Generating an image from 1,000 words. Very excited to release Fibo 😃, the first ever open-source model trained exclusively on long, structured captions. Fibo sets a new standard for controllability and disentanglement in image generation [1/6] 🧵

Generating an image from 1,000 words.

Very excited to release Fibo 😃, the first ever open-source model trained exclusively on long, structured captions.

Fibo sets a new standard for controllability and disentanglement in image generation

 [1/6] 🧵
Yoav HaCohen (@yoavhacohen) 's Twitter Profile Photo

🚀 LTX-2 is now the first AI model to generate 20 seconds of continuous, synchronized audio and video! That’s double the previous 10-second max. With 20s of audio + video, we finally have time to tell much more interesting - and touching - stories. 🎥 Open-sourcing soon! 🧵👇

Ashkan Mirzaei (@ashmrz10) 's Twitter Profile Photo

🚀 We're seeking interns for 2026! Join Snap's Creative Vision Team and help advance the frontiers of generative AI across images, videos, 3D, and 4D. We’re looking for PhD students with strong research backgrounds in these areas. Apply here: 👉 snap-research.github.io/cv-call-for-in…