Taekyung Ki (@taekyungki) 's Twitter Profile
Taekyung Ki

@taekyungki

AI Researcher / KAIST AI / Interested in generative models, machine learning, and computer vision.

ID: 1537455681288404994

linkhttps://taekyungki.github.io calendar_today16-06-2022 15:23:12

108 Tweet

23 Followers

119 Following

Baku (@bk_sakurai) 's Twitter Profile Photo

*動画生成:Sunoで作ったオリジナル曲をComfyUI-FLOATで歌ってもらう #comfyui note投稿しました。 note.com/bakushu/n/n1f8…

Neta Shaul (@shaulneta) 's Twitter Profile Photo

DTM vs FM👇 Lots of interest in how Difference Transition Matching (DTM) connects to Flow Matching (FM). Here is a short animation that illustrates Theorem 1 in our paper: For a very small step size (1/T), DTM converges to an Euler step of FM.

Pika (@pika_labs) 's Twitter Profile Photo

Some news: We're building the next big thing — the first-ever AI-only social video app, built on a highly expressive human video model. Over the past few weeks, we’ve been testing it in private beta. Now, we’re opening early access: download the iOS app to join the waitlist, or

Minki Kang (@mkkang_1133) 's Twitter Profile Photo

Our Agent Distillation paper is accepted at #NeurIPS2025 Spotlight! 🚀 Turn your small LM into a strong agent 💪 Code: github.com/Nardien/agent-…

Jaehyeong Jo (@jaehyeong_jo) 's Twitter Profile Photo

I'll be at NeurIPS to present the final paper of my PhD: Continuous Diffusion Model for Language Modeling (arxiv.org/abs/2502.11564) We present a continuous diffusion model for language modeling using tools from Riemannian geometry, opening up a new direction for diffusion LMs!

Dongki kim (@dongkikim95) 's Twitter Profile Photo

I'm excited to be presenting our work at NeurIPS 2025! 🗓 When: Wed, Dec 3, 2025 11:00 AM - 2:00 PM (PST) 📍Where: Exhibit Hall C, D, E #1606 If you're attending NeurIPS, please stop by. I'd love to chat about the work and AI4Science!

DailyPapers (@huggingpapers) 's Twitter Profile Photo

Avatar Forcing A real-time interactive head avatar generation framework that enables natural conversation with 500ms latency. Uses diffusion forcing for causal motion generation and direct preference optimization for expressive interactions without labeled data.

Avatar Forcing

A real-time interactive head avatar generation framework that enables natural conversation with 500ms latency. Uses diffusion forcing for causal motion generation and direct preference optimization for expressive interactions without labeled data.
Ryan Chan (@ryan_resolution) 's Twitter Profile Photo

We just upgraded XLeRobot 🚀 Built by the MakerMods team Isaac Sin, Mr Thompson and QI LIU. • Easier to build • Improved chassis • Reduced 3D-print time and material • Designed in collaboration with the original author Vector Wang Fully open-source Full build

We just upgraded XLeRobot 🚀

Built by the MakerMods team <a href="/IsaacSin12/">Isaac Sin</a>, <a href="/ThomasSchicksal/">Mr Thompson</a> and <a href="/QILIU9203/">QI LIU</a>.

• Easier to build
• Improved chassis
• Reduced 3D-print time and material
• Designed in collaboration with the original author <a href="/VectorWang2/">Vector Wang</a>

Fully open-source
Full build
Wildminder (@wildmindai) 's Twitter Profile Photo

Self-Refining Video Sampling: inference-time method using a video generator as its own refiner to correct physics and motion. no retraining needed; scores >70% human preference; is validated on Wan2.2 & Cosmos. agwmon.github.io/self-refine-vi…

Sangwon Jang (@jangsangwon7) 's Twitter Profile Photo

What if your video generator could refine itself—at inference time? ❌No new models. ❌No retraining. ❌No external verifier. 💡 Introducing Self-Refining Video Sampling By reinterpreting a pretrained generator (Wan2.2, Cosmos) as a denoising autoencoder, we enable iterative

Saining Xie (@sainingxie) 's Twitter Profile Photo

if you are building video diffusion / world simulators, try this new sampler. temporal consistency pins videos to a low-dimensional manifold in the total pixel space. self-refinement sampling keeps them there.

Ilir Aliu - eu/acc (@iliraliu_) 's Twitter Profile Photo

Learning from robot data? Standard. Direct Video-Action Models (DVA) is different: treat robot control as video generation, then translate the generated video into actions. Built by , the system pre-trains causal video models from scratch and can run complex

Zhikai Zhang (@zhikai273) 's Twitter Profile Photo

🎾Introducing LATENT: Learning Athletic Humanoid Tennis Skills from Imperfect Human Motion Data Dynamic movements, agile whole-body coordination, and rapid reactions. A step toward athletic humanoid sports skills. Project: zzk273.github.io/LATENT/ Code: github.com/GalaxyGeneralR…

Physical Intelligence (@physical_int) 's Twitter Profile Photo

We developed an RL method for fine-tuning our models for precise tasks in just a few hours or even minutes. Instead of training the whole model, we add an “RL token” output to π-0.6, our latest model, which is used by a tiny actor and critic to learn quickly with RL.

Sander Dieleman (@sedielem) 's Twitter Profile Photo

"Diffusability" is all about the spectrum. arxiv.org/abs/2603.14645 If you enjoyed my blog post about diffusion as spectral autoregression, and are wondering how this relates to latent diffusion, give this paper a read!

"Diffusability" is all about the spectrum.
arxiv.org/abs/2603.14645

If you enjoyed my blog post about diffusion as spectral autoregression, and are wondering how this relates to latent diffusion, give this paper a read!