Jack Saunders (@jack_r_saunders) 's Twitter Profile
Jack Saunders

@jack_r_saunders

Talking about everything to do with Facial Avatars | PhD Student | Founder of @realsyncai

ID: 1224673290431684613

linkhttps://jsaunders909.github.io/ calendar_today04-02-2020 12:37:27

231 Tweet

450 Takipçi

330 Takip Edilen

Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

SnapMoGen: Human Motion Generation from Expressive Texts TLDR: This is a method for animating digital characters from text. The model uses multi-scale tokenisation with a masked generative transformer. It is trained on a novel dataset (which should be released) consisting of 44

Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

SoulDance: Music-Aligned Holistic 3D Dance Generation via Hierarchical Motion Modelling TLDR: A large dataset and evaluation framework (with new metrics) for 4D dance generation from music audio. 📽️ Project Page: xjli360.github.io/SoulDance/ 📜 Paper: arxiv.org/abs/2507.14915

Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

⏱️ Turn yourself into a 3D Avatar in real-time with StreamME from Adobe and the University of Rochester (code coming soon) StreamME: Simplify 3D Gaussian Avatar within Live Stream TLDR: This work speeds up Gaussian reconstruction using motion-aware anchor points to prevent the

Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

🧑‍🦳 👩‍🦳 Meta's real-time Codec Avatars now have hair control. Imagine being able to customise your hairstyle in VR! HairCUP: Hair Compositional Universal Prior for 3D Gaussian Avatars TLDR: Trains two separate hyper-networks, one for the face and one for the hair. The data to do

Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

🛫 A double whammy from Microsoft this week, with another synthetics-based paper. This time, you can lift a video stream to 3D with free-viewpoint rendering that tracks the viewer! VoluMe – Authentic 3D Video Calls from Live Gaussian Splat Prediction TLDR: A UNET-based model

Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

📣 Accepted Paper 📣 I'm happy to share that our work DEAD: Data Effecient Audiovisual Dubbing using Neural Rendering Priors (previously Dubbing for Everyone) has been accepted to BMVC25! TLDR: We achive high-quality dubbing with just 4 seconds of personalised data. We seperate

Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

🎶 👗 ByteDance are making amazing progress for Virtual Try-On! DreamVVT: Mastering Realistic Video Virtual Try-On in the Wild via a Stage-Wise Diffusion Transformer Framework. TLDR: This work overcomes the reliance on hard-to-find paired (video, garment) data. By using a

Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

🏎️ We're starting to see some real-time diffusion models for Avatar generation! RAP: Real-time Audio-driven Portrait Animation with Video Diffusion Transformer TLDR: This DiT model is able to perform audio-to-talking head synthesis in real-time. The key is the use of a very

Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

📖 It's always great to get some open-source datasets! Here's one for relighting and novel view synthesis 👇 HumanOLAT: A Large-Scale Dataset for Full-Body Human Relighting and Novel-View Synthesis TLDR: A 21-person dataset where each person does 3 poses and is captured from 40

📖 It's always great to get some open-source datasets! Here's one for relighting and novel view synthesis 👇

HumanOLAT: A Large-Scale Dataset for Full-Body Human Relighting and Novel-View Synthesis

TLDR: A 21-person dataset where each person does 3 poses and is captured from 40
Tavus (@heytavus) 's Twitter Profile Photo

CVI beta launch week feat. 📚Knowledge Base (RAG) Conversations are now powered by your knowledge. Upload docs, link your content, and let CVI answer with speed & precision. Just 30ms response time. 15x faster than other RAG solutions on the market. See it in action ⬇️

CVI beta launch week feat. 📚Knowledge Base (RAG)
Conversations are now powered by your knowledge.

Upload docs, link your content, and let CVI answer with speed & precision.

Just 30ms response time. 15x faster than other RAG solutions on the market.

See it in action ⬇️
Jack Saunders (@jack_r_saunders) 's Twitter Profile Photo

🔥 Facial Puppetry just got a serious upgrade with a new open-source ❗ model. FantasyPortrait: Enhancing Multi-Character Portrait Animation with Expression-Augmented Diffusion Transformers TLDR: This is a DiT based model that uses an implicit representation of facial