I'll be posting more of my implementations and experiments on here from now on
For now, implementation of RF-Inversion for unsampling and editing images using Flux
github.com/logtd/ComfyUI-…
RAVE and FLATTEN were two of the papers that originally got me into diffusion models. They take inverse noise and apply consistency to image models.
Now with RF-Inversion (thanks Litu Rout and Nataniel Ruiz) I can try these on Flux.
Not production quality, but still fun.
Been revisiting Reference-Only Control for Flux. It uses the diffusion model as a pseudo image encoder on a reference image to influence the generation.
Results are somewhere between style and content transfer.
Just published a set of ComfyUI nodes to use Genmo's Mochi to edit videos.
github.com/logtd/ComfyUI-…
It uses rf-inversion, the gift that keeps on giving.
experimenting with regional prompting on the Hunyuan video model, giving some inception vibes
left side prompt: cyberpunk & pan left
right side prompt: steampunk & pan right
Introducing Modify Video. Reimagine any video. Shoot it in post with director-grade control over style, character, and setting. Restyle expressive performances, swap entire worlds, or redesign the frame to your vision. Shoot once. Shape infinitely.