RyanOnTheInside

106 posts

RyanOnTheInside banner
RyanOnTheInside

RyanOnTheInside

@RyanOnTheInside

Yo. I am a software dev and musician. ComfyUI enthusiast

MA Katılım Şubat 2023
104 Takip Edilen331 Takipçiler
Sabitlenmiş Tweet
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
For full technical details, please see arxiv.org/abs/2602.14381
RyanOnTheInside@RyanOnTheInside

YO, I adapted VACE to work with real-time autoregressive video generation. Here's what it can do right now in real time: - Depth, pose, optical flow, scribble, edge maps — all the v2v control stuff - First frame animation / last frame lead-in / keyframe interpolation - Inpainting with static or dynamic masks - Stacking stuff together (e.g. depth + LoRA, inpainting + reference images) - Reference-to-video is in there too but quality isn't great yet compared to batch Getting ~20 fps for most control modes on a 5090 at 368x640 with the 1.3B models. Image-to-video hits ~28 fps. Works with 14b models as well, but doesnt fit on 5090 with VACE. This is all part of Daydream Scope, an open source tool for running real-time interactive video generation pipelines. The demos were created in/with Scope, and is a combination of Longlive, VACE, and Custom LoRA. There's also a very early WIP ComfyUI node pack wrapping Scope. But how is a real-time, autoregressive model relevant to @ComfyUI ? Ultra long video generation. You can use these models distilled from Wan to do V2V tasks on thousands of frames at once, technically infinite length. I havent experimented much more than validating the concept on a couple thousand frames gen. It works! Full technical details on real-time VACE + more examples here (link in comments) Curious what people think. Happy to answer questions. Video + Custom LoRA links also in comments. Love, Ryan p.s. I will be back with a sick update on ACEStep implementation tomorrow (links in first comment)

English
7
3
42
1.7K
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
For full technical details, please see arxiv.org/abs/2602.14381
RyanOnTheInside@RyanOnTheInside

YO, I adapted VACE to work with real-time autoregressive video generation. Here's what it can do right now in real time: - Depth, pose, optical flow, scribble, edge maps — all the v2v control stuff - First frame animation / last frame lead-in / keyframe interpolation - Inpainting with static or dynamic masks - Stacking stuff together (e.g. depth + LoRA, inpainting + reference images) - Reference-to-video is in there too but quality isn't great yet compared to batch Getting ~20 fps for most control modes on a 5090 at 368x640 with the 1.3B models. Image-to-video hits ~28 fps. Works with 14b models as well, but doesnt fit on 5090 with VACE. This is all part of Daydream Scope, an open source tool for running real-time interactive video generation pipelines. The demos were created in/with Scope, and is a combination of Longlive, VACE, and Custom LoRA. There's also a very early WIP ComfyUI node pack wrapping Scope. But how is a real-time, autoregressive model relevant to @ComfyUI ? Ultra long video generation. You can use these models distilled from Wan to do V2V tasks on thousands of frames at once, technically infinite length. I havent experimented much more than validating the concept on a couple thousand frames gen. It works! Full technical details on real-time VACE + more examples here (link in comments) Curious what people think. Happy to answer questions. Video + Custom LoRA links also in comments. Love, Ryan p.s. I will be back with a sick update on ACEStep implementation tomorrow (links in first comment)

English
7
3
42
1.7K
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
Live remixing with ACEStep 1.5
English
0
1
3
164
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
Control ACEStep1.5 with anything. Here I am using motion from an input video to affect the song. Imagine a tiktok dance video that remixes itself.
English
0
0
3
107
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
Control ACEStep1.5 with anything. Here I am using motion from an input video to affect the song. Imagine a tiktok dance video that remixes itself.
English
0
0
2
68
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
YO, I adapted VACE to work with real-time autoregressive video generation. Here's what it can do right now in real time: - Depth, pose, optical flow, scribble, edge maps — all the v2v control stuff - First frame animation / last frame lead-in / keyframe interpolation - Inpainting with static or dynamic masks - Stacking stuff together (e.g. depth + LoRA, inpainting + reference images) - Reference-to-video is in there too but quality isn't great yet compared to batch Getting ~20 fps for most control modes on a 5090 at 368x640 with the 1.3B models. Image-to-video hits ~28 fps. Works with 14b models as well, but doesnt fit on 5090 with VACE. This is all part of Daydream Scope, an open source tool for running real-time interactive video generation pipelines. The demos were created in/with Scope, and is a combination of Longlive, VACE, and Custom LoRA. There's also a very early WIP ComfyUI node pack wrapping Scope. But how is a real-time, autoregressive model relevant to @ComfyUI ? Ultra long video generation. You can use these models distilled from Wan to do V2V tasks on thousands of frames at once, technically infinite length. I havent experimented much more than validating the concept on a couple thousand frames gen. It works! Full technical details on real-time VACE + more examples here (link in comments) Curious what people think. Happy to answer questions. Video + Custom LoRA links also in comments. Love, Ryan p.s. I will be back with a sick update on ACEStep implementation tomorrow (links in first comment)
English
13
17
103
10.8K
Jakob
Jakob@JakobSustersic1·
@RyanOnTheInside @grok is it possible to run this on 60fps and how much woudl it cost
English
2
0
0
106
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
Updated node pack to support LoRA & Prompt blending with AceStep1.5 in Comfyui. Also latent noise masks and a bunch of other stuff. This allows for using LoRAs and other conditioning at distinct times. Imagine Daft Punk chorus, Dr Dre verse. Workflows and links below
English
3
5
20
1K
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
Use the full power of ACE-Step 1.5 in ComfyUI - beyond the native text2music nodes! 🚀 I built custom guiders so you can finally use: • 🎨 Edit (Extend/Repaint): Add audio before/after or regenerate regions while keeping the rest intact • 🔄 Cover: Style transfer that keeps rhythm & melody but changes vibe/instruments • (WIP) Extract: Pull stems (vocals, drums, bass, guitar...) • (WIP) Lego: Generate matching instrument tracks These let you remix, extend, and create with the full capabilities of ACE Step – all local in ComfyUI, with native datatypes. Workflows, examples + tutorial in the reply below 👇
English
1
0
2
202
RyanOnTheInside retweetledi
Daydream
Daydream@DaydreamLiveAI·
Daydream Scope Workshop: VACE, LoRAs & NDI Support! Join @RyanOnTheInside & James for a deep dive into Scope's latest features: • VACE: Guide generations with reference images + control videos • LoRAs: Custom styles & concepts in video • NDI: Real-time AI video streaming between apps 👇👇👇
English
1
1
2
147
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
Hey y'all, first time posting here. I am here to promote an open source hackathon with $5k+ prizes. It's free to enter so not considered shilling I hope! Its real time interactive AI video with Scope and Daydream (I am a dev on the scope team).
English
4
1
5
241
Purz.ai
Purz.ai@PurzBeats·
Image Filter Live does videos now, realtime playback for dialing in fx, and then it renders on next run.
English
13
16
201
10.8K
Jakob
Jakob@JakobSustersic1·
@RyanOnTheInside @grok can i run this on histed gous or do i need locasl ones if so how strong
English
2
0
0
56
RyanOnTheInside
RyanOnTheInside@RyanOnTheInside·
This is StreamDiffusion with SDXL + IPAdapter accelerated with TensorRT. There are also multiple controlnets running, otherwise the FPS hits like 30 on my 5090 at this resolution. Hella fun There's also multi-stage processing we have been working on, allowing for the latent feedback in this example @Livepeer @DaydreamLiveAI
English
5
0
23
2K
el.cine
el.cine@EHuanglu·
omg.. this new AI is unbelievable you can animate objects in image by literally dragging it with your mouse.. it's live, in real time, infinite.. and you can move the camera too this is a huge breakthrough for AI video, link in comment here's why its crazy:
English
63
252
1.8K
136.4K
Xun Huang
Xun Huang@xxunhuang·
We present MotionStream — real-time, long-duration video generation that you can interactively control just by dragging your mouse. All videos here are raw, real-time screen captures without any post-processing. Model runs on a single H100 at 29 FPS and 0.4s latency.
English
35
150
1.1K
96.5K
ATS Vibes
ATS Vibes@ATSVibes·
And I think I'm losing my head again
English
7
7
73
1.1K