Jukka Seppänen

1.8K posts

Jukka Seppänen banner
Jukka Seppänen

Jukka Seppänen

@Kijaidesign

3D modeling/printing artist, AI enthusiast, rookie Python coder.

Finland Katılım Ekim 2017
79 Takip Edilen6.5K Takipçiler
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
WanVideo2video fun, 129 frames with sliding context windows using the 14B model.
English
11
21
176
19.4K
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
@SlipperyGem I2V is bit more complicated, haven't figured out best way to do that yet.
English
1
0
1
163
Brie Wensleydale🧀🐭
Brie Wensleydale🧀🐭@SlipperyGem·
@Kijaidesign Amazing! I've tried this on the 480p I2V model, 257 frames took 30 min. However, it completely burned. Will you try it on the i2v model?
English
1
0
0
291
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
WanVideo sliding context window test on the 480p 14B T2V -model, worked surprisingly well but incredibly slow (50 mins on a 5090, 513 frames at 832x480)
English
10
9
123
13.3K
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
@jqlive Absolutely, I have the rudimentary implementation in my HunyuanVideo wrapper nodes in ComfyUI as well, it needs some updating to be as good though.
English
0
0
1
276
JQ
JQ@jqlive·
@Kijaidesign Couldn’t sliding context window also work with Hunyuan Video? 🤔
English
1
0
1
267
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
@6___0 It's due to the technique itself, each step basically does the whole video in chunks, which include the overlaps, so many many passes though the model required. Actual VRAM use is really low since only the 81 frames are processed at once, as that's what the model does best.
English
1
0
3
292
kfant
kfant@6___0·
@Kijaidesign is the processsing speed due to low mem?
English
1
0
0
321
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
WanVideo has been lot of fun, currently playing with applying sliding context windowing similar to AnimateDiff, the 1.3B model is especially suitable for this due to it's speed, here's 1025 frames in one go, under 5GB VRAM used, but took about 10 mins on a 5090 with 30 steps.
English
9
8
104
13.5K
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
As an impatient person, I often get obsessed with optimizations, tried implementing TeaCache for WanVideo, failed, but in the process accidentally (maybe) succeeded? Results look promising at least!
English
10
22
187
19.6K
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
@zazoum1 Currently only single prompt with sliding context windows, but I plan to test multiple ones if I can figure all that out.
English
0
0
0
68
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
@gjohgj The model is supposed to output 16fps, though people seem to debate that. This video was saved with 16fps.
English
1
0
5
200
GG
GG@gjohgj·
@Kijaidesign Wait what? 1025 frames? on what framerate?
English
1
0
0
188
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
@el_mejnun I actually have both, I couldn't justify upgrading but 2nd rig for all this stuff I couldn't resist, so I splurged. It's considerably faster, around ~40% or so with all optimizations. Slightly difficult to use still as you have to compile many things from source. Runs hot though
English
1
0
4
212
Reverent Elusarca
Reverent Elusarca@el_mejnun·
@Kijaidesign Thanks for your hard work! Also would love to see every benchmark result with 5090. I believe you had 4090 before so how is the performance so far
English
1
0
2
295
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
@YolaoDude Any text2video model can do some level of video2video as long as the VAE encoder is available, and it already was. The process is basically same as the usual img2img.
English
1
0
5
736
Yolao
Yolao@YolaoDude·
@Kijaidesign So this now can do Video2Video?... i though it was just Text2Video at the moment?
English
1
0
1
891
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
HunhuyanVideo -model's vid2vid passes the hippo test! Very promising and versatile model, thanks to the cfg distillation and every possible optimization I could think of this clip of 101 frames at 768x432 took about 2 minutes to sample, fitting slightly under 20GB using my nodes.
English
34
98
563
66.3K
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
Couple more examples, just changing prompt.
English
2
4
65
15.4K
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
@stilfuchs Oh wow that worked surprisingly good, this is on my list of things to explore as well! Thanks for sharing :D
English
1
0
4
490
Ulf
Ulf@stilfuchs·
@Kijaidesign thanks @Kijaidesign for all the work you do, cant wait to get my fingers on this to make some more test! Here is a quick 3dgs test of your footage without cleaning 😋 This stuff will be so much fun in the next years!
English
5
6
38
2.9K
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
I have finally pushed a bigger update to my CogVideoX ComfyUI wrapper nodes, cleaning up most of the bloat that has been accumulating with all these different models. One of the discoveries I made during this is that the orbit -LoRAs work with the "Fun" -models as well!
English
19
26
200
18.1K
Jukka Seppänen
Jukka Seppänen@Kijaidesign·
@OneStrangeW The workflow, among with many (much simpler) others are included with the nodes. I'm afraid I can't help with huggingface issues as all the models are hosted there.
English
1
0
2
358
RubenTainoAI
RubenTainoAI@OneStrangeW·
Amazing! Any link for the workflow and nodes? Please? I messed up, something with the cog. Now it's asking me for a node that links to it on Huggingface, give me an error 404. I want to start from scratch. I am using Mochi, but I was getting better results with cog, and use less resources. I am not an expert, that is why I believe I better start from scratch.
English
1
0
0
393