Zippy

101 posts

Zippy banner
Zippy

Zippy

@AlexanderRedde3

Software engineer working on ML image generation techniques & inference. Obsessed with programming and life.

Here and there เข้าร่วม Mart 2022
120 กำลังติดตาม566 ผู้ติดตาม
Zippy รีทวีตแล้ว
Luma
Luma@LumaLabsAI·
Uni-1 is here! A new kind of model that thinks and generates pixels simultaneously. Less artificial. More intelligent.
English
405
732
4.6K
3.6M
Zippy
Zippy@AlexanderRedde3·
@hisham_artz You can use the code in the official trt repo to build a trt engine, and run it from there. Mine is different- took the Engine class from repo & modified it to behave like a diffusers unet so i can run it in diffusers pipeline. It's under /demo/Diffusion/ github.com/NVIDIA/TensorRT
English
0
0
3
9
💀
💀@hisham_artz·
@AlexanderRedde3 I'm trying to combine SDXL with TensorRT, do you have a github i can follow? I'd like to recreate this
English
1
0
0
25
Zippy
Zippy@AlexanderRedde3·
Now we're real time! 😎 - with SDXL turbo & tensorrt optimized further.
English
6
2
21
1.5K
Zippy
Zippy@AlexanderRedde3·
@oleg__chomp I think you could probably just compile it normally- following the examples from the tensorrt repo. Most of the speedups from tensorrt are a result of it's own internal autotuning. The api for using that autotuning remains the same for most models.
English
0
0
0
22
Oleg Chomp
Oleg Chomp@oleg__chomp·
@AlexanderRedde3 great result! maybe you can share some tips on connecting taesd to tensorrt pipeline?
English
1
0
0
16
Zippy
Zippy@AlexanderRedde3·
@vibeke_udart Also make sure that the scheduler is using "timestep_spacing": "trailing". And depending on the actual trt impl you are using. If you're just using the code in the trt repo as I did, you'll want to make a diffusers compatible unet wrapper which can be used in normal diffusers.
English
0
0
0
89
Zippy
Zippy@AlexanderRedde3·
@vibeke_udart Well sdxl-turbo should use a maximum of guidance scale ~ 1.5, looks like it is using more. I've also found that sometimes the diffusers LCMScheduler works best for details, though can make the image a bit smudgey. Either LCMScheduler or EulerAncestralDiscreteScheduler.
English
0
0
0
71
Zippy
Zippy@AlexanderRedde3·
@rudzinskimaciej Yes I am 😊 I have a serious case of programming addiction haha
English
0
0
1
30
Zippy
Zippy@AlexanderRedde3·
Finally figured out how to speed up my #sdxlturbo frontend! It's so fast that the only way to show the actual speed is to delete the prompt, since I can't type fast enough 😆 .. built with next.js frontend & tensorrt backend.
English
16
28
217
43.7K
Zippy
Zippy@AlexanderRedde3·
@koltregaskes Essentially, yes. Though the lack of inter-frame coherency even with the same seed would make it seem very jittery, similar to the video.
English
0
0
1
230
Kol Tregaskes
Kol Tregaskes@koltregaskes·
@AlexanderRedde3 If you could type fast enough you could create animation/video on the go? 🤔
English
1
0
0
250
Zippy
Zippy@AlexanderRedde3·
@rudzinskimaciej It's an amd 7950x cpu + 2x4090 machine, though this demo is only using 1 4090. Also thanks! 😊
English
1
0
2
288
Zippy
Zippy@AlexanderRedde3·
@ZealotDKD Right now there isn't a way. It's just a web ui / api running on my workstation. 🥹
English
2
0
2
397
Zippy
Zippy@AlexanderRedde3·
@vibeke_udart There's a demo on their github github.com/NVIDIA/TensorRT. It uses diffusers, so can be pulled automatically though huggingface, though getting tensorrt to the state it's in-in my demo requires a lot of tweaks. Also I recommend using docker, tensorrt can be kina finnicky.
English
0
0
0
137
Zippy
Zippy@AlexanderRedde3·
@JonathanSolder3 Not at the moment, it's just a little app running on my workstation at home. Not sure when or if it'll ever be a part of a public thing.
English
2
0
1
30
Zippy
Zippy@AlexanderRedde3·
Made a neat GUI in react for fun & hooked it up to an optimized sdxl-turbo tensorrt api backend I built for image autocomplete! It generates so fast that my browser cant keep up 🥲 - so fun! #sdxlturbo #sdxl #AIart #stablediffusion
English
3
5
17
2.3K
Zippy
Zippy@AlexanderRedde3·
@EMostaque Your prediction of real time image generation came true 😄. It took a little while, but we're here!
English
1
0
13
673
Zippy
Zippy@AlexanderRedde3·
For anyone who may think that this is a fast-forwarded video, or using image caching, I can guarantee it is not- the images are all base64 and it's using a websocket, the terminal below is the output from the tensorrt rest-api server in real time as I ctrl+y / ctrl+z
English
3
0
24
1.4K
Zippy
Zippy@AlexanderRedde3·
@XWAVEart You are so nice friend <3
English
0
0
1
63
Zippy
Zippy@AlexanderRedde3·
@ScottieFoxTTV But my workstation is on the other side of my room. It's too far. 🥴
English
0
0
1
24
Zippy
Zippy@AlexanderRedde3·
Just broke 20 fps!
English
0
0
2
286
Zippy
Zippy@AlexanderRedde3·
Optimized further meaning I'm getting 17 fps for full clip encode -> unet pass -> vae decode -> serialize to jpeg bytes.
English
0
0
5
266