Stable Diffusion Tutorials

2.1K posts

Stable Diffusion Tutorials banner
Stable Diffusion Tutorials

Stable Diffusion Tutorials

@SD_Tutorial

๐Ÿ‘‰ Ai models local installation ๐Ÿ‘‰ Comfy Workflows ๐Ÿ‘‰ Tutorials (Image Gen, Video gen) FOLLOW WEBSITE ๐Ÿ‘‡๐Ÿ‘‡

Bergabung Mayฤฑs 2018
88 Mengikuti1.6K Pengikut
Tweet Disematkan
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
Generate PROMPTS like a PRO ๐Ÿ˜น ๐Ÿ‘‡ #stablediffusion #aipromptgenerator ๐Ÿ”–Bookmark (save) for future reference.... ๐Ÿฉท Like if you loved it.... ๐Ÿ–Š๏ธComment/Suggestions if you are getting any error....
English
1
1
13
3.5K
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
WorldMesh๐Ÿ˜Œ Generating Navigable Multi-Room 3D Scenes via Mesh-Conditioned Image Diffusion To generate a complex, multi-room 3D scene from a text prompt, problem into first constructing the global scene structure as a mesh scaffold (top) github page: mschneider456.github.io/world-mesh/
English
0
0
1
56
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
Foveated Diffusion: ๐Ÿ˜ƒ๐Ÿ˜Œ Efficient Spatially Aware Image and Video Generation iteratively denoise a foveated token sequence of reduced length instead of the full high-resolution sequence. github page:๐Ÿ‘‡ bchao1.github.io/foveated-diffuโ€ฆ
Stable Diffusion Tutorials tweet mediaStable Diffusion Tutorials tweet media
English
0
0
2
61
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
ComfyUI-Wan-VACE-Video-Joiner ๐Ÿ‘‡๐Ÿ˜๐Ÿ˜ƒ - point this workflow at a directory of clips and it will automatically stitch them together - fixing awkward motion and artifacts. - Wan VACE generates new frames guided by context on both sides Github:๐Ÿ‘‡ github.com/stuttlepress/Cโ€ฆ
English
0
0
2
73
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
ComfyUI-DaVinci-MagiHuman ๐Ÿ˜€ -Block-level CPU/GPU swapping -Async CUDA prefetching -Distill mode -1080p super-resolution -TurboVAE decoder -Audio + video Access github repo: ๐Ÿ‘‡ github.com/mjansrud/Comfyโ€ฆ
Adina Yakup@AdinaYakup

daVinci-MagiHuman ๐ŸŽฌ Human Centric Audio-Video Generative Model by GAIR Model: huggingface.co/GAIR/daVinci-Mโ€ฆ Paper: huggingface.co/GAIR/daVinci-Mโ€ฆ โœจ 15B โ€“ Fully open source! โœจ 5-sec 1080p video in 38s on one H100 โœจ Supports 6 languages โœจ Unified model with text + video + audio

English
0
0
1
129
Stable Diffusion Tutorials me-retweet
Adina Yakup
Adina Yakup@AdinaYakupยท
Matrix-Game 3.0๐Ÿ”ฅreal-time interactive world models from @Skywork_ai huggingface.co/Skywork/Matrixโ€ฆ โœจ MIT license โœจ 720p @ 40FPS with a 5B model โœจ Minute-long memory consistency โœจ Unreal + AAA + real-world data โœจ Scales up to 28B MoE
English
10
103
621
41.4K
Stable Diffusion Tutorials me-retweet
Black Forest Labs
Black Forest Labs@bfl_mlยท
Black Forest is focused on visual intelligence as the next AI frontier. On a panel with Jensen at @nvidia GTC, our CEO @robrombach defined visual intelligence as understanding the visual world and simulating it to power content creation, real-time applications, and robotics. Link to the full panel in the comments ๐ŸŽฅ
Black Forest Labs tweet mediaBlack Forest Labs tweet mediaBlack Forest Labs tweet mediaBlack Forest Labs tweet media
English
2
10
80
5.5K
Stable Diffusion Tutorials me-retweet
Adina Yakup
Adina Yakup@AdinaYakupยท
daVinci-MagiHuman ๐ŸŽฌ Human Centric Audio-Video Generative Model by GAIR Model: huggingface.co/GAIR/daVinci-Mโ€ฆ Paper: huggingface.co/GAIR/daVinci-Mโ€ฆ โœจ 15B โ€“ Fully open source! โœจ 5-sec 1080p video in 38s on one H100 โœจ Supports 6 languages โœจ Unified model with text + video + audio
English
8
47
339
41.6K
Stable Diffusion Tutorials me-retweet
ModelScope
ModelScope@ModelScope2022ยท
PrismAudio is open source๐Ÿ‘๐Ÿ‘๐Ÿ‘ a 518M V2A model accepted at ICLR 2026, achieving SOTA across all four perceptual dimensions on both VGGSound and the new AudioCanvas benchmark. ๐Ÿ‘€ Demo video below โฌ‡๏ธ Model: modelscope.ai/models/iic/Priโ€ฆ Demo: modelscope.cn/studios/iic/Prโ€ฆ Paper: modelscope.ai/papers/2511.18โ€ฆ GitHub: github.com/FunAudioLLM/Thโ€ฆ ๐Ÿง  Decomposes V2A reasoning into four specialized CoT modules: Semantic, Temporal, Aesthetic, and Spatial โ€” each with targeted reward functions ๐ŸŽฏ First framework to integrate RL into V2A generation via decomposed CoT planning โšก Fast-GRPO: hybrid ODE-SDE sampling that dramatically reduces RL training overhead ๐Ÿ† VGGSound: tops all baselines on CLAP, DeSync, PQ, and subjective MOS scores โ€” at 0.63s inference, faster than MMAudio (1.30s) and ThinkSound (1.07s) ๐ŸŒ AudioCanvas (out-of-domain): CLAP 0.52, MOS-Q 4.12, beats HunyuanVideo-Foley, MMAudio, ThinkSound ๐Ÿ“Š AudioCanvas benchmark released: 300 single-event classes + 501 multi-event samples
English
4
24
147
22.9K
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
Wan 2.7ย is planned to launchย within March - improvements in: visual quality audio motion dynamics stylization consistency first-frame & last-frame video generation 9-grid image-to-video subject + voice reference instruction-based video editing video recreation / replication
Stable Diffusion Tutorials tweet media
English
0
0
1
316
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
ComfyUI custom nodes for ID-LoRA-2.3 inference โ€” audio+video generation with speaker identity transfer, built on top of LTX-2.3. Supports both one-stage (single resolution) and two-stage (2x spatial upsampling) pipelines. ๐Ÿ‘‡๐Ÿ˜ƒ github.com/ID-LoRA/ID-LoRโ€ฆ
Stable Diffusion Tutorials@SD_Tutorial

ID-Lora: Identity-Driven Audio-Video Personalization with In-Context LoRA ๐Ÿ˜ƒ๐Ÿ˜„ Generate video and audio of a specific person from a single text prompt, a reference image, and a short audio clip โ€” all in one model. Now supporting LTX 2.3. Paper:๐Ÿ‘‡ id-lora.github.io

English
0
0
2
274
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
ID-Lora: Identity-Driven Audio-Video Personalization with In-Context LoRA ๐Ÿ˜ƒ๐Ÿ˜„ Generate video and audio of a specific person from a single text prompt, a reference image, and a short audio clip โ€” all in one model. Now supporting LTX 2.3. Paper:๐Ÿ‘‡ id-lora.github.io
Stable Diffusion Tutorials tweet media
English
0
0
3
496
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
Official reported by Alibaba: ๐Ÿ˜ƒ๐Ÿ˜ New Qwen and Wan models will be Opensource !!!
Stable Diffusion Tutorials tweet media
English
0
0
5
141
Stable Diffusion Tutorials me-retweet
lodestone-rock
lodestone-rock@LodestoneRockยท
z-image to pixel space conversion progress is looking good still need more training time for sure, super crunchy. anyone can try this underbaked model, it's supported in comfy. it should be 3-4x faster than z-image base. weight updated every hour here: huggingface.co/lodestones/Zetโ€ฆ
lodestone-rock tweet medialodestone-rock tweet medialodestone-rock tweet medialodestone-rock tweet media
English
5
27
195
25.7K
Stable Diffusion Tutorials me-retweet
Tongyi Lab
Tongyi Lab@Ali_TongyiLabยท
We are thrilled to see what our community can build! Developer @dx8152 has just dropped a specialized Style Transfer LoRA based on Qwen-Image-Edit-2511, which significantly simplifies the way we remix and transform visual aesthetics. Built using ModelScope's code-free training pipeline, this project is a perfect example of how the right infrastructure empowers creators to turn complex ideas into reality.
Tongyi Lab tweet media
English
9
26
232
12.1K
Stable Diffusion Tutorials
Stable Diffusion Tutorials@SD_Tutorialยท
Z-Image-Turbo-SDA ๐Ÿฅฐโ˜บ๏ธ A highly efficient LoKr (Low-Rank Kronecker Product) adapter designed to rescue the "Diversity Collapse" problem in few-step distilled Flow Matching / Diffusion models. ๐Ÿ‘‡๐Ÿ‘‡ huggingface.co/F16/z-image-tuโ€ฆ
English
0
0
3
273