Bram Wallace

68 posts

Bram Wallace

Bram Wallace

@bram_wallace

Katılım Mayıs 2022
86 Takip Edilen517 Takipçiler
Bram Wallace retweetledi
Sander Dieleman
Sander Dieleman@sedielem·
📢 Another #NeurIPS, another diffusion circle! Join us to talk about diffusion models on Friday Dec 5 at 3:30PM in San Diego! Bayside terrace outside room 11 (upstairs) ☀️🚢🌊 Please help spread the word, tell your friends! No slides, no talks, we just sit down and chat 🗣️
English
7
34
215
63.2K
Bram Wallace retweetledi
ICLR 2026
ICLR 2026@iclr_conf·
That's a wrap for #ICLR2025! See you all next year in Brazil! Please all welcome @BharathHarihar3 as the new Senior Program Chair! (With @cvondrick continuing on as General Chair.)
ICLR 2026 tweet media
English
8
58
682
130.6K
Bram Wallace retweetledi
Sam Altman
Sam Altman@sama·
GPT-4.5 is ready! good news: it is the first model that feels like talking to a thoughtful person to me. i have had several moments where i've sat back in my chair and been astonished at getting actually good advice from an AI. bad news: it is a giant, expensive model. we really wanted to launch it to plus and pro at the same time, but we've been growing a lot and are out of GPUs. we will add tens of thousands of GPUs next week and roll it out to the plus tier then. (hundreds of thousands coming soon, and i'm pretty sure y'all will use every one we can rack up.) this isn't how we want to operate, but it's hard to perfectly predict growth surges that lead to GPU shortages. a heads up: this isn’t a reasoning model and won’t crush benchmarks. it’s a different kind of intelligence and there’s a magic to it i haven’t felt before. really excited for people to try it!
English
3.1K
3.6K
40.5K
5.5M
Bram Wallace retweetledi
Rohan Sahai
Rohan Sahai@rohanjamin·
sora.com signups are fully open
English
39
41
312
849.8K
Bram Wallace retweetledi
will depue
will depue@willdepue·
sora is launching today to all chatgpt pro and plus users! it's been a big effort to make this possible + i think the product is really fun & intuitive. my fav thing to do is generate fake historical found footage. video inpainting is also really strong. have fun!
English
85
77
993
134K
Bram Wallace retweetledi
Bill Peebles
Bill Peebles@billpeeb·
Today's the day! sora.com Included with ChatGPT Plus/Pro!
English
83
99
777
583.2K
Bram Wallace retweetledi
august kamp
august kamp@guskamp·
here's a clip of video set to my new song "sugarize" - every frame created from clips i made using Sora, by @OpenAI
English
5
14
83
50.6K
Bram Wallace retweetledi
Paul Trillo
Paul Trillo@paultrillo·
Made with Sora. The Golden Record - from raw earth material to a time capsule of human life on Earth. Using 11 different generations cut together from Sora, I was able to explore what the odyssey of this record might look like. @OpenAI
English
143
227
1.4K
294.6K
Bram Wallace retweetledi
shy kids
shy kids@shykids·
'air head' is one of the first short films made using #Sora by @OpenAI. the response so far has left us floating.🎈
English
1.1K
555
3.3K
5.6M
Bram Wallace
Bram Wallace@bram_wallace·
Great work! Love the integration of dataset + self learning and the results look sweet. I think this style of method has tons of potential
AK@_akhaliq

Self-Play Fine-Tuning of Diffusion Models for Text-to-Image Generation paper page: huggingface.co/papers/2402.10… Fine-tuning Diffusion Models remains an underexplored frontier in generative artificial intelligence (GenAI), especially when compared with the remarkable progress made in fine-tuning Large Language Models (LLMs). While cutting-edge diffusion models such as Stable Diffusion (SD) and SDXL rely on supervised fine-tuning, their performance inevitably plateaus after seeing a certain volume of data. Recently, reinforcement learning (RL) has been employed to fine-tune diffusion models with human preference data, but it requires at least two images ("winner" and "loser" images) for each text prompt. In this paper, we introduce an innovative technique called self-play fine-tuning for diffusion models (SPIN-Diffusion), where the diffusion model engages in competition with its earlier versions, facilitating an iterative self-improvement process. Our approach offers an alternative to conventional supervised fine-tuning and RL strategies, significantly improving both model performance and alignment. Our experiments on the Pick-a-Pic dataset reveal that SPIN-Diffusion outperforms the existing supervised fine-tuning method in aspects of human preference alignment and visual appeal right from its first iteration. By the second iteration, it exceeds the performance of RLHF-based methods across all metrics, achieving these results with less data.

English
0
0
6
359
Quanquan Gu
Quanquan Gu@QuanquanGu·
Thanks to @_akhaliq for sharing our work! Welcoming a new member, SPIN-Diffusion, into the SPIN family. arxiv.org/abs/2402.10210 Joint work w/ @HuizhuoY @_zxchen_ @Kaixuan_Ji_19
AK@_akhaliq

Self-Play Fine-Tuning of Diffusion Models for Text-to-Image Generation paper page: huggingface.co/papers/2402.10… Fine-tuning Diffusion Models remains an underexplored frontier in generative artificial intelligence (GenAI), especially when compared with the remarkable progress made in fine-tuning Large Language Models (LLMs). While cutting-edge diffusion models such as Stable Diffusion (SD) and SDXL rely on supervised fine-tuning, their performance inevitably plateaus after seeing a certain volume of data. Recently, reinforcement learning (RL) has been employed to fine-tune diffusion models with human preference data, but it requires at least two images ("winner" and "loser" images) for each text prompt. In this paper, we introduce an innovative technique called self-play fine-tuning for diffusion models (SPIN-Diffusion), where the diffusion model engages in competition with its earlier versions, facilitating an iterative self-improvement process. Our approach offers an alternative to conventional supervised fine-tuning and RL strategies, significantly improving both model performance and alignment. Our experiments on the Pick-a-Pic dataset reveal that SPIN-Diffusion outperforms the existing supervised fine-tuning method in aspects of human preference alignment and visual appeal right from its first iteration. By the second iteration, it exceeds the performance of RLHF-based methods across all metrics, achieving these results with less data.

English
3
6
54
15.3K
Bram Wallace
Bram Wallace@bram_wallace·
@EMostaque Love it! Excited to see how DPO alignment can help!
English
0
0
0
94
Emad
Emad@EMostaque·
The #stablecascade output will be even better with DPO (note three stage..) & of course can turbofy it, quantise it etc This is a research preview benchmark/vanilla model but produces great images & solid text out of the box that you can improve with ComfyUI flows
Andrew Carr 🤸@andrew_n_carr

The output is pretty stunning - some highlights: 1. fewer denoising steps 2. cascade improves text generation 3. native image variation generation 4. works with controlnet!

English
4
9
76
12.6K
Bram Wallace
Bram Wallace@bram_wallace·
@Suhail Finetune, nothing here is commercialized and the turbo experiment was just a proof-of-concept run to test how this works on distilled models
English
0
0
1
128
Suhail
Suhail@Suhail·
@bram_wallace Did you finetune sdxl-turbo or did you repro sdxl-turbo and then repro given the licensing restrictions?
English
1
0
0
158
Bram Wallace
Bram Wallace@bram_wallace·
Turboing into 2024! Why let SDXL-Base have all the fun? DPO-tuning 4-step SDXL-Turbo turns out to work pretty well. Original SDXL-Turbo on left, Turb(DP)o on right. The video shown is sped up 2x to stop typing being the bottleneck
English
2
4
13
11.5K
Bram Wallace
Bram Wallace@bram_wallace·
Our research codebase for Diffusion-DPO is now public! github.com/SalesforceAIRe… This is the exact (cleaned up) code we used for the primary results in our paper. It also has an SDXL-Turbo script which works well (see x.com/bram_wallace/s…). Enjoy!
Bram Wallace@bram_wallace

Turboing into 2024! Why let SDXL-Base have all the fun? DPO-tuning 4-step SDXL-Turbo turns out to work pretty well. Original SDXL-Turbo on left, Turb(DP)o on right. The video shown is sped up 2x to stop typing being the bottleneck

English
0
17
78
9.8K