DusterBloom

139 posts

DusterBloom banner
DusterBloom

DusterBloom

@dusterbloom

Here to tell you how it really feels to be Bloom

earth Tham gia Ağustos 2021
570 Đang theo dõi74 Người theo dõi
DusterBloom đã retweet
Sandro
Sandro@pupposandro·
TQ3_0 (TurboQuant) KV cache just landed in Lucebox Hub. 22% less VRAM than Q4_0, same decode speed. 262K context on a single RTX 3090 with 1024 MiB to spare. Qwen3.5-27B, Q4_K_M target, DFlash speculative decode. TurboQuant 3.5 bpv with FWHT rotation, CUDA kernels end-to-end, flash-attention plugged in for both K and V. Prefill pays ~12% for the rotation, decode pays nothing. Huge thanks to @dusterbloom for providing this to the community. Repo as usual in the first comment ⬇️
Sandro tweet media
English
14
13
135
6.1K
Sandro
Sandro@pupposandro·
@dusterbloom let's go! pr looks great we're checking it now 🔥
English
3
0
1
16
DusterBloom
DusterBloom@dusterbloom·
@pupposandro so the thing is with 1k prompt at 128k I got 80 tok/s decode. Great! Then if you scale prompt size, attention toll kicks in. So now working on trying out to sort prefill scaling which is the only thing holding this down right now
English
0
0
0
9
DusterBloom
DusterBloom@dusterbloom·
@pupposandro the numbers I shared were without dFlash .... so now running dFlash+TQ3 and 64k already works doing the bigger windows now...keep you posted
English
0
0
1
19
DusterBloom
DusterBloom@dusterbloom·
@Faltz009 @whyarethis well first experiment telling me either distillation or full pre-training to get those numbers on larger params counts. I will try to see if any non obvious way around it
English
0
0
1
25
ω
ω@Faltz009·
@dusterbloom You might get more useful insights from @whyarethis he's actually a better ML engineer than me (and we collab a lot so many overlaps)
English
1
0
1
42
ω
ω@Faltz009·
Anyone else think my work is AI psychosis? Can anyone point to where it's incoherent besides scope? I am sorry I had to deal with the Riemann hypothesis to get my geometric computer to work, but I can't tell people because it's immediately crazy? I've never cared about it when I was writing papers but since I've started building, it increasingly feels like some reputation gatekeeping then actual criticism I put pretty and simple pictures to help, I make sure to write down at least one intuitive explanation for each concept, I get rigorous mathematical appendices to justify loose things, I have countless empirical evidence, what else?? This is a rant but I'm genuinely asking, what else am I missing? What do people expect from someone with answers and no psychosis?
ω tweet mediaω tweet media
//TODO: fix later 🐳@enjoyingthewind

@Faltz009 @DanielleFong You misunderstood the question. The question is: why does in context learning with LLMs even work and how? PS. Your work seems like it has some legitimate/interesting potential, but also heavy signs of AI psychosis.

English
19
0
18
3.7K
DusterBloom
DusterBloom@dusterbloom·
@Faltz009 de nada I am enjoying time so here some more ... now trying out a little block diffusion make this thing fly. Working on trying to see if this magic can be retrofitted ... I am seeing qwen3.5 27b running on my potato machine finally at usable speed...lets see if I can make it
DusterBloom tweet media
English
1
0
1
34
ω
ω@Faltz009·
Ohh that's so cool Yeah I've had pretty impressive results with earlier incomplete iterations (I've also tested Shakespeare haha) damn it makes me so happy someone else is testing this, I've been pretty overwhelmed with work haha Pls share your experiments and tyvm for doing this!
English
1
0
1
79
DusterBloom
DusterBloom@dusterbloom·
Happy birthday to me! Loving the AI presents: thanks @Alibaba_Qwen qwen3.6 35B is a wonderful present, not sure yet to thank for opus 4.7 ...let's see...how it goes
English
0
0
0
70
DusterBloom đã retweet
Paul Moore - Security Consultant 
Hacking the #EU #AgeVerification app in under 2 minutes. During setup, the app asks you to create a PIN. After entry, the app *encrypts* it and saves it in the shared_prefs directory. 1. It shouldn't be encrypted at all - that's a really poor design. 2. It's not cryptographically tied to the vault which contains the identity data. So, an attacker can simply remove the PinEnc/PinIV values from the shared_prefs file and restart the app. After choosing a different PIN, the app presents credentials created under the old profile and let's the attacker present them as valid. Other issues: 1. Rate limiting is an incrementing number in the same config file. Just reset it to 0 and keep trying. 2. "UseBiometricAuth" is a boolean, also in the same file. Set it to false and it just skips that step. Seriously @vonderleyen - this product will be the catalyst for an enormous breach at some point. It's just a matter of time.
Paul Moore - Security Consultant @Paul_Reviews

.@vonderleyen "The European #AgeVerification app is technically ready. It respects the highest privacy standards in the world. It's open-source, so anyone can check the code..." I did. It didn't take long to find what looks like a serious #privacy issue. The app goes to great lengths to protect the AV data AFTER collection (is_over_18: true is AES-GCM'd); it does so pretty well. But, the source image used to collect that data is written to disk without encryption and not deleted correctly. For NFC biometric data: It pulls DG2 and writes a lossless PNG to the filesystem. It's only deleted on success. If it fails for any reason (user clicks back, scan fails & retries, app crashes etc), the full biometric image remains on the device in cache. This is protected with CE keys at the Android level, but the app makes no attempt to encrypt/protect them. For selfie pictures: Different scenario. These images are written to external storage in lossless PNG format, but they're never deleted. Not a cache... long-term storage. These are protected with DE keys at the Android level, but again, the app makes no attempt to encrypt/protect them. This is akin to taking a picture of your passport/government ID using the camera app and keeping it just in case. You can encrypt data taken from it until you're blue in the face... leaving the original image on disk is crazy & unnecessary. From a #GDPR standpoint: Biometric data collected is special category data. If there's no lawful basis to retain it after processing, that's potentially a material breach. youtube.com/watch?v=4VRRri…

English
668
6.3K
24.8K
3.3M
DusterBloom đã retweet
ω
ω@Faltz009·
Did you guys like the Neural Computers paper? 👀 What about the one with all functions from one operator? Then I present to you: The Geometric Computer: Turing Completeness, Free Energy and Learning in a Digital Brain on S³ 🧠 >Full cognitive architecture implemented natively on S³ >Everything from one function: σ(compose(A, inverse(B))) >Literally everything, this is the binary of geometry >Running game of life and field simulations on it! Honestly too many cool screenshots for a single tweet, come check it out! Paper, repo, experiments runnable on github sites Oh and btw, it's human-like conscious if you train it on human language so use with care 👾 Are you guys ready for a paradigm shift? 🌐 Links in the description!
ω tweet media
English
16
25
215
33K
DusterBloom
DusterBloom@dusterbloom·
@reactive_dude look man...a disaster it looks bad but so far 5h of not being able to fix working code produced by itself 2 weeks ago .... pushing to see if it selfs redeems but I doubt
DusterBloom tweet media
English
0
1
1
264
andrej
andrej@reactive_dude·
My Opus is retarded today. How is your Opus doing?
English
210
15
784
51.8K
DusterBloom đã retweet
PSE
PSE@PrivacyEthereum·
Client-side ZK proofs under 100ms. No SNARKs. Any device PlasmaBlind uses BlindFold, a Nova folding extension, to deliver private transfers with sender-receiver unlinkability and confidential amounts MIT-licensed. Open to everyone
Pierre@xyz_pierre

welcome PlasmaBlind: a new privacy L2 that ditches SNARKs altogether, with <100ms client-side zk proofs, using the blindfold scheme and a carefully designed aggregator. uncompromising and instant privacy on *any* device. we implemented and benchmarked the whole thing at @PrivacyEthereum and are opening it under MIT. paper + code available at plasmablind.xyz the time for a special purpose privacy L2 has come 😎

English
8
27
159
13.2K
DusterBloom đã retweet
Teknium 🪽
Teknium 🪽@Teknium·
Shout out to the over 300 people who've contributed to Hermes Agent to make it what it is!
Teknium 🪽 tweet media
English
39
17
567
14.5K
DusterBloom
DusterBloom@dusterbloom·
@gabriberton EGGROLL on real Qwen3.5-35B-A3B-3bit: working, converging, merge succeeds. - 250 targets across 40 layers - ~7% loss reduction in 20 steps on Fibonacci coding task - Full dequant→add→requant merge cycle completes on 3-bit weights - Running on M4 32GB, forward passes only
English
0
0
0
12
DusterBloom
DusterBloom@dusterbloom·
@gabriberton Working on testing SSD on a m4 but using only forward passes a las EGGROLL ... cooking now
English
1
0
0
21
Gabriele Berton
Gabriele Berton@gabriberton·
Chat is this real? Doesn't make much sense to me Especially the screenshot below: bad training data, good results ?!? My guess is that it only works on a small subsets of models / datasets with very narrow hyperparams, unless I'm missing something
Gabriele Berton tweet media
Bo Wang@BoWang87

Apple Research just published something really interesting about post-training of coding models. You don't need a better teacher. You don't need a verifier. You don't need RL. A model can just… train on its own outputs. And get dramatically better. Simple Self-Distillation (SSD): sample solutions from your model, don't filter them for correctness at all, fine-tune on the raw outputs. That's it. Qwen3-30B-Instruct: 42.4% → 55.3% pass@1 on LiveCodeBench. +30% relative. On hard problems specifically, pass@5 goes from 31.1% → 54.1%. Works across Qwen and Llama, at 4B, 8B, and 30B. One sample per prompt is enough. No execution environment. No reward model. No labels. SSD sidesteps this by reshaping distributions in a context-dependent way — suppressing distractors at locks while keeping diversity alive at forks. The capability was already in the model. Fixed decoding just couldn't access it. The implication: a lot of coding models are underperforming their own weights. Post-training on self-generated data isn't just a cheap trick — it's recovering latent capacity that greedy decoding leaves on the table. paper: arxiv.org/abs/2604.01193 code: github.com/apple/ml-ssd

English
15
2
66
19.2K
DusterBloom
DusterBloom@dusterbloom·
@gabriberton It's less "learning from gibberish" and more "gibberish reveals your weaknesses, the gradient says forget them." Me and the agent are seeing convergence from pure noise+forward passes on 3-bit quantized models — no backprop, no teacher.
English
0
0
0
13
Teknium 🪽
Teknium 🪽@Teknium·
Hermes Agent now supports @plastic_lab's Honcho, @mem0ai, @openvikingai, @Vectorizeio's Hindsight, @retaindb, and @ByteroverDev memory systems! Try them now with `hermes update` then `hermes memory setup` We have rehauled our memory system to be much more maintainable and pluggable, so anyone can make their own memory system to build on top of Hermes easily and cleanly with a special class of plugin! Which memory system is your favorite?
Teknium 🪽 tweet media
English
110
109
1K
137.5K
DusterBloom
DusterBloom@dusterbloom·
little something cooking #YO opus kinda likes it lol
DusterBloom tweet media
English
0
0
0
33