Jonathan Downing

402 posts

Jonathan Downing

Jonathan Downing

@jdowning

Building @QuaiNetwork @PelagusWallet @KipperMoney @BlipPay

Austin Bergabung Ocak 2010
309 Mengikuti537 Pengikut
Jonathan Downing
Jonathan Downing@jdowning·
@katexbt Founders who sell their tokens before their job is done don’t deserve the role in the first place.
English
0
0
0
131
katexbt.hl
katexbt.hl@katexbt·
I used to be sad when I saw protocols shut down and post their goodbyes Then I realized more than half the founders made out like bandits just by being at the right place, right time, and with a dubious sense of morality
katexbt.hl tweet media
English
17
1
94
5.9K
Jonathan Downing me-retweet
Kraken Listings
Kraken Listings@krakenlistings·
Now live: $QUAI $QUAI powers @QuaiNetwork, the world's first energy-based monetary system, built on scalable Proof-of-Work with 50,000+ TPS. Start trading today → app.kraken.com/JDNW/QUAI
Kraken Listings tweet media
English
34
41
160
34K
Jonathan Downing me-retweet
alan ⚡💵
alan ⚡💵@0xalank·
This is for all the haters. The FUDers. Those who said crypto is dead. To those who gave up on the vision. Bitcoin is compromised, Ethereum is institutionalized, and Solana is centralized. For the people, we present you an escape hatch called @QuaiNetwork. The future of money is energy based. Ask your AI about it.
Kraken Listings@krakenlistings

Now live: $QUAI $QUAI powers @QuaiNetwork, the world's first energy-based monetary system, built on scalable Proof-of-Work with 50,000+ TPS. Start trading today → app.kraken.com/JDNW/QUAI

English
15
22
78
3.2K
Vamz
Vamz@Vamzzz93·
@sudoingX Can you tell me how usable qwen 3.5 dense is on M5 Max 128GB unified (I know the bandwidth is lower), but is it tolerable?
English
4
0
0
1.9K
Sudo su
Sudo su@sudoingX·
anyone with an RTX 3090 should be paying attention. this machine is not just a gaming setup in 2026. it can host an intelligence that one-shotted an entire game on a single card that 2x H200 at full precision could not. get your hands dirty. compile llama.cpp with CUDA. run qwen 3.5 27B dense Q4_K_M. undisputed on 24GB. the best model for that card and it's not even close. i haven't tested gemma 4 yet but it's on the queue. the 3090 era is far from over.
Spesh@Speshal_f

@sudoingX gave me confidence that small players can play too

English
56
37
736
55.5K
Jonathan Downing me-retweet
Kraken Listings
Kraken Listings@krakenlistings·
Coming soon: $QUAI $QUAI is the native token of @QuaiNetwork, a scalable Proof-of-Work blockchain merged-mined by the same hardware securing Bitcoin and Litecoin. 50,000+ TPS, sub-penny fees. Trading starts April 8 Get ready → kraken.com/sign-up
Kraken Listings tweet media
English
70
73
307
66.4K
Jonathan Downing
Jonathan Downing@jdowning·
@Steph_Curdy @mechanikalk @0xalank Open source image and video models (stable diffusion) are smaller than open source LLMs, mostly because they are designed differently. Text has sequence, strict syntax, logical reasoning etc but images don't, at least not in the same way. Closed source models are also smaller.
English
2
0
2
48
Steph Curdy
Steph Curdy@Steph_Curdy·
@mechanikalk @0xalank Video/image models, not text, would seem to be higher density for communication than text. But I don’t yet know what energy capacity/hardware is required for the use case that comes from image models rather than text/code models doing the knowledge work use case Maybe space isn’t necessary, maybe it’s a laptop, but I’m pretty sure current architectures for image processing are heavier than text. Would love to be more well informed on this tradeoff
Elon Musk@elonmusk

Video Killed the Radio Star

English
2
1
3
155
Jonathan Downing
Jonathan Downing@jdowning·
@RealTjDunham @jack @michaelneale Well the idea behind paying for inference is that my device can’t do it due to model size. So verification can’t require my device to run the inference. Also sharding large SOTA models over the internet is incredibly slow in my experience
English
1
0
0
45
Tj Dunham
Tj Dunham@RealTjDunham·
verification is one hash comparison. the engine is pure integer math so the same input produces the same hash on any hardware phone, laptop, GPU, doesn't matter. no GPU required. a phone can verify, it just takes longer. but a phone can do more than verify, it can run a piece of the model. a 671B model split across 100 phones, each phone holds a few layers, passes a 14KB hidden state to the next. determinism means every phone's output is provably correct without trusting it. if any phone computes wrong the final hash won't match.
English
1
1
4
129
Tj Dunham
Tj Dunham@RealTjDunham·
great to see this space moving. how can the nodes trust each other compute was done correctly? how can nodes be incentivized to provide compute? built a blockchain that does this.. inference is deterministic on the network so you can shard the compute across any hardware chip github.com/FerrumVir/arc-…
English
8
22
68
7.6K
Jonathan Downing me-retweet
Quai Network ⚡️💵
Quai Network ⚡️💵@QuaiNetwork·
This is how we bring the next billion users onchain. Built compliant with ISO 20022, GDPR, and FATF from day 0.
English
24
24
94
7.3K
Kadir Nar
Kadir Nar@kadirnardev·
@_thomasip Yes, 30B models are very good but running them locally is very difficult. Also, we need very small LLM models to train TTS. The open source community will find new solutions.
English
2
0
6
759
Kadir Nar
Kadir Nar@kadirnardev·
The Qwen team is no longer releasing their models as open source, and this is a big problem for us. We need small models to train many models like TTS, STT, Omni, and others. Previously there was LLaMA, but they're no longer releasing either. The Qwen team won't be releasing anymore either. Our only hope is the LFM models. Minimax, Kimi, and GLM teams are releasing great models for open source, but none of them release small models. And if these companies also stop releasing open source, it's going to be really bad :(
English
72
52
903
74K
Jonathan Downing me-retweet
alan ⚡💵
alan ⚡💵@0xalank·
blockchains are indeed hard to scale although it is not impossible! agreed that lightning is one answer but the entire world cannot onboard to lightning and it also introduces fractured liquidity via hub and spoke we have scaled Proof-of-Work using hierarchical sharding with merge-mined references (think sovereign drivechains / treechains from Peter Todds work) the PoW algorithm to coordinate shards uses what we call Proof of Entropy Minima aka PoEM (paper here: arxiv.org/abs/2303.04305) PoEM uses the intrinsic block weighting to improve cross-shard resolution for conflicts would be interested to hear your thoughts on it here's more on the hierarchical structure: docs.qu.ai/learn/advanced…
English
4
10
29
697
Jonathan Downing
Jonathan Downing@jdowning·
@zawy3 Bitcoiners don’t want to talk about it because they don’t have a good solution other than to censor 30% of the hashrate
English
1
0
2
45
zawy
zawy@zawy3·
It's kind of disappointing to see I'm the only one pointing out the 2-block re-org by Foundry was a selfish mining attack (as indicated by the header arrival times). They also hide their templates. All the news is "this is a normal reorg." Here's the sequence that occured if it were an accident: 1) Network partition prevented Foundry from seeing the new tip and keeps working on it's own. 2) As soon as the honest partition found a new block, in less than a second, everyone can suddenly see Foundry's competing block that's now 1 block behind. The partition was magically fixed at the same time the network found a block. 3) At that time, despite others seeing Foundry's block, Foundry (magically) can't see their block, so it keeps working on it's own tip. 4) Foundry finds another block and magically another partition occurs such that the rest of the network can't see that Foundry has a competing tip. 5) The moment Foundry found another block and is 1 block ahead, the network partition is suddenly fixed and the network can suddenly see both of Foundry's blocks and realize Foundry's tip is in the lead. Just look at the timestamps and the local time when the blocks arrived at each of the nodes. bnoc.xyz/t/two-block-re…
English
6
1
7
379
Jonathan Downing me-retweet
alan ⚡💵
alan ⚡💵@0xalank·
Jason, if you're going to cozy up with the crypto world I suggest you look into what we're building with @QuaiNetwork Quai is the only blockchain that has a truly decentralized form of stability via a flatcoin linked to energy called Qi We're based in Austin (we were here first!) and developed out of UT Austin research Quai offers massive scalability through its hierarchical multi-chain architecture targeting 50k+ TPS, is GPU-mined (ProgPoW alongside SHA/Scrypt), and remains fully decentralized via PoW (unlike $TAO Proof of Authority) We also just built @Entropic_AI, a secure 1-click OpenClaw installer that runs autonomous AI agents locally in a virtualized sandbox right on your device (newly open sourced) Energy-backed money + accessible decentralized compute, could be the way Perfect complement to the $TAO thesis Curious what you think [NFA DYOR ⚡️💵]
English
1
13
51
4.4K
Sudo su
Sudo su@sudoingX·
my DMs are full of this. openclaw users hitting walls and looking for something that actually works on their hardware. hermes agent. local GPU. 35-50 tok/s on a 3060. responds in seconds not minutes. 30+ tools that work on small models without special syntax. if you're migrating from openclaw i will personally help you set up hermes. drop your GPU below.
Daniel Sempere Pico@dansemperepico

My OpenClaw is so unbelievably slow now. I mainly use it for information capture, quick voice note yapping to turn into written posts, and food/workout tracking. I just gave it a very short text to edit and it took 4 minutes to reply. Anybody else experiencing this?

English
31
9
190
17K
JackGK
JackGK@jack_gk·
Austin is the only city where you go shopping for tech glasses and come home with cowboy boots. Good coffee. Great food. Zero RayBan Meta Displays in stock anywhere. Shout out Sarah at Tecovas. 🤠 Austin once again proves it has undeniable motion.
English
15
7
163
14.1K
Jonathan Downing me-retweet
Quasdaq Markets
Quasdaq Markets@quasdaq·
Quasdaq v1 is live on @QuaiNetwork After a week of beta testing with our discord community where we had 10+ markets with 5000+ Quai wagered, we've shipped the full v1 with some updates. Profiles - Full trading stats: win rate, P&L, ROI, best win, total wagered - Activity feed & position history - Look up any wallet address Social Identity - Link your X account to your wallet — your avatar and handle show up across the entire app: comments, leaderboards, activity feed, market creator attribution Public Market Creation - Anyone can create a market with a seed bet - Guided /create wizard with live oracle asset prices Search - Search markets and find traders by @ handle from anywhere in the app Other - Market creator shown on every market detail page - Payout estimates fixed - Cleaner UI throughout Quasdaq is prediction markets on Quai — trade on crypto prices, stocks, forex, and more with real on-chain settlement.👇
Quasdaq Markets tweet media
English
6
10
43
3.9K
Luke Wright
Luke Wright@lukewrightmain·
@mechanikalk @Cellhasher One of the things yes, but mainly AI actually R&D labs and actually running big models, not at amazing speeds but useable for Agents if since its free of course
Luke Wright@lukewrightmain

Here are Android AI @Cellhasher Qwen3.5 + DeepSeek 33B benchmarks. TLDR: its actually worth it if you have old androids for the cheap less than 5w most of these phones operate at. Especially as agents become more deployable and 24/7 hands off or sit and forget. AI model companies will eventually start fine tuning even further to get models into more of a MOE style but tuned directly to the use with Routers, routing each request. Cellhasher is working on this as well. As demand for compute and energy continues, eventually companies will not keep making bigger models and not Company can keep pace with the big ones, they will focus on small local models for retail. We will start with the 5 year old Chipset and Bring you up to speed with some wild results on a newest-chipset. Device: Snapdragon 888 (5 years old chip) CPU only Non-root (28 GB/s memory cap) Rooted devices (56 GB/s) scale ~1.8–1.9x. (using the fastest 4 cores actually outperforms using all 8 cores) Qwen3.5 - 0.8B CPU (4 big cores): 12.54–13.01 tok/s Rooted (1.8–1.9x): 22.6–24.7 tok/s Vulkan GPU (NGL=24): 1.60–1.78 tok/s (7–8x slower) Qwen3.5 - 2B CPU (4 big cores): 9.15–9.36 tok/s Rooted: 16.5–17.8 tok/s Vulkan GPU (NGL=24): 0.78–0.86 tok/s (11–12x slower) Large Models (Non-Root) #p stands for Number of Phones in a parallel pipeline ring made by Cellhasher Swarm AI DeepSeek 33B → 5.89 tok/s (Best was 7.8 tok/s) average is still 5.89 tok/s (12p, d=16, 81% accept) Qwen3.5-35B-A3B → 3.75 tok/s (best was 5.1 tok/s) (3p, d=8, 71.5% accept) Qwen3.5-32B (Coder) → 2.85 tok/s (best 4.8 tok/s) (7p, same-family draft) Rooted Estimate (56 GB/s) DeepSeek 33B → ~10–11 tok/s Qwen3.5-35B → ~6.5–7 tok/s Qwen3.5-32B → ~5–5.5 tok/s Snapdragon 8 Elite Gen 5 plus + 24gb RAM (Android Phone) ~75–85 GB/s memory bandwidth INT4/INT8 NPU usable Well-tuned pipeline Qwen3.5 - 0.8B Model CPU only: 30–45 tok/s CPU + NPU: 70–100+ tok/s Qwen3.5 - 2B Model CPU only: 18–25 tok/s CPU + NPU: 40–60 tok/s DeepSeek 33B CPU only: 15–20 tok/s CPU + NPU (blended): 20–28 tok/s (30 tok/s possible with ideal tuning) Qwen3.5-35B-A3B CPU only: 11–15 tok/s CPU + NPU: 16–22 tok/s Qwen3.5-32B (Coder) CPU only: 10–14 tok/s CPU + NPU: 15–20 tok/s (CPU+NPU is slightly tricky and prefill along with ring pipeline can determine alot, KV cache catching is something i haven't played around with yet) 33B class ~2.5–3.5x over Snapdragon 888 Small models see major NPU uplift Memory bandwidth remains the limiter on large models @Cellhasher has come along way driving inspo from @exolabs over the last few months although things needed to change in order to be correctly managed for android really none of the EXO features are now used, we have our own modification to llama.cpp that overs this ring pipelined parallelism for running LARGE models across however many phones it takes. What's hard is sometimes less phones doesn't always compute to high tokes as some would think less hops will do the trick, in some cases it does other cases like Spec drafting sometimes it doesn't. Regardless Automously running agents 24/7 if i Can run a 80b model at 1-5tok/s on 5 year old Android hardware that runs at 5-15w depending on the amount of phones used, ill take it. If you make the upgrade to the latest generations of phones you get to experience amazing breakthrough of NPU sync and bandwidth optimization that can get you that amazing 20-70 tok/s feel depending per model for every day use. And now im thinking about taking the plunge into 20 of these new phones 20k for 160cores, and 480gb RAM i could possible run the latest and greats at maybe speeds of 10 tok/s.... who knows let me know if you want me to try! (All Phones benchmarked on Ethernet as it offers the best latency over Wifi) (Wifi Still works just slower)

English
1
0
1
105
Dr K.quai⚡💵
Dr K.quai⚡💵@mechanikalk·
I may have just YOLO'd one of these with 128Gb of RAM... Will it be worth it for local AI inference? Or did I just burn $7 grand?
Dr K.quai⚡💵 tweet media
English
15
0
30
2K
shira
shira@shiraeis·
🚨 BREAKING 🚨 Amid reports of Khamenei’s death, Iran unveils its successor: a DeepSeek-powered, YC-backed agentic workflow. Introducing KhamenAI Sources familiar say it’s a $666M seed round, with strategic participation from Qatar and Russia.
English
62
323
6.6K
375.9K