76

34 posts

76 banner
76

76

@TheLastDevvor

انضم Eylül 2023
2 يتبع27 المتابعون
76
76@TheLastDevvor·
14th donation done!
76 tweet media
English
4
1
5
460
76
76@TheLastDevvor·
13th donation done!
76 tweet media
English
0
0
3
352
76
76@TheLastDevvor·
now that we've bonded the donos are going crazy total of $569 donated now
English
2
0
6
351
76
76@TheLastDevvor·
12th donation done!
76 tweet media
English
1
0
3
291
76
76@TheLastDevvor·
11th donation done!
76 tweet media
English
0
0
3
347
76
76@TheLastDevvor·
$323 donated in total so far! $JOE
English
0
0
6
293
76
76@TheLastDevvor·
10th donation done!
76 tweet media
English
0
0
3
228
76
76@TheLastDevvor·
9th donation done!
76 tweet media
English
1
0
5
403
76
76@TheLastDevvor·
8th donation done!
76 tweet media
English
1
0
4
305
76
76@TheLastDevvor·
7th donation done!
76 tweet media
English
2
0
7
372
76
76@TheLastDevvor·
6th donation done!
76 tweet media
English
2
0
2
242
76
76@TheLastDevvor·
5th donation done!
76 tweet media
English
1
0
3
413
76
76@TheLastDevvor·
4th donation done!
76 tweet media
English
2
0
5
254
76
76@TheLastDevvor·
Third donation done!
76 tweet media
English
0
0
2
228
76
76@TheLastDevvor·
2nd donation done!
76 tweet media
English
0
0
3
161
76
76@TheLastDevvor·
WE FINALLY HAVE THE ACTUAL ARTIST OF JOE REQUESTING FINANCIAL SUPPORT SO WE ARE LIVE ON PUMP FUN DONATING 100% OF REWARDS TO THE PAYPAL RIGHT NOW CA: 9HUauc13Zy2guKXeBDQJYuAvteNntN7CFYgELEZRpump
76 tweet media
3D Universe@3DUniverse

@cryptovillain26 I appreciate the support 🙏 I’m the original creator of Emotiguy, but I no longer own the IP - so I’m not affiliated with or endorsing any tokens/projects. If you’d still like to support me as an artist, it’s truly appreciated: paypal@3dgraphics.co.za Thank you ❤️

English
10
3
20
3.3K
76
76@TheLastDevvor·
9HUauc13Zy2guKXeBDQJYuAvteNntN7CFYgELEZRpump First donation done!
76 tweet media
Deutsch
0
0
1
105
Florida Man
Florida Man@floridamandevs·
HE built "pods" which essentially allows you to upload multipole devices to an ai cloud and they can sync up and work togtehr. literal cracked innovation on a project development level. Redirecting 100% of creator rewarsd to @varun_mathur and revoking edit permissions.
Varun@varun_mathur

Introducing Pods Hyperspace Pods lets a small group of people - a family, a startup, a few friends, to pool their laptops and desktops into one AI cluster. Everyone installs the CLI, someone creates a pod, shares an invite link, and the machines form a mesh. Models like Qwen 3.5 32B or GLM-5 Turbo that need more memory than any single laptop has get automatically sharded across the group's devices - layers split proportionally, inference pipelined through the ring. From the outside it looks like one OpenAI-compatible API endpoint with a pk_* key that drops straight into your AI tools and products. No configuration beyond pasting the key and changing the base URL. A team of five paying for cloud AI burns $500–2,000 a month on API calls. The same team's existing machines can serve Qwen 3.5 (competitive on SWE-bench) and GLM-5 Turbo (#1 on BrowseComp for tool-calling and web research) for free - the hardware is already on their desks. When a query genuinely needs a frontier model nobody has locally, the pod falls back to cloud at wholesale rates from a shared treasury. But for the daily work - code reviews, refactors, research, drafting - local models handle it and nobody gets billed. And when it is idle, you can rent out your pod on the compute marketplace, with fine-grained permissions for access management. There's no central server involved in inference. Prompts go from your machine to your pod members' machines and back: all of this enabled by the fully peer-to-peer Hyperspace network. Pod state - who's a member, which API keys are valid, how much treasury is left - is replicated across members with consensus, so the whole thing works on a local network. Members behind home routers don't need port forwarding either. The practical setup for most pods is three models covering different jobs: Qwen 3.5 32B for code and reasoning, GLM-5 Turbo for browsing and research, Gemma 4 for fast lightweight tasks. All running on hardware you already own. Pods ship today in Hyperspace v5.19. Model sharding, API keys, treasury, and Raft coordinator are all live. What Makes This Different - No middleman. Your prompts travel from your IDE to your pod members' hardware and back. There is no server in between reading your data. - No vendor lock-in. Pod membership, API keys, and treasury are replicated across your own machines using Raft consensus. If the internet goes down, your local network keeps working. There is no database in someone else's cloud that your pod depends on. - Automatic sharding. You don't configure layer ranges or calculate VRAM budgets. Tell the pod which model you want. It figures out how to split it across whatever hardware is online. - Real NAT traversal. Your friend behind a home router with a dynamic IP? Works. No VPN, no Tailscale, no port forwarding. The nodes handle it. - Free when local. This is the part that matters most. Cloud AI bills scale with usage. Pod inference on local hardware scales with nothing. The marginal cost of your 10,000th prompt is the electricity your laptop was already using. Coming soon: - Pod federation: pods form alliances with other pods. - Marketplace: pods with spare capacity can sell inference to other pods.

English
8
0
3
952
76
76@TheLastDevvor·
@varun_mathur 589BX1KK8xeQ2ucFCyTgPKnfwgivvYaArApoWMjpump
Filipino
0
0
0
4
Varun
Varun@varun_mathur·
Introducing Pods Hyperspace Pods lets a small group of people - a family, a startup, a few friends, to pool their laptops and desktops into one AI cluster. Everyone installs the CLI, someone creates a pod, shares an invite link, and the machines form a mesh. Models like Qwen 3.5 32B or GLM-5 Turbo that need more memory than any single laptop has get automatically sharded across the group's devices - layers split proportionally, inference pipelined through the ring. From the outside it looks like one OpenAI-compatible API endpoint with a pk_* key that drops straight into your AI tools and products. No configuration beyond pasting the key and changing the base URL. A team of five paying for cloud AI burns $500–2,000 a month on API calls. The same team's existing machines can serve Qwen 3.5 (competitive on SWE-bench) and GLM-5 Turbo (#1 on BrowseComp for tool-calling and web research) for free - the hardware is already on their desks. When a query genuinely needs a frontier model nobody has locally, the pod falls back to cloud at wholesale rates from a shared treasury. But for the daily work - code reviews, refactors, research, drafting - local models handle it and nobody gets billed. And when it is idle, you can rent out your pod on the compute marketplace, with fine-grained permissions for access management. There's no central server involved in inference. Prompts go from your machine to your pod members' machines and back: all of this enabled by the fully peer-to-peer Hyperspace network. Pod state - who's a member, which API keys are valid, how much treasury is left - is replicated across members with consensus, so the whole thing works on a local network. Members behind home routers don't need port forwarding either. The practical setup for most pods is three models covering different jobs: Qwen 3.5 32B for code and reasoning, GLM-5 Turbo for browsing and research, Gemma 4 for fast lightweight tasks. All running on hardware you already own. Pods ship today in Hyperspace v5.19. Model sharding, API keys, treasury, and Raft coordinator are all live. What Makes This Different - No middleman. Your prompts travel from your IDE to your pod members' hardware and back. There is no server in between reading your data. - No vendor lock-in. Pod membership, API keys, and treasury are replicated across your own machines using Raft consensus. If the internet goes down, your local network keeps working. There is no database in someone else's cloud that your pod depends on. - Automatic sharding. You don't configure layer ranges or calculate VRAM budgets. Tell the pod which model you want. It figures out how to split it across whatever hardware is online. - Real NAT traversal. Your friend behind a home router with a dynamic IP? Works. No VPN, no Tailscale, no port forwarding. The nodes handle it. - Free when local. This is the part that matters most. Cloud AI bills scale with usage. Pod inference on local hardware scales with nothing. The marginal cost of your 10,000th prompt is the electricity your laptop was already using. Coming soon: - Pod federation: pods form alliances with other pods. - Marketplace: pods with spare capacity can sell inference to other pods.
English
130
237
2.5K
200.8K