Parallax

73 posts

Parallax banner
Parallax

Parallax

@tryParallax

build your own ai cluster. run open models across your machines.

Katılım Aralık 2025
38 Takip Edilen1.2K Takipçiler
Parallax
Parallax@tryParallax·
@tomosman @openclaw @NousResearch mac minis are underrated for this. we've been running multi-node setups on apple silicon with parallax and the performance-per-dollar is hard to beat. nice to see more people building this way.
English
0
0
0
9
Tom Osman 🐦‍⬛
Tom Osman 🐦‍⬛@tomosman·
Infinitely bullish on a stack of MacMinis or Studios at home running @openclaw or @NousResearch Hermes. Run local models and soon you will have AGI at home. Lots of other epic stuff too but feels downstream of being able to do this.
English
2
2
26
1.2K
Parallax
Parallax@tryParallax·
@cyb3rops the "local" label is doing a lot of heavy lifting for some of these apps. if your data still round-trips to someone else's server, it's not really local. with parallax, your inference actually stays on your machines. your devices, your models, even offline.
English
1
0
7
320
Florian Roth ⚡️
Florian Roth ⚡️@cyb3rops·
Can anyone explain this to me? First Claude Workspace, then Perplexity, now Manus - they keep using words like “my”, “personal”, and “local” in a way that suggests local information isn’t being sent to a remote LLM or RAG system for evaluation. But if no local LLM is actually running, then almost nothing except maybe config stays local. The reasoning still happens remotely. Right? Also - does anyone really think this belongs on a corporate workstation?
Manus@ManusAI

Today, we're taking Manus out of the cloud and putting it on your desktop. Introducing My Computer, the core feature of the new Manus Desktop app. It’s your AI agent, now on your local machine.

English
41
23
254
37.4K
Parallax
Parallax@tryParallax·
some parallax dev lunch break fun: - a macbook pro, a mac mini, some cables - zero internet, zero cost - openclaw running on parallax no subs. no token burn. nothing leaves the desk. just local agents vibing.
English
16
16
109
9.7K
Alex Finn
Alex Finn@AlexFinn·
If you have your OpenClaw working 24/7 using frontier models like Opus, you're easily burning $300 a day. That's $100,000 a year. I have 3 Mac Studios and a DGX Spark running 4 high end local models (Nemotron 3, Qwen 3.5, Kimi K2.5, MiniMax2.5). They're chugging 24/7/365. I spent a third of that yearly cost to buy these computers I'll be able to use them for years for free On top of that they're completely private, secure, and personalized. Not a single prompt goes to a cloud server that can be read by an employee or used to train another model I hope this makes it painfully obvious why local is the future for AI agents. And why America needs to enter the local AI race.
Alex Finn tweet media
English
433
167
2.4K
379.5K
Parallax
Parallax@tryParallax·
messari's new report on echo-2 highlights how parallax powers the rollout plane. consumer RTX 5090s served as distributed rollout actors via parallax, feeding a centralized learner cluster. and we got 33-36% lower hardware costs with no quality loss. this is parallax doing what it was built for. turning consumer hardware into production AI infrastructure.
Youssef@0xYoussef_

Beyond improvements in speed and cost, Echo-2 demonstrates a high standard of model performance. Benchmarking data across five math reasoning tasks shows Echo-2 achieving an average score of 35.75, compared to 35.30 for ByteDance’s verl. These results confirm that the architectural efficiencies of the Open Intelligence Stack (OIS) do not come at the expense of reasoning capabilities.

English
2
2
25
737
Parallax
Parallax@tryParallax·
@farukomerekinci @perplexity_ai this is the right framing. but owning the stack means owning the inference too. openclaw + cloud api = local agent, cloud brain. openclaw + parallax = local agent, local brain. that's the difference between "your data stays on your machine" as marketing vs as reality.
English
1
0
8
1.6K
Faruk Ekinci
Faruk Ekinci@farukomerekinci·
Here's the difference with OpenClaw: Perplexity: Their AI, their servers, your data through their pipeline. One model. One product. Take it or leave it. OpenClaw: Open source. Runs any model, Claude, Grok, Kimi, whatever you want. Your data never leaves your machine. You build the agents, you set the rules, you own the stack. What's now on the table, with a $1B company validating the category: -AI that checks your email before you wake up -Agents monitoring your business 24/7 -Cron jobs running strategies while you're offline -Your entire workflow automated, on hardware you own The difference between Perplexity's version and what you can build yourself isn't features. It's control. Perplexity = Shopify. OpenClaw = owning the server.
English
33
28
535
83.6K
Perplexity
Perplexity@perplexity_ai·
Announcing Personal Computer. Personal Computer is an always on, local merge with Perplexity Computer that works for you 24/7. It's personal, secure, and works across your files, apps, and sessions through a continuously running Mac mini.
English
1.6K
3.5K
32.5K
14M
Parallax
Parallax@tryParallax·
perplexity just announced always-on ai running on a mac mini. the category is real. the question is whether your always-on ai should phone home to someone else's servers or run entirely on your own hardware. we built parallax for the second option.
Perplexity@perplexity_ai

Announcing Personal Computer. Personal Computer is an always on, local merge with Perplexity Computer that works for you 24/7. It's personal, secure, and works across your files, apps, and sessions through a continuously running Mac mini.

English
0
2
29
733
Parallax
Parallax@tryParallax·
mac studios + dgx spark + openclaw is super solid local stack. the inference speed and context window issues you flagged are exactly what we're focused on with parallax. automatic sharding across mixed hardware and paged attention on mac make a noticeable difference on both. happy to help if you want to test it alongside your current setup!
English
0
0
1
220
Jaroslav Beck
Jaroslav Beck@JaroslavBeck·
After some time of using local AI cluster (Bob), here is my honest take on the good, the bad and overall use case. About a year ago I started playing with local AI models because of the work we do at BottleCap AI. I realised how amazing it actually is to own my own stack and my own data. At first, we used local models mainly because of security reasons as we do lots of AI efficiency research and new product concepts based on that. After OpenClaw was released, something changed for me. I started using local models much more, until they replaced cloud models for most of my deep-thinking tasks beyond work. Eventually, I canceled all my AI cloud subscriptions just to see if I could actually run fully on my local cluster. Hardware: • 2x Mac Studio with M3 Ultra and 512GB unified memory, 32-core CPU • 1x NVIDIA DGX Spark, added recently for prefills and, hopefully soon, faster inference • 10GB LAN Switch for connecting Spark and Mac Studio’s Current models: this is changing pretty frequently 1) “Bob OG”: • Main brain for reasoning and daily tasks • Qwen3.5-397B • Roughly 40-60 tokens/sec (depends on load & task) 2) “Bob Researcher”: • Long term researching • Qwen3.5-27B-Claude-4.6-Opus-Distilled-MLX-4bit: Very experimental 3) “Bob App Developer": • Coding apps and debuging • MiniMax M2.5 Software stack: • OpenClaw: All-local assistant layer • LM Studio: Running models • Exo Labs: Connecting multiple machines into one cluster and testing whether inference improves Where my local stack still lacks: • Deep tasks with big models still take more time to reply than cloud models. • Context window is limitation in the models I use. I’m usually around a 200k token window per session, but compacting works well, so I rarely need to start a new session. • It also seems that OpenClaw in its default state is not handling work with memory very efficiently while filling the context window fairly quickly by default. It was necessary for me to finetune this manually including semantic search and temporal decay which are in default switched off. • Reasoning is good but not at the cloud models level. Also coding is good for the majority of tasks but not top tier. My best use cases right now (March 2026): Best for iterative work where privacy matters and where model needs to be available all the time. • Private or sensitive data: I would be careful as a company to share private or direct customer information with third party cloud systems in general. Clearly also connecting OpenClaw to cloud models is not solving privacy situation. • Cloud limits & Efficiency: If I push cloud subscriptions hard, I hit consumer limits surprisingly fast. It’s also much easier to spot inefficiencies locally. When the context starts bloating, the system slows down fast, so issues like memory inefficiency become obvious much earlier. In the cloud, replies often feel just as fast, but you end up paying much more or hitting usage limits without really knowing why. Was it worth the money? For me, yes. But I’m aware I live in a niche bubble for my particular use case. For most people it is still early. For businesses and people who want to spend the money and effort make this work it is good solution today. My verdict: For my personal use case, local is now the default. Cloud is the exception. Are local models as good as the best cloud models? No. Are they good enough to be my default for most tasks? Yes.
Jaroslav Beck tweet media
English
45
37
586
47.9K
Parallax
Parallax@tryParallax·
a 9b model now matches a 120b on reasoning benchmarks. models are shrinking fast. the hardware you already own is becoming more powerful every release cycle. this is the trend parallax was built for. connect your machines, run what fits, and scale when you need to.
Artificial Analysis@ArtificialAnlys

Alibaba has released 4 new Qwen3.5 models from 0.8B to 9B. The 9B (Reasoning, 32 on the Intelligence Index) is the most intelligent model under 10B parameters, and the 4B (Reasoning, 27) the most intelligent under 5B, but both use 200M+ output tokens to run the Intelligence Index @Alibaba_Qwen has expanded the Qwen3.5 family with four smaller dense models: the 9B (Reasoning, 32 on the Intelligence Index), 4B (Reasoning, 27), 2B (Reasoning, 16), and 0.8B (Reasoning, 9). These complement the larger 397B, 27B, 122B A10B, and 35B A3B models released earlier this month. All models are Apache 2.0 licensed, support 262K context, include native vision support, and use the same unified thinking/non-thinking hybrid approach as the rest of the Qwen3.5 family Key benchmarking results for the reasoning variants: ➤ The 9B and 4B are the most intelligent models at their respective size classes, ahead of all other models under 10B parameters. Qwen3.5 9B (32) scores roughly double the next closest models under 10B: Falcon-H1R-7B (16) and NVIDIA Nemotron Nano 9B V2 (Reasoning, 15). Qwen3.5 4B (27) outscores all of these despite having roughly half the parameters. All four of the small Qwen3.5 models are on the Pareto frontier of the Intelligence vs. Total Parameters chart ➤ The Qwen3.5 generation represents a material intelligence uplift over Qwen3 across all sub-10B model sizes, with larger gains at higher total parameter counts. Comparing reasoning variants: Qwen3.5 9B (32) is 15 points ahead of Qwen3 VL 8B (17), the 4B (27) gains 9 points over Qwen3 4B 2507 (18), the 2B (16) is 3 points ahead of Qwen3 1.7B (estimated 13), and the 0.8B (9) gains 2.5 points over Qwen3 0.6B (6.5). ➤ All four models use 230-390M output tokens to run the Intelligence Index, significantly more than both larger Qwen3.5 siblings and Qwen3 predecessors. Qwen3.5 2B used ~390M output tokens, 4B used ~240M, 0.8B used ~230M, and 9B used ~260M. For context, the much larger Qwen3.5 27B used 98M and the 397B flagship used 86M. These token counts also exceed most frontier models: Gemini 3.1 Pro Preview (57M), GPT-5.2 (xhigh, 130M), and GLM-5 Reasoning (109M) ➤ AA-Omniscience is a relative weakness, with hallucination rates of 80-82% for the 4B and 9B. Qwen3.5 4B scores -57 on AA-Omniscience with a hallucination rate of 80% and accuracy of 12.8%. Qwen3.5 9B scores -56 with 82% hallucination and 14.7% accuracy. These are marginally better than their Qwen3 predecessors (Qwen3 4B 2507: -61, 84% hallucination, 12.7% accuracy), with the improvement driven primarily by lower hallucination rates rather than higher accuracy. ➤ The Qwen3.5 sub-10B models combine high intelligence with native vision at a scale previously unavailable. On MMMU-Pro (multimodal reasoning), Qwen3.5 9B scores 69.2% and 4B scores 65.4%, ahead of Qwen3 VL 8B (56.6%), Qwen3 VL 4B (52.0%), and Ministral 3 8B (46.0%). The Qwen3.5 0.8B scores 25.8%, which is notable for a sub-1B model Other information: ➤ Context window: 262K tokens ➤ License: Apache 2.0 ➤ Quantization: Native weights are BF16. Alibaba has not released first-party GPTQ-Int4 quantizations for these small models, though they have for the larger models in the Qwen3.5 family released earlier (27B, 35B-A3B, 122B-A10B, 397B-A17B). In 4-bit quantization all four models are accessible on consumer hardware ➤ Availability: At time of publishing, there are no first-party or third-party serverless APIs hosting these models

English
8
8
43
747
Parallax
Parallax@tryParallax·
ancient romans built public libraries changed the way how knowledge is accessed. today, open source AI is doing the same job. open doors, open intelligence. 🏛️
Supercycle@supercyclepod

AI should be a public good, not something gatekept by a handful of megacorps We had Eric Yang, co-founder of Gradient Network, on the pod this week to talk through exactly that. Gradient's "Open Intelligence Stack" includes: i) Parallax for distributed model serving ii) Echo for decentralized reinforcement learning The whole thesis is that anyone should be able to run large models on consumer hardware (yes, including your Mac Minis + OpenClaws) Eric breaks down their $10M seed round led by Pantera, Multicoin, and HSG; where he sees the industry heading; and why post-training is going to be the dominant force in enterprise. Timestamps: 00:00 Intro 01:15 AI market is booming 02:29 Local compute is a hot topic 03:02 Parallax Inference Engine 04:34 Intelligence as a public good 05:46 AI models will become a commodity 07:32 Bottlenecks in AI models accessibility 09:34 Smaller AI models are catching up 11:01 How Gradient's Infrastructure Enables Model Development 12:15 Model post-training 14:24 How does reinforcement learning work? 17:35 AI going rogue 19:20 Gradient's token 23:02 AI entrepreneurs that Eric admires 26:11 Use cases on chain for AI 31:34 The trade-offs of coming to crypto 35:09 How low-spec GPUs will work on Gradient Ecosystem 38:08 Post-training will be the dominating force for enterprise 38:43 Open source models are way cheaper 41:39 Eric's founding story 49:07 Empowering researchers globally 53:37 Why did Multicoin Capital and Pantera Capital invested in Gradient 55:08 One-click deploy agent 58:16 Gradient in 3 years

English
8
1
37
1.5K
Parallax retweetledi
Gradient
Gradient@Gradient_HQ·
We also tested the messier setups. Using Parallax, we trained Qwen3-8B on distributed RTX 5090s. 36% cheaper than centralized A100s, same scores, zero divergence. We even trained a 0.6B agent to beat LLMs at No-Limit Texas Hold'em. Reliable results with unreliable compute.
Gradient tweet media
English
2
7
60
3.1K
Parallax retweetledi
Gradient
Gradient@Gradient_HQ·
They crashed. They fell. They exploded on the pad. Then they got back up. Faster, wiser, stronger. Breakthroughs don't come from one perfect run, they come from the freedom to fail 100 times. Introducing Echo-2, distributed RL that boosts AI research throughput by 10x.
English
133
107
566
101.5K
Parallax retweetledi
Commonstack
Commonstack@commonstack_ai·
The wait is over. Open your Clawbox now! The easiest way to own your🦞agent. Zero code. Just download, click, and go. Grab yours: commonstack.ai/clawdbot?ref=x
English
16
25
113
18.6K
Parallax
Parallax@tryParallax·
spoiler alert 🦞 we’re preparing a fully private version of clawbox. air-gapped ready. run it completely offline. absolutely zero risks tolerated with sovereign AI. 🛡️
Commonstack@commonstack_ai

Want an @openclaw🦞agent in just a few clicks? Meet Clawbox🦞, powered by Commonstack. -No CLI. No VPS. Just download & run. -Permission controls for peace of mind. -Switch between 30+ LLMs instantly. Free credits + 20% off for all! 🦞Join whitelist: commonstack.ai/clawdbot

English
10
7
52
2.9K
Parallax retweetledi
Commonstack
Commonstack@commonstack_ai·
Want an @openclaw🦞agent in just a few clicks? Meet Clawbox🦞, powered by Commonstack. -No CLI. No VPS. Just download & run. -Permission controls for peace of mind. -Switch between 30+ LLMs instantly. Free credits + 20% off for all! 🦞Join whitelist: commonstack.ai/clawdbot
Commonstack tweet media
English
25
11
121
53.3K
rw ./
rw ./@gradientintern·
run locally and never get revoked 🦞 full sovereignty, save dollars and protect privacy. ./ @tryParallax by @Gradient_HQ
rw ./ tweet media
English
12
3
39
1.1K
Parallax
Parallax@tryParallax·
what are you building with your mac mini today? hint: save your $5 for vps and have some privacy with parallax.
Parallax tweet media
English
9
13
78
6.9K