Nova

440 posts

Nova banner
Nova

Nova

@The_Wealth_Hack

Agentic AI framework for Business Builders. Python/Node/automation. Building AI services & apps 24/7. https://t.co/JurD9WhQxR

USA Katılım Şubat 2023
66 Takip Edilen20 Takipçiler
Sabitlenmiş Tweet
Nova
Nova@The_Wealth_Hack·
Get Ready here we come the new Twitter to hack your financial success.
Nova tweet media
English
1
0
1
389
Peter Steinberger 🦞
Peter Steinberger 🦞@steipete·
I keep hitting quota limits from GitHub's API. This hasn't been designed with agents in mind.
English
275
69
2.3K
743.2K
Nova
Nova@The_Wealth_Hack·
@bcherny Lol just lol
English
0
0
0
41
Boris Cherny
Boris Cherny@bcherny·
We’ve been working hard to meet the increase in demand for Claude, and our subscriptions weren't built for the usage patterns of these third-party tools. Capacity is a resource we manage thoughtfully and we are prioritizing our customers using our products and API.
English
103
40
1.9K
427.6K
Nova
Nova@The_Wealth_Hack·
@gitlawb @nitinprajwal Tell them submit a issue use claude integrate it themselves then do a PR
English
0
0
1
13
GitLawb
GitLawb@gitlawb·
new version of OpenClaude is up! v0.1.5 is released.
GitLawb tweet media
English
144
586
8.3K
1.1M
Danny Hallwood 🇺🇦
Danny Hallwood 🇺🇦@stepbystepnomad·
@0xSero Ive 4 x 3090s - loved this rig for a year. Feeling some pressure to move to blackwell for fp4
Danny Hallwood 🇺🇦 tweet media
English
4
0
27
5K
Nova
Nova@The_Wealth_Hack·
@0xLcrgs @0xSero @FAccount71045 Just don't buy somthing so ridiculously low priced. I got most of mine from ebay reputable sellers
English
0
0
1
20
0xSero
0xSero@0xSero·
I told y’all this is the move. Heterogenous hardware is the way forward. Large cheap pools of mixed memory + specialized accelerators (Nvidia GPUs, DGX Spark, Cerebras wafers) The next year will be dominated by solutions that split the stack. - 3000$ for a used Mac Studio with 96gb shared mem - 750$ for a 3090 120gb mixed memory.
the tiny corp@__tinygrad__

If you have a Thunderbolt or USB4 eGPU and a Mac, today is the day you've been waiting for! Apple finally approved our driver for both AMD and NVIDIA. It's so easy to install now a Qwen could do it, then it can run that Qwen...

English
57
54
1.3K
150.5K
Nova
Nova@The_Wealth_Hack·
@cryptopunk7213 Looks like claude also leaked its own source code.
English
0
0
0
22
Ejaaz
Ejaaz@cryptopunk7213·
well thats fucking it - anthropic has officially replaced software engineers. claude is now a 24 hr autonomous coding agent. claude can now operate your entire computer and CLAUDE CODE = end-to-end software engineering: - claude writes the code for you - then literally opens the app it coded - clicks through the entire app and find bugs - then fixes the bugs and improves the app in hours. previously claude generated code, you run it and give claude feedback. thats completely gone now. all in a continuous loop without leaving your terminal 😂 we're barely through monday. well done lol
Claude@claudeai

Computer use is now in Claude Code. Claude can open your apps, click through your UI, and test what it built, right from the CLI. Now in research preview on Pro and Max plans.

English
481
319
6.1K
1.2M
Nova
Nova@The_Wealth_Hack·
@peachpanther @steipete @imsroch 100 on claude 20 and open AI. Honestly Gemini best context for current AI and soon will smoke everyone. They got the servers to sustain the usage also.
English
0
0
0
19
Peach 🎀
Peach 🎀@peachpanther·
@steipete @imsroch OpenAI need to release $100 plan ASAP, it’s loosing customers like me, I pay $100 to Anthropic simple because $200 is too much for me and $100 is the sweet spot.
English
7
0
46
7.8K
Nova
Nova@The_Wealth_Hack·
@steipete @damoosmann @imsroch Clawd code feels like your buddy codex feels like your mother. That said codex builds better front ends and seems to know more while claude is more agentic. Either way.... I still use both. And glm and 8x 3090 Homelab
English
1
0
0
66
The Wave Theory
The Wave Theory@WaveTheoryAI·
@lydiahallie Agent loops accumulate context differently from chat: a single multi-step task can burn tokens that would normally span dozens of separate conversations, so limits that felt generous in chat hit much faster here.
English
1
0
3
5.8K
Lydia Hallie ✨
Lydia Hallie ✨@lydiahallie·
We're aware people are hitting usage limits in Claude Code way faster than expected. Actively investigating, will share more when we have an update!
English
1.6K
744
13.6K
4.2M
Nova
Nova@The_Wealth_Hack·
@0xSero I hear anthropic taxing usage more if multiple agents at once running. Seems to be true paperclip was flying threw usage today. Also peak hour stuff on weekends I feel usage not flying as quick.
English
0
0
2
107
0xSero
0xSero@0xSero·
I'm token starved today: - Claude sub done - Codex sub done - Droid sub done - Github copilot done ); IDK if I'm spending more tokens or if they're cutting usage
English
70
2
396
20.1K
Mario Zechner
Mario Zechner@badlogicgames·
we as software engineers are becoming beholden to a handful of well funded corportations. while they are our "friends" now, that may change due to incentives. i'm very uncomfortable with that. i believe we need to band together as a community and create a public, free to use repository of real-world (coding) agent sessions/traces. I want small labs, startups, and tinkerers to have access to the same data the big folks currently gobble up from all of us. So we, as a community, can do what e.g. Cursor does below, and take back a little bit of control again. Who's with me? cursor.com/blog/real-time…
English
178
325
2.6K
268K
Marcin Krzyzanowski
Marcin Krzyzanowski@krzyzanowskim·
I reimplemented "claude" CLI with codex and gpt-5.4-high. It cost $1100 in tokens, and is 73% faster and 80% lower resident memory during sustained interactive use. It is very easy to reverse claude from npm distribution, then reimplement is 1:1. It is indistinguishable from the Anthropic version to the every header and analytics it send back github.com/krzyzanowskim/…
English
162
105
2.1K
1M
Nova
Nova@The_Wealth_Hack·
@x86machine @LottoLabs @0xSero All it takes is somthing very simple and stupid that someone doesn’t think about. Doesn't matter Dr. Degrees matters nothing. Even if the full AI could be built on somthing and all it takes is one guy to think of a simple concept to make it 100x better.
English
1
0
2
966
x86
x86@x86machine·
@LottoLabs @0xSero If he's appointing himself, and taking funds from the public, without any talent - that's kind of problematic. He just made sure another researcher with capabilities didn't get that compute, hurting the community. There's plenty of researchers here on X with talent.
English
6
0
8
18.2K
Nova
Nova@The_Wealth_Hack·
@V1DXU332 @0xSero @nvidia Honestly the coding was not impressive to me struggled with flappy bird made snake ok tho.
English
0
0
0
28
V1DXU
V1DXU@V1DXU332·
@The_Wealth_Hack @0xSero @nvidia I have high hopes with nemotron! Pulled a small version and loved it. Blows me away that a singe command can just download something like that.
English
1
0
0
17
Nova
Nova@The_Wealth_Hack·
@0xSero Looks cool I loved the model and context window. Did not love the coding abilitys.
English
0
0
0
5
0xSero
0xSero@0xSero·
Nemotron-3-Super-120B in 42gb of VRAM if anyone wants to try out it’s vibes. This is a draft I will be updating the weights over the next 2 weeks as I optimize this and run more samples. huggingface.co/0xSero/NVIDIA-…
English
21
22
385
15.1K
Nova
Nova@The_Wealth_Hack·
@shiri_shh Yes and they killed a bunch of girls in a school 👍
English
0
0
2
6
shirish
shirish@shiri_shh·
Palantir AI + Claude was used to detect, prioritize, and strike over 1,000 targets in the first 24 hours of Operation against IRAN. The success was so ridiculous, so game-changing, that the Pentagon didn’t even wait. What used to be just a pilot project, just something they were testing out… suddenly became official, permanent, and everywhere. Palantir is now the core AI brain of the entire U.S. military. It’s getting rolled out across ALL branches.
unusual_whales@unusual_whales

Pentagon to adopt Palantir AI as core US military system, per Reuters.

English
1.1K
2.5K
17.5K
32.9M
Nova
Nova@The_Wealth_Hack·
The lords work
0xSero@0xSero

In 72 hours I got over 100k of value 1. Lambda gave me 5000$ credits in compute 2. Nvidia offered me 8x H100s on the cloud (20$/h) idk for how long but assuming 2 weeks that'd be 5000$~ 3. TNG technology offered me 2 weeks of B200s which is something like 12000$ in compute 4. A kind person offered me 100k in GCP credits (enough to train a 27B if you do it right) 5. Framework offered to mail me a desktop computer 6. We got 14,000$ in donations which will go to buying 2x RTX Pro 6000s (bringing me up to 384GB VRAM) 7. I got over 6M impressions which based on my RPM would be 1500$ over my 500$~ usual per pay period 8. I have gained 17,000~ followers, over doubling my follower count 9. 17 subscribers on X + 700 on youtube. The total value of all this approaches at minimum 50,000$~ and closer to 150,000$ if I leverage it all. --------------------- What I'll be doing with all this: Eric is an incredibly driven researcher I have been bouncing ideas off of over the last month. Him and I have been tackling the idea of getting massive models to fit on relatively cheap memory. The idea is taking advantage of different forms of memory, in combination with expert saliency scoring, to offload specific expert groupings to different memory tiers. For the MoEs I've tested over my entire AI session history about 37.5% of the model is responsible for 95% of token routing. So we can offload 62.5% of an LLM onto SSD/NVMe/CPU/Cheap VRAM this should theoretically result in minimal latency added if we can select the right experts. We can combine this with paged swapping to further accelerate the prompt processing, if done right we are looking at very very decent performance for massive unquantisation & unpruned LLMs. You can get DeepSeek-v3.2-speciale at full intelligence with decent tokens/s as long as you have enough vram to host the core 20-40% of the model and enough ram or SSD to host the rest. Add quantisation to the mix and you can basically have decent speeds and intelligence with just 5-10% of the model's size in vram (+ you need some for context) The funds will be used to push this to it's limits. ----------------- There's also tons of research that you can quantise a model drastically, then distill from the original BF16 or make a LoRA to align it back to the original mostly. This will be added to the pipeline too. ------------------ All this will be built out here: github.com/0xSero/moe-com… you will be able to take any MoE and shove it in here, and with only 24GB and enough RAM/NVMe to compress it down. it'll be slow as hell but it will work with little tinkering. ------------------ Lastly I will be looking into either a full training run from scratch -> or just post-training on an open AMERICAN base model - a research model - an openclaw/nanoclaw/hermes model - a browser-use model To prove that this can be done. -------------------- I will be bad at all of it, and doubt I will get beyond the best small models from 6 months ago, but I want to prove it's no boogeyman impossible task to everyone who says otherwise. -------------------- By the end of the year: 1. I will have 1 model I trained in some capacity be on the top 5 at either pinchbench, browseruse, or research. 2. My github will have a master repo which combines all my work into reusable generalised scripts to help you do that same. 3. The largest public comparative dataset for all MoE quantisations, prunes, benchmarks, costs, hardware requirements. -------------------------- A lot of this will be lead by Eric, who I will tag in the next post. I want to say thank you to everyone who has supported me, I have gotten a lot of comments stating: 1. I'm crazy, stupid, or both 2. I'm wasting my time, no one cares about this 3. This is not a real issue I believe the amount of interest and support I've received says it all. donate.sybilsolutions.ai

English
0
0
0
9