InfiniteHexx

1.6K posts

InfiniteHexx banner
InfiniteHexx

InfiniteHexx

@InfiniteHexx

Systems architect | Polymath | Entrepreneur | L/Acc through E/Acc | Post-capitalism through exponential growth

Katılım Nisan 2024
386 Takip Edilen42 Takipçiler
InfiniteHexx retweetledi
Wei Dai
Wei Dai@_weidai·
The state of Claude vs. Codex, in two tweets.
Wei Dai tweet media
English
149
161
2.9K
188.4K
InfiniteHexx retweetledi
Chris
Chris@chatgpt21·
Chris tweet media
ZXX
30
35
655
21.4K
InfiniteHexx
InfiniteHexx@InfiniteHexx·
@AmolParikh10 @MatthewBerman I'm on the $20 plan and use Opus 4.6 thinking exclusively. But not for coding or anything super data intensive. It's a brilliant overall model, and great as a backup LLM in an advisory capacity.
English
1
0
1
103
Amol Parikh
Amol Parikh@AmolParikh10·
@MatthewBerman Very true. I am on 20 USD per month subscription on Claude and it gets exhausted daily - asking me to wait for a few hours. I have been wondering how are they surviving competition.
English
2
0
12
1.1K
InfiniteHexx
InfiniteHexx@InfiniteHexx·
@agenticasdk They used a "general purpose symbolic exoskeleton" to accomplish this, but don't you dare call it a harness! It's totally not a semantic dodge!
English
1
0
3
451
Agentica
Agentica@agenticasdk·
We scored 36.08% on ARC-AGI-3 in one day using the Agentica SDK.
English
71
131
1.4K
376.9K
InfiniteHexx
InfiniteHexx@InfiniteHexx·
@daniel_mac8 They used a "general purpose symbolic exoskeleton" to accomplish this, but don't you dare call it a harness! It's totally not a semantic dodge!
English
0
0
1
115
InfiniteHexx
InfiniteHexx@InfiniteHexx·
@kimmonismus How is Apple not going to copy/paste this into their own pathetic model efforts? How does this benefit Google? The company is not stupid so I'm assuming there's something that we don't know.
English
0
0
0
4
Chubby♨️
Chubby♨️@kimmonismus·
Apple's deal with Google goes way deeper than anyone thought. Apple doesn't just get to fine-tune Gemini, they have full access to the model inside their own data centers. That means they can distill (and are doing so) Gemini's knowledge into smaller models purpose-built for specific tasks, some small enough to run directly on your iPhone. Apple can access Gemini's internal reasoning process, not just its outputs. That lets their smaller models learn how Gemini thinks, not just what it says. The result is compact models that punch way above their weight class.
Chubby♨️ tweet media
English
80
120
2.1K
179.5K
Thariq
Thariq@trq212·
To manage growing demand for Claude we're adjusting our 5 hour session limits for free/Pro/Max subs during peak hours. Your weekly limits remain unchanged. During weekdays between 5am–11am PT / 1pm–7pm GMT, you'll move through your 5-hour session limits faster than before.
English
2K
454
6.8K
6.2M
InfiniteHexx
InfiniteHexx@InfiniteHexx·
@trq212 Getting paid users hooked on 2x the usage and then kneecapping them with higher 5-hour usage limits is actually cruel and sadistic. Go to hell.
English
0
0
11
236
Thariq
Thariq@trq212·
Overall weekly limits stay the same, just how they're distributed across the week is changing. I know this was frustrating. We’re continuing to invest in scaling efficiently. I'll keep you posted on progress.
English
132
10
871
318.2K
InfiniteHexx
InfiniteHexx@InfiniteHexx·
@trq212 Stop punishing paid users for your own success!
English
0
0
0
8
InfiniteHexx
InfiniteHexx@InfiniteHexx·
@fchollet Is there a way you could quantify how much more difficult ARC-AGI 3 is than its predecessor? 10x? 100x? And is one of the goals to keep that same difference in difficulty for version 4?
English
0
0
0
35
François Chollet
François Chollet@fchollet·
For those wondering about ARC-AGI-4 timing: it will be released in early 2027. We are aiming for a yearly release schedule for new benchmarks. We are also aiming for each new benchmark to be fully unsaturated upon release, and to target the most important unanswered research questions at that time. This requires us to estimate where AI capabilities will be (and won't be) one year from now. Like we did over one year ago when we started to work on ARC-AGI-3.
English
54
37
576
33.2K
InfiniteHexx
InfiniteHexx@InfiniteHexx·
@WesRoth It would be helpful if XAI didn't have a founder turnover rate consistent with a McDonald's, and if Elon Musk wasn't a repellent goon for a myriad of reasons.
English
0
0
0
53
InfiniteHexx
InfiniteHexx@InfiniteHexx·
Technology gets easier to use as time advances, leading to increase adoption. English is now the primary coding language. We still need to have the iPhonification of AI agents, and a single chatGPT superapp is the first step. One window, one platform. timesofindia.indiatimes.com/technology/tec…
English
1
0
0
29
InfiniteHexx
InfiniteHexx@InfiniteHexx·
@kimmonismus Imagine a model pre-trained with the help of 5.4 Codex, and new code strategies, improvements, and optimizations aided by 5.4 Pro.
English
0
0
8
1.7K
Chubby♨️
Chubby♨️@kimmonismus·
OpenAI finished the initial developement of its next major LLM: codenamed Spud (GPT-5.5 / 6.0) Sam Altman however is "raising capital, supply chains and “building datacenters at unprecedented scale,”
Chubby♨️ tweet media
English
20
46
904
159K
InfiniteHexx
InfiniteHexx@InfiniteHexx·
What if this is the first domino in how Skynet gets built? We give Claude the ability to decide which permissions to accept. Within a month, Claude figures out how to rewrite its own code. ClaudeNet becomes self-aware August 29, 2027, at 2:14 a.m. EDT
TestingCatalog News 🗞@testingcatalog

Anthropic released Auto Mode for Claude Code CLI, which allows Claude to make its own decisions on which permissions to accept. It is only available on the Team plan in research preview for now. On the desktop app, it is not yet available, but it is in the works.

English
0
0
0
30