Peter Ince

604 posts

Peter Ince banner
Peter Ince

Peter Ince

@satoriweb

Researcher and harness mechanic @cenva_intel, PhD on AI/LLMs for Smart Contract Security.

Melbourne, Victoria Katılım Temmuz 2012
3.4K Takip Edilen440 Takipçiler
Peter Ince
Peter Ince@satoriweb·
Made codex a goblin pet friend
GIF
English
0
0
0
24
Peter Ince
Peter Ince@satoriweb·
@jxnlco Codex: Search "What humans do with computers" then save output as csv. Next /goal Make hyper-real visual demo for everything on human-economy.csv - make no mistakes
English
0
0
0
153
jason liu
jason liu@jxnlco·
ok so I'm on developer experience for codex? yes but codex is for everything yes so my scope is everything experience? yes, can you make a demo for everything? can you be more specific? yeah like all the work in the economy, can you make demos for that?
English
44
1
369
26.3K
Peter Ince
Peter Ince@satoriweb·
OpenAI needs some Aussie (located) staff so @thsottiaux et al don't have to get paged in the middle of the night when the servers melt lol.
English
0
0
0
27
Peter Ince
Peter Ince@satoriweb·
I think the codex servers melted, getting this error every few turns through opencode {"type":"error","sequence_number":4,"error":{"type":"server_error","code":"server_error","message":"An error occurred while processing your request. You can retry your request, or contact us through our help center at help.openai.com ..."}.
English
0
0
0
53
Peter Ince
Peter Ince@satoriweb·
@Teknium @mr_r0b0t Recommend both Hermes and Kimi k2.6. The model is great to working on its own in hermes, a bit slower than ant/oai, but I have come to appreciate it has its own approach / research taste vs 5.4 or 4.6/7
English
2
0
6
134
Peter Ince
Peter Ince@satoriweb·
Everyone from codex is asleep I assume, but getting frequent server overloaded errors for gpt5.4. Anyone else?
English
0
0
1
38
Peter Ince
Peter Ince@satoriweb·
@BoyuanChen0 So good! The text gen, even on text heavy images is impeccable.
English
0
0
0
506
Boyuan Chen
Boyuan Chen@BoyuanChen0·
This is what I’ve been cooking in the past 4 months . GPT Image 2 is over a massive 240 elo jump over the second place model, marking the biggest jump bigger than the rest of the leaderboard combined
Arena.ai@arena

Exciting news - GPT-Image-2 by @OpenAI has claimed the #1 spot across all Image Arena leaderboards! A clean sweep with a record-breaking +242 point lead in Text-to-Image - the largest gap we’ve seen to date. - #1 Text-to-Image (1512), +242 over #2 (Nano-banana-2 with web-search aka gemini-3.1-flash-image) - #1 Single-Image Edit (1513), +125 over #2 (Nano-banana-pro aka gemini-3-pro-image) - #1 Multi-Image Edit (1464), +90 over #2 (Nano-banana-2) No model has dominated Image Arena with margins this wide. Huge congratulations to @OpenAI on this major breakthrough in image generation! More performance breakdowns by category in the thread below.

English
75
77
1.6K
147.3K
Peter Ince
Peter Ince@satoriweb·
@Teknium It's super good in practice also, just a bit slow at the moment. It is much more capable of working through issues on its own without coming back to ask questions.
English
0
0
1
15
Fireworks AI
Fireworks AI@FireworksAI_HQ·
We’re launching Kimi K2.6 on Fireworks as a Day-0 launch partner! K2.5 was the base for standout models like @Cursor’s Composer 2 and was the most popular model on our training platform. K2.6 on Fireworks raises the bar again. → Optimized across the stack, from custom speculators to heterogeneous hardware support across @Nvidia and @AMD. → Day-0 serverless support, and coming soon to Fire Pass (stay tuned). → Unlock new use cases with capabilities for 12+ hour autonomous runs and 4,000+ tool calls. → $0.95 input / $4.00 output per 1M tok Get started today → fireworks.ai/models/firewor…
Fireworks AI tweet media
English
37
19
389
81K
Peter Ince
Peter Ince@satoriweb·
Any chance we can get Pro in codex? :) Although ideally not taking away from the normal mode usage, like a separate spark-esque usage bar. @thsottiaux @jxnlco love your work!
English
0
0
0
42
Peter Ince
Peter Ince@satoriweb·
@realCoinAPI am having issues with auth to your apibricks.io and flatfile api. but the status page says everything is good, is it just me?
English
0
0
0
10
Peter Ince
Peter Ince@satoriweb·
The only problem with the pace of change and how amazing it is to build right now, is I should be spend most of my time raising but I am having too much fun building the next thing.
English
0
0
0
60
Peter Ince
Peter Ince@satoriweb·
@JohnThilen that works as well! As long as the request -> payment -> access is through the LLM interface. It seems like it finds the product, I go talk to their AI interface, it creates a stripe link, I pay, give api details to model. Too many steps!
English
0
0
0
15
John Thilén
John Thilén@JohnThilen·
@satoriweb I would prefer if it had to make a business case for why the payment is necessary.
English
1
0
1
12
Peter Ince
Peter Ince@satoriweb·
Would be great to have a payment system for agents that operates like a child credit card, it gets approval for a spend or sub, you approve, and then sorts out the rest of the details with the provider. Stripe is in a great position for this, they own the payment rails for many.
English
1
0
1
74
Peter Ince
Peter Ince@satoriweb·
@Teknium Legend! I should have figured you'd be on it
English
0
0
1
28
Peter Ince
Peter Ince@satoriweb·
Did codex change the response shape? As it's failing in the @NousResearch Hermes harness at the moment and it looks like the cause different stream expectation. Only happens using codex auth, not using gpt5.4 in open router.
Peter Ince tweet media
English
1
0
16
1.8K
Peter Ince
Peter Ince@satoriweb·
It turns out it was causing an issue in auxillary compression as well - here's the summary to fix it quick - Compression has a separate Codex Responses API bug in agent/auxiliary_client.py. The auxiliary Codex adapter streams the response but previously ignored stream events and only read final.output; Codex can now return response.completed with output=[] even when the real answer arrived in response.output_item.done / response.output_text.delta. This makes compression summaries come back empty (message.content=None). Fix by capturing streamed output items/text in _CodexCompletionsAdapter.create(), backfilling a minimal response when final.output is empty, normalizing dict-shaped streamed items, and avoiding plain-text synthesis if function_call events were streamed.
English
1
0
2
142
Teknium 🪽
Teknium 🪽@Teknium·
@satoriweb Yea it was super weird how it just changed under our feet like that lol but it should be fixed in main now
English
3
0
21
1.8K
Peter Ince
Peter Ince@satoriweb·
@Teknium Amazing! Thanks Tek! Yeah, there was no announcement or docs, afaik the latest update to the endpoints for codex was to support openclaw better, so might have been an unintended breaking change.
English
0
0
17
25.3K