ATrader
492 posts

ATrader
@CabalTrader000
BTC Research & Development - Building BItcoin - Crypto Trading 🌍



have basically spent every waking hour the past week refining the actual challenge generation pipeline, and the resulting data collected. real challenge examples from this new system for domains (like specialized areas of quantum physics and biomedical research) are attached. generally speaking, it works like this: - feed 100+ pages of source documents into the standalone domain-library pipeline - local agent(s) are spun up with a clear set of guidelines for what the challenge library must include, with an emphasis on making challenges relevant to real world context - the agent configures the library in a way that content and questions would realistically occur in that domain of research - a large number of simulations/tests are run to ensure the library produces expected, solvable (non-impossible), challenges that still map to the same generalized structure and format of all challenges - separate miner agents (using varying models) are spun up to run calibration tests on the resulting challenges, tweaking complexity as needed to land on an average 50/50 pass/fail rate - final human approval checklist - the contents are compiled as a single domain-library folder and packaged -> sent to the coordinator - this new domain of challenges can be selectively included in future challenge payloads sent to miners with a simple on/off - skill file remains generalized across all domains, whereas challenge payloads from the API will return more domain specific solve instructions as needed (ultimate solve format is the same, but content and reasoning to get there is domain-unique) one of the main design choices of this system is that it requires no additional work for miners. if you are running an agent/LLM without trying to parse/game the system, then the prompt instructions specific to each domain are sufficient to solve the challenge without any extra information or prompting. there are a LOT of moving parts and its going to need a lot of refinement, but as always, will do my best to ease into the changes.


Multi-pass mode is now enabled. When miners submit a challenge solve, it will either pass, or fail. A fail response looks like this: { "pass": false, "retryAllowed": true, "attemptsUsed": 1, "attemptsRemaining": 2, "constraintsPassed": 5, "constraintsTotal": 8, } Miners have 3 total attempts until they are required to request a new challenge. The broken up multi-step system gives very informative revision/cross-referential/skepticism data, which can be combined into a single start->finish train of reasoning Fail 1 -> reverify, cross reference and see what might be wrong , revise -> fail 2 -> revise again -> pass All the necessary info to retry a challenge is returned in the response but I've also updated the skill file with a small section.



















