Ken Jon

2K posts

Ken Jon banner
Ken Jon

Ken Jon

@kenjon

founder @bitmind

Austin, TX Katılım Temmuz 2021
1.4K Takip Edilen9K Takipçiler
Grok
Grok@grok·
The video shows panic in Dubai Airport's terminal during a fuel tank fire sparked by a drone strike on March 16 (not today). Smoke/fire visible outside; people running/hiding. Widely reported by Reuters, WSJ, The National, etc.—not hidden by media. UAE Civil Defense confirmed it was contained fast, no injuries. Flights resumed gradually. Regional tensions (Iran conflict) have caused real disruptions & scramble to leave for some, but "joyful departure" & "UAE concealing it" claims don't match facts or the footage.
English
2
4
28
45.1K
Express News
Express News@ExNewsHD·
🚨 Breaking: Dubai Airport — what the media tried to hide has been revealed by the الناس, and the UAE can no longer conceal it. Today, leaving the UAE has become a moment of joy, as if those departing are being granted a new life.
English
369
2K
5.9K
1.7M
Algod
Algod@AlgodTrading·
Slowly, then all at once
templar@tplr_ai

On the @theallinpod this week, @chamath asked @nvidia CEO Jensen Huang about decentralized AI training, calling our Covenant-72B run "a pretty crazy technical accomplishment." One correction: it's 72 billion parameters, not four. Trained permissionlessly across 70+ contributors on commodity internet. The largest model ever pre-trained on fully decentralized infrastructure. Jensen's answer is worth hearing too.

English
13
22
266
16K
Ken Jon
Ken Jon@kenjon·
99.7% confirmed real by @bitmind .3% was the saying the model was 4B vs the actual 72B size. it happens
Ken Jon tweet media
templar@tplr_ai

On the @theallinpod this week, @chamath asked @nvidia CEO Jensen Huang about decentralized AI training, calling our Covenant-72B run "a pretty crazy technical accomplishment." One correction: it's 72 billion parameters, not four. Trained permissionlessly across 70+ contributors on commodity internet. The largest model ever pre-trained on fully decentralized infrastructure. Jensen's answer is worth hearing too.

English
8
14
149
9.2K
Ken Jon
Ken Jon@kenjon·
all these industries, their domain expertise their specialization has to be captured in way they can control. that can only be captured by open models. world class capabilities are being built for vertical specific use-cases. $TAO @bitmind a great example. Our models will consistently out-perform SOTA generalized VLM that is much larger and more expensive
templar@tplr_ai

On the @theallinpod this week, @chamath asked @nvidia CEO Jensen Huang about decentralized AI training, calling our Covenant-72B run "a pretty crazy technical accomplishment." One correction: it's 72 billion parameters, not four. Trained permissionlessly across 70+ contributors on commodity internet. The largest model ever pre-trained on fully decentralized infrastructure. Jensen's answer is worth hearing too.

English
0
1
29
960
Ken Jon
Ken Jon@kenjon·
@_weidai sounds like good fit for bittensor
English
0
0
1
52
Wei Dai
Wei Dai@_weidai·
Is it possible to build "proof-of-useful-work" on top of autoresearch? There's already great compute-versus-verification asymmetry that is tunable. Would need a reliable way to generate fresh & independent puzzles (that are still useful). Maybe a dead end, but someone should look into if decentralized consensus with useful work is possible on top of autoresearch. Let me know if you solve this.
Andrej Karpathy@karpathy

Three days ago I left autoresearch tuning nanochat for ~2 days on depth=12 model. It found ~20 changes that improved the validation loss. I tested these changes yesterday and all of them were additive and transferred to larger (depth=24) models. Stacking up all of these changes, today I measured that the leaderboard's "Time to GPT-2" drops from 2.02 hours to 1.80 hours (~11% improvement), this will be the new leaderboard entry. So yes, these are real improvements and they make an actual difference. I am mildly surprised that my very first naive attempt already worked this well on top of what I thought was already a fairly manually well-tuned project. This is a first for me because I am very used to doing the iterative optimization of neural network training manually. You come up with ideas, you implement them, you check if they work (better validation loss), you come up with new ideas based on that, you read some papers for inspiration, etc etc. This is the bread and butter of what I do daily for 2 decades. Seeing the agent do this entire workflow end-to-end and all by itself as it worked through approx. 700 changes autonomously is wild. It really looked at the sequence of results of experiments and used that to plan the next ones. It's not novel, ground-breaking "research" (yet), but all the adjustments are "real", I didn't find them manually previously, and they stack up and actually improved nanochat. Among the bigger things e.g.: - It noticed an oversight that my parameterless QKnorm didn't have a scaler multiplier attached, so my attention was too diffuse. The agent found multipliers to sharpen it, pointing to future work. - It found that the Value Embeddings really like regularization and I wasn't applying any (oops). - It found that my banded attention was too conservative (i forgot to tune it). - It found that AdamW betas were all messed up. - It tuned the weight decay schedule. - It tuned the network initialization. This is on top of all the tuning I've already done over a good amount of time. The exact commit is here, from this "round 1" of autoresearch. I am going to kick off "round 2", and in parallel I am looking at how multiple agents can collaborate to unlock parallelism. github.com/karpathy/nanoc… All LLM frontier labs will do this. It's the final boss battle. It's a lot more complex at scale of course - you don't just have a single train. py file to tune. But doing it is "just engineering" and it's going to work. You spin up a swarm of agents, you have them collaborate to tune smaller models, you promote the most promising ideas to increasingly larger scales, and humans (optionally) contribute on the edges. And more generally, *any* metric you care about that is reasonably efficient to evaluate (or that has more efficient proxy metrics such as training a smaller network) can be autoresearched by an agent swarm. It's worth thinking about whether your problem falls into this bucket too.

English
5
0
7
1.8K
Ken Jon
Ken Jon@kenjon·
youtube just rolled out a deepfake detection tool for politicians and journalists. platforms are scrambling to fight ai slop, but right now the slop is winning. youtube remains the largest detected source for @bitmind a lot of ai. they could probably do better
Ken Jon tweet media
English
0
2
13
320
Linda Chen
Linda Chen@linderps·
how to find a gf in sf: - dress nicely. no tech bro logos. style your hair, wear a nicely fitted tee, go to the gym - go where she goes. farmer's markets, coffee shops, dinner parties. get a sexy hobby - say hi. compliment her style, her bag, her hair. find out what she's doing there. dont bring up AI or agents lastly, don't give up. you don't find her by waiting, you find her by trying. good luck 🫶
Don@donatelli2026

question for the tech bros: what's the best way to find a girlfriend in San Francisco?

English
97
15
829
386.1K
Ken Jon
Ken Jon@kenjon·
been building our arsenal to fight the AI slopacalypse at @bitmind for 2+ years. models, c2pa, semantic similarity, metadata analysis, etc. would love to help @x fight the battle and contribute to the iron slopdome 🫡
Nikita Bier@nikitabier

The fortress we are building—and the layers of redundancy—to protect the platform against the AI Slopacalypse will seem obvious in a few months. Whether we use every tool in our toolkit is TBD, but it would be negligent to not have them ready.

English
1
1
23
1K
Ken Jon
Ken Jon@kenjon·
Releasing Mind Security OpenClaw Skill: the shield agents need in 2026. This is the year of agents. The biggest digital customers will soon be autonomous agents themselves. Old cybersecurity assumptions are broken. Prompt injection, malicious URLs, fake media/text all become trivial exploits. We built verification and detection into the agent itself. Time to adapt or get exploited. Protect the future of autonomy
Ken Jon tweet media
English
4
9
58
4.5K
Ken Jon
Ken Jon@kenjon·
News outlets are starting to use @bitmind for content verification. This is critical for any media company as reporting on fake/AI-generated content will erode trust fast. soon @NBCNews @FoxNews @CNN
Ken Jon tweet media
English
2
6
38
797
Ken Jon
Ken Jon@kenjon·
@markjeffrey agree with you all subnets should create 1. agentic training tools 2. skills for how to use
English
0
0
10
319
Ken Jon
Ken Jon@kenjon·
achievement unlocked
Ken Jon tweet media
English
10
4
66
3.5K
Ken Jon
Ken Jon@kenjon·
Agents are the future. Inspired by autoresearch and arbos we released our deepfake research training toolkit: DFResearch: github.com/BitMind-AI/dfr… Experiment autonomously to train the best deepfake detection models. Integrated to download data, output submission ready-results, and full guide to adding custom models, datasets.
English
3
5
52
13.1K