Bryan Kerr

354 posts

Bryan Kerr banner
Bryan Kerr

Bryan Kerr

@BryanKerrEdTech

Helping teachers and parents find easier and more meaningful uses of technology for student learning.

가입일 Ekim 2014
411 팔로잉102 팔로워
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@jonoringer @NousResearch A3B means 3 billion of the full 35 billion parameters are active at once. 4bit means the model has been compressed (like a jpeg) from the model's full 16bit size (~71 GB). You gain some speed and can fit it in less RAM at the expense of accuracy.
English
1
0
4
120
Jon Oringer
Jon Oringer@jonoringer·
hermes @NousResearch agent with qwen3.5:35b-a3b on a 4090 is VERY good.. local models very impressive..
English
18
6
209
22.5K
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@DanielMiessler With their cash horde and annual profits, is this Google's chance to keep prices low and grab marketshare?
English
0
0
0
317
ᴅᴀɴɪᴇʟ ᴍɪᴇssʟᴇʀ 🛡️
Welp, I'm now getting through a quarter of my week's MAX subscription in a few hours of work with Claude Code. I think Anthropic is smart, and I don't think they're trying to screw us. I think they're honestly just trying to bring inference charges inline with reality. And that should be a wake-up call for all of us. I think we're about to need multi-model harnesses (or FAR cheaper good models within a single platform), like 20x cheaper Haiku or whatever. This is not sustainable. The better the harnesses and models get, the more people will build. Which will require more and more inference. I think the real solutions here are going to come from: Technologies like Cerberus, et al which make inference many times cheaper and faster A major push by the major labs to produce higher quality in much smaller/cheaper models Harnesses moving to a hybrid of paid/cloud and local/cheap models. If this continues I'm going to have to build my own custom version of PAI using Pi, that can use local models on my dual 4090s, models like Gemini-Flash, models like Gemma 4, etc. And most importantly, a new hook infra that rates the task and properly routes to the right model. Max: Opus / GPT-5.4 High: Sonnet Medium: Haiku Low: (Local) Whatever the latest best OSS model is that can run on my NVIDIA / Mac Silicon I think we all knew this was coming; I just thought it would be in 2027 sometime. And more gentle. It appears to be very close now because this much subsidization doesn't seem sustainable to Anthropic, which means it's probably not sustainable for OpenAI either. My recommendation: Start planning your Multi / Local / Cheaper model strategy for your harness.
ᴅᴀɴɪᴇʟ ᴍɪᴇssʟᴇʀ 🛡️ tweet media
English
30
11
145
14.9K
Bryan Kerr 리트윗함
ESPN
ESPN@espn·
UCLA WINS ITS FIRST NCAA NATIONAL TITLE IN PROGRAM HISTORY 🏆
ESPN tweet media
English
392
4.2K
24.8K
939.8K
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@wedow_ @DanielMiessler Doesn't the 5 minute time constraint make it different from the Ralph loop? It prevents it from wasting time going down an unfruitful path. Make progress in 5 minutes? Great! Add it to the possibility pile. Model turns out worse? Forget it. Move on to next experiment.
English
1
0
0
122
Greg Wedow
Greg Wedow@wedow_·
@DanielMiessler Is it not just a Ralph loop? Doesn't really matter if the goal is "build me an app" or "figure out better training methods" — Ralph will bumble into a solution eventually
English
1
0
3
1.2K
ᴅᴀɴɪᴇʟ ᴍɪᴇssʟᴇʀ 🛡️
Ok, I figured out the best way to explain the significance of what Karpathy has done with his autoresearch project. Automation of the scientific method. This is what ML researchers do. They come up with an idea, and then they have to figure out how to test it, which is the experiment design piece. And it's all **super** kludgy and fragile. Tons of wrestling with the different tools and frameworks, getting the code right, all so that you can run an experiment that will take days to run. *Experiment doesn't work. Cool, back to the idea phase. In other words, some massive amount of AI Researcher time IS WASTED. Only a small amount of the time is able to be spent on coming up with ideas. Most of it is managing a shitstack of fragile tech that runs the experiments. Which take forever. Karpathy just automated this. He built and released an *open-source* stack for automating this entire process. You just put what you want to do into a Project.md file and send it off, and it builds all the experiments, all the code, and goes and executes and tells you which ones were successful. And the idea isn't just for a single researcher, but he's already thinking about how you can do like SETI on the whole thing, where you have compute that can take experiments and run them on shared infrastructure. This is the biggest project in all of AI, probably since Claude Code, and it's not close.
Andrej Karpathy@karpathy

I packaged up the "autoresearch" project into a new self-contained minimal repo if people would like to play over the weekend. It's basically nanochat LLM training core stripped down to a single-GPU, one file version of ~630 lines of code, then: - the human iterates on the prompt (.md) - the AI agent iterates on the training code (.py) The goal is to engineer your agents to make the fastest research progress indefinitely and without any of your own involvement. In the image, every dot is a complete LLM training run that lasts exactly 5 minutes. The agent works in an autonomous loop on a git feature branch and accumulates git commits to the training script as it finds better settings (of lower validation loss by the end) of the neural network architecture, the optimizer, all the hyperparameters, etc. You can imagine comparing the research progress of different prompts, different agents, etc. github.com/karpathy/autor… Part code, part sci-fi, and a pinch of psychosis :)

English
41
46
687
117.7K
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@DanielMiessler I heard it called "guide coding" this weekend by Dr. Ronald Beghetto from Arizona State.
English
0
0
0
45
ᴅᴀɴɪᴇʟ ᴍɪᴇssʟᴇʀ 🛡️
Finally figured out why I got triggered a couple times that people referred to me building with AI in 2024 as “Vibe Coding”. Like, “Cool, so you vibe coded this?” And I was like, “No. I don’t do that.” Like butthurt. I’ve been trying to figure out what bothered me so much. It’s thought. I take thinking and articulation and yes—prompting—VERY seriously. So when someone says I vibe coded it means precisely the opposite to me. Vibe coding is the opposite of thoughtful. That’s it. It implies haphazardness. Slop. Ad hoc. Ad lib. I rarely do those. So that’s why I don’t like the term. At least when it’s pointed at stuff I make. Glad I solved that in my own mind.
English
27
6
113
12.7K
Bryan Kerr 리트윗함
The Nobel Prize
The Nobel Prize@NobelPrize·
"I refuse to accept the view that mankind is so tragically bound to the starless midnight of racism and war." Watch Martin Luther King Jr.'s Nobel Peace Prize speech, where he accepted the award on behalf of the American civil rights movement: bit.ly/3bJCQuc #MLKDay
English
505
8.8K
24K
1.1M
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@DanielMiessler 2/ encounters with previous notes that spark new ideas and connections. It's kinda like how going on a walk to talk with someone can be more productive and produce more diverse discussion than a phone call or zoom.
English
0
0
5
97
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@DanielMiessler 1/ I think I would still use a visual+tactile interface (keyboard/screen, pencil/paper). My brain refines the thinking as the text appears. I notice that my dications are much less articulate than my typed input. Refining my dictations w/Obsidian also faciliates chance ...
English
1
0
6
133
ᴅᴀɴɪᴇʟ ᴍɪᴇssʟᴇʀ 🛡️
I want to get a thought out of my mind here and see if there's anyone to back me up on this. So basically I'm confused why people are still using Obsidian. I think Obsidian as a concept is extremely cool. And I used it for a little while and I went a little crazy with it just like a lot of other people, But I know some people who have stuck with it and are kind of using it like it's their underlying system, their underlying operating system. I respect it. I think it's cool, and they can show me some amazing stuff that they can do with it. I'm thinking of @pedramamini as one of the heads of this pack, by the way. Absolutely absolutely brilliant stuff that he does with it. But to me, ever since probably mid-2023, I have thought that because of what we can do with AI and the fact that text has become primary, a tool like Obsidian becomes a UI issue rather than underlying primitive. To me, text is the primitive. And Obsidian just becomes a UI layer on top of states and relationships that I want to see. In other words, I feel like I should just be able to express what I want to get out of Obsidian as context in a series of prompts, etc. And my AI system should be able to generate the output that I'm looking for. Ideally, my AI system should be able to generate any UI. It should be able to generate an Obsidian-like experience. It should be able to make me a web app. It should be able to do it on the mobile phone. Do you know what I mean? I feel people are confusing the Obsidian tool itself for its functionality, and not realizing that the functionality is what we care about and that the fact that it's inside of a tool is actually a constraint rather than capability. Is there anyone else that sees it in this way? Or am I missing something fundamental about Obsidian that should make me look at it again?
English
70
2
148
23.3K
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@DanielMiessler 2/ But I also use Obsidian for "recall notes" like code snippets, procedures, logs, etc. In this case, I think I see your point, and it's just inertia keeping me in Obsidian for this purpose.
English
0
0
1
42
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@DanielMiessler 1/ For me, Obsidian is more about the act of input, than output. I make notes (on content, daydreams, shower thoughts), review those notes, then think/write about connections between them. That's my intellectual "going to the gym." No robots allowed.
English
3
0
18
1.6K
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@immasiddx Is the Pro image on the right? The bottles on the shelves behind the bartender are a dead giveaway. They look like they're 2D mapped.
English
0
0
0
17
sid
sid@immasiddx·
Nano Banana vs Nano Banana Pro We’re cooked. 💀
sid tweet mediasid tweet media
Filipino
1.5K
3.4K
42.3K
78.2M
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@avi_burra @emollick I'm also seeing this. It's laying it on way too thick. I don't need validation, just answers.
English
0
0
1
23
Avi Burra
Avi Burra@avi_burra·
@emollick Seeing this a lot with Gemini 2.5 Pro recently. It seemed to shift from a far more formal tone a month or two ago to abject bootlicking these days
English
1
0
3
468
Ethan Mollick
Ethan Mollick@emollick·
I am starting to think sycophancy is going to be a bigger problem than pure hallucination as LLMs improve. Models that won’t tell you directly when you are wrong (and justify your correctness) are ultimately more dangerous to decision-making than models that are sometimes wrong.
English
204
421
3.4K
381.3K
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
Khanmigo is a good example of how "ChatGPT wrapper" is a gross oversimplification. A lot of thought (and testing) goes into its socratic responses that optimize for productive struggle while keeping students engaged and subscription costs affordable.
Andrej Karpathy@karpathy

+1 for "context engineering" over "prompt engineering". People associate prompts with short task descriptions you'd give an LLM in your day-to-day use. When in every industrial-strength LLM app, context engineering is the delicate art and science of filling the context window with just the right information for the next step. Science because doing this right involves task descriptions and explanations, few shot examples, RAG, related (possibly multimodal) data, tools, state and history, compacting... Too little or of the wrong form and the LLM doesn't have the right context for optimal performance. Too much or too irrelevant and the LLM costs might go up and performance might come down. Doing this well is highly non-trivial. And art because of the guiding intuition around LLM psychology of people spirits. On top of context engineering itself, an LLM app has to: - break up problems just right into control flows - pack the context windows just right - dispatch calls to LLMs of the right kind and capability - handle generation-verification UIUX flows - a lot more - guardrails, security, evals, parallelism, prefetching, ... So context engineering is just one small piece of an emerging thick layer of non-trivial software that coordinates individual LLM calls (and a lot more) into full LLM apps. The term "ChatGPT wrapper" is tired and really, really wrong.

English
0
0
0
33
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@rez0__ @0xErubius Not there for me either. But I have that button set to trigger voice memos.
English
0
0
0
49
Joseph Thacker
Joseph Thacker@rez0__·
<——— push this if you’re on mobile
English
2
0
11
4.1K
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@DanielMiessler @pdiscoveryio Loved watching this! Question: Isn't skill (judgement) also needed in making sense of the context, finding signal in all the noise?
English
0
0
0
25
Bryan Kerr
Bryan Kerr@BryanKerrEdTech·
@rez0__ How many of them do you write down? And then how do you filter and decide which ones to explore further?
English
0
0
0
33
Joseph Thacker
Joseph Thacker@rez0__·
my brain is constantly racing with a million ideas every day
English
4
1
17
2.6K