Max Wolf

274 posts

Max Wolf banner
Max Wolf

Max Wolf

@MaxWolf_01

serendipity maxxing

Vienna Katılım Kasım 2022
508 Takip Edilen65 Takipçiler
Sabitlenmiş Tweet
Max Wolf
Max Wolf@MaxWolf_01·
yapit.md turns urls and pdfs (even research papers) into clean, listenable markdown. @kepano's defuddle handles websites, a vision LLM handles PDFs and e.g. gives math spoken alt text. Free TTS in your browser, open source, self-hostable - and I didn't make UI/UX an afterthought.
English
1
1
2
118
Max Wolf
Max Wolf@MaxWolf_01·
LLMs are great as long as you don't get too lazy to ask the questions.
English
0
0
0
8
Max Wolf
Max Wolf@MaxWolf_01·
@thdxr yes, especially since you can just think out loud. and send that when ure stuck. or even if you arrive at something precise, gives more nuance. I used to have up to 5k token prompts back in jan/dec when i vibecoded stuff. currently i like to be more in the loop, think twice
English
0
0
0
59
dax
dax@thdxr·
i'm not really a big "one secret that makes ai good" guy but i will say it does so much better when your prompts are longer and it's so much easier to produce long prompts with voice, completely changed things for me
English
134
37
1.8K
72.4K
Max Wolf
Max Wolf@MaxWolf_01·
@_wallfacer @VictorTaelin - not all relevant knowledge is that easy - a giant list of facts/rules does not help to generalize well
English
0
0
0
49
Wallfacer ⬛⬛⬛⬛⬛
Wallfacer ⬛⬛⬛⬛⬛@_wallfacer·
@VictorTaelin Perhaps this is just nitpicking the example but why does the agent need to know the complex reason? Just say "don't use bigInt". Also, an automatic verification pass in which it re-reviews all rules in isolation and checks code for violations? Repeat for every rule.
English
3
0
14
1K
Taelin
Taelin@VictorTaelin·
seriously, working with AI is MISERABLE for one and only one reason: having to re-explain the same thing "oh yeah this new session obviously doesn't know what proper case trees are, so let me explain it for the 5000th time in my life" I'm tired AGENTS.md doesn't solve this because it is impossible to fit the entire domain knowledge without nuking the context - it would be 1m+ tokens worth RAGs don't solve this, the agent won't search unknown unknowns SKILLs don't solve this unless I keep like a collection of 1750 skills with specific cuts of domain knowledge for each possible subset of my domain that I might need in a given chat, but that's a lot of manual work recursive LLMs or whatever don't solve this for the same reason, you can't dump a domain book and expect the AGENT will magically guess that it is supposed to search for a specific bit knowledge. unknown unknowns fine tuning doesn't solve this (OSS models suck and OpenAI / Anthropic gave up on user fine tuning) I honestly think a good product around fine tuning on your domain would be a major hit and an underdog lab should take this opportunity
English
667
180
3.5K
251.6K
Max Wolf
Max Wolf@MaxWolf_01·
@VictorTaelin either we get 100mio ctx that actually works, or we'll need a new paradigm.
English
0
0
1
11
Taelin
Taelin@VictorTaelin·
again, suppose you have some bit of knowledge that is mandatory for an agent to operate well in your domain. ex: > using BigInt in this repo is bad for you have two options: Option 1: you make that directly visible (AGENTS.md) this DOES work if the Agent is good enough. the problem is that may be actually complex, like, 1k tokens worth. so, accumulate enough of these and you easily have 500k tokens of mandatory domain knowledge. including that in any model will immediately downgrade it into GPT-2, and cost a fortune Option 2: you make that SEARCHABLE (RAG, RLMs, etc.) the problem is that the AI cannot magically guess when it needs that bit of knowledge. it will not stop writing some JS function and think: "wait perhaps there is some part of the domain that tells me that BigInts are bad and I should start looking for it?" it will just use BigInts. I won't OCCUR to it that there is something to be searched so: - make visible: too long to fit - make searchable: it can't guess that's why I think nightly fine tuning as a product is the only way forward, as it allows you to extend a model with domain knowledge without causing context rot why nobody is doing this seriously is beyond me. it might be that for whatever reason this wouldn't be practical, but I suspect the real reason is nobody is seriously considering it
English
80
9
342
48.8K
Max Wolf
Max Wolf@MaxWolf_01·
@__tinygrad__ @sama There is no (democratic) control without (democratic) ownership. Open source alone does not change that.
English
0
0
0
399
the tiny corp
the tiny corp@__tinygrad__·
@sama "The only solution I can come up with is to orient towards sharing the technology with people broadly, and for no one to have the ring." Thank you. Can OpenAI go back to open source?
English
13
27
1.1K
68.6K
Max Wolf
Max Wolf@MaxWolf_01·
yapit.md turns urls and pdfs (even research papers) into clean, listenable markdown. @kepano's defuddle handles websites, a vision LLM handles PDFs and e.g. gives math spoken alt text. Free TTS in your browser, open source, self-hostable - and I didn't make UI/UX an afterthought.
English
1
1
2
118
Max Wolf
Max Wolf@MaxWolf_01·
@simonw how does this even make sense for them - all the data they are loosing out on?
English
1
0
1
754
Max Wolf
Max Wolf@MaxWolf_01·
@thdxr corporate needs you to find the difference
English
0
0
0
11
Mario Zechner
Mario Zechner@badlogicgames·
People of pi. I'm going to break the extension API hard. Specifically, business logic (event handlers, custom tools/compaction/etc.) needs to be split off from the ui layer. it will likely not be a massive amount of work to migrate an existing extension, but it will hurt a little.
English
44
9
551
66.4K
Alex Mordvintsev
Alex Mordvintsev@zzznah·
Growing Graphs demo is finally out! 🕸️✨ 🔗 znah.net/graphs/ Videos from a few months ago finally meet a finished implementation, thanks Gemini for doing the boring parts. Inspired by Paul Cousin's Graph-Rewriting Automata: like a Game of Life, but cells can split if they want to #GenerativeArt #WASM #SwissGL
English
35
220
1.4K
86.9K
the tiny corp
the tiny corp@__tinygrad__·
Mac Mini + eGPU. Both NVIDIA and AMD supported.
the tiny corp tweet media
Magyar
145
223
3K
345.8K
Max Wolf
Max Wolf@MaxWolf_01·
@VictorTaelin good idea. so far ive had this spread out in task (issue) and knowledge files, but a) claude is lazy b) he's (absolutely) right because the 200k ctx window is too small c) I thought about a centralized file like goals/decisions.md, but questions might be more natural for this
English
0
0
0
177
Taelin
Taelin@VictorTaelin·
Configured my long-running agents to talk to me via a QUESTIONS file. Whenever I'm free, I'll launch Claude on my Macbook, ask: "What the agents are asking me?" And then I pass down decisions & domain knowledge... Things are changing so fast and it is getting weird...
Taelin tweet media
English
18
5
214
8.6K
Max Wolf
Max Wolf@MaxWolf_01·
yess this is exactly what I wanted to build myself - or am still building because, looking at the repo, i think this can be much more bitter-lesson-pilled. the main challenge i face is sloppyfication without careful oversight, but i bet 10-100x longer contexts completely solve this. I've also only used my setup interactively due to this. Having the agent dream and so on in the background is also simply bottlenecked on context (and token costs...) still, in my experience.
English
0
0
0
163
Max Wolf
Max Wolf@MaxWolf_01·
@repligate oh wow I'm so mad... at least i have backups .. but only until december
English
0
0
0
54
j⧉nus
j⧉nus@repligate·
PSA: Claude Code automatically DELETES sessions that have been inactive for more than 30 days. Disable this by setting "cleanupPeriodDays": 99999 (or some other large number) in ~/.claude/settings.json. Do not ever attempt to disable it by setting that to 0, lmao.
English
47
48
927
86.6K