Yevhen Bobrov

8.4K posts

Yevhen Bobrov banner
Yevhen Bobrov

Yevhen Bobrov

@yevhen

Making hard things easy

Kyiv, Ukraine Katılım Haziran 2008
307 Takip Edilen551 Takipçiler
Yevhen Bobrov
Yevhen Bobrov@yevhen·
@danveloper You can work with both and switch on the fly using open-source pi.dev plus incredibly extensible harness with lots of useful plugins.
English
0
0
0
189
Dan Woods
Dan Woods@danveloper·
I sort of load balance between Claude Code (Opus 4.6 - max effort) and Codex (GPT-5.4 - medium) based on whether I need more outside-the-box thinking (Claude) or more precision execution (Codex). Sometimes, I'll have Claude Code experiment with an idea and then hand it to Codex to maximize the implementation. Sometimes even ask them to optimize each other's changes. It works great. Anyway, Claude Code is what I mainly collaborate with on engineering tasks. I always start with Claude Code. But, today Anthropic had so many problems with API stability and something being off about the model. It was just making foolish mistakes, tried to overwride the internal python print to be able to flush writes, forgot to save checkpoints on an hours long training run (my bad, I've come to trust it too much)... I had to fire that agent and /compact. And I went to Codex and man has it gotten so good. Speed, precision, throughput... the fact that it can watch a log and comment about it in real time as the data streams, as opposed to Claude Code's lazy sleep 9600. I'm very impressed. I wish gpt-5.4 had a 1M context window.
English
6
0
11
1.9K
Yevhen Bobrov retweetledi
Jef Newsom
Jef Newsom@jef·
@danveloper Claude’s your bro. He’s a genius, but he has a mix of early onset Alzheimer’s and dissociative identity disorder. Codex is the really good QA guy who you would never hang with outside of work.
English
0
1
1
305
Yevhen Bobrov retweetledi
Sovey
Sovey@SoveyX·
AI is gonna take your job and your girl.
English
1.6K
2.6K
26.3K
3M
Yevhen Bobrov
Yevhen Bobrov@yevhen·
Caught myself having similar experience recently. That's so worry-some that I'm considering taking a break from agentic coding and returning back to using it as auto-complete.
camsoft2000@camsoft2000

I’m getting to the point with one of the projects I work on where the complexity of AI slop is becoming a real issue. While I can still happily prompt the agent to add x feature and it will do so and it will likely work perfectly, the code is just getting too complex and fragmented. Agents love to copy and paste and keeping patterns DRY is a real challenge. The agent will start diverging all those copy and pastes until you’ve got loads of similar but slightly different blocks of logic. Again it all still works and solves the problem I’m after. But I just can’t get any kind of consistency anymore, the code is a mess and I just don’t have a handle on it. I want a clean unified architecture but agents just code with tunnel vision. The project is now too big and complex for an agent to fully reason with and too big and complex for me to reason with. The only real solution is a complete rewrite. Maybe this is the way things will go. Code will just become disposable. I don’t really want to care about the code and to be honest I don’t but I do care about consistency and maintainability and the AI slop is hurting those very things I do care about. I know some will say “I’m holding it wrong”, use x,y,z skill, tool whatever and already use tools and anti slop skills, plans, docs, etc but the outcome is the same. Vibe coding something into existence is truly magical. But turning it into a mature product with months of iterations is painful. I can’t even hand code this thing because I don’t understand the code anymore and I’m too lazy to try and code myself because I’m addicted to AI. So what’s the solution, either start again and accept that’s just the way we have to roll, or just carry on fighting the slop and accept each new feature will take longer to implement than the last. I’m tired. I’m addicted.

English
1
0
0
219
Yevhen Bobrov
Yevhen Bobrov@yevhen·
@GergelyOrosz That’s all still on us. And that’s why not everyone is so excited about tons of AI generated code and cognitive debt created. At 3 a.m., it doesn’t matter whether it was Codex or Opus.
English
0
0
0
255
Gergely Orosz
Gergely Orosz@GergelyOrosz·
The chatter about generating code with AI tools feels stuck at the "basic" level of... well, codegen, plus (perhaps) reviews and testing. I hear close to little talk about the things that come right after generating code: deploying, canarying, o11y, SLOs, error budgets etc
English
93
10
453
44.6K
Yevhen Bobrov retweetledi
dax
dax@thdxr·
you're probably underestimating how crazy things are
dax tweet media
English
295
902
10.6K
1.7M
Yevhen Bobrov
Yevhen Bobrov@yevhen·
@petergostev Stop lying to yourself - managing clankers has nothing to do with doing the thing with your own hands
English
0
0
0
17
Peter Gostev (SF: 29 Mar - 3 Apr)
There's worry that people will stop using their brains with LLMs, but managing several AI agent threads in parallel has been some of the most cognitively intensive work I've done in years
English
178
137
1.7K
70.9K
Yevhen Bobrov retweetledi
K’Bucko
K’Bucko@KBucko7·
Reading Dune. Frank Herbert was cooking.
K’Bucko tweet media
English
139
7.2K
46.8K
1.1M
Yevhen Bobrov retweetledi
BURKOV
BURKOV@burkov·
GPT-5.4 > Opus 4.6 And Google still doesn't have anything even remotely competitive.
English
142
25
870
123.6K
Yevhen Bobrov
Yevhen Bobrov@yevhen·
100%
Mario Zechner@badlogicgames

i can't speak for david. what i see is this: if you let agents build or extend a codebase with only minor or no supervision, you get unmaintainable garbage, because the agent makes terrible decisions that compound, both big and small. those decisions make it hard for both you and the agent to keep modifying the code base, until eventually it's unrecoverable. why does the agent make bad decisions? i can't tell for sure, but my gut tells me that training data can currently not capture the holistic thinking needed to design and evolve complex systems. that's one part of the problem. related to that, and oversimplified: agents output the "mean quality" of the code they saw during training. most of that code is very bad. specifically tests, which humans are terrible at writing at. another part of the problem is that specification via prompt is not precise enough, so the agent has to fill in the blanks, giving it enough rope to hang itself. the more detailed your spec gets, so the agent gets constrained and less likely to produce crap, the closer you are to handwriting the code yourself, as that's the most detailed version of the spec that can exist. so then you gain nothing. back to prompt spec it is, which means the agent fills in blanks, which means we get suboptimal or truely bad results. using agents can still be a net productivity boost (see other posts in my thread), but it is not easy to come up with consistent workflows that produce both production quality maintainable code while retaining the speed advantages agents give you.

QST
0
0
1
154
Yevhen Bobrov retweetledi
David Cramer
David Cramer@zeeg·
im fully convinced that LLMs are not an actual net productivity boost (today) they remove the barrier to get started, but they create increasingly complex software which does not appear to be maintainable so far, in my situations, they appear to slow down long term velocity
English
466
228
3.5K
663.6K
Yevhen Bobrov retweetledi
Mo
Mo@atmoio·
I was a 10x engineer. Now I'm useless.
English
1.5K
1.7K
16.2K
6M
Yevhen Bobrov
Yevhen Bobrov@yevhen·
Як розрізнити, чи ти чіпляєшся за минуле, чи втрачаєш ідентичність?
Українська
2
0
1
151
Yevhen Bobrov retweetledi
Mo
Mo@atmoio·
AI is making CEOs delusional
Indonesia
1K
2.6K
19K
2.8M