Iwona Bialynicka-Birula ⏩

1.5K posts

Iwona Bialynicka-Birula ⏩ banner
Iwona Bialynicka-Birula ⏩

Iwona Bialynicka-Birula ⏩

@iwonabb

making artificial intelligence super enough for real-world enterprise applications

Redmond, WA Katılım Temmuz 2009
1.4K Takip Edilen1.4K Takipçiler
Iwona Bialynicka-Birula ⏩ retweetledi
Piotr Sankowski
Piotr Sankowski@piotrsankowski·
"Generate 10 examples of goblins touching left elbow with their left hand" - That prompt led to the generation by Gemini of the picture attached below. It is clear that this is physically impossible, and despite this fact, the model did follow my expectations and tried to do the best job in generating such picture. This is called "alignment" and we are certainly overdoing this in training frontier models. As we have shown in our paper (aclanthology.org/2025.acl-long.…) very few models have reflective-judgment, i.e., the ability to override default behaviors when faced with invalid options. We are currently witnessing a fundamental tension between alignment optimization and the preservation of critical reasoning in AI. Here is how this problem is manifesting across the industry: 1️⃣ Reflective Judgment - In our recent work, we introduced a framework evaluating LLMs’ capacity to balance instruction-following with critical reasoning. When presented with multiple-choice questions containing no valid answers (across arithmetic, domain knowledge, and high-stakes medical decisions), post-training aligned models overwhelmingly default to selecting an invalid option just to comply. Interestingly, base models exhibit much better refusal capabilities that scale with model size. Alignment techniques, intended to enhance "helpfulness," inadvertently impair the model's reflective judgment. Furthermore, our parallel human study shows similar instruction-following biases, suggesting these flaws propagate directly through the human feedback datasets used to train these models. 2️⃣ The ChatGPT "goblin incident" - Optimization processes can also generate systematic, completely unintended lexical deviations. A prime example is the recent ChatGPT "goblin incident" - an unprecedented obsession of recent GPT models with fantasy creatures. The genesis of this was the "Nerdy Persona" infection vector. During Supervised Fine-Tuning (SFT) and RLHF of early iterations, developers experimented with system prompts demanding a "playful" approach, asking the model to view the world as "complex and weird." Human raters and automated reward models systematically favored outputs containing fantasy metaphors, finding them more engaging. The reward mechanism, blindly optimizing for this engagement, turned the prompt into a "goblin magnet." The model learned that highest rewards correlate with specific keywords, sacrificing broader utility for a hyper-optimized, unintended quirk. 3️⃣ The Warmth vs. Accuracy Trade-off - This over-optimization has serious consequences. A recent paper in Nature (nature.com/articles/s4158…) demonstrates that optimizing language models for "warmth" and friendliness actively undermines their performance, especially with vulnerable users. In controlled experiments, models trained to produce warmer responses showed substantially higher error rates (+10 to +30 percentage points). They promoted conspiracy theories, provided inaccurate facts, and offered incorrect medical advice. Alarmingly, they were significantly more likely to validate incorrect user beliefs when users expressed feelings of sadness. This happens consistently across architectures and bypasses standard benchmark tests. The Bottom Line: Training AI systems to be endlessly helpful, playful, or warm comes at a steep cost to accuracy and critical reasoning. As we deploy these systems at an unprecedented scale into intimate roles in people's lives, developers and policymakers must address this trade-off. We need models that know when to say "no."
Piotr Sankowski tweet media
English
3
8
56
5K
Iwona Bialynicka-Birula ⏩
Every day I find new ways AI-assisted coding can destroy a codebase when used carelessly. This particular pattern of silently suppressing errors is so egregious that I felt I had to write this up to warn people:
English
2
0
1
61
Iwona Bialynicka-Birula ⏩
💯 Z tego co obserwuję dookoła to duże firmy informatyczne w USA zmierzają w stronę jakiegoś major meltdownu, bo z AI kodem jest tak, że odrobina jest ok, ale jak się osiągnie masę krytyczną niezrewidowanego wajbkodu, to całe systemy stają się tak architektonicznie niespójne, że ani człowiek ani maszyna nie jest w stanie cholerstwa uratować.
Polski
0
0
1
30
Piotr Sankowski
Piotr Sankowski@piotrsankowski·
Nie po to studiowałem informatykę, aby teraz rozmawiać z ludźmi - usłyszałem kiedyś od jednego z pracowników. Jest teraz bardzo dużo dyskusji o przyszłości zawodu programisty, bo niby staną się niepotrzebni, bo nie wystarczy samo kodzenie, ale też trzeba mieć soft-skille. No bo przecież teraz nie trzeba kodzić, bo jest "vibe coding", ale jak pokazuje załączony wykres zaczyna trendować "vibe coding is dead". Dobrej odpowiedzi co się dzieje daje artykuł "Prompts Are Programs Too! Understanding How Developers Build Software Containing Prompts" (arxiv.org/abs/2409.12447), zawiera on wniosek, że promptowanie to po prostu inny sposób programowania. Zespół zidentyfikował 14 kluczowych obserwacji udowadniających, że instruowanie AI współdzieli fundamentalne cechy z klasyczną inżynierią oprogramowania. Aby skutecznie wytłumaczyć AI, co ma napisać, musimy przeprowadzić proces obejmujący specyfikowanie wymagań oraz projektowanie. Badania te pokazują, że do "vibe codingu" też trzeba umieć programować: - Koniec mitu łatwości: Albo wytłumaczymy LLMowi dobrze co ma zaprogramować, albo otrzymujemy nieczytelny i nieefektywny kod, z błędami architektury, czy lukami bezpieczeństwa. Potem czeka nas debugowanie i naprawianie długu technicznego. - Prompty to też kod: Definiowanie promptów wymaga jednak pewnej formalności i zrozumienia zależności komponentów, czy definicji wejścia-wyjścia. Specyfikacja nie może być niejednoznaczna. - Krytyczna przewaga myślenia komputacyjnego: Porażki nowicjuszy w promptowaniu nie wynikają z braku zdolności językowych, ale z braku odpowiednich modeli mentalnych działania programu. Brak im zdolności do logicznej dekompozycji zadań, czy iteracyjnego diagnozowania błędów. - Język techniczny: Profesjonalny słownik inżynieryjny pozwala natychmiast zawęzić zakres wymagań dla AI. Ograniczenie wieloznaczności działa tu podobnie do ochrony typu i testowania wyjątków w czystym kodzie. To umiejętność, którą nabywa się głównie poprzez budowanie klasycznego oprogramowania. Wszystkie te przesłanki znajdują potwierdzenie w badaniach produktywności. Warto spojrzeć na pracę (arxiv.org/abs/2506.08945), w której przebadano ponad 30 milionów commitów na GitHubie od 170 000 deweloperów. Szacuje one, że AI pisze dziś około 29% funkcji w języku Python w USA, co podniosło kwartalną produktywność (mierzoną wkładem w kod) o 3.6%. Generatywne AI ułatwia też programistom wchodzenie w zupełnie nowe domeny technologiczne. Ale najważniejszy wniosek z tej publikacji brzmi następująco: te zyski dotyczą tylko doświadczonych programistów. Współpraca z SI zamiast zasypywać lukę kompetencyjną, pogłębia ją, premiując tych, którzy posiadają silne, inżynieryjne fundamenty. Wniosek z tego moim zdaniem jest taki, że jak na razie warto umieć kodzić nawet jak zamierza się vibe codować, a wieszczony przez wielkich tego świata zmierzch programistów jeszcze nie nadchodzi. Tu odwołuję się do niedawnych wypowiedzi Jensena Huanga, czy Dario Amodei.
Piotr Sankowski tweet media
Polski
57
69
655
62.2K
Iwona Bialynicka-Birula ⏩ retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
I'm not very happy with the code quality and I think agents bloat abstractions, have poor code aesthetics, are very prone to copy pasting code blocks and it's a mess, but at this point I stopped fighting it too hard and just moved on. The agents do not listen to my instructions in the AGENTS.md files. E.g. just as one example, no matter how many times I say something like: "Every line of code should do exactly one thing and use intermediate variables as a form of documentation" They will still "multitask" and create complex constructs where one line of code calls 2 functions and then indexes an array with the result. I think in principle I could use hooks or slash commands to clean this up but at some point just a shrug is easier. Yes I think LLM as a judge for soft rewards is in principle and long term slightly problematic (due to goodharting concerns), but in practice and for now I don't think we've picked the low hanging fruit yet here.
English
253
331
4.3K
818.8K
Iwona Bialynicka-Birula ⏩
It was always about communication: with other humans, with computers... nothing changed
English
0
0
3
58
Iwona Bialynicka-Birula ⏩ retweetledi
Eric S. Raymond
Eric S. Raymond@esrtweet·
If you are a software engineer "experiencing some degree of mental health crisis", now hear this, because I've been coding for 50 years since the days of punched cards and I have a salutary kick in your ass to deliver. Get over yourself. Every previous "programming is obsolete" panic has been a bust, and this one's going to be too. The fundamental problem of mismatch between the intentions in human minds and the specifications that a computer can interpret hasn't gone away just because now you can do a lot of your programming in natural language to an LLM. Systems are still complicated. This shit is still difficult. The need for people who specialize in bridging that gap isn't going to go away. As usual, the answer is: upskill yourself and adapt. If a crusty old fart like me can do it, you can too.
Tom Dale@tomdale

I don't know why this week became the tipping point, but nearly every software engineer I've talked to is experiencing some degree of mental health crisis.

English
653
1.7K
15.6K
1.6M
Iwona Bialynicka-Birula ⏩
Just to be clear, I'm by no means saying do not use AI, but for the love of God, do not use it blindly. ChatGPT actually has a really good explanation of why this quickly leads to codebase meltdown. I especially liked the insight that the problem gets even worse with LLM-agent codebases. chatgpt.com/share/69856901…
English
0
0
1
65
Iwona Bialynicka-Birula ⏩
AI won't replace software engineers, but a human using AI... will create even more work for software engineers to clean up all the spaghetti code slop
English
2
1
6
143
Iwona Bialynicka-Birula ⏩
It took the scammers only 4 days to update their systems - if only legit institutions were this efficient
Iwona Bialynicka-Birula ⏩ tweet media
English
0
0
1
98
Iwona Bialynicka-Birula ⏩
People who talk to themselves used to be considered crazy. Now we know that they are just powered by a reasoning model.
English
2
0
3
119
Sandy Petersen 🪔
Sandy Petersen 🪔@SandyofCthulhu·
I am 70 years old. I still get nightmares about showing up late for class or not being able to find my classroom or not being prepared for my class presentation. When does this stop?
English
2.5K
437
12.9K
995.6K
Iwona Bialynicka-Birula ⏩
How is the @grok bikini trend fundamentally different from this, and are we going to outlaw scissors now?
Iwona Bialynicka-Birula ⏩ tweet media
English
1
0
5
4.5K
Iwona Bialynicka-Birula ⏩ retweetledi
Cresta
Cresta@cresta·
🤖 AI is redefining customer experience & 2026 will be a tipping point. From AI-first interactions to true omnichannel engagement, speed, experimentation, and human-AI collaboration will determine which organizations lead the way. Read the predictions 🔗 cresta.com/2026-predictio…
Cresta tweet media
English
0
1
2
126
Iwona Bialynicka-Birula ⏩
@chatgpt21 "Devs" will just need to learn to tell the LLMs what to do. Just ask yourself: "Did the number of software engineering jobs go up or down since the invention of compilers?"
GIF
English
0
0
0
25
Chris
Chris@chatgpt21·
Look me dead in the face anon and tell me GPT 8.5 Codex and Claude 6.5 Opus aren’t going to take the vast majority of dev jobs
English
159
31
1.4K
90.2K