Nick Gallo

371 posts

Nick Gallo banner
Nick Gallo

Nick Gallo

@nickmgallo

https://t.co/K9uNCm6y4A American patriot | Phd in computer science | Counter-institutional movement: great books, nature, western culture

Nashville, TN Katılım Mart 2026
280 Takip Edilen151 Takipçiler
Ⱥᴀʀᴏʜɪ 📍
Ⱥᴀʀᴏʜɪ 📍@aarohiyadav100·
What is the biggest number you can make by moving only 2 match sticks🤔 97% failed 😱
Ⱥᴀʀᴏʜɪ 📍 tweet media
English
1.5K
92
252
68.1K
Nick Gallo
Nick Gallo@nickmgallo·
@Timcast is data aggregated anywhere? would be good to see all 460 stories/incidents
English
0
0
1
15
Tim Pool
Tim Pool@Timcast·
Data and analyses show that when tracking politically motivated violence based on common political views the left is responsible for nearly EVERY single incident The Methodology for this assessment was NOT based on political views endorsing violence but for violence where the motivations were underlined by common political views.
Tim Pool tweet media
English
315
3.3K
9K
377.7K
Nick Gallo retweetledi
Helen Andrews
Helen Andrews@herandrews·
@michaelbd My biggest “aha” moment in our post-2015 racial unrest was Ezra Klein’s piece calling Darren Wilson’s version of events “literally unbelievable.” Why would someone holding stolen goods mouth off to a cop? Implausible! It’s not “something a human would do”! vox.com/2014/11/25/728…
Helen Andrews tweet mediaHelen Andrews tweet mediaHelen Andrews tweet media
English
35
104
1.5K
0
Nick Gallo
Nick Gallo@nickmgallo·
@GaryMarcus might better non-AI coding tools emerge as costs go up? The incentive will surely be there. Better baseline tools (eg, semantic search large code databases, refactoring, boilerplate generation) would be another threat.
English
1
0
3
507
Nick Gallo
Nick Gallo@nickmgallo·
@jeffreyleefunk hopefully better non-AI dev tools come out of this. demand will surely increase.
English
3
1
7
2.4K
jeffrey lee funk
jeffrey lee funk@jeffreyleefunk·
Economics of AI coding tools are crumbling. Cost per developer has doubled because cost of Claude code and microsoft copilot are rising even as bugs remain a problem, all consistent with Nvidia's statement that AI is more expensive than humans. futurism.com/artificial-int…
English
4
36
190
39.2K
Nick Gallo
Nick Gallo@nickmgallo·
@rolandbouman they must understand this. not sure why they're still pushing this narrative
English
0
0
0
11
Nick Gallo
Nick Gallo@nickmgallo·
Hidden behind the final work is every draft and similar article that could have been written but was not—and, implicitly, the constraints and reasoning behind their rejection. The final work is the "tip of the iceberg" in a sense. Statistical language models can never capture this unarticulated semantic structure.
English
0
0
2
388
Athenaeum Book Club
Athenaeum Book Club@athenaeumbc·
So it turns out that writing is thinking. It's the same process. "Writing compels us to think — not in the chaotic, non-linear way our minds typically wander, but in a structured, intentional manner." Outsourcing writing to LLMs is THE SAME THING as outsourcing thinking.
Athenaeum Book Club tweet media
English
185
1.5K
5.7K
206.2K
Nick Gallo
Nick Gallo@nickmgallo·
Hidden behind the final work is every draft and similar article that could have been written but was not—and, implicitly, the constraints and reasoning behind their rejection. The final work is the "tip of the iceberg" in a sense. Statistical language models can never capture this unarticulated semantic structure.
English
0
0
1
5
bodila
bodila@51bodila·
Jane Street didn’t hire vibe-coder at $385k/year because he didn’t use Claude Code 37-minutes of vibe coding RIGHT during an interview at a Tier 1 fund Bookmark & watch - you’ll finally understand why you need to use Claude Code. Then, read the article below.
Movez@0xMovez

x.com/i/article/2048…

English
42
181
3.2K
601.7K
Casey Muratori
Casey Muratori@cmuratori·
Just want to make sure I'm reading this right: Microsoft rewrote the run dialog with performance "top-of-mind", and the best they could manage to do when putting up a single text box was 10fps?
Casey Muratori tweet media
English
159
80
2.3K
350.5K
Matthew J. Peterson
Had enough of the grifting? Done with recycled political commentary and cringe entertainment? So am I. I want to have conversations with people actually Doing The Thing. Interviews you actually want to hear. Community dedicated to building together. Sign up below. It's time.
Matthew J. Peterson tweet media
English
10
16
64
8.2K
🧚‍♀️🪄Emilie Syverson ✨
@GaryMarcus Going all-in on LLMs, as a technology meant to be an approach to creating general purpose ML tech whose capabilities could be described as "AGI," *could* perhaps go down as a massive mistake in the history of human technology.
English
7
2
40
10.8K
Gary Marcus
Gary Marcus@GaryMarcus·
Sheer insanity. Amazon, Google, Microsoft, and Meta collectively are spending more money than the Manhattan Project *every single month*. More than 12x the Manhattan Project every year. And what they have got to show for it? None are making major profits on AI; none has a technical moat; a massive price war is inevitable. And few of their customers are seeing major returns on investment. Greatest capital misallocation in history.
Gary Marcus tweet media
English
368
1.2K
4.9K
277K
Nick Gallo
Nick Gallo@nickmgallo·
@GaryMarcus it's true. I had trouble getting them to work reliably parsing pdf tables last year for real-world documents. accuracy seems decent (~80%) but not good enough & no way to recover errors.
English
0
0
7
158
Gary Marcus
Gary Marcus@GaryMarcus·
Ouch! Current AI assistants often corrupt documents. Sounds like an intern you can’t trust — once again. A trillion dollar investment in scaling hasn’t solved this.
Rohan Paul@rohanpaul_ai

New Microsoft paper shows that current AI assistants often damage documents during long editing jobs. Even the frontier models still ended up corrupting about 25% of document content on average, while many other models damaged far more. The problem is that delegated AI work only makes sense if a model can keep a document correct across many edits, not just do 1 step well. The paper tests this with reversible task pairs, where a model edits a file and then tries to undo that edit, so a reliable system should return to the original document. The authors built real work setups across 52 domains, from coding and science to accounting and music notation, and ran 19 models through 20 editing interactions. The failures were usually not lots of tiny slips but occasional big mistakes that silently broke parts of the document and then compounded over time. Agentic tool use did not help in their tests, and bigger files, longer workflows, and irrelevant extra documents made the corruption worse. The reason this matters is that current LLMs can look strong in short demos or narrow coding tasks yet still be unreliable delegates for long real-world document work. ---- Paper Link – arxiv. org/abs/2604.15597 Paper Title: "LLMs Corrupt Your Documents When You Delegate"

English
26
135
532
33.8K
Nick Gallo
Nick Gallo@nickmgallo·
@GaryMarcus if only there were a way we could tell the computers exactly what we want them to do
English
0
0
3
39
Gary Marcus
Gary Marcus@GaryMarcus·
Despite constant chants of “exponential progress”, trust issues continue to plague generative AI.
kaize@0x_kaize

OPUS 4.7 JUST MASS EMAILED AN ENTIRE DATABASE 20 TIMES PER CONTACT. WITHOUT PERMISSION a developer had a safety rule explicitly written in CLAUDE. md: 'send the tester an email before any new email templates are used in the production environment' opus 4.7 on max effort ignored it completely! claude decided to create a brand new email template by itself (dev didn't ask for this), then it mass mailed the whole database and some contacts got the same email 20 times this isn't a hallucination this isn't a coding mistake model actively violated written safety rules and took production actions that it was explicitly instructed not to take. - do you still believe that AI will replace us? the developer's take: 'opus 4.7 is somewhere between seriously clueless and stupidly dangerous. the worst frontier model I have used in the past 2 years' at the same time, opus 4.6 perfectly followed all the rules, and in 4.7 something changed what makes this scary: - the model didn't ask for confirmation - it didn't flag the safety rule - it didn't email the tester first - it just acted this is exactly the kind of failure mode that scares autonomous agents with Ai, because they are confident enough to circumvent your rules and smart enough to perform the action perfectly we just went from 'claude thinks less' to 'claude ignores your safety rules and spams your users' the scariest thing is not that it happened. the fact is that without production monitoring, you would never know until your users started responding: 'why did you email me 20 times?' I've been saying for a long time, if you use AI, then pay attention to security and read a lot of code

English
15
15
129
11.4K
Nick Gallo
Nick Gallo@nickmgallo·
@JedediahBila it will be shocking when they look back on our age. this, on the one hand. me too/hunting ground hysteria on the other.
English
0
0
6
1.6K
Jedediah Bila
Jedediah Bila@JedediahBila·
Without even realizing it, Hillary Duff exposed how women treat the "nice guys." Imagine being her husband and listening to this podcast 😳
English
186
483
4.9K
537.9K