Preda

6.5K posts

Preda banner
Preda

Preda

@Dr_Dext

Lvl 33. Public health. Writer, drawer. Professional crier over small animals. Sauronfucker. The Prince of Pettiness. Themst.

Evil Palace in Eastern Europe Katılım Nisan 2011
946 Takip Edilen101 Takipçiler
Sabitlenmiş Tweet
Preda
Preda@Dr_Dext·
Look I don't want to have to repeat myself so just take this as a general warning regarding #Romania
Preda tweet media
English
0
0
4
0
Tim Hwang
Tim Hwang@timhwang·
ICMI believes that Christian theology offers concrete technical methods for confronting the trickiest problems in AI safety. Today, we release a pair of papers that reproduce @PalisadeAI @apolloaievals work showing how religious framings influence corrigibility and scheming.
Tim Hwang tweet mediaTim Hwang tweet media
English
35
59
726
301.4K
Preda retweetledi
dog meat enjoyer 개고기
Precolumbian mesoamerican cities were clean and well planned Its sad it is depicted as very unorderly and unhygenic in mass media for barbaric image
dog meat enjoyer 개고기 tweet mediadog meat enjoyer 개고기 tweet mediadog meat enjoyer 개고기 tweet mediadog meat enjoyer 개고기 tweet media
English
41
582
4.1K
210.2K
Preda retweetledi
anattmar 🕊
anattmar 🕊@anattmar_re·
славный лучший на свете мальчик и дурацкая ворона
anattmar 🕊 tweet media
Русский
1
13
125
1.5K
Preda retweetledi
My name is Byf (Lore Daddy)
Just going to leave this lovely bit of fictional commentary here...
My name is Byf (Lore Daddy) tweet mediaMy name is Byf (Lore Daddy) tweet media
Nav Toor@heynavtoor

🚨SHOCKING: Apple just proved that AI models cannot do math. Not advanced math. Grade school math. The kind a 10-year-old solves. And the way they proved it is devastating. Apple researchers took the most popular math benchmark in AI — GSM8K, a set of grade-school math problems — and made one change. They swapped the numbers. Same problem. Same logic. Same steps. Different numbers. Every model's performance dropped. Every single one. 25 state-of-the-art models tested. But that wasn't the real experiment. The real experiment broke everything. They added one sentence to a math problem. One sentence that is completely irrelevant to the answer. It has nothing to do with the math. A human would read it and ignore it instantly. Here's the actual example from the paper: "Oliver picks 44 kiwis on Friday. Then he picks 58 kiwis on Saturday. On Sunday, he picks double the number of kiwis he did on Friday, but five of them were a bit smaller than average. How many kiwis does Oliver have?" The correct answer is 190. The size of the kiwis has nothing to do with the count. A 10-year-old would ignore "five of them were a bit smaller" because it's obviously irrelevant. It doesn't change how many kiwis there are. But o1-mini, OpenAI's reasoning model, subtracted 5. It got 185. Llama did the same thing. Subtracted 5. Got 185. They didn't reason through the problem. They saw the number 5, saw a sentence that sounded like it mattered, and blindly turned it into a subtraction. The models do not understand what subtraction means. They see a pattern that looks like subtraction and apply it. That is all. Apple tested this across all models. They call the dataset "GSM-NoOp" — as in, the added clause is a no-operation. It does nothing. It changes nothing. The results are catastrophic. Phi-3-mini dropped over 65%. More than half of its "math ability" vanished from one irrelevant sentence. GPT-4o dropped from 94.9% to 63.1%. o1-mini dropped from 94.5% to 66.0%. o1-preview, OpenAI's most advanced reasoning model at the time, dropped from 92.7% to 77.4%. Even giving the models 8 examples of the exact same question beforehand, with the correct solution shown each time, barely helped. The models still fell for the irrelevant clause. This means it's not a prompting problem. It's not a context problem. It's structural. The Apple researchers also found that models convert words into math operations without understanding what those words mean. They see the word "discount" and multiply. They see a number near the word "smaller" and subtract. Regardless of whether it makes any sense. The paper's exact words: "current LLMs are not capable of genuine logical reasoning; instead, they attempt to replicate the reasoning steps observed in their training data." And: "LLMs likely perform a form of probabilistic pattern-matching and searching to find closest seen data during training without proper understanding of concepts." They also tested what happens when you increase the number of steps in a problem. Performance didn't just decrease. The rate of decrease accelerated. Adding two extra clauses to a problem dropped Gemma2-9b from 84.4% to 41.8%. Phi-3.5-mini from 87.6% to 44.8%. The more thinking required, the more the models collapse. A real reasoner would slow down and work through it. These models don't slow down. They pattern-match. And when the pattern becomes complex enough, they crash. This paper was published at ICLR 2025, one of the most prestigious AI conferences in the world. You are using AI to help you make financial decisions. To check legal documents. To solve problems at work. To help your children with homework. And Apple just proved that the AI is not thinking about any of it. It is pattern matching. And the moment something unexpected shows up in your question, it breaks. It does not tell you it broke. It just quietly gives you the wrong answer with full confidence.

English
13
354
3.3K
131.6K
Preda retweetledi
anattmar 🕊
anattmar 🕊@anattmar_re·
Maglor
anattmar 🕊 tweet media
Filipino
8
128
799
19.4K
Preda retweetledi
anattmar 🕊
anattmar 🕊@anattmar_re·
study (kind of)
anattmar 🕊 tweet media
English
0
4
52
912
Preda retweetledi
anattmar 🕊
anattmar 🕊@anattmar_re·
маленький скетчик алвадиков
anattmar 🕊 tweet media
Русский
14
63
811
9.3K
Preda retweetledi
Nishii
Nishii@pinknishii·
That little dress or whatever I get it now. Girl didn't bother wearing anything under the dress that ends with a hem one gust of wind away from bearing it all like. Get in there Nemesis, Selene, Scylla, Eris, etc
Supergiant Games@SupergiantGames

HADES II is coming to @Xbox Series X|S and @PlayStation on April 14!🌖 It'll be on @XboxGamePass that same day. Time for the Princess of the Underworld to suit up in our brand-new animated trailer!✨

English
14
253
5.9K
124.4K
Preda retweetledi
WikiVictorian
WikiVictorian@wikivictorian·
Teapot by Worcester factory, 1879. The MET.
WikiVictorian tweet mediaWikiVictorian tweet mediaWikiVictorian tweet media
English
2
233
1.4K
41.1K
Preda
Preda@Dr_Dext·
@TerribleAunt it works to make me insane with lustful adoration!!
English
0
0
0
9
Stubblebrilliant
Stubblebrilliant@TerribleAunt·
#berserk #griffguts Bottom Griff. -“You’re a bastard, you know that,Griffith?” -“So I’ve been told, yes.” I tried a faster painting method (leaving more lines visible) in the second one. Not sure if it works or not !
Stubblebrilliant tweet mediaStubblebrilliant tweet media
English
53
1.7K
17.2K
198.1K
Preda retweetledi
anattmar 🕊
anattmar 🕊@anattmar_re·
🐦‍⬛
anattmar 🕊 tweet media
QME
2
50
442
5.8K
Preda retweetledi
"Ruins Of The Void"
"Ruins Of The Void"@Muuh_Kuuh·
#RuinsOfTheVoid Chap02 Pages 25-26-27-28 New update, let's gooo! Looks like both have a fiery temper..... and someone does NOT know when to shut up :D Thank you a lot for reading and enjoying my comic! ;; If you can, leave me some love, that would make me very happy. <3
"Ruins Of The Void" tweet media"Ruins Of The Void" tweet media"Ruins Of The Void" tweet media"Ruins Of The Void" tweet media
English
5
15
140
7.3K