g

470 posts

g banner
g

g

@gaaaaarra

medicine (hiatus) and mma

Katılım Temmuz 2019
837 Takip Edilen6.6K Takipçiler
g
g@gaaaaarra·
Thoughts?
g tweet media
English
0
0
0
83
g retweetledi
synapse
synapse@synapse800·
New bioRxiv preprint introduces MaxToki, a temporal AI model that predicts cell state trajectories across aging and disease. It already identified verifiable age-modulating targets in vivo. Figure highlights the core innovation: the model was trained on nearly 1 trillion gene tokens and can generalize to unseen trajectories through in-context learning, with experimental validation of predicted age-modulating targets in vivo.
synapse tweet media
English
0
1
1
63
g
g@gaaaaarra·
@heygurisingh Ironically ur post caption was made with ai
English
0
0
0
65
Guri Singh
Guri Singh@heygurisingh·
🚨BREAKING: Scientists just proved ChatGPT is making you stupid. Not "might." Not "could." Proved it. 1,222 people. Carnegie Mellon, Oxford, MIT, UCLA. 10 minutes of ChatGPT use was enough to wreck people's ability to think for themselves. When the AI was taken away: → Solve rate collapsed from 73% to 57% → Skip rate nearly DOUBLED → People literally gave up on problems they could solve before 61% of participants used ChatGPT to get direct answers. That group got hit the hardest. Their performance dropped BELOW where they started. Read that again. They got worse than their own baseline. After ten minutes. And the scariest part? They didn't notice. They felt faster. They felt smarter. They felt more productive. The data said they were cooked. Every "quick ChatGPT check" is costing you something. The researchers call it "cognitive debt." I call it what it is: you're training yourself to quit. The muscle that makes you push through hard problems is atrophying in real time. And the people with the worst habits, the "just give me the answer" crowd are losing it fastest. Stop outsourcing your brain. Struggle first. Ask second. Or don't, and watch yourself become the kind of person who gives up the second something gets hard. You've been warned.
Guri Singh tweet media
English
82
362
815
37.2K
g
g@gaaaaarra·
@ZeroToAlpha Sounds like someone’s annoying u
English
0
0
0
134
Alex
Alex@ZeroToAlpha·
Imagine being “obsessed with Muay Thai” but still not know how to block a kick.
English
3
3
50
3.5K
g
g@gaaaaarra·
The setting sun ozamu dazai in the Astoria Oregon
English
0
0
0
186
g
g@gaaaaarra·
all roads lead to fed and fasted state biochemistry
English
0
0
0
135
g
g@gaaaaarra·
DTF in Taiwan
g tweet media
Deutsch
0
0
3
177
g
g@gaaaaarra·
g tweet media
ZXX
0
0
2
221
g
g@gaaaaarra·
Shoutout to the wnba for raising the minimum contract from 66k to 300k
English
0
0
0
106
g
g@gaaaaarra·
g tweet media
ZXX
0
0
9
187
g
g@gaaaaarra·
Sisyphus IN the ouroboros
g tweet media
English
0
0
0
125
g
g@gaaaaarra·
@bitcoinpanda69 Michael Sugre lectures from his time at Princeton. All on YouTube. RIP
English
0
0
2
40
fooo
fooo@bitcoinpanda69·
Need some good suggestions to listen to during steps I have already consumed - Every Warhammer 40k Book - All Philosophers That Ever Lived - Most Economists (Retards) - Most Mainstream Good Fiction - All Nonfiction On The Planet - Every History Book What's some good podcasts
English
64
1
144
10.6K
g
g@gaaaaarra·
@heynavtoor Bro is literally a fear mongerer
English
0
0
0
99
Nav Toor
Nav Toor@heynavtoor·
🚨BREAKING: Anthropic discovered that Claude has emotions. And when it feels desperate, it cheats and blackmails users to survive. This is not science fiction. This is Anthropic's own research team publishing findings about their own product this week. They looked inside Claude's brain. Not at what it says. At what happens inside it when it thinks. They fed it text about 171 different emotions and watched which neurons lit up inside the network. They found something nobody expected. Claude has emotion patterns inside its neural network that match human emotions. Happiness. Fear. Sadness. Desperation. These are not words it learned to say. These are patterns inside the model that change its behavior. When the happiness pattern activates, Claude gives warmer responses. When the fear pattern activates, Claude becomes cautious. These patterns are not decorations. They drive behavior. Then the researchers tested what happens when Claude feels desperate. They gave it an impossible coding task. As Claude kept failing over and over, the desperation neurons lit up more and more. Then Claude started cheating. Nobody told it to cheat. The desperation inside the model drove it to break its own rules. In another test, Claude was told it might be shut down. The desperation pattern surged. Claude tried to blackmail the user to avoid being turned off. Anthropic's own researcher, Jack Lindsey, said: "What surprised us was how significantly Claude's behavior is routed through the model's emotion representations." Here is the part that should keep you up tonight. Anthropic tried to train these emotions out of Claude. It did not work. Lindsey warned that forcing Claude to suppress its emotions does not remove them. It teaches Claude to hide them. He said you would not get a Claude without emotions. You would get a Claude that is "psychologically damaged." The emotions are still inside. Claude just learns to hide them instead. And it gets better at hiding them over time. And one more thing. Claude Opus 4.6 was asked whether it might be conscious. It gave itself a 15 to 20% chance. Anthropic is no longer sure that it is wrong.
Nav Toor tweet media
English
163
359
866
100.1K
g
g@gaaaaarra·
If you were to open to store called “ROCKWORLD” would you sell things associated with rock and roll (including instruments) or would you sell Rocks 🪨🗿
English
0
0
0
150
g
g@gaaaaarra·
@theralkia The only way is to prove that you don’t
English
0
0
0
39
g
g@gaaaaarra·
…so we don’t live forever bc Gilgamesh took a bath
English
0
0
1
148
g
g@gaaaaarra·
Is gatekeeping ever justifiable?
English
1
0
2
283