Angelo Poerio

2.3K posts

Angelo Poerio banner
Angelo Poerio

Angelo Poerio

@angelo_p8

Katılım Ocak 2016
833 Takip Edilen123 Takipçiler
Angelo Poerio retweetledi
Math Cafe
Math Cafe@Riazi_Cafe_en·
Most computer users never realize their typing is bound to a legacy layout (called QWERTY), mathematically optimized to MINIMIZE their type speed. See more below 👇
Math Cafe tweet media
English
5
8
78
8.7K
Angelo Poerio retweetledi
tmuxvim
tmuxvim@tmuxvim·
I put a prompt injection into my LinkedIn bio and recruiters are messaging me in Old English and calling me Lord.
tmuxvim tweet mediatmuxvim tweet media
English
645
7.5K
91.9K
4.2M
Angelo Poerio retweetledi
Probability and Statistics
Probability and Statistics@probnstat·
One theorem every ML engineer should know: The Bellman Optimality Principle. It states that the optimal solution to a decision problem can be constructed recursively from optimal subproblems. In reinforcement learning, this becomes: Why it matters: • Foundation of Q-learning and dynamic programming • Enables sequential decision-making under uncertainty • Central to robotics, game AI, and autonomous systems • Connects optimization with learning The profound idea: Intelligence can emerge from recursively improving future decisions. Almost every modern RL algorithm — from DQN to AlphaGo — builds on Bellman’s insight. Reinforcement learning is ultimately the mathematics of long-term consequences. Image: share.google/AIBaxXi8u61KVl…
Probability and Statistics tweet media
English
7
68
518
25.6K
Angelo Poerio retweetledi
Tom Jøran Sønstebyseter Rønning
Tom Jøran Sønstebyseter Rønning@L1v1ng0ffTh3L4N·
Microsoft Edge loads all your saved passwords into memory in cleartext — even when you’re not using them.
English
251
1.4K
8.9K
1.5M
Angelo Poerio retweetledi
Avi Chawla
Avi Chawla@_avichawla·
The most comprehensive RL overview I've ever seen. Kevin Murphy from Google DeepMind, who has over 128k citations, wrote this. What makes this different from other RL resources: → It bridges classical RL with the modern LLM era: There's an entire chapter dedicated to "LLMs and RL" covering: - RLHF, RLAIF, and reward modeling - PPO, GRPO, DPO, RLOO, REINFORCE++ - Training reasoning models - Multi-turn RL for agents - Test-time compute scaling → The fundamentals are crystal clear Every major algorithm, like value-based methods, policy gradients, and actor-critic are explained with mathematical rigor. → Model-based RL and world models get proper coverage Covers Dreamer, MuZero, MCTS, and beyond, which is exactly where the field is heading. → Multi-agent RL section Game theory, Nash equilibrium, and MARL for LLM agents. I have shared the arXiv paper in the replies!
Avi Chawla tweet media
English
11
182
1.3K
86.3K
Angelo Poerio retweetledi
tetsuo
tetsuo@tetsuoai·
how CNNs see images 16 boxes covering the core CNN stack. tensors, filters, feature maps, stride, padding, channels, pooling, receptive fields, mental model
tetsuo tweet mediatetsuo tweet mediatetsuo tweet mediatetsuo tweet media
English
12
146
840
33.7K
Angelo Poerio retweetledi
Suni
Suni@suni_code·
97% vibe coder will fumble during interviews if asked what this is.
Suni tweet media
English
452
326
8.2K
2M
Angelo Poerio retweetledi
Yohan
Yohan@yohaniddawela·
A single GPU can now calculate hundreds of global weather scenarios in under 60 seconds. The exact same task requires a supercomputer and hours of brute-force physics. Google DeepMind recently released WeatherNext 2. The model beats the previous state-of-the-art system on 99.9% of weather variables across a 15-day forecast window. It achieves this massive jump in accuracy using a new modelling approach called a Functional Generative Network. Meteorologists categorise weather data into two buckets: 1. Marginals are isolated data points, like the precise temperature at a specific location or the wind speed at a certain altitude. 2. Joints are the massive, interconnected systems that form when all those individual elements interact. The researchers hid the joint systems from the model during training. They only taught it the isolated marginals. When they turned it on, the model skillfully predicted the massive, complex systems anyway. The architecture forces an 87-million-dimensional output distribution through a 32-dimensional mathematical bottleneck. To survive this severe constraint and still produce accurate individual data points, the neural network has no choice but to learn the underlying physics linking everything together. It figures out the weather because that’s the most efficient way to solve the maths. The practical results are immediate. The model gives forecasters a full 24-hour advantage in tropical cyclone tracking compared to the previous leading system. It maps extreme wind speeds and heatwaves with unprecedented precision. We’re watching a pretty big shift in predictive capabilities. The machine is deducing the structural reality of planetary weather from isolated fragments of data.
Yohan tweet media
English
39
360
2.3K
223.4K
Angelo Poerio retweetledi
How To AI
How To AI@HowToAI_·
Someone just built a doomsday computer that runs without the internet. It's called Project N.O.M.A.D. It packs a local AI, all of Wikipedia, offline maps, medical guides, and full Khan Academy courses into a solar-powered mini PC. Runs on 15 watts. 100% Open Source.
How To AI tweet media
English
16
117
635
21.6K
Angelo Poerio retweetledi
Terminal Trove
Terminal Trove@terminaltrove·
netwatch is an all in one network diagnostics tool that monitors connections in real time. It has a live traffic timeline, ASCII network map, latency heat-maps and more. Matt Hartley (matthart1983 on GitHub) made netwatch using @ratatui_rs and is Terminal Tool of the Week! ⭐️
English
26
223
1.9K
112.2K