ME|META|

2.1K posts

ME|META| banner
ME|META|

ME|META|

@interMemetica

Weaving a tighter fabric of universal self-discovery. || M-Mbresia Strategy || NomisMeta || InterMemetica

Everywhere at once Katılım Temmuz 2015
384 Takip Edilen109 Takipçiler
Sabitlenmiş Tweet
ME|META|
ME|META|@interMemetica·
The human psyche, concentrated in the individual, interacts clumsily with itself in human relations not specifically pyche-centric.
English
1
6
11
0
ME|META| retweetledi
Camille Roux
Camille Roux@CamilleRoux·
Algorithm Visualizer devrait être dans la boîte à outils de tout·e développeur·se qui apprend ou enseigne des algos — la visualisation en direct à partir du code rend les mécanismes abstraits vraiment lisibles. algorithm-visualizer.org
Français
11
420
2.7K
104.3K
ME|META|
ME|META|@interMemetica·
@nakadai_mon @HowToAI_ Each sub-agent can still be prompted to consider its interactions into a recursive pre-output stage.
English
0
0
0
4
ИΛKΛDΛI👒🦁#FREESAMOURAI
@HowToAI_ If you want a team of experts you let the agent summon sub-agents with different roles, let them argue with eachother and let the best solution rise to the top.
English
1
0
0
96
How To AI
How To AI@HowToAI_·
🚨 MIT just dropped a technique that makes ChatGPT reason like a team of experts instead of one overconfident intern. It’s called “Recursive Meta-Cognition” and it outperforms standard prompts by 110%. Right now, if you want an AI to solve a problem, you ask one question. The AI gives you one answer. If it’s wrong, you never know. It’s like asking a random person on the street for medical advice and just trusting them.. MIT researchers realized this is a massive structural flaw. They asked a simple question. What if AI could check its own work from multiple angles before giving you the final answer? Not just "think step by step." Actually verify. Score confidence. Flag uncertainty. They built a framework called Recursive Meta-Cognition. And the results rewrite the rules of prompt engineering. Instead of a single, blind output, the technique forces the AI to break a complex problem into smaller pieces and interrogate its own logic from multiple perspectives. But here is the secret sauce. It assigns a strict confidence score to every single reasoning path it takes, from 0.0 to 1.0. Paths below 0.4 are instantly rejected. Paths above 0.8 are trusted. Anything in between? The AI stops and tells you: "I am not sure, and here is why." No more hallucinations. No more fake confidence. If you are running a business and relying on AI to outline a SaaS marketing strategy, analyze data, or drive operational efficiency, a hallucination isn't just an error. It is a loss of profit. With a standard prompt, the AI gives you ONE answer. It sounds confident. It might be completely wrong for your specific situation. With Recursive Meta-Cognition, the AI spots its own hidden assumptions. It tests its logic. It catches its own errors before they ever reach your screen. We've been treating AI like a vending machine. Put a prompt in, get a result out. But true intelligence doesn't just answer questions. It questions its own answers.
How To AI tweet media
English
63
165
762
45.5K
whowouldathunk
whowouldathunk@whowouldathunk2·
@bswrpray @JustinEchterna9 You are of course free to believe whatever you think explains reality. I can't logically see how consciousness came first. Physical reality existed before biology showed up. Consciousness is a property of biology, not of rocks.
English
2
0
1
30
Justin Echternach
Justin Echternach@JustinEchterna9·
Orch OR specifies a physical substrate (microtubules), a physical process (quantum coherence in aromatic networks), a physical trigger (objective reduction at a specific threshold), and physical perturbations that should disrupt it (anesthetics at specific binding sites). Whether it's correct is an empirical question. But it's more mechanistically specific than any competing theory of consciousness ...
English
11
3
27
2.5K
ME|META| retweetledi
Bartosz Naskręcki
Bartosz Naskręcki@nasqret·
I am almost halfway through the semester, and this term I am teaching two heavy courses (60 hours each). One is about cryptanalysis: all the way from classical cryptology of substitution ciphers, through block ciphers, up to post-quantum cryptography. The second course is a tutorial: a "zero to hero" introduction to neural networks, starting from the classic topics of the 1940s up to the transformer revolution. I have decided to rely heavily on agentic tools to help me write lecture notes, slides, and interactive applets. It's a very systematic process where I give lectures, do exercises with students, and fine-tune the materials. I have noticed that the boost is incredible. Now we can really dive deep into the topics, adding a lot of custom-made help and examples (full Enigma breaking or high-level implementations in JAX). I see that the students enjoy having such complete notes. I put a lot of effort into the structure and scaffolding of those notes. Many texts are generated from good sources, but I read every single page and apply corrections. I think I have saved a lot of hours on the tedious fine-tuning of pictures and examples. I also had the courage to test and cover much more experimental content that I had never explored before. We spend three active hours each week with the students, and the work is very intense. But at the same time, the courses feel very rewarding, and I should admit that I have learned a ton (especially about the classic papers on AI). I will definitely keep building complete lecture notes for future courses, but I already see that the main issue is proper internalization of the knowledge, both by the students and by me. Overall, the process is much smoother because we can always ask an LLM to provide a more accurate answer.
Bartosz Naskręcki tweet mediaBartosz Naskręcki tweet media
English
17
58
710
117.2K
ME|META| retweetledi
Mehtaab Sawhney
Mehtaab Sawhney@mehtaab_sawhney·
We’ve just released another paper solving five further Erdős problems with an internal model at OpenAI: arxiv.org/abs/2604.06609. Several of the proofs were especially enjoyable to digest while writing the paper. My personal favorite was the solution to Erdős Problem 1091. The question asks: if a graph G has chromatic number 4, while every small subgraph has chromatic number at most 3, must it contain an odd cycle with many diagonals? The internal model gives a very enlightening counterexample to this conjecture, and the proof was a pleasure to understand. For those so inclined, a really fun exercise is to try to reconstruct the proof from Figure 5 of the paper, which was of course produced by Codex.
Mehtaab Sawhney tweet media
English
20
150
888
195.3K
ME|META| retweetledi
Vincent Abbott
Vincent Abbott@vtabbott_·
The lack of formalism wrt broadcasting in deep learning models annoyed me so much I learned category theory. Weaves, Wires, and Morphisms is now out on arXiv! First step to using the Yoneda lemma to automatically derive fused kernels. arxiv.org/abs/2604.07242
Vincent Abbott tweet media
English
7
37
275
12.9K
ME|META| retweetledi
Antonio Lupetti
Antonio Lupetti@antoniolupetti·
This is Algebrica. A mathematical knowledge base I’ve been building for 2.5 years. 215+ entries, carefully written and structured. 400k+ views over this time. Not much in absolute terms, but meaningful to me. No ads. No courses to sell. No gamification. No distractions. Just essential pages, aiming to explain mathematics as clearly as possible, for a university-level audience. Built simply for the pleasure of sharing knowledge. Content licensed under Creative Commons (BY-NC). Best experienced on desktop. If it helps even a few people understand something better, it’s worth it.
Antonio Lupetti tweet media
English
188
784
5.5K
221.2K
ME|META| retweetledi
Zo
Zo@zo_zhuo_wang·
When Kumiko (組子) Meets Computational Fluid Dynamics (CFD): vortex shedding. Just beautiful, at every scale.
Zo tweet mediaZo tweet mediaZo tweet mediaZo tweet media
English
3
20
124
7.4K
ME|META| retweetledi
Teknium (e/λ)
Teknium (e/λ)@Teknium·
We are now the 6th largest ai app in the world on Open Router and growing!
Teknium (e/λ) tweet media
English
95
54
1.1K
29.9K
ME|META| retweetledi
Elijah Meeks
Elijah Meeks@Elijah_Meeks·
I don't know what it mean anymore that I can just spin up something like this in 15 minutes by asking @claudeai to integrate Semiotic and Pretext using the wikipedia streaming API. This would have been such an extraordinary thing but now it's just... Tuesday.
GIF
English
17
42
387
18.9K
ME|META| retweetledi
missjenny
missjenny@missjenny·
“the tensor field” by gemini micron 005 as the mesh with a single stroke of metallic gelly roll as the observer I FUCKING LOVE AI
missjenny tweet media
English
6
5
43
1.7K
ME|META| retweetledi
Dr. Paul Wilhelm | Advanced Rediscovery
🏗️ Your textbook operates at the ground floor. The building has five stories. 🔸Level 0: Scalar wave equation — the seed 🔸Level 1: Hertz potential — Lorenz gauge is AUTOMATIC here 🔸Level 2: Whittaker scalars — two functions generate ALL of EM 🔸Level 3: Four-potential — where textbooks start 🔸Level 4: Fields (E, B) — where textbooks end Each level generates the next by differentiation. Each derivative is lossy. The Lorenz gauge, which at Level 3 is an imposed constraint physicists argue about, is at Level 1 an algebraic tautology. Not a law. A tautology of deeper structure. Standard EM operates at Levels 3-4. Two out of five. My paper "The Deleted Degrees of Freedom" maps the full hierarchy in Section 4.7. Pinned on my profile.
Dr. Paul Wilhelm | Advanced Rediscovery tweet media
English
3
15
77
3.2K
ME|META|
ME|META|@interMemetica·
@oprydai -kapnet replicates the brain
English
0
0
1
16
Mustafa
Mustafa@oprydai·
the difference between a biological neuron and an artificial neuron is massive. they share the idea. not the reality. what a biological neuron does: • receives signals through dendrites → thousands of inputs, not just a few • integrates over time → signals accumulate, decay, interact • fires spikes → discrete events, not smooth numbers • adapts continuously → plasticity rewires connections based on experience • runs on chemistry + electricity → noisy, slow, but incredibly efficient what an artificial neuron does: • takes weighted inputs → simple numbers in, number out • applies an activation function → relu, sigmoid, etc • updates via backprop → global optimization, not local biology • runs on silicon → fast, precise, but power-hungry • no real “memory” → unless explicitly designed (rnn, transformers) why it matters: • ai today is inspired by the brain, not a replica of it. • biological neurons are dynamic, adaptive, and energy-efficient. • artificial neurons are simplified, scalable, and mathematically convenient. we’re not building brains. we’re building approximations that work well enough to be useful.
Mustafa tweet media
English
33
68
345
15.3K
ME|META| retweetledi
Autism Capital 🧩
Autism Capital 🧩@AutismCapital·
Imagine a place where people could communicate around one shared idea and that only attracted people interested in that one specific topic thus cutting down on unnecessary noise and voices irrelevant to the topic being discussed.
Autism Capital 🧩 tweet media
English
246
171
2.3K
99.1K
ME|META| retweetledi
uncle-sam killed-the-clown
uncle-sam killed-the-clown@KilledSam59564·
@mathemetica physical constants and dimensions are not "inputs" to the universe, but emergent "outputs" of a 12-state discrete phase lattice attempting to describe an analog process. The "cost" of this discretization—the mathematical overhead—is what we measure as the "Laws of Physics."
English
0
1
1
57
ME|META| retweetledi
aeon
aeon@aeonBTC·
Abracadabra. From the ancient Aramaic: *avra kehdabra* - “I create as I speak.” That isn’t metaphor anymore. Words become worlds. Prompt the machine, and reality bends.
English
0
3
11
667
ME|META| retweetledi
Anish Moonka
Anish Moonka@anishmoonka·
Every time you get a cancer biopsy, the lab makes a tissue slide that costs about $5. It shows the shape of your cells under a microscope, and every cancer patient already has one on file. There’s a much fancier version of that test called multiplex immunofluorescence (basically a protein-level map showing which immune cells are near your tumor and what they’re doing). It costs thousands of dollars per sample, takes specialized equipment most hospitals don’t have, and barely scales. But it’s the kind of data oncologists need to figure out whether immunotherapy will actually work for you. Right now, only about 20 to 40% of cancer patients respond to immunotherapy, and one of the biggest reasons is that doctors can’t easily tell whether a tumor is “hot” (immune cells actively fighting it) or “cold” (immune system ignoring it). Microsoft, Providence Health, and the University of Washington trained an AI to analyze the $5 slide and predict what the expensive test would show across 21 different protein markers. They called it GigaTIME, trained it on 40 million cells in which both the cheap slide and the expensive test coexisted, and then turned it loose on 14,256 real cancer patients across 51 hospitals in 7 US states. The results landed in Cell, one of the most selective journals in biology. The model generated about 300,000 virtual protein maps covering 24 cancer types and 306 subtypes. It found 1,234 real, verified connections between immune cell behavior, genetic mutations, tumor staging, and patient survival that were previously invisible at this scale. When they tested it against a completely separate database of 10,200 cancer patients, the results matched up almost perfectly (0.88 out of 1.0 agreement). Nature Methods named spatial proteomics (mapping where specific proteins sit inside your tissue) its Method of the Year in 2024, and specifically cited GigaTIME in a March 2026 update as a model that “democratizes” this kind of analysis. The full model is open-source on Hugging Face. Any cancer research lab with archived biopsy slides, and most of them have thousands, can now run virtual immune profiling without buying a single piece of new equipment.
Satya Nadella@satyanadella

We’ve trained a multimodal AI model to turn routine pathology slides into spatial proteomics, with the potential to reduce time and cost while expanding access to cancer care.

English
104
1.8K
11.1K
942K