Andrés Anaya Isaza - PhD

46 posts

Andrés Anaya Isaza - PhD banner
Andrés Anaya Isaza - PhD

Andrés Anaya Isaza - PhD

@AndresAI_

Scientist in artificial intelligence and bioengineering, passionate about knowledge and technology creation, as well as a musical artist at heart.

Medellín - Colombia Katılım Mart 2017
156 Takip Edilen59 Takipçiler
Craft Gallery
Craft Gallery@5min__crafts·
Try this 😜
English
13
701
2.7K
169.4K
Freddy Vega
Freddy Vega@freddier·
Ser un gran diseñador con conocimientos básicos de código y Linux, junto con Claude Code/Codex, tiene que ser la combinación de skills más poderosa del momento.
Español
78
98
1.6K
85.1K
Andrés Anaya Isaza - PhD
Andrés Anaya Isaza - PhD@AndresAI_·
@elonmusk @Yuchenj_UW Singularity ≠ bigger models. It implies qualia + human-level emergent cognition. Today we don’t even know how to measure that let alone implement it...
English
0
0
0
14
Yuchen Jin
Yuchen Jin@Yuchenj_UW·
Roban (Gemini co-author, now at Anthropic) and Lucas (ex-DeepMind/OpenAI, now at Meta) understand models like Claude Opus down to the atomic level. And yet, Claude Code still blew their minds. This happens all the time in AI. Even the people who build these systems are shocked by the intelligence that emerges, just like Demis being stunned by AlphaGo’s “Move 37.” That mystery is what pulled me into AI research during my PhD. AI is mysterious and unpredictable, yet deeply attractive. The models just want to learn. Buckle up. Fast takeoff ahead!
Yuchen Jin tweet media
Yuchen Jin@Yuchenj_UW

ex-Google and Meta distinguished engineer, Gemini co-author @_arohan_: “if I had agentic coding and particularly opus, I would have saved myself first 6 years of my work compressed into few months.” This matches my experience. AI collapses the learning curve, and turns junior engineers into senior engineers dramatically fast. New-hire onboarding on large codebases shrinks from months to days. What used to take hours of Googling and Stack Overflow is now a single prompt. AI is also a good mentor and pair programmer. Agency is all you need now.

English
157
254
3K
1.2M
Andrés Anaya Isaza - PhD
Andrés Anaya Isaza - PhD@AndresAI_·
@lexfridman Review the mathematics of "infinite memory o Nested Learning", Especially because of the border problems with the creation of drugs and vaccines.
English
0
0
0
15
Lex Fridman
Lex Fridman@lexfridman·
Doing a long, super-technical podcast on the state-of-the-art in AI. Let me know if you have question, topic suggestions. Everything from details of LLM training pipeline & architectures, to coding, robotics, scaling, compute, business, geopolitics, etc. Besides topics & questions... add papers, blogs, posts, rants, perspectives that you'd like to see covered.
English
1.1K
271
5K
757.7K
Aryan Mahajan
Aryan Mahajan@aryanXmahajan·
TikTok + n8n + Claude = AI Content Machine (this system monitors Instagram, YouTube, TikTok, LinkedIn, Twitter, Facebook for viral content 24/7) → No manual scrolling. → No $12K agency research fees. → No guessing what content works. Just automated intelligence extraction across 6 platforms. Perfect for brands tired of content guesswork. Here's how it works: → Platform monitoring engine (tracks LinkedIn, Twitter, IG, TikTok, YouTube, Reddit) → Comment analysis system (extracts pain points + desires automatically) → Psychological pattern mapping (identifies emotional triggers driving shares) → Content strategy generator (creates platform-specific playbooks instantly) → Trend detection loop (updates as market psychology shifts) Built with competitive intelligence architecture. Runs 24/7 without manual work. 10-minute setup. Enterprise-grade insights. Want the complete system blueprint? Like + comment "MACHINE" + repost, and I'll DM it to you. (must be following)
English
1.3K
717
2.6K
232.8K
Ernest Ryu
Ernest Ryu@ErnestRyu·
I used ChatGPT to solve an open problem in convex optimization. *Part I* (1/N)
English
86
351
2.2K
1.3M
Elon Musk
Elon Musk@elonmusk·
Sometimes I say things so profound that people that it’s stupid. And sometimes I say things so stupid that people thinks it’s profound.
English
32K
22.3K
315.2K
38.7M
Andrés Anaya Isaza - PhD
Andrés Anaya Isaza - PhD@AndresAI_·
@sama They had delayed team... As they are also a little delayed in expanding the context window...
English
0
0
0
11
Sam Altman
Sam Altman@sama·
today we are introducing codex. it is a software engineering agent that runs in the cloud and does tasks for you, like writing a new feature of fixing a bug. you can run many tasks in parallel.
English
1.2K
2.6K
36.2K
6.1M
AdrIAna
AdrIAna@adrianaia_·
Si lo de ayer en USA te sorprendió… espera a ver los MEMES que están arrasando hoy 😂🔥 ¡Prepárate para reír! 😂🇺🇸
AdrIAna tweet media
Español
398
3K
23.3K
8.4M
Andrés Anaya Isaza - PhD
Andrés Anaya Isaza - PhD@AndresAI_·
Deseo el mejor de los éxitos y bendiciones a @MauricioLizcano. Sin temor a equivocarme, debo decir que se consolidó como el Ministro de la IA. Nosotros los científicos y empresarios de IA, estaremos siempre agradecidos por tú gestión. #MinTic 2025
Español
1
2
4
336
Andrés Anaya Isaza - PhD retweetledi
Alex Reibman 🖇️
Alex Reibman 🖇️@AlexReibman·
OpenAI’s biggest rival is shaking things up. Anthropic invited 200+ elite hackers to their SF headquarters to see what’s possible with Claude Here’s what we saw at the @AnthropicAI x @MenloVentures Builder Day Hackathon (🧵):
Alex Reibman 🖇️ tweet mediaAlex Reibman 🖇️ tweet media
English
94
667
7.5K
1.2M
Andrés Anaya Isaza - PhD
Andrés Anaya Isaza - PhD@AndresAI_·
¡Premio Nobel para Geoffrey Hinton! Increíble!!! Esto es un claro mensaje que reafirma la importancia de la ciencia en Inteligencia Artificial 🙌🏼 Es un merecido reconocimiento que valida elementos teóricos en física y matemáticas, al más alto honor en la historia. @geoffreyhinton
Andrés Anaya Isaza - PhD tweet media
Español
0
0
2
70
Andrés Anaya Isaza - PhD retweetledi
Jim Fan
Jim Fan@DrJimFan·
OpenAI Strawberry (o1) is out! We are finally seeing the paradigm of inference-time scaling popularized and deployed in production. As Sutton said in the Bitter Lesson, there're only 2 techniques that scale indefinitely with compute: learning & search. It's time to shift focus to the latter. 1. You don't need a huge model to perform reasoning. Lots of parameters are dedicated to memorizing facts, in order to perform well in benchmarks like trivia QA. It is possible to factor out reasoning from knowledge, i.e. a small "reasoning core" that knows how to call tools like browser and code verifier. Pre-training compute may be decreased. 2. A huge amount of compute is shifted to serving inference instead of pre/post-training. LLMs are text-based simulators. By rolling out many possible strategies and scenarios in the simulator, the model will eventually converge to good solutions. The process is a well-studied problem like AlphaGo's monte carlo tree search (MCTS). 3. OpenAI must have figured out the inference scaling law a long time ago, which academia is just recently discovering. Two papers came out on Arxiv a week apart last month: - Large Language Monkeys: Scaling Inference Compute with Repeated Sampling. Brown et al. finds that DeepSeek-Coder increases from 15.9% with one sample to 56% with 250 samples on SWE-Bench, beating Sonnet-3.5. - Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters. Snell et al. finds that PaLM 2-S beats a 14x larger model on MATH with test-time search. 4. Productionizing o1 is much harder than nailing the academic benchmarks. For reasoning problems in the wild, how to decide when to stop searching? What's the reward function? Success criterion? When to call tools like code interpreter in the loop? How to factor in the compute cost of those CPU processes? Their research post didn't share much. 5. Strawberry easily becomes a data flywheel. If the answer is correct, the entire search trace becomes a mini dataset of training examples, which contain both positive and negative rewards. This in turn improves the reasoning core for future versions of GPT, similar to how AlphaGo’s value network — used to evaluate quality of each board position — improves as MCTS generates more and more refined training data.
Jim Fan tweet media
English
135
1.1K
6.1K
799.2K
Carlos Alarcón
Carlos Alarcón@alarcon7a·
@AndresAI_ Se abrió la discusión sobre los benchmarks, sobre si es cierto o solo se vende humo, sobre cómo medir realmente el impacto de un modelo etc etc etc y pues, atacan fuertemente al tipo que publicó el modelo
Español
1
0
1
30
Carlos Alarcón
Carlos Alarcón@alarcon7a·
Que tóxico se está volviendo todo al rededor de reflection 70b 👀
Español
4
0
3
456
Andrés Anaya Isaza - PhD
Andrés Anaya Isaza - PhD@AndresAI_·
Cordialmente invitados. Platzi Conf 2024. Donde hablaremos de "Nuestro paradójico destino con la IA" Se analizará el RoadMap de la Inteligencia Artificial General - AGI, la evolución de nuestra especie, un análisis de los AVANCES, singularidad, la civilización humana y mucho más!
Andrés Anaya Isaza - PhD tweet media
Español
0
0
1
1.5K