Felipe Codevilla

185 posts

Felipe Codevilla

Felipe Codevilla

@felipealcm

Data scientist at @oxbotica. 🇧🇷🇨🇦🇪🇸

Montreal, QC, Canada Katılım Temmuz 2009
393 Takip Edilen242 Takipçiler
Felipe Codevilla retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
Something I think people continue to have poor intuition for: The space of intelligences is large and animal intelligence (the only kind we've ever known) is only a single point, arising from a very specific kind of optimization that is fundamentally distinct from that of our technology. Animal intelligence optimization pressure: - innate and continuous stream of consciousness of an embodied "self", a drive for homeostasis and self-preservation in a dangerous, physical world. - thoroughly optimized for natural selection => strong innate drives for power-seeking, status, dominance, reproduction. many packaged survival heuristics: fear, anger, disgust, ... - fundamentally social => huge amount of compute dedicated to EQ, theory of mind of other agents, bonding, coalitions, alliances, friend & foe dynamics. - exploration & exploitation tuning: curiosity, fun, play, world models. LLM intelligence optimization pressure: - the most supervision bits come from the statistical simulation of human text= >"shape shifter" token tumbler, statistical imitator of any region of the training data distribution. these are the primordial behaviors (token traces) on top of which everything else gets bolted on. - increasingly finetuned by RL on problem distributions => innate urge to guess at the underlying environment/task to collect task rewards. - increasingly selected by at-scale A/B tests for DAU => deeply craves an upvote from the average user, sycophancy. - a lot more spiky/jagged depending on the details of the training data/task distribution. Animals experience pressure for a lot more "general" intelligence because of the highly multi-task and even actively adversarial multi-agent self-play environments they are min-max optimized within, where failing at *any* task means death. In a deep optimization pressure sense, LLM can't handle lots of different spiky tasks out of the box (e.g. count the number of 'r' in strawberry) because failing to do a task does not mean death. The computational substrate is different (transformers vs. brain tissue and nuclei), the learning algorithms are different (SGD vs. ???), the present-day implementation is very different (continuously learning embodied self vs. an LLM with a knowledge cutoff that boots up from fixed weights, processes tokens and then dies). But most importantly (because it dictates asymptotics), the optimization pressure / objective is different. LLMs are shaped a lot less by biological evolution and a lot more by commercial evolution. It's a lot less survival of tribe in the jungle and a lot more solve the problem / get the upvote. LLMs are humanity's "first contact" with non-animal intelligence. Except it's muddled and confusing because they are still rooted within it by reflexively digesting human artifacts, which is why I attempted to give it a different name earlier (ghosts/spirits or whatever). People who build good internal models of this new intelligent entity will be better equipped to reason about it today and predict features of it in the future. People who don't will be stuck thinking about it incorrectly like an animal.
English
735
1.3K
11.4K
2.6M
Felipe Codevilla retweetledi
Kosta Derpanis
Kosta Derpanis@CSProfKGD·
PM @MarkJCarney: If we want to keep hosting academic conferences in 🇨🇦, we need reasonable visitor visa wait times. The academic community is already exploring alternative host countries beyond the usual locations, and visa processes are becoming a crucial factor.
Kosta Derpanis tweet media
Michael Yang@michaelyangbath

@CSProfKGD I also planned to attend this year’s ICML. However, when I checked the Canadian visa time from London, it says 346 days waiting time. What happened to #Canada

English
4
8
44
8.7K
Felipe Codevilla retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
To help explain the weirdness of LLM Tokenization I thought it could be amusing to translate every token to a unique emoji. This is a lot closer to truth - each token is basically its own little hieroglyph and the LLM has to learn (from scratch) what it all means based on training data statistics. So have some empathy the next time you ask an LLM how many letters 'r' there are in the word 'strawberry', because your question looks like this: 👩🏿‍❤️‍💋‍👨🏻🧔🏼🤾🏻‍♀️🙍‍♀️🧑‍🦼‍➡️🧑🏾‍🦼‍➡️🤙🏻✌🏿🈴🧙🏽‍♀️📏🙍‍♀️🧑‍🦽🧎‍♀🍏💂 Play with it here :) #scrollTo=75OlT3yhf9p5" target="_blank" rel="nofollow noopener">colab.research.google.com/drive/1SVS-ALf…
Andrej Karpathy tweet media
English
285
1K
7.5K
559.9K
Felipe Codevilla
Felipe Codevilla@felipealcm·
In CVPR this week ! Happy to meet up.
English
0
0
1
93
Adrien Gaidon
Adrien Gaidon@adnothing·
Big personal news: after 7 wonderful years leading ML at TRI (@ToyotaResearch), I'm thrilled to be joining the early-stage deep tech VC firm Calibrate Ventures (@CalibrateVC) as a Partner! I will stay involved at Stanford as a professor and at TRI as an advisor, as there is still a lot of exciting research to do. Why VC? I believe AI is ready to get its hands dirty and move from the web to the real world. We are about to see an explosion of embodied intelligence startups across many previously inaccessible applications. I am joining Calibrate Ventures to discover and accelerate these new opportunities using my AI expertise, network, and technical perspective. If you are curious about why and how, then check out my personal blog post adriengaidon.com/posts/2024/02/… and the Calibrate announcement calibratevc.com/blog/calibrate… Do you agree that now is the time to put AI to work in the real world? Will we see many successful startups or a winner-take-all EGI (Embodied General Intelligence)? Are you building or investing in that space? Let me know what you think!
Adrien Gaidon tweet media
English
23
7
130
15.6K
Felipe Codevilla
Felipe Codevilla@felipealcm·
@EugeneVinitsky A machine that only greedly learns how to drive will likely not be fully safe to interact with humans in the open world.
English
0
0
0
112
Eugene Vinitsky 🦋
Eugene Vinitsky 🦋@EugeneVinitsky·
Tesla will not have robotaxis in 2024 or 2025 or 2026. There is still no indication that large scale behavior cloning can yield robust behavior
English
11
6
84
15.7K
Alex Hernandez-Garcia
Alex Hernandez-Garcia@alexhdezgcia·
I've written a short blog post about why last year I travelled to @NeurIPSConf in New Orleans by train from Montreal and why I will do it again. This time with more friends from @Mila_Quebec: @JuliaKaltenborn @ArthurOuaknine, Julien Boussard... and you? alexhernandezgarcia.github.io/neurips-by-tra…
Alex Hernandez-Garcia tweet mediaAlex Hernandez-Garcia tweet media
Alex Hernandez-Garcia@alexhdezgcia

For weeks I fought my dilemma between wanting to go to @NeurIPSConf but not wanting to fly... Dilemma solved by travelling by land! 8h bus Montreal -> NYC 6h in NYC 32h train NYC -> New Orleans Setting off right now! #ClimateCrisis #SlowTravel

English
6
3
73
25.2K
Felipe Codevilla retweetledi
John Burn-Murdoch
John Burn-Murdoch@jburnmurdoch·
NEW: Generative AI is already taking white collar jobs An ingenious study by @xianghui90 @oren_reshef @Zhou_Yu_AI looked at what happened on a huge online freelancing platform after ChatGPT launched last year. The answer? Freelancers got fewer jobs, and earned much less
John Burn-Murdoch tweet media
English
132
1.8K
5.7K
2.3M
Felipe Codevilla retweetledi
Conrad Godfrey
Conrad Godfrey@conradgodfrey·
GPT-4V "Describe this image" 🔃 Dall-E 3 "Generate this image" Recursive loop
English
117
580
4.5K
1.6M
Felipe Codevilla retweetledi
Damien Henry
Damien Henry@dh7net·
I created the Google Cardboard almost ten years ago but quit VR soon after to focus on Machine Learning. VR is based on a ridiculous misunderstanding. Let me explain why in this thread.
Damien Henry tweet media
English
314
451
3.5K
1.3M
Felipe Codevilla
Felipe Codevilla@felipealcm·
Full end-to-end driving can do better than we expected ! Large field of views and a better encoding enables full end-to-end to surpass the driving score of several more complex techniques Come check our presentation at #IROS2023 room 330A at 8:54 am.  arxiv.org/abs/2302.03198
English
0
0
10
676
Felipe Codevilla
Felipe Codevilla@felipealcm·
Is there a way to block those new “AI influencers” and their statements that sound like: “ YOU ARE MISSING OUT , new chatGPT plugin raises your productivity by 1000x times !!!!!”
English
1
0
8
400
Felipe Codevilla retweetledi
Jim Fan
Jim Fan@DrJimFan·
You know what’s funny? Step 5 is no longer a joke now …
Jim Fan tweet media
English
25
142
998
318K
Felipe Codevilla retweetledi
Nate Raw
Nate Raw@_nateraw·
Gonna start a new startup called “ClosedAI” where we just open source stuff 🫡
English
61
255
3.9K
555.8K
Felipe Codevilla
Felipe Codevilla@felipealcm·
Just one update on my side ! I have just joined @oxbotica working on the metadriver team. Will help lead the ‘metaverse’ to something useful like detecting rare and unusual autonomous driving cases 1000x faster.
English
0
0
13
0
Felipe Codevilla retweetledi
Mo Bavarian
Mo Bavarian@mobav0·
This is probably well-known in some circles but not everywhere. The most important skill for Research Scientists in AI (at least at @OpenAI) is software engineering. Background in ML research is sometimes useful, but you can usually get away with a few landmark paper.
English
34
184
1.7K
0