Steven

2.5K posts

Steven banner
Steven

Steven

@ascendingfn

22

Still Inside Yash Patel Katılım Ağustos 2019
584 Takip Edilen135 Takipçiler
Sabitlenmiş Tweet
Steven
Steven@ascendingfn·
Great work is happiness.
English
0
0
7
1.7K
Steven retweetledi
Greg Brockman
The world is transitioning to a compute-powered economy. The field of software engineering is currently undergoing a renaissance, with AI having dramatically sped up software engineering even over just the past six months. AI is now on track to bring this same transformation to every other kind of work that people do with a computer. Using a computer has always been about contorting yourself to the machine. You take a goal and break it down into smaller goals. You translate intent into instructions. We are moving into a world where you no longer have to micromanage the computer. More and more, it adapts to what you want. Rather doing work with a computer, the computer does work for you. The rate, scale, and sophistication of problem solving it will do for you will be bound by the amount of compute you have access to. Friction is starting to disappear. You can try ideas faster. You can build things you would not have attempted before. Small teams can do what used to require much larger ones, and larger ones may be capable of unprecedented feats. More and more, people can turn intent into software, spreadsheets, presentations, workflows, science, and companies. People are spending less energy managing the tool and more energy focusing on what they are actually trying to create. That shift brings a kind of joy back into work that many people haven’t felt in a long time. Everyone can just build things with these tools. This is disruptive. Institutions will change, and the paths and jobs that people assumed were stable may not hold. We don’t know exactly how it will play out and we need to take mitigating downsides very seriously, as well as figuring out how to support each other as a society and world through this time. But there is something very freeing about this moment. For the first time, far more people can become who they want to become, with fewer barriers between an idea and a reality. OpenAI’s mission implies making sure that, as the tools do more, humans are the ones who set their intent and that the benefits are broadly distributed, rather than empowering just one or a small set of people. We're already seeing this in practice with ChatGPT and Codex. Nearly a billion people are using these systems every week in their personal and work lives. Token usage is growing quickly on many use-cases, as the surface of ways people are getting value from these models keeps expanding. Ten years ago, when we started OpenAI, we thought this moment might be possible. It’s happening on the earlier side, and happening in a much more interesting and empowering way for everyone than we’d anticipated (for example, we are seeing an emerging wave of entrepreneurship that we hadn’t previously been anticipating). And at the same time, we are still so early, and there is so much for everyone to define about how these systems get deployed and used in the world. The next phase will be defined by systems that can do more — reason better, use tools better, plan over longer horizons, and take more useful actions on your behalf. And there are horizons beyond, as AI starts to accelerate science and technology development, which have the potential to truly lift up quality of life for everyone. All of this is starting to happen, in small ways and large, today, and everyone can participate. I feel this shift in my own work every day, and see a roadmap to much more useful and beneficial systems. These systems can truly benefit all of humanity.
English
410
656
5.2K
548.2K
HackerRank
HackerRank@hackerrank·
LeetCode is dead. Developers don't write code line-by-line anymore. They orchestrate AI agents working in parallel, review AI-generated code, and make architectural decisions. That's the job now. But most interview processes haven't caught up. They still test algorithm memorization instead of AI fluency, code review, and judgment. We're building assessments for next-gen hiring that mirror how developers actually work. Here's how we think about it:
English
128
203
2.4K
334.7K
Steven retweetledi
Reads with Ravi
Reads with Ravi@readswithravi·
“No matter how isolated you are and how lonely you feel, if you do your work truly and conscientiously, unknown friends will come and seek you.” — Carl Jung
English
50
1.6K
12K
322.8K
Steven retweetledi
Chamath Palihapitiya
When your working life rewards you, it’s easy to ratchet up the complexity: homes, cars, travel, possessions etc. I have found that all that complexity comes at the sake of your most fleeting asset: your time. Instead of building things, all of a sudden you’re dealing with minutiae and logistics. Instead of talking mostly to engineers, you’re talking mostly to non-engineers. The building stops…the business of managing self inflicted complexity begins. It’s worth noting that the best players in the game (Buffett, Elon) have kept their life extremely basic, almost monastic/nomadic, as success ratcheted them ever higher. I think it’s the biggest secret hiding in plain sight: When the world upgrades your status, downgrade your complexity.
English
543
1K
12K
953.6K
Steven
Steven@ascendingfn·
compilers are the bridge
English
0
0
0
2
Steven retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
Yes it's the tractable form of brain upload. There's a ton of scifi on brain uploads that requires way too exotic tech (scanning and simulating brains etc), when we're about to get a lossy and approximate version of that *a lot* sooner via LLM simulators. You can easily imagine a "brain upload" startup - you show up for a few days to carry out detailed video interviews, then they use all that data with an LLM finetuning process to "upload" you and give you an API endpoint of your simulation that you can talk to. Look at what's already possible with HeyGen as an example, but combine it with an LLM model that has deep knowledge and personality. Trippy and admittedly kind of dystopian but in principle quite possible around now.
English
203
204
3.3K
545.8K
Steven retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
Judging by my tl there is a growing gap in understanding of AI capability. The first issue I think is around recency and tier of use. I think a lot of people tried the free tier of ChatGPT somewhere last year and allowed it to inform their views on AI a little too much. This is a group of reactions laughing at various quirks of the models, hallucinations, etc. Yes I also saw the viral videos of OpenAI's Advanced Voice mode fumbling simple queries like "should I drive or walk to the carwash". The thing is that these free and old/deprecated models don't reflect the capability in the latest round of state of the art agentic models of this year, especially OpenAI Codex and Claude Code. But that brings me to the second issue. Even if people paid $200/month to use the state of the art models, a lot of the capabilities are relatively "peaky" in highly technical areas. Typical queries around search, writing, advice, etc. are *not* the domain that has made the most noticeable and dramatic strides in capability. Partly, this is due to the technical details of reinforcement learning and its use of verifiable rewards. But partly, it's also because these use cases are not sufficiently prioritized by the companies in their hillclimbing because they don't lead to as much $$$ value. The goldmines are elsewhere, and the focus comes along. So that brings me to the second group of people, who *both* 1) pay for and use the state of the art frontier agentic models (OpenAI Codex / Claude Code) and 2) do so professionally in technical domains like programming, math and research. This group of people is subject to the highest amount of "AI Psychosis" because the recent improvements in these domains as of this year have been nothing short of staggering. When you hand a computer terminal to one of these models, you can now watch them melt programming problems that you'd normally expect to take days/weeks of work. It's this second group of people that assigns a much greater gravity to the capabilities, their slope, and various cyber-related repercussions. TLDR the people in these two groups are speaking past each other. It really is simultaneously the case that OpenAI's free and I think slightly orphaned (?) "Advanced Voice Mode" will fumble the dumbest questions in your Instagram's reels and *at the same time*, OpenAI's highest-tier and paid Codex model will go off for 1 hour to coherently restructure an entire code base, or find and exploit vulnerabilities in computer systems. This part really works and has made dramatic strides because 2 properties: 1) these domains offer explicit reward functions that are verifiable meaning they are easily amenable to reinforcement learning training (e.g. unit tests passed yes or no, in contrast to writing, which is much harder to explicitly judge), but also 2) they are a lot more valuable in b2b settings, meaning that the biggest fraction of the team is focused on improving them. So here we are.
staysaasy@staysaasy

The degree to which you are awed by AI is perfectly correlated with how much you use AI to code.

English
1.1K
2.4K
19.9K
4M
Steven
Steven@ascendingfn·
@pmarca too much caffeine bro
English
0
0
0
4
Marc Andreessen 🇺🇸
I'm calling it. AGI is already here – it's just not evenly distributed yet.
English
1.6K
1.2K
13.7K
2.6M
Steven retweetledi
Nick shirley
Nick shirley@nickshirleyy·
@SenWarren Why don’t you guys eliminate fraud before asking for more money?
English
2.6K
15.4K
162.8K
1.5M
Steven
Steven@ascendingfn·
overgeneralization never works
English
0
0
0
5
Steven
Steven@ascendingfn·
@kareem_carr and yes, i prefer to be completely out of the loop, but understand the concepts being implemented, not read every line of code anymore
English
0
0
0
1
Steven
Steven@ascendingfn·
Right now, the biggest downsides are drift from code spec (arch, cleanliness, paradigm, patterns, etc. specified) and domain spec. What i do is, at the start of building a feature, i pair program with codex to converge to domain invariants which i code in a docs/domain when we agree with the shape. Then, I implement the code (assuming i have grounding specs for the repo/service such as the arch, patterns, etc.) which can be done by parts (e.g., domain files (policies), use cases, db tables first) or all at once (including networking, db read/write, etc.). Finally, i do a quick policing check which audits all uncommitted changes to see if there are drifts from arch, patterns, etc. which reduces drift by a lot. Now, for systems, i consider it mostly diffusive when compared to tradition programming. This is because the speed doesn’t let me read every single line of code, so with enough code in the repo/service, i can start sensing something might need to be streamed, idempotent, cached, async in protocols or transaction, etc.; and this works best with observability
English
1
0
0
5
Dr Kareem Carr
Dr Kareem Carr@kareem_carr·
I keep hearing that software engineers don’t write much code anymore and it’s mostly AI now. Can any software engineers confirm how true this is? Do you just drink coffee and watch Claude code all day now?
English
534
12
581
171.9K
Steven retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.
English
2.8K
6.7K
56K
19.9M
Steven retweetledi
NASA
NASA@NASA·
We see our home planet as a whole, lit up in spectacular blues and browns. A green aurora even lights up the atmosphere. That's us, together, watching as our astronauts make their journey to the Moon.
NASA tweet media
English
4.8K
66K
312.7K
76.4M
John Coogan
John Coogan@johncoogan·
TBPN has been acquired by OpenAI! The show is staying the same and we’ll continue to go live at 11am pacific every weekday. This is a full circle moment for me as I’ve worked with @sama for well over a decade. He funded my first company in 2013. Then helped us fix a serious logjam during a critical funding round a few years later. When I took my second company through YC, he was president at the time, and then when I joined Founders Fund, the first deal I saw in motion was the post-ChatGPT round in late 2022. And as we started growing TBPN last year, he was the very first lab lead to join the show. Thank you to everyone that has been a part of TBPN until now. The last year has been the most fun and rewarding part of my career and we’re excited to have more resources than ever going forward.
English
1.3K
410
8.7K
3M
Steven
Steven@ascendingfn·
@sleepingbvby b/c latinas are the prettiest and latinas like dating latinos
English
0
0
0
91
༺☆༻cin✩‧
༺☆༻cin✩‧@sleepingbvby·
those mexicans that wear skinny ripped jeans with a t-shirt & cap & some sneakers with a chain. how do you honestly pull the most prettiest girls
English
20
13
1.9K
76.6K
Steven retweetledi
Greg Brockman
Greg Brockman@gdb·
if you can imagine it, you can build it
English
985
1.2K
9K
524.1K