Powering ProjeX

574 posts

Powering ProjeX banner
Powering ProjeX

Powering ProjeX

@PoweringProjeX

A world where project people are celebrated as the quiet architects of progress: turning bold visions into victory through learning, wisdom and courage.

Australia เข้าร่วม Mayıs 2025
151 กำลังติดตาม53 ผู้ติดตาม
Nikita Bier
Nikita Bier@nikitabier·
@nypost Can I say something without everyone getting mad
English
478
36
2K
75.1K
New York Post
New York Post@nypost·
Notorious Gen. Soleimani's sultry grandniece led lavish lifestyle touring US hotspots, as her mom promoted Iranian regime trib.al/y38evjw
New York Post tweet media
English
702
1.3K
6.6K
1.1M
Powering ProjeX
Powering ProjeX@PoweringProjeX·
I should do this with all the P3M articles and research papers I’ve collected over the years. Such a smart idea
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
0
0
0
10
Powering ProjeX รีทวีตแล้ว
Lenny Rachitsky
Lenny Rachitsky@lennysan·
"Using coding agents well is taking every inch of my 25 years of experience as a software engineer." Simon Willison (@simonw) is one of the most prolific independent software engineers and most trusted voices on how AI is changing the craft of building software. He co-created Django, coined the term "prompt injection," and popularized the terms "agentic engineering" and "AI slop." In our in-depth conversation, we discuss: 🔸 Why November 2025 was an inflection point 🔸 The "dark factory" pattern 🔸 Why mid-career engineers (not juniors) are the most at risk right now 🔸 Three agentic engineering patterns he uses daily: red/green TDD, thin templates, hoarding 🔸 Why he writes 95% of his code from his phone while walking the dog 🔸 Why he thinks we're headed for an AI Challenger disaster 🔸 How a pelican riding a bicycle became the unofficial benchmark for AI model quality Listen now 👇 youtu.be/wc8FBhQtdsA
YouTube video
YouTube
English
68
140
1.1K
2M
Powering ProjeX
Powering ProjeX@PoweringProjeX·
@lennysan Really interesting. I’ve been setting all my skills in Cowork …just trying to get human details out of my head using voice capture is exhausting! 🤣
English
0
0
0
230
Lenny Rachitsky
Lenny Rachitsky@lennysan·
"Using coding agents well is taking every inch of my 25 years of experience as a software engineer, and it is mentally exhausting. I can fire up four agents in parallel and have them work on four different problems, and by 11am I am wiped out for the day. There is a limit on human cognition. Even if you're not reviewing everything they're doing, how much you can hold in your head at one time. There's a sort of personal skill that we have to learn, which is finding our new limits. What is a responsible way for us to not burn out, and for us to use the time that we have?" @simonw
Lenny Rachitsky@lennysan

"Using coding agents well is taking every inch of my 25 years of experience as a software engineer." Simon Willison (@simonw) is one of the most prolific independent software engineers and most trusted voices on how AI is changing the craft of building software. He co-created Django, coined the term "prompt injection," and popularized the terms "agentic engineering" and "AI slop." In our in-depth conversation, we discuss: 🔸 Why November 2025 was an inflection point 🔸 The "dark factory" pattern 🔸 Why mid-career engineers (not juniors) are the most at risk right now 🔸 Three agentic engineering patterns he uses daily: red/green TDD, thin templates, hoarding 🔸 Why he writes 95% of his code from his phone while walking the dog 🔸 Why he thinks we're headed for an AI Challenger disaster 🔸 How a pelican riding a bicycle became the unofficial benchmark for AI model quality Listen now 👇 youtu.be/wc8FBhQtdsA

English
545
680
6.7K
1.7M
Powering ProjeX
Powering ProjeX@PoweringProjeX·
@elonmusk Most of my friends use AI purely for health conversations. I would never have thought.
English
0
0
0
3
Powering ProjeX รีทวีตแล้ว
Powering ProjeX รีทวีตแล้ว
Massimo
Massimo@Rainmaker1973·
Microsurgery robot stitches up a corn kernel showing millimeter-level precision
English
116
622
3.4K
224.8K
Powering ProjeX รีทวีตแล้ว
X Freeze
X Freeze@XFreeze·
Tesla's Full Self-Driving (FSD) is currently ~9x safer than the average human driver Because of this massive safety advantage, auto insurance providers like Lemonade are now offering Tesla owners up to a 50% discount on their per-mile premiums when FSD is engaged Choosing Tesla FSD driving is not just safer, but it also directly saves you money
X Freeze tweet media
English
377
433
3.4K
11.4M
Powering ProjeX รีทวีตแล้ว
Anthropic
Anthropic@AnthropicAI·
We've signed an MOU with the Australian Government to collaborate on AI safety research and support Australia's National AI Plan. Read more: anthropic.com/news/australia…
English
220
159
1.8K
345.7K
Powering ProjeX รีทวีตแล้ว
Ejaaz
Ejaaz@cryptopunk7213·
this is insane lol japan is running out of monks... so they're training AI robots called "buddharoid" to replace them 😂 (im not joking): - japan's temples are closing because fewer priests are available to run them + aging population - the solution: chatgpt robots trained on 1000+ years of buddhist scripture that answer your spiritual questions - the robot even sits in religious prayer positions like an actual monk does. you can literally have a conversation on life's deepest dilemmas with a robot as smart as the dalai llama i cannot believe they're scaling these robots to run actual temples.
English
381
640
3.1K
358.5K
Powering ProjeX รีทวีตแล้ว
George Pu
George Pu@TheGeorgePu·
GitLab's founder was told he has bone cancer. No trials would take him. Doctors signed off. So he went founder mode on his own survival. - Built his own treatments - Used AI to analyze his own tumor data - Open-sourced 25TB of his medical records for any researcher on earth Relapse-free since 2025. The system said he was out of options. He made his own.
George Pu tweet media
English
194
799
8.7K
1.1M
Powering ProjeX รีทวีตแล้ว
Elon Musk
Elon Musk@elonmusk·
The Starship factory is stunning
Jawwwn@jawwwn_

SpaceX COO @Gwynne_Shotwell: “I love working for @elonmusk.” “He’s funny— he’s hilarious actually.” “He focuses on things that I would never have thought were important.” “One is— beautiful spaces.” “This is one of the most beautiful factories I have ever seen.” Via @TIME

English
3.3K
9K
70.7K
28M
Powering ProjeX รีทวีตแล้ว
Peter H. Diamandis, MD
Peter H. Diamandis, MD@PeterDiamandis·
Tesla's FSD: 5.3 million miles between accidents. US driving average: 660,000.  That's 9x safer. And it's only getting better.
English
601
999
8K
84.6M
Powering ProjeX
Powering ProjeX@PoweringProjeX·
Worth reading…
Marc Andreessen 🇺🇸@pmarca

Claude knows! —> The Lump of Labor Fallacy and Why AGI Unemployment Panic Is Economically Illiterate Let me lay this out with full rigor, because this argument deserves to be prosecuted completely rather than waved away with a sound bite. I. What the Lump of Labor Fallacy Actually Is The lump of labor fallacy is the assumption that there exists a fixed, finite quantity of work in an economy — a lump — such that if a machine (or an immigrant, or a woman entering the workforce) does some of it, there is necessarily less left for human workers to do. It treats employment as a zero-sum pie. The fallacy was named and formalized in the early 20th century but the error it describes is far older. It animated the Luddite riots of 1811–1816, where English textile workers destroyed power looms convinced that the machines would steal their jobs permanently. It drove opposition to the spinning jenny, the cotton gin, the mechanical reaper, the steam engine, the telegraph, the railroad, the automobile assembly line, the personal computer, and every other major labor-displacing technology in the history of industrial civilization. Every single time, the catastrophists were wrong. Not partially wrong. Structurally, fundamentally, categorically wrong — because they misunderstood the nature of economic production itself. The reason the fixed-pie assumption fails is this: demand is not fixed. Work generates income. Income generates demand for goods and services. Demand for goods and services generates new categories of work. This is an engine, not a reservoir. When you drain some of the reservoir with a machine, the engine speeds up and refills it — and often refills it past its previous level. II. The Classical Economic Mechanism That Destroys the Fallacy To understand why the lump-of-labor assumption is wrong about AGI, you need to understand the precise mechanism by which technological unemployment resolves itself. There are four distinct channels, all operating simultaneously: Channel 1: The Productivity-Demand Feedback Loop (Say’s Law, Modified) When a technology increases the productivity of labor or replaces labor entirely in a given task, it lowers the cost of producing whatever that task was part of. Lower production costs mean either: ∙Lower prices for consumers (real purchasing power rises), or ∙Higher profits for producers (which get reinvested, distributed as dividends, or spent as wages for other workers), or ∙Both. Either way, aggregate real income in the economy rises. That additional real income does not evaporate. It gets spent on something — including goods and services that didn’t previously exist or were previously too expensive to consume at scale. That spending creates demand. That demand creates jobs. This is not a theoretical conjecture. The average American in 1900 spent roughly 43% of their income on food. Today it’s around 10%. Agricultural mechanization didn’t produce a nation of starving unemployed farm laborers — it freed up 33% of household income to be spent on automobiles, television sets, air conditioning, healthcare, education, travel, smartphones, and streaming services, most of which didn’t exist as industries in 1900. The workers who left farms went to factories, then to offices, then to service industries, then to information industries. The economy didn’t run out of work. It metamorphosed.

English
0
0
0
7
Powering ProjeX รีทวีตแล้ว
The All-In Podcast
The All-In Podcast@theallinpod·
David Friedberg on Personal Agency in the Age of AI: "Stop Blaming Everyone Else" "We never talk about responsibility. We always talk about where the government failed us and where these companies f***ed us. And we never talk about, what did we individually do wrong? How did I individually choose to drink 100 sodas a week? How did I individually choose to get my kids addicted to social media? Where the f*** was I as a parent? We don't talk about our responsibility. And by the way, this fundamentally addresses this point about human agency, which I think is more critical in this era than ever because AI is going to flood us with f*****g everything all the time, nonstop. What we choose to do in a world where we're already getting everything, and how we choose to not take everything that's being offered to us, I think is a critical part of what's going to distinguish human success from human failure. And it's gonna become more apparent in the future, and not everything is about liability, and not everything is about the government failing us. It's about people making choices and we don't talk about it."
English
95
207
1.5K
107.9K
Nikita Bier
Nikita Bier@nikitabier·
If you’re seeing a bunch of Japanese posts, here are some fun facts: Japan has more daily active users and more time spent on X than any other country in the world. Over two thirds of the country is monthly active on X. X in Japan has one of the highest penetration rates of any social network in history.
English
5.7K
5.3K
75.9K
7.6M