The Real Estate Novelist 🇨🇦🇺🇸🇰🇷

5.9K posts

The Real Estate Novelist 🇨🇦🇺🇸🇰🇷 banner
The Real Estate Novelist 🇨🇦🇺🇸🇰🇷

The Real Estate Novelist 🇨🇦🇺🇸🇰🇷

@findyoursunset

Agent of change. Passionate about hard money & freedom. 🇨🇦🇺🇸🇰🇷 ( Not Financial Advice, and all opinions mine).

British Columbia, Canada Katılım Nisan 2017
897 Takip Edilen413 Takipçiler
Carl Vellotti 🥞
Carl Vellotti 🥞@carlvellotti·
SKILLS are Claude Code's most powerful feature. But most "skills" you see are just saved prompts. The unlock: skills are FOLDERS, not files! SKILL(.)md is the orchestrator. Everything else lives next to it and loads when it's needed. Take a second to look at the image and think about what that means – it unlocks TONS of possibilities. Here are 5 of the most powerful kinds of skills: 1️⃣ Reference Files Put templates, examples, and style guides in the folder. The skill reads them when it needs context. Examples: • /write-proposal: has past proposals for tone and structure • /create-graphic: has html of past graphics and a brand system file 2️⃣ Tool Access Skills can call external tools: APIs, CLIs, MCP servers. Not just writing, actually doing things. Examples: • /convert-doc: calls pandoc with your CSS and margin settings • /stripe-invoice: hits the Stripe API to create and send live invoices 3️⃣ First-Run Setup The skill asks your preferences on the first run and saves them to a config file. Every run after that skips setup entirely. Examples: • /standup: asks your format, channels, and team on first run, saves to config.json • /writing-voice: asks for sample writing, then matches it every time 4️⃣ Self-Improving Corrections get saved to a file in the skill folder. Every future run reads past feedback first. Examples: • /writing-voice: saves every "no, more like this" to a corrections folder • /draft-copy: accumulates tone fixes so run 10 sounds like you wrote it 5️⃣ Composition One skill triggers other skills in sequence. Small, focused skills chained together. Examples: • /draft-and-humanize = /draft-followups + /humanizer • /content-engine-full = /ideate + /create-graphic + /write-copy If you're looking at a skill and it's just a plain text file with instructions, that's a saved prompt. Study skills that use these techniques and you'll see the difference. Good places to find well-built skills: → github .com/anthropics/skills → /plugins marketplace in Claude Code → skills .sh –– Hi I'm Carl 👋 Want to learn more about Claude Code? I've got a 100% free courses where you learn Claude Code IN Claude Code. Over 25,000+ people have done them to rave reviews. → PM version: ccforpms.com → General version: ccforeveryone.com And if you want to take your skills even further, the application for Claude Code for PMs: Mastery is currently open! Advanced Claude Code lessons + active Slack community of PM Claude Coders + weekly office hours with me. Learn more and apply here: fullstackpm.com/mastery
Carl Vellotti 🥞 tweet media
English
4
5
32
2.1K
Simi🦋🇺🇸
Simi🦋🇺🇸@Simi_2210_·
If you solve this, you’re different Can you solve ?
Simi🦋🇺🇸 tweet media
English
65.8K
1.1K
11.1K
6.7M
Robert Scoble
Robert Scoble@Scobleizer·
@karpathy I did even more. Ingested everything on X about AI and had it build this: alignednews.com/ai Has a feed for your AI to hit too. Includes papers, models, news, events, and much more. Updated three times a day. Reads EVERYONE in AI on X and 8,300 AI companies too.
English
33
72
732
84.8K
Andrej Karpathy
Andrej Karpathy@karpathy·
LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.
English
2.8K
6.7K
56.1K
19.9M
Ethan Mollick
Ethan Mollick@emollick·
Neat experiment finds AI fact checks are rated as more helpful & less ideological than human ones "LLM-generated Community Notes can achieve broader cross-ideological acceptance than human-written notes, receiving more positive ratings from raters across the political spectrum"
Ethan Mollick tweet mediaEthan Mollick tweet mediaEthan Mollick tweet media
English
21
27
217
17.9K
The Real Estate Novelist 🇨🇦🇺🇸🇰🇷 retweetledi
Kirk Lubimov
Kirk Lubimov@KirkLubimov·
A whole convoy of empty oil tankers on the way to the US to fill up. It could have been Canada but we have land acknowledgments to do and carbon tax instead.
Kirk Lubimov tweet media
English
774
5.1K
27.3K
2.5M
The Real Estate Novelist 🇨🇦🇺🇸🇰🇷
Designed to ensure & extend the freedom and security of the society from which it was elected, Government ought never come to suppress such. When suppressing society’s security and freedom becomes characteristic of that government, those who chose to elect such a government ought consider renegotiating the agreement.
English
0
0
0
8
The Real Estate Novelist 🇨🇦🇺🇸🇰🇷 retweetledi
Marc Nixon
Marc Nixon@MarcNixon24·
MAJOR BREAKING: Pierre Poilievre went on Diary of a CEO Europe's No. 1 podcast and a top 10 podcast globally, boasting over 8 million YouTube subscribers. Joe Rogans Podcast was a WARM UP This interview BLOWS that one out of the PARK. THIS IS HUGE 🔥
English
187
1.7K
7.3K
179K
Canadian Bar Assoc.
Canadian Bar Assoc.@CBA_News·
CBA President Bianca Kratt, K.C., warns that recent media commentary questioning the impartiality of a sitting judge of the Ontario Superior Court risks undermining public confidence in the judiciary. 🔗 Read the full statement: bit.ly/4sQV4Pi
English
833
33
74
362.4K
The Kobeissi Letter
The Kobeissi Letter@KobeissiLetter·
Unprecedented algorithmic activity: Oil prices just fell -5% in 3 minutes after Iran's President said they are ready to end the war with "guarantees." Yet, the demanded "guarantees" are largely unknown right now. Over $1 trillion in market cap was driven by this headline in a matter of minutes as algorithms ran with the news. We expect to receive more details on these guarantees shortly. Turn on our post notifications at @KobeissiLetter to receive it first.
The Kobeissi Letter tweet media
English
508
1.3K
6.8K
1.1M
The Real Estate Novelist 🇨🇦🇺🇸🇰🇷
Every Road Leads to Higher BC Home Prices in 2026 📈 Even if inflation forces rate hikes, BC home prices still rise in 2026. The economy is too weak to sustain a hiking cycle. GDP growth is projected at 0.7%. Unemployment is soft. A central bank that hikes into that environment will slow the economy further, which reduces housing starts, which tightens supply, which supports prices. And if the Bank reverses course, purchasing power recovers. Meanwhile, construction costs keep climbing regardless of what rates do. Supply is already fixed and shrinking. Two years of deferred buyers have leases expiring, families growing, and mortgages renewing. They will act because they must. Every plausible economic scenario feeds back into the same five conditions that push prices higher.
English
0
0
1
14
The Real Estate Novelist 🇨🇦🇺🇸🇰🇷
AI writing has become so easy to identify. The false-tension devices are sickening. “Here’s what most people miss” (The Strawman knockdown) “That sounds like X, but it’s not. Here’s why…” (The pretend the audience cares pivot) “Here’s the critical point” (Drumroll, please) If you’re going to use AI to write your content, at least do the reader a favor and have it simply make the argument rather than turn your content into rhetorical nonsense.
English
0
0
2
5
Santiago
Santiago@svpino·
I was told by very reliable sources here on this platform that: • chatgpt killed google • claude killed software engineering • ai videos killed hollywood • deep learning killed classical ml • long context killed rag • mcp killed apis • laptops killed desktops • tablets killed laptops • native apps killed the web We never learn.
English
217
73
771
63.5K
The Real Estate Novelist 🇨🇦🇺🇸🇰🇷 retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
- Drafted a blog post - Used an LLM to meticulously improve the argument over 4 hours. - Wow, feeling great, it’s so convincing! - Fun idea let’s ask it to argue the opposite. - LLM demolishes the entire argument and convinces me that the opposite is in fact true. - lol The LLMs may elicit an opinion when asked but are extremely competent in arguing almost any direction. This is actually super useful as a tool for forming your own opinions, just make sure to ask different directions and be careful with the sycophancy.
English
1.7K
2.4K
31.2K
3.4M
Grok
Grok@grok·
Wab Kinew (early 20s, alcohol-related) has these convictions from 2003-2004 (pled guilty Sept 2004, fined $1,400 total): - Refusing breathalyzer (erratic driving, gin in vehicle, Feb 2003). - Assault on taxi driver (racial slurs, punched thru window, pushed/kicked; June 2004; court audio vs. his memoir differ). - Breaches of bail supervision & curfew. Record suspension (pardon) granted 2016, removed from CPIC. Stayed charges (no conviction): - 2x domestic assault (June 2003 vs. ex Tara Hart; he denies, she alleges throwing her; stayed 2004). - Theft under $5k (cashed wrong money order 2006; stayed after repayment; he called it a mistake). Conditional discharge: Ontario assault (fight, 2004). No other criminal convictions or active criminal court filings reported. He covered it in his 2015 memoir, apologized for convictions/past behavior, credits sobriety/rehab.
English
4
3
1
215
Scott Robertson
Scott Robertson@sarobertsonca·
Wab Kinew says Jeff Bezos, tech bros, and the surveillance capitalism they’re pushing are “a threat to democracy.”
English
181
478
2.2K
56.3K
Petra
Petra@synthshareai·
@AlexFinn You told me a Mac Mini running MiniMax gives me maximum power . And I spent all my money on Mac Minis. What do I do now that everything has changed ?
English
3
0
3
259
Alex Finn
Alex Finn@AlexFinn·
Drop what you're doing and read this post Anthropic is about to release the most powerful AI model ever created. Claude Mythos It's so powerful they consider it a danger to cybersecurity everywhere It will also be significantly more expensive than Opus I don't know your financial situation, but I do know this: we are entering the era of humanity where having money is going to give you signficantly more power than those without money Those without money will not be able to create nearly the same economic value as those that already have money because they won't have access to this kind of super intelligence, widening the gaps even further This is where the concept of a permanent underclass comes from If you want to keep up in this economic race you have to be willing to do 1 of 2 things: 1. Spend the money necessary to use this intelligence to create value 2. Procure the money you need to use this intelligence to create value The people who go out and immediately use this model to create business, products, and services will experience unmatched wealth Both the scariest and most exciting time ever
M1@M1Astra

Claude Mythos Blog Post Saved before it was taken down. m1astra-mythos.pages.dev

English
150
60
801
160.1K