Michal Krcmar

20.1K posts

Michal Krcmar banner
Michal Krcmar

Michal Krcmar

@Krcmic

Entrepreneur: https://t.co/Ii45eqhaee and some other projects - check: https://t.co/WGALkfWAYZ

Prague, Czech Republic Katılım Mayıs 2011
1.5K Takip Edilen1.6K Takipçiler
Michal Krcmar retweetledi
Krishna Agrawal
Krishna Agrawal@Krishnasagrawal·
If you create content, this is a no-brainer. You’re not just saving time, you’re cutting production costs massively. And with plans starting at ~$2.5 (sale 👀), it’s probably the most accessible pro-level voice tool right now. Try it: minimax.io/audio #MiniMax #MiniMaxAudio #voiceclone
English
0
1
12
205
Michal Krcmar retweetledi
Sophia Yang, Ph.D.
Sophia Yang, Ph.D.@sophiamyang·
How to clone your voice in @MistralAI Studio? Check out my video for a demo and tips:
English
7
10
105
10.2K
Michal Krcmar retweetledi
Sajeel Purewal 🇨🇦 🇵🇰
Build Robots Build Drones Build Hexapods Build Glasses Build Radios Build Clocks Build Rovers Build Wearables Build Rockets Build Exoskeletons Build Sensors Build it all blueprint.am
English
43
412
3.3K
194.8K
Michal Krcmar retweetledi
Jiri Hynek
Jiri Hynek@JiriHynek·
Areál u Karlových Varů postavený bez povolení. Už v roce 2001 ho začal stavět Nikolaj Stěpanov bývalý voják Ruské armády. Dnes ho vlastní vdova po oficírovi KGB. @odkryto.cz herohero.co/ruskysvet
Čeština
17
149
601
11.8K
Michal Krcmar retweetledi
Luděk Niedermayer
Luděk Niedermayer@LudekNie·
Pokud někdo označí @HnutiDUHA bez jakékoli větší nadsázky za teroristickou organizaci, pak je buď opravdu hodně, ale hodně pomatený, anebo zcela vědomě roztáčí naplno kola nenávisti proti těm, kteří mají jiné názory než on sám. Doufám, že Hnutí Duha předloží tento výrok soudu, aby zazněl jasný výrok o tom, kde leží hranice komunikace politiků, která vyvolává nenávist.
Luděk Niedermayer tweet media
Čeština
98
112
972
19.2K
Michal Krcmar retweetledi
Pavel Hertl
Pavel Hertl@PavelHertl·
Dám ti to tedy obsáhle vedle sebe: Fiala - nafta za 46 Babiš - nafta za 51 Fiala - kauza Dozimetr u soudu Babiš - před soudem se schoval za poslaneckou imunitu Fiala - kampelička a nepřiznaný uložený 1 milion bez škody pro stát Babiš - dotační podvod se škodou 8 miliard (to je 8000x víc)
Čeština
33
301
1.9K
25.7K
Michal Krcmar retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
Wow, this tweet went very viral! I wanted share a possibly slightly improved version of the tweet in an "idea file". The idea of the idea file is that in this era of LLM agents, there is less of a point/need of sharing the specific code/app, you just share the idea, then the other person's agent customizes & builds it for your specific needs. So here's the idea in a gist format: gist.github.com/karpathy/442a6… You can give this to your agent and it can build you your own LLM wiki and guide you on how to use it etc. It's intentionally kept a little bit abstract/vague because there are so many directions to take this in. And ofc, people can adjust the idea or contribute their own in the Discussion which is cool.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
744
2.1K
20.7K
4.2M
Michal Krcmar retweetledi
Rimsha Bhardwaj
Rimsha Bhardwaj@heyrimsha·
🚨BREAKING: Block (Jack Dorsey's company) just open-sourced a local AI agent that goes way beyond code suggestions. It's called Goose and it installs, executes, edits, and tests with any LLM fully on your machine. 100% Opensource.
Rimsha Bhardwaj tweet media
English
76
437
3.2K
199.7K
Michal Krcmar retweetledi
Wildminder
Wildminder@wildmindai·
Netflix dropped some useful stuff. VOID -video object and interaction deletion. - removes objects while realistically simulating physical consequences; - beats Runway/ProPainter; - CogVideoX-5B + SAM 2; looks good, no smudges/artifacts void-model.github.io
English
54
345
3K
335.3K
Michal Krcmar retweetledi
0xMarioNawfal
0xMarioNawfal@RoundtableSpace·
MOST COMPLETE CLAUDE CODE SETUP OPEN SOURCED - 27 agents, 64 skills, 33 commands + built-in AgentShield with 1,282 security tests - Handles planning, code review, fixes, TDD, token optimization & more - Works on Cursor, OpenCode, Codex CLI — one repo replaces weeks of setup, 100% free/open-source Repo: github.com/affaan-m/every…
0xMarioNawfal tweet media
English
82
490
3.6K
370.6K
Michal Krcmar retweetledi
Alvaro Cintas
Alvaro Cintas@dr_cintas·
Google Stitch has released 𝗗𝗘𝗦𝗜𝗚𝗡.𝗺𝗱 🤯 One markdown file that teaches your AI coding agent your entire design system. → No Figma exports → No JSON schemas → Nothing to configure The part that saves the most time: A free collection of 40+ pre-built files already exists, extracted from real products. Stripe, Vercel, Linear, Notion, Lovable, Claude, ElevenLabs, Cursor, Warp, Zapier, and more. Drop it in your project root. Claude Code, Cursor, Gemini CLI, and GitHub Copilot all read it natively. 100% Free and Open-Source.
Alvaro Cintas tweet media
English
72
459
3.8K
260.6K
Michal Krcmar retweetledi
scaredofboobs🪲
scaredofboobs🪲@boobs_scary·
I hacked the engine out of comfyUI's image-to-3d to utilize llama.cpp and now it can run quantized models and this workflow is down to 4g of VRAM. You're paying $400 a month in meshy credits for something you can run on a GPU from 2018.
scaredofboobs🪲@boobs_scary

I locked myself in a room for 48 hours and didn't let myself leave until I had 3D meshes that rivaled meshy running on 12gb of VRAM. LET HIM COOK.😤

English
50
127
1.9K
73.8K
Michal Krcmar retweetledi
FriendlyCosmonaut
FriendlyCosmonaut@friendly_cosmo·
I HIGHLY recommend this tool; you can import any 3D model - human, avian, quadruped, and now spider and snake!? - easily rig & animate it, and export the animations you want. Similar to Adobe's Mixamo but entirely Adobe-free... please support this creator!
Scott Petrovic@scottpetrovic

🎉 Mesh2Motion Release 10 Out 🎉 - Spider and Snake rigs with 16 animations - Simple bone positioning to snap to mesh instead using a 3d transform widget - Many other small improvements Check out the full release 10 notes with all changes: mesh2motion.org/news Try it out and any feedback welcome: app.mesh2motion.org

English
5
168
1.7K
62.4K
Michal Krcmar retweetledi
Skywork
Skywork@Skywork_ai·
Skywork Matrix-Game 3.0 is here! FULLY OPEN SOURCE! Real-Time and Streaming Interactive World Model with Long-Horizon Memory - Fully open source: code, model, and technical report - 720p @ 40FPS with a 5B model - Minute-long memory consistency - Trained on Unreal Engine + AAA games + real-world data - Scales up to 28B MoE for quality, dynamics, and generalization Homepage 👉 matrix-game-v3.github.io Code 👉 github.com/SkyworkAI/Matr… Model 👉 huggingface.co/Skywork/Matrix… Tech report 👉 github.com/SkyworkAI/Matr… Create. Explore. Play. With Matrix-Game 3.0
English
21
94
655
47.9K
Michal Krcmar retweetledi
Nicolas Zullo
Nicolas Zullo@NicolasZu·
For all gamedevs and complex app designers out there. I think I found a really clever way to scaling production. 1. Mirror your game systems in a spreadsheet (ask your AI to do it) 2. One tab per system to scale. For instance I have one tab for "buildings", one for "missions", "items", "maps", "enemies", "recipes", etc. 3. Use an MCP, CLI or direct connector to allow Codex or Claude Code to read and write 4. Then I built specific skills like /add-building, /add-mission, /add-map The workflow becomes 1. I design my game in the spreadsheet, like a new "Nuclear Reactor" building: I type everything I want: size, goal, 3D prompt, etc. 2. I run "/add-building nuclear reactor" 3. The skill uses what's in the spreadsheet as context 4. Review and done. Makes it super easy to have scalable systems and the spreadsheet is so nice to have an overview and design your game. I hope it helps
Nicolas Zullo tweet media
English
26
26
522
39K
Michal Krcmar retweetledi
Nav Toor
Nav Toor@heynavtoor·
🚨 Professional 3D artists are going to hate this. Tencent open sourced an AI that turns a single photo into a production-ready 3D model with full PBR textures. The kind that ships in AAA games. The kind that renders in Unreal Engine without post-processing. The kind that takes a senior 3D artist 3 to 5 days to build by hand. It's called Hunyuan3D 2.1. The first fully open source production-ready 3D generation model. Full weights. Full training code. Free. Upload one photo of anything. Get a complete 3D model with Albedo, Normal, Roughness, and Metallic maps. Ready for Blender. Ready for Unity. Ready for Unreal Engine. No 3D modeling skills required. Not a toy. Not a research demo. Production-ready assets that work in real game engines and film pipelines. No Maya license. No Substance Painter subscription. No $125/hour artist retainer. Here's what it does: → Image to 3D. Upload one photo, get a full 3D mesh in seconds. → Text to 3D. Describe any object in words, get a 3D model. → PBR texture synthesis. Metallic reflections, subsurface scattering, physically accurate materials. → Outperforms every open source AND closed source 3D generation model in benchmarks. → Full model weights released. Fine-tune it for your own use case. → Full training code released. First time ever for a model at this level. → Runs on 10GB VRAM for shape generation. → Supports macOS, Windows, and Linux. → Blender addon included. ComfyUI integration available. Here's the wildest part: This solves the "Janus problem." Where AI-generated 3D models have a different face on every side. Hunyuan3D 2.1 generates a clean mesh first, then projects PBR textures from multiple angles simultaneously. Coherent geometry. Seamless textures. Every angle looks correct. 3D artists charge $500 to $5,000 per model. Game studios pay $50,000 to $200,000 per year for 3D asset pipelines. Film VFX houses spend millions on modeling teams. Indie developers can't afford professional 3D assets at all. This generates them from a photo. On your GPU. For free. 3.1K GitHub stars. 447 forks. Built by Tencent. Published on arXiv. Models on Hugging Face. 100% Open Source. Apache 2.0 License.
Nav Toor tweet media
English
70
133
1K
73.7K
Michal Krcmar retweetledi
Alex Finn
Alex Finn@AlexFinn·
Do you even understand what this means? An open source model just released that is: • Outperforms models 20x its size • Can run on a base model Mac Mini • Is AMERICAN 🇺🇸 If you have a base model Mac Mini you can have unlimited super intelligence on your desk. For free. Sonnet 4.5 was released 5 months ago In 5 months that level of intelligence went from frontier to free on your desk And not only that, can run on any basically any computer out there If you have even a remotely modern computer, do the following immediately: 1. Download LM Studio 2. Go to your OpenClaw and ask which of these new Gemma 4 models is best for your hardware 3. Have it walk you through downloading and loading it 4. Build apps with it knowing you are using your own personal, private super intelligence on your desk The people denying this is the future are so beyond lost.
Google DeepMind@GoogleDeepMind

Meet Gemma 4: our new family of open models you can run on your own hardware. Built for advanced reasoning and agentic workflows, we’re releasing them under an Apache 2.0 license. Here’s what’s new 🧵

English
286
561
7.1K
1.3M
Michal Krcmar retweetledi
Mayank Vora
Mayank Vora@aiwithmayank·
Say goodbye to video editors. This open-source tool turns a news headline into a published YouTube Short in one command. It's called YouTube Shorts Pipeline and it chains Claude, Gemini Imagen, ElevenLabs, and Whisper together into one pipeline. The cost breakdown is brutal: Claude script: $0.02 Gemini visuals: $0.03 ElevenLabs voice: $0.05 Total: $0.10 per video Type a topic. Get a live YouTube link. 3–5 minutes. Supports multiple languages. Custom voice IDs. Dry-run mode to preview before producing. Manual script override before rendering. Everything stored locally. No cloud dependency. 100% Open Source. MIT License.
Mayank Vora tweet media
English
50
187
1.6K
162.2K
Michal Krcmar retweetledi
Prasenjit
Prasenjit@Star_Knight12·
This project is insane and most devs still don't know about it. A full Chromium browser that runs entirely in your terminal. → supports WebGL, WebGPU, audio, video → starts in less than a second → runs at 60 FPS → idles at 0% CPU → works through SSH → no window server needed → you can literally watch YouTube in your terminal it's called Carbonyl, 15.9K stars on GitHub built with Rust btw. 100% open source. github.com/fathyb/carbonyl
English
32
157
967
80.5K
Michal Krcmar retweetledi
Alvaro Cintas
Alvaro Cintas@dr_cintas·
You can now generate AI videos, images, speech and sound effects directly from your terminal. PixVerse CLI just launched and in one command you get: - Sora 2, Veo 3.1, Nano Banana Pro - Automated workflows & scripts - Install in seconds. Zero web interface Full step-by-step ↓
English
10
21
138
15.1K