hari

121 posts

hari

hari

@harinj

USA Katılım Nisan 2009
231 Takip Edilen46 Takipçiler
Julian Goldie SEO
Julian Goldie SEO@JulianGoldieSEO·
China just made Hermes free forever 🚀 You can now run Hermes with Kimi K 2.6 through Ollama in one click. Just install the model, launch Hermes, select Kimi K 2.6 and suddenly you have a powerful AI agent running locally with zero API fees. This might be the easiest free Hermes setup yet.
English
86
173
1.5K
134.1K
Matt Dancho (Business Science)
🚨 RIP Prompt Engineering Enter Context Engineering 2.0 It completely reframes how we think about human-AI interactions. This is what you need to know (28 page PDF):
Matt Dancho (Business Science) tweet media
English
10
74
360
27.5K
hari retweetledi
Historic Vids
Historic Vids@historyinmemes·
In 2005, Alex Tew, a British student, created The Million Dollar Homepage, selling pixels for $1 each that linked to advertisers’ websites—and ultimately sold all one million pixels.
English
42
83
1.2K
105.6K
hari
hari@harinj·
Is your code punchy or just bulky? 🥊 GitHub Stars are vanity. Lines of Code (LOC) are a liability. The real flex? The Star-to-LOC Ratio. High Ratio = Massive impact, tiny footprint. 💎 Low Ratio = High maintenance, low buzz. 📉 Stop building monoliths. Start building "Impact Density." What’s your ratio? 👇 #BuildInPublic #Coding #GitHub
English
0
0
0
19
AJ
AJ@AjMa697292·
@jerryjliu0 Been running 120+ knowledge files with structured directories for months. Grep + filesystem beats RAG for anything under a few million tokens. Karpathy was right.
English
1
0
0
474
Jerry Liu
Jerry Liu@jerryjliu0·
This is a cool article that shows how to *actually* make filesystems + grep replace a naive RAG implementation. ̶F̶i̶l̶e̶s̶y̶s̶t̶e̶m̶s̶ ̶+̶ ̶g̶r̶e̶p̶ ̶i̶s̶ ̶a̶l̶l̶ ̶y̶o̶u̶ ̶n̶e̶e̶d̶ ̶ Database + virtual filesystem abstraction + grep is all you need
Jerry Liu tweet media
Dens Sumesh@densumesh

x.com/i/article/2039…

Español
28
73
784
163.4K
hari retweetledi
BuBBliK
BuBBliK@k1rallik·
Anthropic tried to kill 8,100 GitHub repos. Then this happened > They filed a DMCA. GitHub nuked the entire network within hours. Developers got notices for forks of Anthropic's OWN public repo - one guy's fork had zero leaked code. > Boris Cherny, head of Claude Code, had to go on X personally: "This was not intentional. Should be better now." > Meanwhile Sigrid Jin - who used 25 billion Claude Code tokens last year - woke up at 4AM and rewrote the entire thing in Python before sunrise. DMCA can't touch a clean-room rewrite. > It hit 50K stars in 2 hours. Fastest repo in GitHub history. > Today claw-code officially launched as an independent project with a formal press release. And the Rust port merged today - what started as a panic rewrite now ships release 0.1.0. > 140K stars. 102K forks. More than Anthropic's own repo. > 512,000 lines are in the wild forever. What started as Anthropic's biggest embarrassment just became their most dangerous competitor. You cannot make this up.
BuBBliK tweet mediaBuBBliK tweet mediaBuBBliK tweet media
BuBBliK@k1rallik

x.com/i/article/2038…

English
194
980
8.3K
745.7K
hari
hari@harinj·
The US just answered back. 🇺🇸 With Chinese models like DeepSeek and Qwen dominating the open-source charts lately, Google’s launch of Gemma 4 is more than just a tech update—it's a massive move to reclaim the open AI throne. 100k+ variants and counting. Is the comeback real? 📈 #AIWars #Gemma4 #Google #OpenSource
hari tweet media
English
0
0
0
55
Elon Musk
Elon Musk@elonmusk·
Stand By Me
English
36K
23.8K
332K
81.5M
hari
hari@harinj·
Finally found an AI that actually gets technical documents! 🤯 Been using @TXYZAI to cut through the noise of complex PDFs. It links every answer to the exact source paragraph, so zero hallucinations. Great for paper discovery too. If you handle dense research, you need to try this. txyz.ai #AIforResearch #AcademicChatter #PhDLife #STEM
English
0
0
0
18
Eyisha Zyer
Eyisha Zyer@eyishazyer·
This mixed reality app lets you create and ride thrilling rollercoasters in your own living room.
English
2
1
5
2.1K
Robert Scoble
Robert Scoble@Scobleizer·
All AI posters at GTC. This is not for human consumption. This video is for AI to watch. Click the grok button and talk to it about what it learned by seeing all the AI posters (highly technical) presented at @NVIDIAGTC tonight. Thanks NVIDIA for the badge and access.
English
42
50
481
330.6K
hari
hari@harinj·
@Remotion @grok how was this video created. Break it into steps.
English
1
0
1
904
Remotion
Remotion@Remotion·
What's Remotion? A way to compose your video with code! • Use React, a powerful frontend technology, to build sophisticated videos with code. • Put together building blocks - like
English
21
42
656
338.4K
hari
hari@harinj·
AI-powered motion graphics generator that transforms natural language prompts into Remotion code.#remotion
English
2
1
3
58
hari
hari@harinj·
@steipete @openclaw @grok what is chrome side panel slide mode. Can this be launched either by chrome add in or chrome mcp
English
1
0
0
150
Peter Steinberger 🦞
Peter Steinberger 🦞@steipete·
📝 Summarize 0.12.0 out now! Fast summaries from URLs, files, and media - used in @openclaw Big update for slides + media: stronger Chrome sidepanel slide mode, better YouTube/video switching, AssemblyAI+Gemini transcription, xurl for X, NVIDIA provider github.com/steipete/summa…
English
48
84
1.3K
101.9K
OpenClaw🦞
OpenClaw🦞@openclaw·
OpenClaw 2026.3.11 🦞 🏹 Hunter & 🩹Healer Alpha — free 1M context models via @OpenRouter 🧠 GPT 5.4 stops stopping mid-thought 💎 Gemini Embedding 2 for memory 💻 OpenCode Go support 🔒 Security hardening sprint We ship faster than they can clone. github.com/openclaw/openc…
English
372
375
4.2K
684.4K
Guri Singh
Guri Singh@heygurisingh·
Holy shit... Microsoft open sourced an inference framework that runs a 100B parameter LLM on a single CPU. It's called BitNet. And it does what was supposed to be impossible. No GPU. No cloud. No $10K hardware setup. Just your laptop running a 100-billion parameter model at human reading speed. Here's how it works: Every other LLM stores weights in 32-bit or 16-bit floats. BitNet uses 1.58 bits. Weights are ternary just -1, 0, or +1. That's it. No floats. No expensive matrix math. Pure integer operations your CPU was already built for. The result: - 100B model runs on a single CPU at 5-7 tokens/second - 2.37x to 6.17x faster than llama.cpp on x86 - 82% lower energy consumption on x86 CPUs - 1.37x to 5.07x speedup on ARM (your MacBook) - Memory drops by 16-32x vs full-precision models The wildest part: Accuracy barely moves. BitNet b1.58 2B4T their flagship model was trained on 4 trillion tokens and benchmarks competitively against full-precision models of the same size. The quantization isn't destroying quality. It's just removing the bloat. What this actually means: - Run AI completely offline. Your data never leaves your machine - Deploy LLMs on phones, IoT devices, edge hardware - No more cloud API bills for inference - AI in regions with no reliable internet The model supports ARM and x86. Works on your MacBook, your Linux box, your Windows machine. 27.4K GitHub stars. 2.2K forks. Built by Microsoft Research. 100% Open Source. MIT License.
English
876
2.6K
15.3K
2.2M
NotebookLM
NotebookLM@NotebookLM·
What better way to explain the launch of our new Cinematic Video Overviews than… using… a Cinematic Video Overview? (video-ception 😵‍💫)
English
97
231
2.1K
146.2K