iamrobotbear (bk)

27.7K posts

iamrobotbear (bk) banner
iamrobotbear (bk)

iamrobotbear (bk)

@iamrobotbear

Product Manager & AI Engineer working on Gen AI & ML. Opinions are my own, not my employer's. RT !=endorsement

Seattle, WA Katılım Kasım 2021
7K Takip Edilen4.1K Takipçiler
iamrobotbear (bk)
iamrobotbear (bk)@iamrobotbear·
@trq212 Overall I really like it. I'd really love to see something similar to @FactoryAI's codebase scoring and something that helps generate a plan to improve CC's ability to work effectively in a codebase. IE: If your codebase doesn't have tests, is poorly organized, etc it should assign a score and then offer to create a branch / worktree to improve CC's ability to work in the codebase and then generate a plan that the user approves that then writes automated test coverage, etc. BTW - I really need some feedback on something I emailed you and @AnthropicAI support on to ensure we're in compliance, can you please help??
English
1
0
1
69
iamrobotbear (bk) retweetledi
Jerry Liu
Jerry Liu@jerryjliu0·
We’ve created an agents skill that gives all of your agents the power to understand the most complex PDFs - with dense tables, unlabeled charts, messy handwriting and more. Our LlamaParse agents skill can be installed in one-line thanks to @vercel’s skills utility. LlamaParse orchestrates VLMs to deliver best-in-class accuracy over 40+ document types. The skills file allows agents to invoke it when needed to translate complex PDFs into agent-readable plaintext markdown. `npx skills add run-llama/llamaparse-agent-skills --skill llamaparse` If you prefer something fast, free, and local, you can similarly install liteparse as a skill. Come check it out: developers.llamaindex.ai/python/cloud/l… Sign up to LlamaParse: cloud.llamaindex.ai/?utm_source=xj…
LlamaIndex 🦙@llama_index

LlamaParse now has an official Agent Skill you can use across 40+ agents. With built-in instructions for parsing complex documents, including different formats, tables, charts, and images, your agents gain access to deeper document understanding, not just raw text extraction. 👇 Watch the demo 📖 Read the docs: developers.llamaindex.ai/python/cloud/l… 🚀 Get started with LlamaCloud: cloud.llamaindex.ai/signup?utm_sou…

English
14
39
292
31.5K
iamrobotbear (bk) retweetledi
ℏεsam
ℏεsam@Hesamation·
your .claude folder is what separates you from every other developer under the sun who uses Claude. it’s your taste, rules, skills, and workflows. this article is the most complete and simple explanation of all its components.
ℏεsam tweet media
Akshay 🚀@akshay_pachaar

x.com/i/article/2034…

English
27
103
920
149.2K
iamrobotbear (bk) retweetledi
Noah Zweben
Noah Zweben@noahzweben·
You can now schedule recurring cloud-based tasks on Claude Code. Set a repo (or repos), a schedule, and a prompt. Claude runs it via cloud infra on your schedule, so you don’t need to keep Claude Code running on your local machine.
English
279
554
7.4K
1.9M
iamrobotbear (bk) retweetledi
Sai Rajeswar
Sai Rajeswar@RajeswarSai·
🔥 𝗘𝗻𝘁𝗲𝗿𝗽𝗿𝗶𝘀𝗲𝗢𝗽𝘀-𝗚𝘆𝗺 𝗶𝘀 𝘁𝗮𝗸𝗶𝗻𝗴 𝗼𝗳𝗳 𝗵𝘂𝗴𝗲: 2K downloads in 3 days (trending #6 dataset + #3 paper of the day) 🏆. So we re-ran the leaderboard on the 𝗹𝗮𝘁𝗲𝘀𝘁 𝗳𝗿𝗼𝗻𝘁𝗶𝗲𝗿 𝗰𝗹𝗼𝘀𝗲𝗱 𝗺𝗼𝗱𝗲𝗹𝘀… and the results were promising. ✅ Claude versions show a meaningful jump in reliability on enterprise tasks. ✅ Gemini 3.1 Pro is catching up fast, now much closer to Sonnet 4.6 than earlier releases. And yet, the bigger takeaway is still the same: - Big room for improvement on enterprise-grade agentic tasks. - These workflows punish "seemingly correct." One wrong default, one policy miss, one unintended side effect.. and the task fails. 📢 𝗖𝗮𝗹𝗹𝗼𝘂𝘁 (especially if you’re working on agents): As we prepare our next NeurIPS/COLM submissions, try your agents on EnterpriseOps-Gym and see how they hold up on realistic, policy-constrained, long-horizon tasks. 🌐 Website: enterpriseops-gym.github.io 🤗 Dataset: huggingface.co/datasets/Servi… @ServiceNowRSRCH , @sagardavasam , @turingcom , @turingcomdev , @Mila_Quebec , @shiva_malay @PShravannayak
Sai Rajeswar tweet media
English
2
16
49
3.5K
iamrobotbear (bk)
iamrobotbear (bk)@iamrobotbear·
@nlw But don’t forget the groceries, too. Lol. Completely agree on this framing. Aligned on your opportunity ai as well.
English
0
0
0
20
Nathaniel Whittemore
Is AI going to take everyone’s job, or is the best use case for your new AI product helping me book flights, because it can’t be both.
English
10
2
18
2.1K
iamrobotbear (bk) retweetledi
Jerry Liu
Jerry Liu@jerryjliu0·
Introducing LiteParse - the best model-free document parsing tool for AI agents 💫 ✅ It’s completely open-source and free. ✅ No GPU required, will process ~500 pages in 2 seconds on commodity hardware ✅ More accurate than PyPDF, PyMuPDF, Markdown. Also way more readable - see below for how we parse tables!! ✅ Supports 50+ file formats, from PDFs to Office docs to images ✅ Is designed to plug and play with Claude Code, OpenClaw, and any other AI agent with a one-line skills install. Supports native screenshotting capabilities. We spent years building up LlamaParse by orchestrating state-of-the-art VLMs over the most complex documents. Along the way we realized that you could get quite far on most docs through fast and cheap text parsing. Take a look at the video below. For really complex tables within PDFs, we output them in a spatial grid that’s both AI and human-interpretable. Any other free/light parser light PyPDF will destroy the representation of this table and output a sequential list. This is not a replacement for a VLM-based OCR tool (it requires 0 GPUs and doesn’t use models), but it is shocking how good it is to parse most documents. Huge shoutout to @LoganMarkewich and @itsclelia for all the work here. Come check it out: llamaindex.ai/blog/liteparse… Repo: github.com/run-llama/lite…
LlamaIndex 🦙@llama_index

We've spent years building LlamaParse into the most accurate document parser for production AI. Along the way, we learned a lot about what fast, lightweight parsing actually looks like under the hood. Today, we're open-sourcing a light-weight core of that tech as LiteParse 🦙 It's a CLI + TS-native library for layout-aware text parsing from PDFs, Office docs, and images. Local, zero Python dependencies, and built specifically for agents and LLM pipelines. Think of it as our way of giving the community a solid starting point for document parsing: npm i -g @llamaindex/liteparse lit parse anything.pdf - preserves spatial layout (columns, tables, alignment) - built-in local OCR, or bring your own server - screenshots for multimodal LLMs - handles PDFs, office docs, images Blog: llamaindex.ai/blog/liteparse… Repo: github.com/run-llama/lite…

English
40
232
1.8K
227.4K
iamrobotbear (bk) retweetledi
Thariq
Thariq@trq212·
We just released Claude Code channels, which allows you to control your Claude Code session through select MCPs, starting with Telegram and Discord. Use this to message Claude Code directly from your phone.
English
1.7K
2.4K
25.6K
7.3M
iamrobotbear (bk) retweetledi
SkalskiP
SkalskiP@skalskip92·
spent some time today playing with MolmoPoint it's pretty crazy that we can use VLMs for multi-object tracking now instead of spelling out coordinates as text, it points by directly selecting parts of its own visual features prompt: "Track blue players."
Jae Sung Park@jjaesungpark

VLMs today—including our own Molmo—point via raw text strings (e.g. ""). What if pointing meant directly selecting the visual tokens instead? 🤔 Introducing MolmoPoint: Better Pointing for VLMs with Grounding Tokens 🎯 🔓models, code, data, demo all OPEN 🧵👇 Paper: allenai.org/papers/molmopo…

English
5
19
252
24.3K
iamrobotbear (bk) retweetledi
Muratcan Koylan
Muratcan Koylan@koylanai·
Many people shared this with me but I feel like something is off. Not that the Google Skills article is wrong but they sucked the joy out of it. Skills are interesting because they're a new way for us to encode how agents operate. The Anthropic team recently published an article about their experiences in developing with Skills. I learned so many things from it because they shared: - what kept breaking - what turned out to matter - weird tricks that actually helped - unique examples adn use cases (e.g. the video one) - how the team discovered the pattern Google’s article is just like here are 5 boxes, here is the standard form, here is the taxonomy... None of these feels like they came from production experience. I don't know why, but Google always gives everything formal names, which doesn't add much, like a platform team turning taste into a corporate framework. If someone is not already deep in skills, this post could make the space feel more complicated, not clearer. Maybe I'm wrong tho...
Google Cloud Tech@GoogleCloudTech

x.com/i/article/2033…

English
8
4
32
4.7K
iamrobotbear (bk) retweetledi
Simon Willison
Simon Willison@simonw·
Dan says he's got Qwen 3.5 397B-A17B - a 209GB on disk MoE model - running on an M3 Mac at ~5.7 tokens per second using only 5.5 GB of active memory (!) by quantizing and then streaming weights from SSD (at ~17GB/s), since MoE models only use a small subset of their weights for each token
Dan Woods@danveloper

x.com/i/article/2034…

English
97
184
1.9K
250.3K
iamrobotbear (bk) retweetledi
Paul Adams
Paul Adams@Padday·
How we build software @intercom is unrecognisable vs 12 months ago. We're fully Claude Code pilled and seeing enormous productivity gains. Excellent thread here from Brian on some things we're doing.
Brian Scanlan@brian_scanlan

We've been building an internal Claude Code plugin system at Intercom with 13 plugins, 100+ skills, and hooks that turn Claude into a full-stack engineering platform. Lots done, more to do. Here's a thread of some highlights.

English
15
14
381
137.1K
iamrobotbear (bk) retweetledi
Felix Rieseberg
Felix Rieseberg@felixrieseberg·
By popular demand, Dispatch can now launch Claude Code sessions. Ask it to build, make, or improve something! To use it, update your Claude desktop app and make sure you have Code enabled.
Felix Rieseberg tweet media
English
180
132
2.5K
214.2K
iamrobotbear (bk) retweetledi
Latent.Space
Latent.Space@latentspacepod·
🆕 Claude Cowork, Skills, and the Future of AI Coworkers latent.space/p/felix-anthro… @felixrieseberg has spent years working at the interface layer, from Electron and the Slack desktop app to now helping build @claudeai Cowork. In this episode, Felix explains why execution is getting so cheap that teams can “build all the candidates,” why Anthropic is betting on local-first agent workflows, and why the future of AI products may belong less to chatbots and more to systems that can actually do knowledge work.
English
12
11
98
57K
iamrobotbear (bk) retweetledi
Dropbox
Dropbox@Dropbox·
How we used DSPy to turn our relevance judge into a measurable optimization loop, making it more reliable and scalable in Dropbox Dash.
English
10
40
234
96.1K
iamrobotbear (bk) retweetledi
Morph
Morph@morphllm·
Introducing FlashCompact - the first specialized model for context compaction 33k tokens/sec 200k → 50k in ~1.5s Fast, high quality compaction
English
86
136
2.2K
210.4K