山麻楂菜角子
3K posts


This Chinese developer launched Llama 70B locally on a MacBook on a plane and for a full 11 hours without internet ran client projects. He was sitting by the window on a transatlantic flight with a MacBook Pro M4 with 64 GB of memory. WiFi on board cost $25 for the flight. He declined. No cloud API, no connection to Anthropic or OpenAI servers, no internet at all. Just a local Llama 3.3 70B on bf16 and his own orchestrator script. The model runs through llama.cpp. Generation speed, 71 tokens per second. Context around 60,000 tokens. Memory usage, 48.6 GiB out of 64. Battery at takeoff, 3 hours 21 minutes. And he gave the orchestrator this system prompt before takeoff: "You are an offline orchestrator running on a single MacBook. There is no network. The only resources you have are local files in /Users/dev/work, the Llama 70B inference server at localhost:8080, and a battery budget of 3 hours 21 minutes. Process the queue at /Users/dev/work/queue.jsonl (one client task per line). For each task: draft → run local evals → save artefact to /Users/dev/work/done/. Save context checkpoints every 12 tasks so you can resume after a battery swap. Stop only on empty queue or when battery drops below 5%." So the system knows exactly what resources it is running on. It knows it has no connection to the outside world for the next 11 hours. It knows it has finite memory and a finite battery. It knows the human will not intervene until the plane lands. The system runs in 1 loop. Takes a task from the queue, runs it through inference, saves the artifact, writes a checkpoint. Task after task, just like that. And only when the battery drops below 5% does the orchestrator automatically pause, waits for the laptop to switch to the backup power bank, and continues from the last checkpoint. Here is what the system actually writes in his log during the flight: "saved context checkpoint 8 of 12 (pos_min = 488, pos_max = 50118, size = 62.813 MiB)" "restored context checkpoint (pos_min = 488, pos_max = 50118)" "prompt processing progress: n_tokens = 50 / 60 818" "task 37016 done | tps = 71 s tokens text → /Users/dev/work/done/proposal_westside.md" Outside the window, clouds, blue sky, and no WiFi. On the tray, 1 MacBook, an open terminal on 2 screens, and an inference server on localhost. From what I have observed, this is the cleanest offline AI workflow I have seen in the past year: 11 hours of flight, $0 for WiFi, and the entire client queue closed before landing.

听说 Hermes 最近迭代了很多,稳定了很多,下午我让openclaw 去安装 hermes,还挺顺利,有惊无险,先用几天看看 分享下整个对话过程 sharethis.chat/s/shr_d2d366fd…






真的感谢any公益大善人,我几乎每天都在用 opus 4.6 模型 来迭代我的项目代码,用过这个模型以后觉得国产模型真的没法比。 这里我再说明下,这个any目前已经关闭了普通邮箱注册和github账号注册,只保留了一种注册方式:通过linuxdo论坛账号注册,可以让身边有do账号的朋友去邀请你,或者去找闲鱼,反正不亏感觉。 其次我是通过 CC Switch 这个工具来快速切换配置信息的,用起来一直很丝滑。 注册入口👉 anyrouter.top/register?aff=a… 温馨提示: 1、建议新手拿来学习 Vibe Coding 练手用,公司商业项目在意数据价值的慎用(听说数据可能被拿去训练模型)。 2、偶尔会爆网络满载拥堵错误,过会儿就好了,大部分时间是可用的。也会存在封号风险,建议合规项目学习技术为主。



给中国用户的好消息:Hermes Agent 现在原生支持个人微信了 微信扫码即可连接,私聊群聊都支持。图片、视频、文件、语音消息全覆盖,长轮询直连,不需要公网 IP。 运行 'hermes update' 即可体验 文档:hermes-agent.nousresearch.com/docs/user-guid… 感谢 @Bravohenry_ 的贡献



OpenClaw 2026.4.11 is out ✨ big polish drop for stability 🛡️ safer provider transport/routing 🤖 more reliable subagents + exec approvals 💬 lots of Slack / WhatsApp / Telegram / Matrix fixes 🌐 browser + mobile cleanup a chunky cleanup pass 😎github.com/openclaw/openc…














