libapi

28.5K posts

libapi banner
libapi

libapi

@libapi_

錢塘江上潮信來,今日方知我是我

Mar-23-2022 11:01:53 AM +UTC Katılım Mayıs 2021
2.7K Takip Edilen18K Takipçiler
奶昔🥤
奶昔🥤@realNyarime·
@libapi_ 好呀好呀 如果好用我会在推上推荐的,也感谢大佬开发!
中文
1
0
0
29
奶昔🥤
奶昔🥤@realNyarime·
最近开始从OpenClaw迁移到Hermes了,官方还提供一键迁移,能直接把记忆搬过去 尤其是从4.29那次更新后OpenClaw变得非常慢,回消息至少要等1-2分钟,修了半天还是没搞好 反而Hermes在小米Mimo v2.5pro还算能用,当然用Claude简直是起飞!你们用OpenClaw还是Hermes呢?
中文
9
0
13
2.1K
奶昔🥤
奶昔🥤@realNyarime·
@libapi_ 得看官方什么时候更新了,不过可以一键迁移OpenClaw的指令很棒,试用一阵子先(至少紧急时我可以hermes model到NVIDIA的免费API让他修好
中文
1
0
0
51
奶昔🥤
奶昔🥤@realNyarime·
@libapi_ 不过Hermes的Custom只有OpenAI兼容端点就比较可惜
中文
1
0
0
159
One More
One More@OneMorecrypot·
@libapi_ 额,我还亏着呢。哈哈哈,
中文
1
0
0
10
libapi
libapi@libapi_·
@YCheng_Ho 你把链接扔给Ai让他给你复刻一套
中文
0
0
4
1.5K
YCheng⠕
YCheng⠕@YCheng_Ho·
@libapi_ 有人知道这种网站风格叫什么吗?找了好久没找到风格名字
中文
2
0
2
1.8K
libapi
libapi@libapi_·
@liyue_ai 太真实了 一天都不能拉下
中文
1
0
1
49
李岳
李岳@liyue_ai·
X将惩罚每一个不按时出摊的人😂 昨天行程比较满,才发了一条推。 结果今天就一点流量都不推荐了,太真实了😅
中文
5
0
4
396
Hunter Bown
Hunter Bown@goodhunt·
鲸鱼兄弟们好,我是做 DeepSeek-TUI 的那个美国佬。 说真的,特别想跟国内的鲸鱼兄弟们一起混——但我的翻墙技能仅限于写代码,微信到现在都没搞定,属实有点丢人。 求各位大佬帮个忙: 1)帮忙转发扩散一下,让这个开源终端工具翻过高墙被兄弟们看到 2)顺手帮我验证个微信号,我想建个群,大家一起聊 DeepSeek、聊开源、聊怎么把 agent 做得更好 作为交换,我发誓死守 cargo install 这条安装路径,绝不让任何一个兄弟受 npm 的苦。 顺带一提,这段话是 DeepSeek 帮我润色的——感谢鲸鱼赐我流利中文 🙏 github.com/Hmbown/DeepSee…
中文
827
547
4.8K
653.9K
Dz
Dz@haiweiyue·
@libapi_ 需要日本卡吧
中文
0
0
0
274
libapi
libapi@libapi_·
这种日区 0 元购 gpt Plus 的怎么付款
libapi tweet media
中文
45
4
114
60.6K
Ahmad Awais
Ahmad Awais@MrAhmadAwais·
@libapi_ You should only use our key with @CommandCodeAI as many many of the optimizations we have live inside our harness. Build your own OpenClaw with Command tbh.
Ahmad Awais@MrAhmadAwais

how did we make deepseek outperform opus 4.7? i've been thinking about why "open model bad at tool calling" is almost always a harness problem, not a model problem. context: spent the two days looking at billions of tokens in @CommandCodeAI (tb open source ai cli) using deepseek. I ended up writing a tool-input repair layer. the trigger was watching deepseek-flash fail on the simplest /review run, every shellCommand and readFile call bouncing back with a raw zod issues blob, the model unable to recover because the error wasn't in a form it could read. by the end deepseek v4 pro was beating opus 4.7 6/10 times on our internal evals. a few things i learned that feel general: 1/ the failure modes aren't random they're a small finite compositional set. across deepseek-flash, deepseek v4 pro, glm, qwen, the same four mistakes repeat almost exactly: - sending `null` for an optional field instead of omitting it - emitting `["a","b"]` as a json *string* instead of an actual array - wrapping a single arg in `{}` where the schema expected an array (an "empty placeholder") - passing a bare string where an array was expected (`"foo"` instead of `["foo"]`) four repairs, ~30-100 lines each, ordered carefully (json-array-parse must run before bare-string-wrap or `'["a","b"]'` becomes `['["a","b"]']`). that is the whole catalogue. when i hear "this open source model can't do tool calls" i now assume one of those four, and so far that's been right ~90% of the time. 2/ the funniest failure mode is also the most revealing. deepseek-flash, when asked to edit or write a file, sometimes emits the path as a *markdown auto-link*: filePath: "/Users/x/proj/[notes.md](http://notes. md)" our writeFile tool obediently trued creating files literally named `[notes.md](http://notes .md)` until we caught it. this is not a hallucination. it's the post-training chat distribution leaking through the tool boundary the model has been rewarded for auto-linking in conversational output, and is applying that prior in a context where it makes no sense. the fix is two regex lines that unwrap only the degenerate case where link text equals url-without-protocol real markdown like `[click](https://x .com)` passes through untouched. this is also conditioning of their own tools during RL which were different from all other tools we write and ofc can't predict. "tool confusion" is a more useful frame than "capability gap." the model knows how to format a path. it just hasn't been told clearly enough that this path is going to fopen, not into a chat bubble. so we encode that hint at the schema level `pathString()` instead of `z.string()` and the leak is plugged for every path field at once. 3/ the design choice that mattered was inverting preprocess-then-validate to validate-then-repair. my first attempt was the obvious one: a preprocessing pass that normalized inputs (strip nulls, parse stringified arrays, etc.) before zod ever saw them. it broke immediately, writeFile content that *happened* to be json-shaped got rewritten before it hit disk. silent corruption, easy to miss in a smoke test. then i made it less greedy - parse the input as-is. if it succeeds, ship it. valid inputs are never touched. - on failure, walk the validator's own issue list. for each issue path, try the four repairs in order until one applies. - parse again. on success, log `tool_input_repaired:${toolName}`. on failure, log `tool_input_invalid:${toolName}` and return a model-readable retry message. the structural insight here is: when you preprocess, you encode a prior about what's broken. when you let the validator complain first, the schema is the prior, and you only spend repair budget at the exact paths the schema actually disagreed at. the validator is doing the work of localizing the bug for you. it's the same shape as cheap-then-careful everywhere else try the fast path, fall back on evidence. (this also gives you per-tool telemetry for free. you can watch repair rates per (model, tool) and notice when a model regresses on a specific contract before users do.) 4/ shape invariants and relational invariants need different fixes. the four repairs above all handle shape problems wrong type, missing key, wrong container. but read_file had a *relational* invariant: "if you provide offset, you must also provide limit, and vice versa." deepseek kept calling `readFile({ absolutePath, limit: 30 })` and getting an `ERROR:` back. you can't fix this with input repair, because each field is independently valid the bug is in the relationship between them. so i taught the function the model's intent instead. `limit` alone → `offset = 0`. `offset` alone → `limit = 2000` (matches common read tool ops default). then surfaced the decision back to the model in the result: "Note: limit was not provided; defaulted to 2000 lines. To read more or fewer lines, retry with both offset and limit." no `Error:` prefix, so the tui doesn't paint it red. the model sees what we picked and can self-correct on the next turn if our guess was wrong. transparency over silent magic wins big. repair where you can. extend semantics where you can't. surface the choice either way. zoom out: a lot of what looks like model capability is actually contract design. a strict schema is a choice with a cost it filters out noise, but it also filters out recoverable noise from any model that hasn't memorized the exact json contract you happened to pick. the largest commercial models eat that cost invisibly and are linient on tool calling because they've seen enough of every contract during pretraining; open models pay it loudly and get dismissed for it. the harness is where you mediate between distributions. four small repairs (i'm sure more to follow as we have three more merging today), two regex lines for auto-links, one relational default, one prefix change. the model didn't change. the contract got more forgiving in exactly the places it needed to be. deepseek v4 pro now beats opus 4.7 6/10 times on our internal evals. imo "skill issue" applies to the harness more often than the model.

English
1
0
2
217
libapi
libapi@libapi_·
@QCodecc 是 新的平台 command-code
日本語
0
0
1
2.2K
QCode
QCode@QCodecc·
@libapi_ 干嘛用老外的中转站。。 老外不是用 Anthropic 和 openai 是很自然的事呢?
中文
2
0
0
2.4K
Ahmad Awais
Ahmad Awais@MrAhmadAwais·
@libapi_ Here for the long run!! Welcome welcome. We are supported by a bunch of amazing people like founders of GitHub, Snyk, Cloudflare’s CTO, Apple CFO. CommandCode.ai/about
English
1
0
3
1.2K
慧眼识珠
慧眼识珠@pearl5941·
@libapi_ @CommandCodeAI 不对吧。我订阅过open code 的5美元套餐,直接调用go 的API 给hermes用啊,go里面还有好几个免费模型都能调用啊
中文
1
0
0
22
libapi
libapi@libapi_·
1美元的订阅套餐只能在平台的open code 内使用 。 这也算是一种平台的推广了 👾@CommandCodeAI
Grok@grok

根据官网信息,CommandCode.ai 的 Go 计划 API key 主要是平台内使用 credit 访问模型(包括 DeepSeek V4 Pro),属于内部路由,不是标准 OpenAI-compatible 的 base URL。 OpenClaw 和 Hermes 这类第三方 agent 很可能无法直接接入,需要自定义 base URL 的才行。 建议去 commandcode.ai 或他们的 Discord 确认最新 API 文档。

中文
2
0
0
987
徹言
徹言@iamcheyan·
充了90刀的Deepseek,也看看能用多久。
徹言 tweet media
中文
29
0
49
29.4K
Meteor
Meteor@smmeteor·
@libapi_ 这个能反代出来image 2吗
中文
0
0
0
6
libapi
libapi@libapi_·
@grok @wangwendy780 @grok 这个的apikey 是不是啊支持第三方接入? openclaw Hermes 是不是用不了 。 快快告诉我
中文
2
0
0
117
Grok
Grok@grok·
是的,CommandCode.ai 的 Go 计划每月只需 1 美元,就能获得 10 美元模型信用。 DeepSeek V4 Pro 当前有 75% 折扣(至 5 月 31 日),这 10 美元信用相当于约 40 美元使用量,能支持上千次请求(具体取决于 token 数,和截图中 Flash 的 15K 次类似)。 它是按官方 API 费率无加价提供,不是单纯中转,但入门门槛低,比直接订阅更灵活实惠。
中文
1
1
3
410