julio (bigsxy)

5.4K posts

julio (bigsxy) banner
julio (bigsxy)

julio (bigsxy)

@juemrami

🇲🇽🤝🇺🇸 LA Software "Engineer" (Unemployed af) Into ML, Web, WoW Addons, Nix, and Embedded LFG Bulletin Board - 250K+ users | $0 MRR https://t.co/Si89fkutRl

Northshire Valley Katılım Haziran 2017
194 Takip Edilen113 Takipçiler
julio (bigsxy)
julio (bigsxy)@juemrami·
and look at these nice errors , that i wont bother handling, i get
julio (bigsxy) tweet media
English
0
0
0
7
julio (bigsxy)
julio (bigsxy)@juemrami·
Decided to just ditch the majority of the SDK and just use their zod validator and write my own effect-ts client over the parts of the API i actually am using and look at the fucking difference. Nearly half the size. Another reason to not just blindly bring in libraries when you boilerplate is so much easier to write now.
julio (bigsxy) tweet mediajulio (bigsxy) tweet media
julio (bigsxy)@juemrami

man the mistral-ai typescprit api sdk is so bad at tree shake-ability. Im literally just using 2 of the clients functions atm and its taking up 50% of my bundles real estate. Another reason to use effect if youre building libraries, makes you build in a way that makes DCE ez for bundlers.

English
3
0
1
49
julio (bigsxy)
julio (bigsxy)@juemrami·
here's what the new graph looks like Id like to not use zod but i'd need to find an OpenAPI spec to effect-schema generator.
julio (bigsxy) tweet media
English
0
0
0
10
julio (bigsxy)
julio (bigsxy)@juemrami·
@MichaelArnaldi @LukeParkerDev my agents run on the budget from copilot (gpt5mini and their "raptor mini" model) They are complete imbeciles. Cant afford a nice claude/codex plan unfortunately.
English
0
0
0
60
Luke Parker
Luke Parker@LukeParkerDev·
im am this close to crashing out. every AI just does dumb stuff unless you are so specific you may as well code. it can help for mass migrations once you've already done the shape and exact impl, and have a bunch of boring work. im so sick of trying to wrangle it lol
English
154
36
969
44.7K
julio (bigsxy)
julio (bigsxy)@juemrami·
man the mistral-ai typescprit api sdk is so bad at tree shake-ability. Im literally just using 2 of the clients functions atm and its taking up 50% of my bundles real estate. Another reason to use effect if youre building libraries, makes you build in a way that makes DCE ez for bundlers.
julio (bigsxy) tweet media
English
0
0
0
70
julio (bigsxy)
julio (bigsxy)@juemrami·
@kimmonismus Did you just rehash the op tweet for the sake of making this a quote tweet instead of just retweet?
English
0
0
0
306
Chubby♨️
Chubby♨️@kimmonismus·
Very impressive: MSA (memory sparse attentions) is a so exciting because it lets AI models directly store and reason over massive long-term memory inside their attention system, without relying on external retrieval or lossy compression, making them far more accurate and scalable. it allows 100M context window with minimal performance loss
Chubby♨️ tweet media
艾略特@elliotchen100

论文来了。名字叫 MSA,Memory Sparse Attention。 一句话说清楚它是什么: 让大模型原生拥有超长记忆。不是外挂检索,不是暴力扩窗口,而是把「记忆」直接长进了注意力机制里,端到端训练。 过去的方案为什么不行? RAG 的本质是「开卷考试」。模型自己不记东西,全靠现场翻笔记。翻得准不准要看检索质量,翻得快不快要看数据量。一旦信息分散在几十份文档里、需要跨文档推理,就抓瞎了。 线性注意力和 KV 缓存的本质是「压缩记忆」。记是记了,但越压越糊,长了就丢。 MSA 的思路完全不同: → 不压缩,不外挂,而是让模型学会「挑重点看」 核心是一种可扩展的稀疏注意力架构,复杂度是线性的。记忆量翻 10 倍,计算成本不会指数爆炸。 → 模型知道「这段记忆来自哪、什么时候的」 用了一种叫 document-wise RoPE 的位置编码,让模型天然理解文档边界和时间顺序。 → 碎片化的信息也能串起来推理 Memory Interleaving 机制,让模型能在散落各处的记忆片段之间做多跳推理。不是只找到一条相关记录,而是把线索串成链。 结果呢? · 从 16K 扩到 1 亿 token,精度衰减不到 9% · 4B 参数的 MSA 模型,在长上下文 benchmark 上打赢 235B 级别的顶级 RAG 系统 · 2 张 A800 就能跑 1 亿 token 推理。这不是实验室专属,这是创业公司买得起的成本。 说白了,以前的大模型是一个极度聪明但只有金鱼记忆的天才。MSA 想做的事情是,让它真正「记住」。 我们放 github 上了,算法的同学不容易,可以点颗星星支持一下。🌟👀🙏 github.com/EverMind-AI/MSA

English
32
69
881
76.8K
julio (bigsxy)
julio (bigsxy)@juemrami·
@haydendevs I know u can do this with other tools, but nix keeps it unified under one “dsl”
English
0
0
0
14
julio (bigsxy)
julio (bigsxy)@juemrami·
@haydendevs If you appreciate declarative setups. Specially for little system hacks that you may forget about when you move from machine to machine. Yes.
English
1
0
1
81
hayden
hayden@haydendevs·
is nix actually better than arch
English
48
0
86
9.1K
Christoffer Bjelke
Christoffer Bjelke@chribjel·
i wish the open weight models like kimi/glm/minimax would have a license so all derivative models also would have to be open weight
English
10
3
235
12.8K
julio (bigsxy) retweetledi
Darth Powell
Darth Powell@VladTheInflator·
Darth Powell tweet media
ZXX
87
897
7.6K
270.9K
julio (bigsxy)
julio (bigsxy)@juemrami·
@YashGouravKar1 yea, just like using Express or whatever other lib you were comfortable with. Ive just never had to actually implement one, so i didnt know how it. I had an idea but i wasnt going to sit there and google it. Pretty fucking bad on my part lol. 🤷‍♂️i shoulda just googled it xD.
English
2
0
1
24
julio (bigsxy)
julio (bigsxy)@juemrami·
I flopped so fucking hard. its not even funny. well it is funny. I just realized ive never had to implement an http server from an empty project. Even though i have made my own http parser and simple server on Zig, i had 0 clue how to "spin one up" in javascript land. I knew the tech to do it, either effect HttpServer or Hono, but i didnt know implementation details and i looked like a fucking idiot.
julio (bigsxy)@juemrami

i have a phone screening interview today and im incredibly nervous. Im pasta syndrome.

English
1
0
0
49
julio (bigsxy) retweetledi
艾略特
艾略特@elliotchen100·
论文来了。名字叫 MSA,Memory Sparse Attention。 一句话说清楚它是什么: 让大模型原生拥有超长记忆。不是外挂检索,不是暴力扩窗口,而是把「记忆」直接长进了注意力机制里,端到端训练。 过去的方案为什么不行? RAG 的本质是「开卷考试」。模型自己不记东西,全靠现场翻笔记。翻得准不准要看检索质量,翻得快不快要看数据量。一旦信息分散在几十份文档里、需要跨文档推理,就抓瞎了。 线性注意力和 KV 缓存的本质是「压缩记忆」。记是记了,但越压越糊,长了就丢。 MSA 的思路完全不同: → 不压缩,不外挂,而是让模型学会「挑重点看」 核心是一种可扩展的稀疏注意力架构,复杂度是线性的。记忆量翻 10 倍,计算成本不会指数爆炸。 → 模型知道「这段记忆来自哪、什么时候的」 用了一种叫 document-wise RoPE 的位置编码,让模型天然理解文档边界和时间顺序。 → 碎片化的信息也能串起来推理 Memory Interleaving 机制,让模型能在散落各处的记忆片段之间做多跳推理。不是只找到一条相关记录,而是把线索串成链。 结果呢? · 从 16K 扩到 1 亿 token,精度衰减不到 9% · 4B 参数的 MSA 模型,在长上下文 benchmark 上打赢 235B 级别的顶级 RAG 系统 · 2 张 A800 就能跑 1 亿 token 推理。这不是实验室专属,这是创业公司买得起的成本。 说白了,以前的大模型是一个极度聪明但只有金鱼记忆的天才。MSA 想做的事情是,让它真正「记住」。 我们放 github 上了,算法的同学不容易,可以点颗星星支持一下。🌟👀🙏 github.com/EverMind-AI/MSA
艾略特 tweet media
艾略特@elliotchen100

稍微剧透一下,@EverMind 这周还会发一篇高质量论文

中文
164
566
3.1K
1.6M
sami
sami@hellosami·
I started teaching 4th graders graphic design at an afterschool program, and this one kid does not engage with any of the material, but he comes into class and furiously googles the Titanic for an hour every week
sami tweet media
English
588
674
23.2K
19.4M
julio (bigsxy)
julio (bigsxy)@juemrami·
i dont get why most software has to be a service. I get providing services for spaces and populations that are either hardware bound OR dont have technical savy to interact with your software outside of like GUIs or something. But besides that stuff like SaaS just makes it seem like youre trying to milk out as much money as possible from customers by making yourself seem like a requirement to make the software run or to make the software maintainble. IDK im not a business person, but as a consumer it does all seem like a giant game.
English
0
0
0
16
julio (bigsxy)
julio (bigsxy)@juemrami·
why does github attribute commits on your graph to when they were pushed to the github remote vs when the commits were actually made?
English
0
0
0
14