N 🟩

791 posts

N 🟩 banner
N 🟩

N 🟩

@corang_ee

All roads lead to Rome

Seoul, Korea Katılım Nisan 2025
817 Takip Edilen1.2K Takipçiler
RJ 🟩
RJ 🟩@rj_aligned·
check out the latest amazing work by @fede_intern and @diego_aligned. they have found a way to make verifiable ai affordable and usable with open source models. of course already with a paper and code for anyone to review. i'm incredibly lucky to be able to work with this crazy and brilliant people. and there is more to come!
Fede’s intern 🥊@fede_intern

LLMs now make critical decisions in hospitals, defense, banks, and governments. Yet nobody can verify which model actually ran, or whether the output was tampered with. A provider or middleman can swap weights, silently requantize the model, alter decoding, inject hidden prompts, do supply chain attacks, or change the deployment surface without the user knowing. This problem is already serious. It will become critical. We think this needs a practical solution, not just a theoretically clean one. CommitLLM is designed to be deployable on existing serving stacks now: the provider keeps the normal GPU serving path, does not need a proving circuit, does not need a kernel rewrite, and does not generate a heavy proof for every response. In practice, two families of approaches dominated the conversation before this work: fingerprinting, which can be gamed, and proof-based systems, which are theoretically strong but too expensive for production inference. We built CommitLLM to target the middle ground. The core idea is to keep the verification discipline of proof systems, but specialize it to open weight LLM inference. The cryptographic core is simple: Freivalds style randomized checks for the large linear layers, plus Merkle commitments for the traced execution. Then a lot of engineering work is needed to make that line up with real GPU inference. The key trick is this. A provider claims `z = W × x` for a massive weight matrix. Normally you would verify that by redoing the multiply. Instead, the verifier samples a secret random vector `r`, precomputes `v = rᵀ × W`, and later checks whether `v · x = rᵀ · z`. Two dot products instead of a full matrix multiply. In the current implementation, a wrong result passes with probability at most `1 / (2^32 - 5)` per check. A full matrix multiply, audited with two dot products. Most of the transformer can then be checked exactly or canonically from committed openings. Nonlinear operations such as activations and layer norms are canonically re executed by the CPU verifier. The one honest caveat is attention: native FP16/BF16 attention is not bit reproducible across hardware. CommitLLM verifies the shell around attention exactly, then independently replays attention and checks that the committed post attention output stays within a measured INT8 corridor. So attention is bounded and audited, not proved exactly. That means the protocol already gives very strong exact guarantees on the parts that matter operationally most. If an audited response used the wrong model, the wrong quantization/configuration, or a tampered input/deployment surface, the audit catches that exactly. That includes things like model swaps, silent requantization, and provider side prompt or system prompt injection. Today the implementation and measurements are strongest on Qwen and Llama. But the protocol itself is not meant to be Qwen or Llama specific: we expect it to generalize across open weight decoder only families. What still has to be done is the engineering work to integrate and validate more families explicitly, and we are already working on that. On the measured path, online generation overhead is about 12 to 14% with the provider staying on the normal GPU serving path. The heavier receipt finalization cost is separate and can be deferred off the user facing path. The main systems costs are RAM and bandwidth, not proof generation. The full response is always committed, but only a random fraction of responses are opened for audit. Individual audits are much larger, roughly 4 MB to 100 MB depending on audit depth. The important number is the amortized one: under a reasonable audit policy, the added bandwidth averages to roughly 300 KB per response. After too many weeks without sleep, I’m proud to show what I built with @diego_aligned: CommitLLM. Thanks Diego for your patience. I've been calling you at random hours. The code and paper still need some cleaning and formalization. We’re already in talks with multiple providers and teams that have cryptography related ideas on how to improve it even more. We’re really excited about this and we will continue doubling down on building products in AI, cryptography and security with my company @class_lambda. If governments, hospitals, defense and financial systems are going to run on LLMs, verifiable inference is not optional. It is infrastructure. I will be explaining this in more details in the days to come and I will show how to test it and run it.

English
3
1
7
653
Pedgy Penguins
Pedgy Penguins@Pedgypenguin·
For those who missed the first mint, another chance is coming. 1111 Penguin Z Free Mint soon on $ETH Drop your EVM wallets
Pedgy Penguins tweet media
English
485
180
723
8.4K
Wonnie🔺
Wonnie🔺@wonnie·
GM Legends Have a lovely day
Wonnie🔺 tweet media
English
214
7
318
5.2K
N 🟩
N 🟩@corang_ee·
협상 진전, 공격 연기 트럼프: “미국과 이란은 지난 이틀 동안 중동 내 적대 행위를 완전히 해소하기 위한 생산적인 논의를 진행해왔습니다. 이번 주 협상이 계속되는 가운데, 진전이 있다는 전제하에 이란의 에너지 인프라에 대한 모든 군사 공격을 5일간 중단하라고 지시했습니다. 감사합니다. — 도널드 J. 트럼프 대통령”
N 🟩 tweet media
한국어
0
0
0
205
N 🟩
N 🟩@corang_ee·
오늘 새벽 이란 방공망이 이란 중부 상공에서의 전투 임무 중인 미국 F-35를 격추했습니다. 제트기는 손상되었지만 중동의 미국/동맹 기지에서 비상 착륙을 했습니다. 조종사는 안전하다고 합니다. 이번 전쟁에서 이란이 미국 항공기를 타격한 첫 번째 사례입니다. #미국 #이란 #F35
한국어
0
0
0
38
N 🟩
N 🟩@corang_ee·
모건 스탠리, 현물 비트코인 ETF 신청 제출 @MorganStanley SEC에 현물 비트코인 ETF를 위한 S-1 신청서를 제출하고 업데이트했으며, 모건 스탠리 비트코인 트러스트를 출시할 계획을 추진중. #모건스탠리 #현물 #비트코인
N 🟩 tweet media
한국어
0
0
0
15
N 🟩
N 🟩@corang_ee·
트럼프는 진짜 대단하네요... 🇯🇵 일본 기자 이란과의 전쟁을 시작하기 전에 왜 우리에게 말하지 않았나요? 🇺🇸 @realDonaldTrump 우리는 기습을 원했어요. 일본만큼 기습에 대해 잘 아는 나라는 또 있을까요? 진주만에 대해 왜 나에게 말하지 않았나요? 마지막에 다카이치 표정이 ㅋㅋㅋㅋ #트럼프 #다카이치
한국어
0
0
0
17
N 🟩
N 🟩@corang_ee·
예측시장 카테고리별 거래량 (2월 16일~3월 15일) 총 명목 거래량: 191억 8,000만 달러 1 스포츠 53.2% 2, 암호화폐 19.9% 1 정치 16.1% 2 기타 5.3% 3 문화 1.7% 4 비즈니스 1.6% 5 경제 1.1% 6 날씨 0.8% 7 기술 0.3% #예측시장 #거래량
N 🟩 tweet media
한국어
0
0
0
22
MAGIC (theo/acc)🕯️🦅
인생 함박 스테이크 맛집 찾음 일본에서 살다온 친구가 일본에서 핫한 함박 집이라고 가자고함. 개인적으로 함박을 별로 선호하지 않아서 시큰둥 했지만 친구픽을 믿는 편이라 속는셈치고 가봄 근데 오늘 먹어보고 ㄹㅇ 감동함. 우리가 지금까지 먹었던건 진짜 함박이 아니었음. 한국인들은 그동안 사기를 당했던거임 한줄요약 : 꼭 가보세요 히키니쿠토코메 도산 서울 강남구 선릉로155길 21 2층 naver.me/531ezIH6
MAGIC (theo/acc)🕯️🦅 tweet media
한국어
16
4
45
2.5K
Javi🥥.eth
Javi🥥.eth@jgonzalezferrer·
I reply to every DM. Every day. It's exhausting. Sometimes painful But it's been one of the most valuable things I've done for building communities Ok, some days I rest and don't answer DMs But you get the point!
English
472
25
967
22.9K
Kate
Kate@mynameiskatekim·
강아지 토이스토리 옷 입혀봤는데 귀엽🐶 버즈 우디 옷을 살걸그랬나! 👀
GIF
Kate tweet media
한국어
3
0
6
213
Kate
Kate@mynameiskatekim·
이제 USDe는 @Compound_xyz 메인넷의 USDC 및 USDT Comet에서 담보 자산으로 추가되었습니다!
Kate tweet media
한국어
3
0
19
793
N 🟩
N 🟩@corang_ee·
@Pedgypenguin 0x26112ad891f1a64ba2722bae303b40dff85a843b Please Please Please pudgy Penguins ~~~
English
0
0
2
298
Pedgy Penguins
Pedgy Penguins@Pedgypenguin·
4,444 Pedgys are coming to ETH for FREE drop your EVM wallets
Pedgy Penguins tweet media
English
3.4K
839
3.5K
164.9K
N 🟩
N 🟩@corang_ee·
@CENTWT ㅋㅋㅋㅋㅋㅋㅋ
한국어
0
0
0
47
CEN
CEN@CENTWT·
현재 코인시장 한장으로 요약 - 위 : 프로젝트,VC - 아래 : 에어드랍 노동자
CEN tweet media
한국어
17
1
34
1.9K