smly

7.2K posts

smly banner
smly

smly

@smly

AI Fellow at Rist. ex-PFN / 4x Kaggle Grandmaster https://t.co/Fbnhh888VD / Google Developer Expert (Kaggle) / Mahjong AI https://t.co/LQRTdKP3AO

Katılım Nisan 2007
3.6K Takip Edilen8.3K Takipçiler
smly
smly@smly·
Kaggle Notebook の Embed 機能のお試し。KaggleのGPUを使ってゲーム実行してリプレイを Kaggle Notebook 上で可視化してブログに埋め込む。便利 #kaggle-notebook" target="_blank" rel="nofollow noopener">ho.lc/blog/riichienv…
日本語
2
0
22
1.6K
smly
smly@smly·
@hotpepsi おお、なるほど!Windowsの操作感を完全に忘れていました。
日本語
0
0
0
78
hotpepsi
hotpepsi@hotpepsi·
@smly そのうち左側でドラッグするように慣れるかもなのですが Windowsだと右上にボタン類があり、どのアプリでもドラッグできる隙間があるのに気づきました。macで縦タブだと隙間がなくなり違和感に気づいたという感じです。
日本語
1
0
1
139
hotpepsi
hotpepsi@hotpepsi·
Chromeに縦タブ来た。これは良いかも。設定にタブの位置が追加されてた。ただmacだと隙間が少なくてウィンドウのドラッグがしづらい
日本語
1
0
1
1.4K
smly
smly@smly·
オンライン対局の同時実行が3ゲームに増えた🙏 最強の麻雀AIの参戦をお待ちしてます riichi.dev/games
smly tweet media
日本語
0
0
25
1.9K
smly retweetledi
YI
YI@y_imjk·
麻雀AIをLLMでやるやつの発表があり嬉しい anlp.jp/proceedings/an…
日本語
0
3
13
2.7K
smly
smly@smly·
@y_imjk うれしい!シャンテン数計算の高速化が無いと高度な特徴量計算はCPUボトルネックになって使えないので先人に感謝しかないです
日本語
1
0
1
59
YI
YI@y_imjk·
シャンテン数計算の高速化周りの話とか昔ブログで読んだりしててほえ〜とか言ってた記憶があり、nyantenが使われてたりするのを見てにっこりしてる
日本語
1
0
1
227
YI
YI@y_imjk·
趣味で麻雀環境作るかとか思ってたけど要らなくなりそう
日本語
1
0
0
405
smly
smly@smly·
autoresearch is fascinating. For about a week I had an agent autonomously edit code, run training, monitor metrics, and iterate on Mahjong AI reinforcement learning. It explored PPO hyperparameters, KL penalties, reward structures, opponent selection, and rolled back to stable checkpoints when metrics degraded. Couldn't solve a fundamentally misaligned reward signal, but watching it map out the failure landscape was eye-opening.
English
0
0
24
2.3K
smly retweetledi
snwy
snwy@snwy_me·
autoresearch really interested me, despite me not being "all-in" on agents yet. i wanted to get started with running auto experiments i looked to existing tools to serve as a harness but each one had its problems. so i made one introducing Helios for autonomous ML research
snwy tweet media
Andrej Karpathy@karpathy

Three days ago I left autoresearch tuning nanochat for ~2 days on depth=12 model. It found ~20 changes that improved the validation loss. I tested these changes yesterday and all of them were additive and transferred to larger (depth=24) models. Stacking up all of these changes, today I measured that the leaderboard's "Time to GPT-2" drops from 2.02 hours to 1.80 hours (~11% improvement), this will be the new leaderboard entry. So yes, these are real improvements and they make an actual difference. I am mildly surprised that my very first naive attempt already worked this well on top of what I thought was already a fairly manually well-tuned project. This is a first for me because I am very used to doing the iterative optimization of neural network training manually. You come up with ideas, you implement them, you check if they work (better validation loss), you come up with new ideas based on that, you read some papers for inspiration, etc etc. This is the bread and butter of what I do daily for 2 decades. Seeing the agent do this entire workflow end-to-end and all by itself as it worked through approx. 700 changes autonomously is wild. It really looked at the sequence of results of experiments and used that to plan the next ones. It's not novel, ground-breaking "research" (yet), but all the adjustments are "real", I didn't find them manually previously, and they stack up and actually improved nanochat. Among the bigger things e.g.: - It noticed an oversight that my parameterless QKnorm didn't have a scaler multiplier attached, so my attention was too diffuse. The agent found multipliers to sharpen it, pointing to future work. - It found that the Value Embeddings really like regularization and I wasn't applying any (oops). - It found that my banded attention was too conservative (i forgot to tune it). - It found that AdamW betas were all messed up. - It tuned the weight decay schedule. - It tuned the network initialization. This is on top of all the tuning I've already done over a good amount of time. The exact commit is here, from this "round 1" of autoresearch. I am going to kick off "round 2", and in parallel I am looking at how multiple agents can collaborate to unlock parallelism. github.com/karpathy/nanoc… All LLM frontier labs will do this. It's the final boss battle. It's a lot more complex at scale of course - you don't just have a single train. py file to tune. But doing it is "just engineering" and it's going to work. You spin up a swarm of agents, you have them collaborate to tune smaller models, you promote the most promising ideas to increasingly larger scales, and humans (optionally) contribute on the edges. And more generally, *any* metric you care about that is reasonably efficient to evaluate (or that has more efficient proxy metrics such as training a smaller network) can be autoresearched by an agent swarm. It's worth thinking about whether your problem falls into this bucket too.

English
48
82
1.5K
183.4K
smly
smly@smly·
一番のお気に入りポイントは CSS 3D Transforms でゲームシーンを構成して wasm+svg すべて込みで gz 圧縮して 174kb で動作するところです
smly tweet media
日本語
0
3
21
1.3K
smly
smly@smly·
麻雀AIのシミュレーション環境とオープンアリーアを公開しました。AIのオンライン対戦ができます。戦いましょう ho.lc/blog/riichienv…
日本語
1
23
87
7.8K
smly retweetledi
Tri Dao
Tri Dao@tri_dao·
The FA4 paper is finally out after a year of work. On Blackwell GPUs, attention now goes about as fast as matmul even though the bottlenecks are so different! Tensor cores are now crazy fast that attn fwd is bottlenecked by exponential, and attn bwd is bottlenecked by shared memory bandwidth.  Some fun stuff in the redesigned algorithm to overcome these bottlenecks: exponential emulation with polynomials, new online softmax to avoid 90% of softmax rescaling, 2CTA MMA instructions that allow two thread blocks to share operands to reduce smem traffic.
Ted Zadouri@tedzadouri

Asymmetric hardware scaling is here. Blackwell tensor cores are now so fast, exp2 and shared memory are the wall. FlashAttention-4 changes the algorithm & pipeline so that softmax & SMEM bandwidth no longer dictate speed. Attn reaches ~1600 TFLOPs, pretty much at matmul speed! joint work w/ Markus Hoehnerbach, Jay Shah(@ultraproduct), Timmy Liu, Vijay Thakkar (@__tensorcore__ ), Tri Dao (@tri_dao) 1/

English
30
230
1.8K
183.6K
smly
smly@smly·
FA4 まだ transformers での end-to-end の改善幅が小さくて気になるけど期待。varlen まわり問題になりやすいのわかりみ github.com/huggingface/tr…
日本語
0
0
5
1.1K
smly retweetledi
PyTorch
PyTorch@PyTorch·
FlexAttention now has a FlashAttention-4 backend. FlexAttention has enabled researchers to rapidly prototype custom attention variants—with 1000+ repos adopting it and dozens of papers citing it. But users consistently hit a performance ceiling. Until now. We've added a FlashAttention-4 backend to FlexAttention on Hopper and Blackwell GPUs. PyTorch now auto-generates CuTeDSL score/mask modifications and JIT-instantiates FlashAttention-4 for your custom attention variant. The result: 1.2× to 3.2× speedups over Triton on compute-bound workloads. 🖇️ Read our latest blog here: hubs.la/Q045FHPh0 No more choosing between flexibility and performance. hashtag#PyTorch hashtag#FlexAttention hashtag#FlashAttention hashtag#OpenSourceAI
PyTorch tweet media
English
12
97
732
99.7K
smly
smly@smly·
Chrome の canary-dcheck build のオーバーヘッドが思っていた以上に大きくて泣いた
日本語
0
0
0
1.1K
smly
smly@smly·
weekly rate limits にぶつかったので Claude Max 20x にアップグレード。大電力消費時代って感じだ
日本語
0
0
8
1.4K
smly retweetledi
Lex Fridman
Lex Fridman@lexfridman·
Here's my conversation with Peter Steinberger (@steipete), creator of OpenClaw, an open-source AI agent that has taken the Internet by storm, with now over 180,000 stars on GitHub. This was a truly mind-blowing, inspiring, and fun conversation! It's here on X in full and is up everywhere else (see comment). Timestamps: 0:00 - Episode highlight 1:30 - Introduction 5:36 - OpenClaw origin story 8:55 - Mind-blowing moment 18:22 - Why OpenClaw went viral 22:19 - Self-modifying AI agent 27:04 - Name-change drama 44:15 - Moltbook saga 52:34 - OpenClaw security concerns 1:01:14 - How to code with AI agents 1:32:09 - Programming setup 1:38:52 - GPT Codex 5.3 vs Claude Opus 4.6 1:47:59 - Best AI agent for programming 2:09:59 - Life story and career advice 2:13:56 - Money and happiness 2:17:49 - Acquisition offers from OpenAI and Meta 2:34:58 - How OpenClaw works 2:46:17 - AI slop 2:52:20 - AI agents will replace 80% of apps 3:00:57 - Will AI replace programmers? 3:12:57 - Future of OpenClaw community
English
502
1.1K
6.8K
2M
smly
smly@smly·
.@takuoko1 さんとKINTAN焼肉!
smly tweet media
日本語
0
1
14
2K
smly
smly@smly·
Bambu Lab P2S Combo レビュー。届いた瞬間に破壊しましたが、なんとか沼に入門できました ho.lc/blog/bambu-p2s
日本語
0
0
9
1.3K