F.Mackenzie 约克.小汽车. 嘟嘟

37.5K posts

F.Mackenzie 约克.小汽车. 嘟嘟 banner
F.Mackenzie 约克.小汽车. 嘟嘟

F.Mackenzie 约克.小汽车. 嘟嘟

@FMackenzie7

🇬🇧 AI, LLM interpretation, Maths, Information geometry, Manifold hypothesis, SAE and EV battery safety

London Katılım Kasım 2019
3K Takip Edilen1.9K Takipçiler
Sabitlenmiş Tweet
F.Mackenzie 约克.小汽车. 嘟嘟
bookmarked
马东锡 NLP@dongxi_nlp

@FMackenzie7 我个人必读论文list有: trasnformer:attention is all you need, encoder blocks:BERT encoder-decoder blocks: BART decoder-blocks: GPT-1, 2, 3 prompt-based learning instruct tuning cot, react YouTube的Mu Li老师讲的非常好 我自己也会逐步把这些论文写成thread

English
5
4
32
77.8K
F.Mackenzie 约克.小汽车. 嘟嘟
AXPLORER Democratizing the search for interesting mathematical constructions Open sourced 🙇‍♂️🙇‍♂️🙇‍♂️🙇‍♂️🙇‍♂️ 🥰🥰🥰🥰🥰 🙌🙌🙌🙌🙌
Axiom@axiommathai

We open-sourced Axplorer. Axplorer builds on PatternBoost; it discovers outlier math constructions to attack open problems. On Turán 4-Cycles, No 5 Points on Sphere, and Isosceles-Free Sets, Axplorer matched SOTA w/ a fraction of compute cost and time. It's now in your hands.

English
0
0
0
1
F.Mackenzie 约克.小汽车. 嘟嘟 retweetledi
F.Mackenzie 约克.小汽车. 嘟嘟
Towards end-to-end automation of AI research 🙇‍♂️🙇‍♂️🙇‍♂️🙇‍♂️🙇‍♂️
Sakana AI@SakanaAILabs

The AI Scientist: Towards Fully Automated AI Research, Now Published in Nature Nature: nature.com/articles/s4158… Blog: sakana.ai/ai-scientist-n… When we first introduced The AI Scientist, we shared an ambitious vision of an agent powered by foundation models capable of executing the entire machine learning research lifecycle. From inventing ideas and writing code to executing experiments and drafting the manuscript, the system demonstrated that end-to-end automation of the scientific process is possible. Soon after, we shared a historic update: the improved AI Scientist-v2 produced the first fully AI-generated paper to pass a rigorous human peer-review process. Today, we are happy to announce that “The AI Scientist: Towards Fully Automated AI Research,” our paper describing all of this work, along with fresh new insights, has been published in @Nature! This Nature publication consolidates these milestones and details the underlying foundation model orchestration. It also introduces our Automated Reviewer, which matches human review judgments and actually exceeds standard inter-human agreement. Crucially, by using this reviewer to grade papers generated by different foundation models, we discovered a clear scaling law of science. As the underlying foundation models improve, the quality of the generated scientific papers increases correspondingly. This implies that as compute costs decrease and model capabilities continue to exponentially increase, future versions of The AI Scientist will be substantially more capable. Building upon our previous open-source releases (github.com/SakanaAI/AI-Sc…), this open-access Nature publication comprehensively details our system's architecture, outlines several new scaling results, and discusses the promise and challenges of AI-generated science. This substantial milestone is the result of a close and fruitful collaboration between researchers at Sakana AI, the University of British Columbia (UBC) and the Vector Institute, and the University of Oxford. Congrats to the team! @_chris_lu_ @cong_ml @RobertTLange @_yutaroyamada @shengranhu @j_foerst @hardmaru @jeffclune

English
0
0
0
78
F.Mackenzie 约克.小汽车. 嘟嘟
Manifold geometry underlies a unified code for category and category-independent features 🙇‍♂️🙇‍♂️🙇‍♂️🙇‍♂️🙇‍♂️ 🥰🥰🥰🥰🥰 🙌🙌🙌🙌🙌
David Klindt@klindt_david

I like @HSompolinsky and @s_y_chung manifold capacity theory, but I always wondered how it avoids neural collapse (tinyurl.com/neuralcollapse), where each category manifold collapses to a point. That would clearly be at odds with identifiability theory and all the empirical work finding linearly decodable features in neural representations. Excited to see this new paper combining the two! biorxiv.org/content/10.648… *also, shameless plug, here is our theory why/when doing classification necessitates learning a linear representation of *all* (task-relevant) latent variables: arxiv.org/abs/2410.21869

English
1
0
0
89
荷必如此
荷必如此@hbrc72299588·
周四好🖼️
荷必如此 tweet media
日本語
2
0
11
118
荷必如此
荷必如此@hbrc72299588·
周三好🪮
荷必如此 tweet media
日本語
3
0
10
185
荷必如此
荷必如此@hbrc72299588·
周二好🔫
荷必如此 tweet media
日本語
2
0
10
208
F.Mackenzie 约克.小汽车. 嘟嘟 retweetledi
Ksenia_TuringPost
Ksenia_TuringPost@TheTuringPost·
Must-read AI research of the week: ▪️ Complementary Reinforcement ▪️ Efficient Exploration at Scale ▪️ MetaClaw ▪️ Online Experiential Learning for LMs ▪️ A Subgoal-driven Framework for Improving Long-Horizon LLM Agents ▪️ When AI Navigates the Fog of War ▪️ Attention Residuals ▪️ Mixture-of-Depths Attention ▪️ Efficient Reasoning on the Edge ▪️ Beyond Single Tokens: Distilling Discrete Diffusion Models via Discrete MMD ▪️ Unified Spatio-Temporal Token Scoring for Efficient Video VLMs ▪️ HopChain: Multi-Hop Data Synthesis for Generalizable Vision-Language Reasoning ▪️ Cognitive Mismatch in MLLMs for Discrete Symbol Understanding ▪️ LoopRPT: Reinforcement Pre-Training for Looped LMs ▪️ AI Can Learn Scientific Taste Find the full list and the main AI news here: turingpost.com/p/fod145
Ksenia_TuringPost tweet media
English
5
48
145
7.6K
F.Mackenzie 约克.小汽车. 嘟嘟 retweetledi
Claude
Claude@claudeai·
You can now enable Claude to use your computer to complete tasks. It opens your apps, navigates your browser, fills in spreadsheets—anything you'd do sitting at your desk. Research preview in Claude Cowork and Claude Code, macOS only.
English
4.9K
14.6K
139.3K
74.7M