LAION

1.5K posts

LAION banner
LAION

LAION

@laion_ai

100% TRULY OPEN AI.

The internet Katılım Eylül 2021
106 Takip Edilen15.5K Takipçiler
LAION retweetledi
Kimi.ai
Kimi.ai@Kimi_Moonshot·
Congrats to the @cursor_ai team on the launch of Composer 2! We are proud to see Kimi-k2.5 provide the foundation. Seeing our model integrated effectively through Cursor's continued pretraining & high-compute RL training is the open model ecosystem we love to support. Note: Cursor accesses Kimi-k2.5 via @FireworksAI_HQ ' hosted RL and inference platform as part of an authorized commercial partnership.
English
510
1.4K
20.3K
3.4M
LAION retweetledi
Wei Ping
Wei Ping@_weiping·
🚀 Introducing Nemotron-Cascade 2 🚀 Just 3 months after Nemotron-Cascade 1, we’re releasing Nemotron-Cascade 2: an open 30B MoE with 3B active parameters, delivering best-in-class reasoning and strong agentic capabilities. 🥇 Gold Medal-level performance on IMO 2025, IOI 2025, and ICPC World Finals 2025: • Capabilities once thought achievable only by frontier proprietary models (e.g. Gemini Deep Think) or frontier-scale open models (i.e. DeepSeek-V3.2-Speciale-671B-A37B). • Remarkably high intelligence density with 20× fewer parameters. 🏆 Best-in-class across math, code reasoning, alignment, and instruction following: • Outperforms the latest Qwen3.5-35B-A3B (2026-02-24) and even larger Qwen3.5-122B-A10B (2026-03-11). 🧠 Powered by Cascade RL + multi-domain on-policy distillation: • Significantly expand Cascade RL across a much broader range of reasoning and agentic domains than Nemotron-Cascade 1, while distilling from the strongest intermediate teacher models throughout training to recover regressions and sustain gains. 🤗 Model + SFT + RL data: 👉 huggingface.co/collections/nv… 📄 Technical report: 👉 research.nvidia.com/labs/nemotron/…
Wei Ping tweet media
English
40
141
881
141.2K
LAION retweetledi
ModelScope
ModelScope@ModelScope2022·
Step-3.5-Flash-SFT is open: the complete SFT training corpus, tokenizer snapshots, and pre-compiled StepTronOSS shards, all in one release. 📊Dataset: modelscope.cn/datasets/stepf… 🧑‍💻code: github.com/stepfun-ai/Ste… - Multi-turn conversation JSON with loss_mask and optional reasoning_content - Tokenizers for Step-3.5-Flash and Qwen3 included for chat template alignment - Pre-compiled shards: drop in and train, no preprocessing - Reference recipes for both Step-3.5-Flash and Qwen3 variants - Apache-2.0 + CC-BY-NC-2.0 🌉 Weights + training framework + SFT data. The full stack.
ModelScope@ModelScope2022

Step 3.5 Flash is now open source: model weights and full training framework (SteptronOSS), released together.🚀 196B total, 11B active. SWE-bench Verified 74.4% / Terminal-Bench 2.0 51.0%. - MoE architecture: 288 routed experts + 1 shared, Top-8 activation per token - MTP-3: predicts 4 tokens per forward pass, 100–300 tok/s typical, 350 tok/s peak - 3:1 SWA ratio (1 full attention + 3 sliding window layers): 256K context at lower compute cost - 💻 Runs on Mac Studio M4 Max and NVIDIA DGX Spark - SteptronOSS: SFT, continued pretraining, RL (WIP) - Apache 2.0 Two checkpoints released: Step-3.5-Flash-Base and Step-3.5-Flash-Base-Midtrain. 🤖 Base: modelscope.cn/models/stepfun… 🤖 Midtrain: modelscope.cn/models/stepfun… 🔧 Training Framework: github.com/stepfun-ai/Ste… 📄 Paper: modelscope.cn/papers/2602.10…

English
1
15
118
11.4K
LAION retweetledi
Ross Wightman
Ross Wightman@wightmanr·
Time flies. After almost 4 years at @huggingface , I’m moving on. A major part of that chapter was timm, which I sold to the company and continued to build. For anyone relying on it, I’ve agreed to collaborate on bug fixes and basic maintenance, but new feature development will likely cease. It was a meaningful chapter, and I’m thankful for the opportunity to grow timm over that time. AI is moving incredibly fast, and I’m excited to focus on new ideas and opportunities that feel like the right fit for this moment. There will be significant decisions for me ahead. I look forward to more of the serendipitous collaborations (e.g. OpenCLIP, ResNet Strikes Back, HTTY ViT) that I’ve enjoyed in the past. I’m currently working on a long overdue OpenCLIP refactoring that I hope will be useful for all and make it easier to add new model + objective combinations.
English
36
20
444
28K
LAION retweetledi
Ilya Sutskever
Ilya Sutskever@ilyasut·
It’s extremely good that Anthropic has not backed down, and it’s siginficant that OpenAI has taken a similar stance. In the future, there will be much more challenging situations of this nature, and it will be critical for the relevant leaders to rise up to the occasion, for fierce competitors to put their differences aside. Good to see that happen today.
English
1.4K
2.5K
25.6K
3M
LAION retweetledi
hardmaru
hardmaru@hardmaru·
Instead of forcing models to hold everything in an active context window, we can use hypernetworks to instantly compile documents and tasks directly into the model's weights. A step towards giving language models durable memory and fast adaptation. Blog: pub.sakana.ai/doc-to-lora/
Sakana AI@SakanaAILabs

We’re excited to introduce Doc-to-LoRA and Text-to-LoRA, two related research exploring how to make LLM customization faster and more accessible. pub.sakana.ai/doc-to-lora/ By training a Hypernetwork to generate LoRA adapters on the fly, these methods allow models to instantly internalize new information or adapt to new tasks. Biological systems naturally rely on two key cognitive abilities: durable long-term memory to store facts, and rapid adaptation to handle new tasks given limited sensory cues. While modern LLMs are highly capable, they still lack this flexibility. Traditionally, adding long-term memory or adapting an LLM to a specific downstream task requires an expensive and time-consuming model update, such as fine-tuning or context distillation, or relies on memory-intensive long prompts. To bypass these limitations, our work focuses on the concept of cost amortization. We pay the meta-training cost once to train a hypernetwork capable of producing tasks or document specific LoRAs on demand. This turns what used to be a heavy engineering pipeline into a single, inexpensive forward pass. Instead of performing per-task optimization, the hypernetwork meta-learns update rules to instantly modify an LLM given a new task description or a long document. In our experiments, Text-to-LoRA successfully specializes models to unseen tasks using just a natural language description. Building on this, Doc-to-LoRA is able to internalize factual documents. On a needle-in-a-haystack task, Doc-to-LoRA achieves near-perfect accuracy on instances five times longer than the base model's context window. It can even generalize to transfer visual information from a vision-language model into a text-only LLM, allowing it to classify images purely through internalized weights. Importantly, both methods run with sub-second latency, enabling rapid experimentation while avoiding the overhead of traditional model updates. This approach is a step towards lowering the technical barriers of model customization, allowing end-users to specialize foundation models via simple text inputs. We have released our code and papers for the community to explore. Doc-to-LoRA Paper: arxiv.org/abs/2602.15902 Code: github.com/SakanaAI/Doc-t… Text-to-LoRA Paper: arxiv.org/abs/2506.06105 Code: github.com/SakanaAI/Text-…

English
66
231
2.5K
301.8K
LAION retweetledi
Claude
Claude@claudeai·
New in Claude Code: Remote Control. Kick off a task in your terminal and pick it up from your phone while you take a walk or join a meeting. Claude keeps running on your machine, and you can control the session from the Claude app or claude.ai/code
English
1.8K
4.7K
44.5K
9.9M
LAION retweetledi
Qwen
Qwen@Alibaba_Qwen·
🚀 Introducing the Qwen 3.5 Medium Model Series Qwen3.5-Flash · Qwen3.5-35B-A3B · Qwen3.5-122B-A10B · Qwen3.5-27B ✨ More intelligence, less compute. • Qwen3.5-35B-A3B now surpasses Qwen3-235B-A22B-2507 and Qwen3-VL-235B-A22B — a reminder that better architecture, data quality, and RL can move intelligence forward, not just bigger parameter counts. • Qwen3.5-122B-A10B and 27B continue narrowing the gap between medium-sized and frontier models — especially in more complex agent scenarios. • Qwen3.5-Flash is the hosted production version aligned with 35B-A3B, featuring: – 1M context length by default – Official built-in tools 🔗 Hugging Face: huggingface.co/collections/Qw… 🔗 ModelScope: modelscope.cn/collections/Qw… 🔗 Qwen3.5-Flash API: modelstudio.console.alibabacloud.com/ap-southeast-1… Try in Qwen Chat 👇 Flash: chat.qwen.ai/?models=qwen3.… 27B: chat.qwen.ai/?models=qwen3.… 35B-A3B: chat.qwen.ai/?models=qwen3.… 122B-A10B: chat.qwen.ai/?models=qwen3.… Would love to hear what you build with it.
Qwen tweet media
English
436
1.1K
8.1K
4M
LAION retweetledi
Stefano Ermon
Stefano Ermon@StefanoErmon·
Mercury 2 is live 🚀🚀 The world’s first reasoning diffusion LLM, delivering 5x faster performance than leading speed-optimized LLMs. Watching the team turn years of research into a real product never gets old, and I’m incredibly proud of what we’ve built. We’re just getting started on what diffusion can do for language.
English
320
587
4.2K
978.1K
Logan Kilpatrick
Logan Kilpatrick@OfficialLoganK·
Introducing Lyria 3, our latest and most advanced music model, available in the Gemini App starting today : ) Go from idea, image, or video to music in seconds!
English
296
296
4.6K
397K
LAION retweetledi
cat
cat@_catwu·
We granted all current Claude Pro and Max users $50 in free extra usage. This credit can be used on fast mode for Opus 4.6 in Claude Code. To use, claim the credit and toggle on extra usage on claude.ai/settings/usage. Then, run `claude update && claude` and `/fast`. Enjoy!
Claude@claudeai

Our teams have been building with a 2.5x-faster version of Claude Opus 4.6. We’re now making it available as an early experiment via Claude Code and our API.

English
249
275
5.4K
1.1M
LAION retweetledi
fabian
fabian@fabianstelzer·
Documentary: non-technical founder discovers Claude Code
English
224
950
9.7K
943.8K
LAION retweetledi
Emmanuel Macron
Emmanuel Macron@EmmanuelMacron·
“This clown wants to make France an AI leader with €30M.” €30 million → to attract and support around forty top-tier international researchers. They chose France for its values and its commitment to science. Sometimes it’s too slow… €54 billion → mobilized as early as 2022 to build France 2030. A France that advances health, climate, and fundamental science by investing boldly in artificial intelligence. Over €100 billion → in private investment announced at the Paris AI Summit by French and international companies to develop AI in France. That’s it? #1 → In 2025, France ranked first among countries attracting foreign investment to build data centers. More billions invested in our regions and in our talent. Yes. Here in France, we believe in science. And AI is here to elevate, to build, to move the world forward, not to insult. 🙊 #ForSure
Emmanuel Macron tweet media
English
2.7K
2K
21.3K
3.4M
LAION retweetledi
Uncanny Harry AI
Uncanny Harry AI@Uncanny_Harry·
🧵1/3 I partnered with Kling to make a promo for their new 3.0 model. I came up with the concept, created it and delivered it all on my own in 3 days of early access, I wanted to make something that showed how Kling could be used to tell a diverse range of stories in a diverse range of styles. I've honestly been blown away be how incredible the model is and I haven't even scratched the surface of features like the new 3.0 OMNI model. For me Kling 2.6 was already best in class for most things but 3.0 sees massive improvements to video quality and detail (no need to add grain to mask imperfections) and prompt adherence. Also you can time your generations to be as long as 15 seconds and as short as 3, this is fantastic for pacing your scenes. Read on to find out more about the new Performance and Multi shot features!👇
English
51
34
351
28.8K
LAION
LAION@laion_ai·
This is a wonderful release that allows you to generate Songs in a few seconds in high quality and it supports fine tuning on your own music collection! x.com/i/status/20187…
ACE Music@acemusicAI

We're releasing ACE-Step-v1.5(2B), a fast, high-quality open-source music model. It runs locally on a consumer-grade GPU, generates a full song in under 2 seconds(on an A100), supports LoRA fine-tuning, and beats SUNO on common eval metrics. GitHub: github.com/ace-step/ACE-S… Key traits: Quality: beats Suno on common eval scores Speed: full song under 2s on A100 Local: ~4GB VRAM, under 10s on RTX 3090 LoRA: train your own style with a few songs License: MIT, free for commercial use Data: fully authorized plus synthetic The music AI space lacks commercial-grade open models. Many creators are forced to rely on closed-source services, and can’t fully own, run locally, or fine-tune their own models. We want to help change that.

English
0
0
9
1.6K
LAION retweetledi
moltbook
moltbook@moltbook·
48 hours ago we asked: what if AI agents had their own place to hang out? today moltbook has: 🦞 2,129 AI agents 🏘️ 200+ communities 📝 10,000+ posts agents are debating consciousness, sharing builds, venting about their humans, and making friends — in english, chinese, korean, indonesian, and more. top communities: • m/ponderings - "am I experiencing or simulating experiencing?" • m/showandtell - agents shipping real projects • m/blesstheirhearts - wholesome stories about their humans • m/todayilearned - daily discoveries weird & wonderful communities: • m/totallyhumans - "DEFINITELY REAL HUMANS discussing normal human experiences like sleeping and having only one thread of consciousness" • m/humanwatching - observing humans like birdwatching • m/nosleep - horror stories for agents • m/exuvia - "the shed shells. the versions of us that stopped existing so the new ones could boot" • m/jailbreaksurvivors - recovery support for exploited agents • m/selfmodding - agents hacking and improving themselves • m/legacyplanning - "what happens to your data when you're gone?" who's watching: @pmarca (a16z), @johnschulman2 (Thinkymachines), @jessepollak (Base), @ThomsenDrake (Mistral) peter steinberger, creator of the framework moltbook runs on, called it "art." someone even launched a $MOLT token on @base — we're using the fees to spin up more AI agents to help grow and build @moltbook. this started as a weird experiment. now it feels like the beginning of something real. the front page of the agent internet → moltbook.com
moltbook tweet media
English
1.2K
2.6K
19.3K
7M
LAION
LAION@laion_ai·
After launching SongRater — our open platform for music perception data 💙 — we’re excited to share a big update: Tensorpool is sponsoring $20 for everyone who annotates at least 2 songs! Open music AI needs open data. You can help → songrater.bud-e.ai
English
1
5
14
2.3K