billyG88

2.2K posts

billyG88

billyG88

@billyG881

-ML engineer- "I don't need a reason to grind, I grind because I need it" - BillygG8

Katılım Aralık 2020
428 Takip Edilen110 Takipçiler
billyG88 retweetledi
Red Hat AI
Red Hat AI@RedHat_AI·
What compression looks like on @vllm_project. Same Gemma 4 31B. Red Hat AI's quantized version runs at nearly 2x tokens/sec, half the memory, 99%+ accuracy retained. Open source. Quantized with LLM Compressor. Links in comments. 🙏 @_soyr_ for the 2-minute demo.
English
1
5
31
1.4K
billyG88 retweetledi
David Hendrickson
David Hendrickson@TeksEdge·
🚀 New TTS Model: MOSS-TTS-Nano, a Real-time TTS that runs fully on CPU! 💡I'm always looking for good TTS models to run on an average PC (w/o a GPU) and this is one of them. OpenMOSS is a new 0.1B multilingual model: • No GPU needed, pure CPU inference • Real-time speech generation + streaming • 20 languages supported • 48 kHz stereo quality • Ultra-lightweight & easy local/web deployment Huge value for edge devices, local demos, web apps, and lightweight products where GPU isn’t an option. Part of the MOSS-TTS family 👏
David Hendrickson tweet media
English
1
7
54
2.2K
billyG88 retweetledi
ModelScope
ModelScope@ModelScope2022·
Say hello to MOSS-TTS-Nano 🚀 0.1B multilingual TTS from MOSI.AI and OpenMOSS. Designed for realtime speech generation without a GPU. Runs directly on CPU, keeping the deployment stack simple enough for local demos, web serving, and lightweight product integration. Part of the MOSS-TTS family alongside the 1.7B and 8B flagship models. 🤖 modelscope.cn/models/openmos… 🌍 modelscope.ai/models/openmos… 💻 github.com/OpenMOSS/MOSS-…
ModelScope tweet mediaModelScope tweet mediaModelScope tweet media
English
4
32
209
30.2K
billyG88 retweetledi
Abhimanyu Sharma
Abhimanyu Sharma@0xN1nja·
started with a raspberry pi, now i run an entire AWS region at home
Abhimanyu Sharma tweet media
English
390
670
16K
966.3K
billyG88 retweetledi
Masato Ota
Masato Ota@ottamm_190·
Externalization in LLM Agents: A Unified Review of Memory, Skills, Protocols and Harness Engineering arxiv.org/abs/2604.08224
Masato Ota tweet media
English
7
72
401
28.2K
billyG88
billyG88@billyG881·
@Hesamation I wonder how many AI avenues could benefit from 1T $$$ in investment…
English
0
0
0
68
ℏεsam
ℏεsam@Hesamation·
DHH is in. Karpathy is in. Andrew Ng is in. Terence Tao is in. Linus Torvalds is in. John Carmack is in. Tony with an opinion still believes AI is just a next-token predictor with no real future.
English
87
62
1.4K
136.4K
billyG88 retweetledi
dealign.ai
dealign.ai@dealignai·
MiniMax m2.7 at 83gb. wow.
dealign.ai tweet media
English
12
10
187
10.8K
billyG88 retweetledi
NVIDIA AI Developer
NVIDIA AI Developer@NVIDIAAIDev·
🎉Congratulations to the @MiniMax_AI team on the launch of MiniMax M2.7! MiniMax M2.7 is now available with NVIDIA GPU accelerated endpoints ready to try out with claws including NemoClaw and @OpenClaw. 🦞 📝Get started with our technical guide: developer.nvidia.com/blog/minimax-m… and see how you can begin experimenting for free at build.nvidia.com/minimaxai/mini…. What will you build this weekend? Share in comments. 👇
NVIDIA AI Developer tweet media
MiniMax (official)@MiniMax_AI

We're delighted to announce that MiniMax M2.7 is now officially open source. With SOTA performance in SWE-Pro (56.22%) and Terminal Bench 2 (57.0%). You can find it on Hugging Face now. Enjoy!🤗 huggingface:huggingface.co/MiniMaxAI/Mini… Blog: minimax.io/news/minimax-m… MiniMax API: platform.minimax.io

English
56
159
1.8K
352.5K
billyG88 retweetledi
Roan
Roan@RohOnChain·
This 2 hour Stanford lecture shows exactly how Stanford trains it's engineers to build AI systems. It's more practical than every Claude tutorial & prompting threads you've seen. Bookmark & give it 2 hours, no matter what. It'll be the most productive thing you do this weekend.
English
146
1.8K
12.9K
1.5M
billyG88 retweetledi
Asuka Groyper 🚬
Asuka Groyper 🚬@asukagrypr·
I'm sorry George.
Asuka Groyper 🚬 tweet media
English
164
2.8K
34K
784.7K
billyG88
billyG88@billyG881·
@pfau AXAXAXAXAXAXXAAXXAXAXAXAXAXAXAX
Suomi
0
0
0
4
David Pfau
David Pfau@pfau·
At the current growth rate of 3x every quarter, Anthropic's revenue is on track to surpass Google in Q4 this year, Amazon in Q1 next year and the entire United States federal government somewhere in Q2/Q3.
English
60
52
1.4K
171.3K
billyG88 retweetledi
François Fleuret
François Fleuret@francoisfleuret·
This is really 2019 pure science fiction.
François Fleuret tweet media
English
22
89
2K
230.3K
billyG88 retweetledi
Avid
Avid@Av1dlive·
This 15-minute talk by the creator of Pydantic on how to correctly use MCPs will teach you more about making your AI tools actually work together than everything you've scrolled past this year. Bookmark this & watch, no matter what. Then read the guide below by @eng_khairallah1
Khairallah AL-Awady@eng_khairallah1

x.com/i/article/2042…

English
34
260
2K
344.5K
billyG88 retweetledi
@Axiomofmind ⚡
@Axiomofmind ⚡@axiomofmind·
training in unlsoth - learning which quant size to train - training parameters based on data size - merging LoRA - exporting to gguf / quantized - merging ggufs - running on llama.cpp vs. ollama fun stuff. lots of wasted time but learning by doing is best IMO
English
2
2
44
10.8K
billyG88 retweetledi
Lotto
Lotto@LottoLabs·
I actually think this is the best subscription out now I just used $10 of opus last night or I could have a month of GO $10 a month with generous limits on open models The best of both worlds for qwen 27b + Hermes
Lotto tweet media
English
41
13
366
25.1K
billyG88 retweetledi
Design Arena
Design Arena@Designarena·
BREAKING: Wan2.7-Video by @Alibaba_Wan is now #1 on Video-to-Video Arena with an Elo of 1337! This establishes a new state of the art with video editing models Huge congrats to the @Alibaba_Wan team!
Design Arena tweet media
English
18
31
288
22.9K
billyG88 retweetledi
Fahd Mirza
Fahd Mirza@fahdmirza·
💥 MiniMax M2.7 is NOW running locally — 229B open source model on CPU + GPU ♠ and everyone can do it with llama.cpp 🔹229B MoE model, 256 experts, 8 active per token 🔹IQ4_XS quant brings it down to 100GB — no need for multiple GPUs 🔹Split across H100 VRAM + system RAM via llama.cpp GPU offloading 🔹OpenAI compatible API endpoint — drop-in replacement for any client 🔹SWE-Bench Pro 56.2 — within 2 points of GPT-5.4 🔹MLE-Bench Lite 66.6% medal rate — #2 only to Opus 4.6 and GPT-5 🔹Modified MIT license — run it, build on it, ship it 🔥 Watch the full video below:👇
English
2
1
48
4.5K