ollama

7.7K posts

ollama banner
ollama

ollama

@ollama

https://t.co/1JpLwJ93nX

California, USA Beigetreten Ağustos 2023
10 Folgt136K Follower
Angehefteter Tweet
ollama
ollama@ollama·
Ollama is now updated to run the fastest on Apple silicon, powered by MLX, Apple's machine learning framework. This change unlocks much faster performance to accelerate demanding work on macOS: - Personal assistants like OpenClaw - Coding agents like Claude Code, OpenCode, or Codex
English
242
599
4.9K
579.2K
ollama
ollama@ollama·
Ollama is now updated to run the fastest on Apple silicon, powered by MLX, Apple's machine learning framework. This change unlocks much faster performance to accelerate demanding work on macOS: - Personal assistants like OpenClaw - Coding agents like Claude Code, OpenCode, or Codex
English
242
599
4.9K
579.2K
ollama
ollama@ollama·
@D_Twitt3r Need to wait for now. Sorry! We are getting it in shape.
English
0
0
3
371
D.
D.@D_Twitt3r·
@ollama Will this updated Ollama support other mlx and/or nvfp4 models downloaded from Hugging Face? Or do we need to wait for you to do more adjustments to them and post in your own catalog?
English
2
0
0
441
ollama
ollama@ollama·
@crashev Will be accelerated across all Apple silicon devices.
English
0
0
4
102
Pawel
Pawel@crashev·
@ollama What about M4 Pro ?
English
1
0
0
52
ollama
ollama@ollama·
This results in a large speedup of Ollama on all Apple Silicon devices. On Apple’s M5, M5 Pro and M5 Max chips, Ollama leverages the new GPU Neural Accelerators to accelerate both time to first token (TTFT) and generation speed (tokens per second). note: test was conducted on using Alibaba’s Qwen3.5-35B-A3B model quantized to nvfp4 and Ollama’s previous implementation quantized to q4_K_M using Ollama 0.18. Ollama 0.19 will see even higher performance (1851 token/s prefill and 134 token/s decode when running with int4).
ollama tweet media
English
8
20
298
30.5K
ollama
ollama@ollama·
@John7Istheman Try Ollama's cloud! Water can't touch it. ❤️
English
1
0
12
514
Jonathan Rudderham
Jonathan Rudderham@codeRunnerUK·
@ivanfioravanti @ollama I stopped using Ollama because it wouldn’t download models. It would do a few %, then drop back down, do a few %, then drop back down, rinse and repeat. At least with LM Studio I can just manually download the models and point it at the folder.
English
1
0
0
30
ollama
ollama@ollama·
@tinyblue_dev I don't know your specific usage patterns - but Ollama's cloud's $100 plan offers significantly more usage than Anthropic's 20x Max plan.
English
1
0
2
546
nick
nick@tinyblue_dev·
@ollama You missed the question. The free plan is going to hit the usage limit in 10 minutes for my workload. I have 2x Anthropic's 20x Max plan. For your $100/ month plan IF I never hit your usage limit? I will transfer those funds directly into your pocket and cancel Anthropic.
English
1
0
1
541
nick
nick@tinyblue_dev·
Hey @ollama - give me a 1 day trial of your max plan, the $100.00 a month plan, if I never hit a usage limit today, I will change all of my subscriptions to you.
English
2
0
9
7.2K
ollama
ollama@ollama·
@RamanduLight so sorry! We are working to make MiniMax experience good.
English
1
0
9
1.2K
Radu
Radu@RamanduLight·
Trying to use minimax cloud @ollama - I'm running into lots of errors
Radu tweet media
English
1
0
4
1.2K
Ziwen
Ziwen@ziwenxu_·
Peak hour limits in Claude are brutal now. Used to push 2 hours straight. Now I'm tapped out in under 1. Sonnet blocked. Opus blocked. What's the play here? Only move left is running Codex to survive those 3-4 peak hours daily.
Ziwen tweet media
English
22
2
100
6.8K
ollama
ollama@ollama·
Qwen 3.5 35B will be great! Works well for - coding (Claude Code, Codex, VS Code, etc. ) - building assistants / agents (Pi for excel, OpenClaw, etc.) - general chat // with docs - we are seeing developers building their own integrations for Ollama (over 50k+) now Claude Code: ollama launch claude --model qwen3.5:35b-a3b-coding-nvfp4 OpenClaw: ollama launch openclaw --model qwen3.5:35b-a3b-coding-nvfp4 Chat with the model: ollama run qwen3.5:35b-a3b-coding-nvfp4
English
0
1
1
63
Kweku Amoah
Kweku Amoah@KwekuOnX·
@ollama what models can I run on a M5 Pro with 48GB and what are some use cases
English
1
0
0
18
Ivan Fioravanti ᯅ
Ivan Fioravanti ᯅ@ivanfioravanti·
If you are not yet following @Prince_Canuma do it now! He is the man behind many of the engines powering local AI on your Apple Silicon, leveraging Apple MLX framework. 🚀
English
7
7
86
4.3K
ollama
ollama@ollama·
@purea1go okay, there is only a single model for MLX right now as we add more! ❤️
English
0
0
2
157
Mostafa Adel
Mostafa Adel@purea1go·
@ollama That's good to know. I plan to benchmark inference speeds in Ollama as well.
English
1
0
1
169
ollama
ollama@ollama·
@purea1go This is super cool! Thank you for sharing. Ollama is built on top of MLX, and doesn't use MLX-LM.
English
2
0
4
1.3K
ollama retweetet
John O'Reilly
John O'Reilly@joreilly·
Just tried out new qwen3.5:4b-nvfp4 @ollama model on M1 Max here (in project where it's used with Koog AI agent).....38% faster than qwen3.5:4b (averaged over 5 runs of the agent).
English
5
3
44
8.5K
ollama
ollama@ollama·
@keter_slater Ollama also offers hosted models via Ollama’s cloud. It’s the best place to use open models. Give it a try!!
English
0
0
6
1.9K
Keter Slater
Keter Slater@keter_slater·
solid update but real talk the Apple silicon speed gap between ollama and cloud APIs is still massive for anything beyond casual use. "fastest on Mac" is a different benchmark than production inference at scale. this is great for tinkering tho not quite replacing hosted APIs yet fr
English
1
0
0
2.1K
ollama
ollama@ollama·
@harrycblum No. It means Ollama didn’t detect it installed on your computer.
English
1
0
6
1.4K
ollama
ollama@ollama·
@urieli17 Yes, technically across the board. We still need to enable more model architectures to run
English
1
0
13
4.3K
Uri Eliabayev
Uri Eliabayev@urieli17·
@ollama Wait, what about Neo? You can run 4B models on it. Would this support A18 Pro?
English
1
0
3
4.7K
marcelo
marcelo@zidszopers·
@ollama why are you ignoring hermes-agent?
English
1
0
1
3.2K