J

91 posts

J banner
J

J

@jfgdes

Ignorance is bliss

Katılım Şubat 2014
527 Takip Edilen147 Takipçiler
J
J@jfgdes·
@private_llm 4.99 WEEKLY to use a model locally is outrageous
English
1
0
1
116
Private LLM
Private LLM@private_llm·
We’re flattered by all our copycats. One underhanded trick we repeatedly see a lot of them use is the free forever schtick. Translation: Free until they get enough reviews, and then hoodwink users into usurious weekly subscriptions. We’ll never do that. Rugpulls are not cool.
Private LLM tweet media
English
1
2
18
1.3K
Scott Walsh
Scott Walsh@mallocblock·
@captain0001 @pflodin @pronounced_kyle The audio is streamed over AVDTP. A good audio quality stream is +380kb/s. In such a noisy location that can probably be reduced by 100kb and not be easily noticed.
English
1
0
9
732
Pelucho
Pelucho@basedn7·
Apunten
Pelucho tweet media
Indonesia
18
122
1.1K
29.2K
J
J@jfgdes·
@iknovusnucleo @Franquistaaa @basedn7 Pues como venezolano en España te digo que todos mis familiares (venezolanos hijos de españoles) y amigos venezolanos nacionalizados votan derecha. La putada es q aún así hay mucho venezolano retrasado que no se da cuenta de que el PSOE nos lleva por el mismo camino, es demencial
Español
1
0
0
34
Iknovus
Iknovus@iknovusnucleo·
@Franquistaaa @basedn7 Salvo contadas excepciones (cubanos), el 90% de los panxos votan a izquierda. El otro día mismamente un venezolano me dijo: "Yo sé que Sánchez es un hijoputa pero estoy aquí gracias a él, así que le tengo que votar". No se podía saber.
Español
1
2
9
596
J
J@jfgdes·
@DePasqualeOrg Oh yes, I’m sorry for not clarifying it, I thought it was clear. The app and all the other models work perfectly!
English
0
0
1
21
Anthony
Anthony@DePasqualeOrg·
@jfgdes It looks like you're referring to DeepSeek R1. This will be fixed soon in the next version of Local Chat. I became aware of the error in the current version only after enabling the model.
English
1
0
1
51
Sebastiaan de With
Sebastiaan de With@sdw·
A crisp fall night in Amsterdam. snapped on iPhone 16 Pro.
Sebastiaan de With tweet mediaSebastiaan de With tweet mediaSebastiaan de With tweet mediaSebastiaan de With tweet media
English
27
67
2.5K
176K
J
J@jfgdes·
@dessatel @mweinbach TBF, I’ve seen some implementations (I think it was MLX?) that disregard the NPU and use the GPU and CPU exclusively for better speeds. Would love to see the comparison between mlc-llm, MLX, and llama.cpp, running on both M3 and M4 chips!
English
0
0
0
96
dessatel
dessatel@dessatel·
@mweinbach Not surprising. A lot of cookie-cutter reviews. GPU AI reviews will come first. Apple Neural Engine will be the last to come. I miss AnandTech.
English
1
0
3
704
Max Weinbach
Max Weinbach@mweinbach·
I’m disappointed not a single MacBook Pro review of M4 Max I’ve seen tries AI (transformers, MLX, llama.cpp, CoreML, Diffusers) performance Reviewers, rightfully, are too focused on their workflows but not what I think is the biggest and most useful one for the M4 Max chip!
English
96
44
1.7K
238.2K
Apple Hub
Apple Hub@theapplehub·
Apple Intelligence will be available in the EU starting in April 2025
Apple Hub tweet media
English
211
536
9.4K
626.9K
J
J@jfgdes·
@Depthperpixel @flyosity Exactly! Some don’t seem to realize those exact results are the ones people will love to use
English
0
0
1
34
Lee Higgins
Lee Higgins@Depthperpixel·
@flyosity But that IS ready! I want comically bad AI emojis so bad!
English
1
0
8
592
Mike Rundle
Mike Rundle@flyosity·
When the PM for Genmoji asks if it’s ready to ship
Mike Rundle tweet media
English
10
14
473
34.6K
J
J@jfgdes·
@DePasqualeOrg Will it be recorded? Sadly, I can’t attend even though I’m in Spain😕
English
1
0
0
106
Anthony
Anthony@DePasqualeOrg·
I'll be giving a presentation on MLX for Swift developers at Glovo in Barcelona on November 5. Come by if you're in town! meetup.com/nsbarcelona/ev…
English
2
6
20
4.5K
J
J@jfgdes·
@rossetate Awesome and impressive insight! Additionally it seems to be written in LaTeX?!
English
0
0
0
532
Ross Tate
Ross Tate@rossetate·
As the author of this PDF, it's been interesting seeing people guess at the rationale behind its design. However, the rationale had nothing to do with theory vs practice, and everything to do with pragmatically coping with an unaccommodated disability in academia. (1/16)
Deedy@deedydas

Compilers was was known to be the hardest CS class at Cornell which was hard as it is. We were handed a 8-page PDF at the start of sem for a language spec we'd be implementing by the end of sem, split into 6 parts. On part 5, the median was a 0/100 and most the class failed.

English
127
1.5K
16.3K
3.4M
J
J@jfgdes·
@simonbs Thinking of open-sourcing it? I’ve been looking to improve small/local LLMs’ performance with math queries, and your function implementation seems perfect (alas, I lack programming skills to implement it). Cool stuff!
English
0
0
2
85
Simon B. Støvring
Simon B. Støvring@simonbs·
I'm exploring running LLMs locally on iPhones and Macs. I’ve got Llama 3.2 running locally using Apple’s MLX and with support for tool calling. This example runs two LLMs: one to identify tools to call based on the query and one to generate responses based on the tools’ outputs.
English
20
11
376
48.3K
J
J@jfgdes·
@samhenrigold My 2020 Intel MBP is on 82%, so this could take a while😅
English
0
0
0
27
sam henri gold
sam henri gold@samhenrigold·
alright gang, we got 27 days to get this macbook battery health below 80% so i can get a free battery/bottom case replacement within warranty
sam henri gold tweet mediasam henri gold tweet media
English
228
226
25.1K
1.4M
J
J@jfgdes·
@LMStudioAI Are there plans for Intel Mac support?
English
0
0
0
86
LM Studio
LM Studio@lmstudio·
LM Studio 0.3.4 ships with Apple MLX 🚢🍎 Run on-device LLMs super fast, 100% locally and offline on your Apple Silicon Mac! Includes: > run Llama 3.2 1B at ~250 tok/sec (!) on M3 > enforce structured JSON responses > use via chat UI, or from your own code > run multiple models simultaneously > download any model from Hugging Face Video at 1x speed.
English
42
128
855
171.6K
J
J@jfgdes·
@le_chuck_melee @literallydenis AFAIK, it’s not that simple, as Apple has been designated as a gatekeeper by the EU (Samsung hasn’t). Still sucks, and I think the EU has overstepped its boundaries once again.
English
0
0
1
13
Denis Shiryaev 💙💛
Denis Shiryaev 💙💛@literallydenis·
I’m concerned: ChatGPT Advanced Voice Mode – not available in the EU AI features in iOS – not available in the EU Llama 3.2 model – not available in the EU It looks like this is becoming a standard practice. Does the EU government understand that if they continue to delay introducing AI features to their citizens, there will be: a) Not enough time for EU citizens to learn new technologies, making them less competitive in the world-wide job market b) Long-term negative effects on the IT sector, as AI development will happen elsewhere, hitting the EU economy hard
English
12
17
129
14.1K
Philipp Schmid
Philipp Schmid@_philschmid·
We have GPT-4 for coding at home! I looked up @OpenAI GPT-4 0613 results for various benchmarks and compared them with @Alibaba_Qwen 2.5 7B coder. 👀 > 15 months after the release of GPT-0613, we have an open LLM under Apache 2.0, which performs just as well. 🤯 > GPT-4 pricing is $30/$60 while a ~7-8B model is at $0.09/$0.09 that's a cost reduction of ~333-666x times, or if you run it on your machine, it's “free”.💰 Still Mindblown. Full post about Qwen 2.5 tomorrow. 🫡
Philipp Schmid tweet media
English
19
81
587
109.3K
J
J@jfgdes·
@NathanielIStam @foley2k2 @tosho @mattshumer_ What’s your inference speed (tok/s) when splitting the model between RAM and VRAM? Is it usable? I’ve read about memory swapping (admittedly, to SSD) being super slow but haven’t read much about swapping between VRAM and RAM
English
2
0
0
34
StrayVolts
StrayVolts@StrayVoltage47·
I've run quantized versions of 70B on my dual 3090s, with about half the layers on 64 GB of ram. I'd probably need 128 GB RAM to do full precision. The reason I replied with use LMStudio is you can slit the model between your ram and your Vram, I run Llama3.1 Q8 on my 1660 Ti 6GB with that very often
English
1
0
0
70
Matt Shumer
Matt Shumer@mattshumer_·
We’re looking for a compute sponsor for our 405B run. Happy to give a shout out when we launch it include you in the report, first access for inference, etc. Ideally 64x H100s. Please reach out if you’re serious.
Matt Shumer@mattshumer_

I'm excited to announce Reflection 70B, the world’s top open-source model. Trained using Reflection-Tuning, a technique developed to enable LLMs to fix their own mistakes. 405B coming next week - we expect it to be the best model in the world. Built w/ @GlaiveAI. Read on ⬇️:

English
74
72
1.2K
150.6K
J
J@jfgdes·
@RobDenBleyker They probably should just fine-tune the model so it uses python to count the letters at this point lol
English
0
0
0
11