Emanuil Rusev

272 posts

Emanuil Rusev banner
Emanuil Rusev

Emanuil Rusev

@erusev

Entrepreneur, former child, 1 in 8 billion.

Katılım Mart 2008
552 Takip Edilen430 Takipçiler
Emanuil Rusev retweetledi
Georgi Gerganov
Georgi Gerganov@ggerganov·
Introducing LlamaBarn — a tiny macOS menu bar app for running local LLMs Open source, built on llama.cpp
Georgi Gerganov tweet media
English
21
64
784
53.3K
Xuan-Son Nguyen
Xuan-Son Nguyen@ngxson·
WIP: using multiple models at the same time with llama-server 🦙
English
3
2
24
2.8K
Anthem
Anthem@Artemu78Artem·
@iddar @ggerganov @erusev I've downloaded and run a quantized version of model with LlamaBarn, now i want to run llama server with "api" key. I need to know a model name for that, to run the same model without new download?
English
1
0
0
38
Georgi Gerganov
Georgi Gerganov@ggerganov·
LlamaBarn v0.10.0 (beta) is out - feedback appreciated
Georgi Gerganov tweet media
English
16
15
214
35.7K
Venkat Mamilla
Venkat Mamilla@MamillAI·
@erusev @fishright @ggerganov I thought the mention of the API endpoint http://127.0.0.1:8080/v1 should be helpful for newbies like me. Okay, maybe it's not necessary.
English
1
0
0
32
Emanuil Rusev
Emanuil Rusev@erusev·
@MamillAI @fishright @ggerganov I'm afraid I still don't understand. Can you share a specific use case that you have in mind? What is it that you would like to achieve? Thanks!
English
1
0
0
105
Emanuil Rusev
Emanuil Rusev@erusev·
@iddar @ggerganov macOS often deletes files in /tmp, perhaps this is what happened. Can you try to run it again and see if the .log file appears? Also, is this Qwen3-VL-specific or does it freeze for other models as well? Thanks 🙏
English
1
0
0
50
psyv
psyv@psyv282j9d·
@ggerganov @erusev “Curated list” I see the model downloader, and the memory calculator, but nothing jumps out at that says “don’t worry, you won’t fall into the ollama Modelfile/hf repo disconnect hell”
English
1
0
0
207
Emanuil Rusev
Emanuil Rusev@erusev·
@fishright @ggerganov Just pushed a fix for this — this is what first launch is going to look like in the next version.
Emanuil Rusev tweet media
English
1
1
12
3.4K
Emanuil Rusev
Emanuil Rusev@erusev·
@iddar @ggerganov Could you please take a look at /tmp/llama-server.log and see if you can spot any issues there? It might give us some clues about why it froze.
English
1
0
0
44
Iddar Olivares
Iddar Olivares@iddar·
@ggerganov @erusev I installed it from brew and when I tried to launch Qwen3-VL 2B on a 24GB Macbook Pro M4, it froze. Using llama.cpp compiled from Git works fine.
English
3
0
0
326
Emanuil Rusev
Emanuil Rusev@erusev·
@jayrodge15 @ggerganov LlamaBarn doesn't replace webUI, it builds on top of it — it's a thin wrapper of llama.cpp — when you run a model in LlamaBarn it starts the llama.cpp server and the llama.cpp webUI.
English
1
0
3
262
Emanuil Rusev
Emanuil Rusev@erusev·
@kanwisher @ggerganov The idea is to make it easy to run an LLM on your device and then connect that LLM to whatever you want — similar to how you connect to a Wi-Fi network and use that connection in any app you want.
English
1
0
0
46
Emanuil Rusev
Emanuil Rusev@erusev·
@kanwisher @ggerganov > Can we hook in some tools Can you elaborate on that? It runs the llama.cpp server and while running, you can connect to it via the same oai-compatible api.
English
1
0
0
80
Matthew Campbell
Matthew Campbell@kanwisher·
A couple quick feedback 1) lots of people mentioned finding it 2) An initial onboarding screen that tells the user what it does and installs their first model 3) For the web gui there a few missing features that are in librechat.ai I would copy 3a) Artifacts for code, so if I have 10 files they group together 3b) forking discussion 4) Can we hook in some tools? 4a) Memory 4b) Websearch 4c) Rag 5) Mcps? It looks like a good first start. Look forward to it taking over other webuis
English
2
0
2
1.2K