
Allen Vailliencourt
5.2K posts

Allen Vailliencourt
@Valien
Solutions Engineer @ Tailscale











LM Studio emailed me yesterday and invited me to test this — I've been using Wireguard and llama.cpp to run large models on my studio workstations, from anywhere. (I run a few *local* LLMs to help with MicroPython, debugging PTP, that sort of thing...) I'm attaching a video showing LM Studio running on my low-spec Mac mini at home, using a large model running on a Dell GB10 at the studio. Also tested it running through my Mac Studio at the studio. They use Tailscale on the backend, but it's separate from Tailscale (managed through LM Studio)—works well, but I still prefer llama.cpp since it's open source. The llama.cpp Web GUI makes it easy enough (for me) to use on the road via VPN, through a browser. I haven't really messed with LM Studio or vLLM before, but I can see the appeal, especially if you link LLMs to external tools (I don't, at least not at this time).



Red Robin has lost ~90% of its value over the last 5 years You can now buy the ENTIRE company for just ~$60 million They used to be one of the most beloved spots for kids, teens, and families... Where did it go wrong for them and can they turn it around?







🔐Tailscale’s at #GartnerIO NA in Vegas (Dec 9–11) 🎲 Come by Booth #515 (Networks & Security). 🎙️Hear Allen Vailliencourt: 12/11 11:00–11:30 · Zeno 4610 🍸Then happy hour w/ @mezmodata + @catchpoint: 12/10 8–10pm at The Venetian! RSVP: mezmo.com/event/the-suga…





