
Fastino Labs
42 posts

Fastino Labs
@fastinoAI
Building the first foundational model for agent personalization.
Palo Alto, CA Katılım Temmuz 2024
78 Takip Edilen272 Takipçiler
Sabitlenmiş Tweet

Introducing GLiNER-2 - Fastino’s next-gen open-source model for unified entity extraction, classification & structured parsing.
• NER, classification & JSON in 1 blazing-fast pass
• ⚡ <150 ms CPU latency
• 🧩 Apache-2.0 + hosted API
Built by @fastino_ai — unveiled live at #EMNLP2025 by @urchadeDS
🔗 github.com/fastino-ai/GLi…
English

Last night we sat down with @l2k, @vanpelt, @scottcjohnston, @george_onx, and @ash_csx to talk SLMs, agents, and inference. Production AI needs specialist models, not one giant generalist.

English

Join us next Monday evening for a conversation with our co-founders, @george_onx and @ash_csx wuth @l2k, co-founder of @wandb (and now @CoreWeave).
We’ll have free food and drinks and a chance to win an NVIDIA Jetson. Hope to see you there.
luma.com/yz9eq2em?utm_s…

English

Most teams treat inference as the last step in deployment.
We think that's backwards.
Join us today at Coding Agents to hear @ash_csx talk about optimizing models in production from the moment they ship.
📍 Compute History Museum, Mountain View
🔗 luma.com/codingagents

English

GLiNER redefined modern Named Entity Recognition.
Our latest blog post explains how.
🔗: fastino.ai/our-research/g…
English
Fastino Labs retweetledi

We’re kicking off an SF hackathon with our friends at @GoSenso this Saturday!
A massive warehouse is buzzing with 400 hackers building all day.
Always a pleasure to feel this kind of on-the-ground energy!
cc @fastinoAI @gladly @Meet_campfire



English
Fastino Labs retweetledi


Early preview from from our researcher @var6595 of a new foundational model for personalization we've been building at Fastino. At the self-evolving agents hack luma.com/agentshack

English
Fastino Labs retweetledi

Fastino Labs retweetledi

#developers, Building an #LLM-powered app, but no GPU? No problem! Check-out @fastinoAI 's latest OSS model!
Fastino Labs@fastinoAI
Introducing GLiNER-2 - Fastino’s next-gen open-source model for unified entity extraction, classification & structured parsing. • NER, classification & JSON in 1 blazing-fast pass • ⚡ <150 ms CPU latency • 🧩 Apache-2.0 + hosted API Built by @fastino_ai — unveiled live at #EMNLP2025 by @urchadeDS 🔗 github.com/fastino-ai/GLi…
English
Fastino Labs retweetledi

Big milestone for our team - GLiNER-2 is live!
One model for NER, classification & structured parsing in a single pass.
<150 ms CPU latency, open-source (Apache-2.0) + hosted API.
🔗 github.com/fastino-ai/GLi…
Incredible work by @urchadeDS and the @fastinoAI research crew after EMNLP 2025 🙌
English
Fastino Labs retweetledi

Fastino Labs retweetledi

The Fastino team hiking around the peak district after a day of coding 🥾🏔️ @fastinoAI #HiringNow #TechJobs #aijobs #SanFranciscoJobs


English
Fastino Labs retweetledi
Fastino Labs retweetledi

@wolfejosh you might want to take a look at @fastinoAI
Josh Wolfe@wolfejosh
i am convinced on-device inference will dominate next and my theory of these case is memory (flash/NAND,etc) players will move in here (SK, Micron) jensen/nvda want u to believe u need giant clusters for inference and that may be true for 50% of your off-device search queries
English

Huge thanks to @mspiro3 and @InsightPartners for helping make our NYC rooftop happy hour a success! Missed @fastinoAI and @george_onx this time?
Sign up for first dibs on our next event:
forms.gle/tCPR2ibRKsWEvn…

English

Fastino trains cutting-edge language models on <$100K of gaming GPUs.
No racks of H100s. No $100M burn.
Just smart engineering.
A new path for enterprise AI—accurate, fast, and cost-effective.
🔗 Tom’s Hardware feature: bit.ly/fastino-news
cc: @tomshardware
English

Not nearly enough people are talking about the implications of Klarna rolling back some of their AI bets.
Not knowing any of the details, I can guess why:
Replacing determinism or humans with probabilistic code is fraught with edge cases and require new ways of software development and process engineering that aren't well solved yet.
The implications to an entire generation of AI "apps" will be severe as more companies come to terms with the difficulty in getting products to work reliably in production with AI in the loop. Customer Service may be the first funeral signpost.
The result will be that many startups will need to pivot to simply use AI for narrow use cases and otherwise act deterministically. So what have people funded then? An AI company? Not really. Just a really overpriced SaaS company at AI valuations.
English

We just dropped our first deep dive on Fastino's TLMs which are purpose-built to outperform generalist LLMs like GPT-4o on high scale enterprise tasks.
🦊 Millisecond latency
🦊 Benchmarked against real-world use cases
🦊 Inference on CPU and low-end GPU
Read the full launch blog here ⬇️
fastino.ai/blog/introduci…
English
Fastino Labs retweetledi

#developers, Need to run an LLM locally, but the usual suspects are too big to fit on your laptop? Check-out @fastinoAI 's task-specific language models (TLMs): they fit on a laptop, no GPU required ... and FREE ;-)!
fastino.ai/blog/introduci…
fastino.ai
English

