Lamini

308 posts

Lamini banner
Lamini

Lamini

@LaminiAI

The LLM tuning & inference platform for enterprises. Factual LLMs. Deployed anywhere.

Katılım Nisan 2023
9 Takip Edilen6.4K Takipçiler
Lamini
Lamini@LaminiAI·
🎯 Aiming for 90%+ accuracy on your Text-to-SQL agent, but can't get past 50%? With our proven methodology, our customers have cracked the code and hit 9s of accuracy! We're spilling the tea 🍵 in our upcoming webinar. Bring your toughest Text-to-SQL questions—we’ve got answers! 💪 🎯 Build high-accuracy Text-to-SQL BI agents 📅 March 20, 2025 🕘 10:00 - 10:45 AM PT 🔗 Register here today: bit.ly/41qIycU
English
1
1
6
2.3K
Lamini
Lamini@LaminiAI·
Join us for a webinar on building Text-to-SQL BI agents. We’ll show how to finetune any open LLM to reach 90%+ accuracy. Register now bit.ly/41qIycU 🎯 Build high-accuracy Text-to-SQL BI agents 📅 March 20, 2025 🕘 10:00 - 10:45 AM PT
English
0
0
3
2.1K
Lamini
Lamini@LaminiAI·
🙌Introducing Memory RAG—a simpler approach to RAG that leverages embed-time compute to create more intelligent, validated data representations. Build mini-agents with a simple prompt. Get the paper: hubs.la/Q0333d5c0
Lamini tweet media
English
0
4
9
1.9K
Noorie
Noorie@nooriefyi·
@LaminiAI intent classification is the key to unlocking truly conversational ai
English
2
0
1
60
Lamini
Lamini@LaminiAI·
Have you seen our Classifier Agent Toolkit 😺 demo yet? Learn how to use our SDK to build a highly accurate Classifier Agent for a customer service chatbot. The agent categorizes customer interactions by intent so it can respond appropriately. You can run multiple evaluations until you reach your desired level of accuracy. youtube.com/watch?v=-wadT3…
YouTube video
YouTube
English
1
2
6
3.5K
Sharon Zhou
Sharon Zhou@realSharonZhou·
I'm so excited to launch @LaminiAI’s Classifier Agent Toolkit, aka. CAT! 🚀🐱 CAT hunts & tags the important signals 🐭 in a vast amount of data — so devs can easily create agentic classifiers. ❌ Manual data labeling ❌ Large, slow general LLM calls that can only handle 20-30 categories with mid accuracy ✅ CAT has helped our customers tag 2,000 pages across 1,000 categories in just 3.6 seconds with 99.9% accuracy. Dev time? A few hours to a few days. Hallucinations? Approaching zero. Meow. Some common agentic classifiers with CAT: ◽️Customer service agents that extract user intent ◽️Finding high severity tickets, so your teams can prioritize urgent issues ◽️Triage legacy application code based on importance, to prioritize development ◽️Analyze sentiment in earnings calls, reviews, posts, surveys, etc. More on it 👉 lamini.ai/classifier Demo from one of our amazing architects Scott youtube.com/watch?v=-wadT3… Happy holidays, hope you like our gift 🎁 Reach out anytime to fill our inbox with cheer at info@lamini.ai (we read, we respond!) This was a huge effort by the entire Lamini team 🎀
YouTube video
YouTube
English
1
7
42
5K
Lamini
Lamini@LaminiAI·
🎁 Our new Classifier Agent Toolkit (CAT 🐱) is here! No more extensive manual data labeling or heavy ML systems. 😻 Build classifier agents that can quickly categorize large volumes of data at 95%+ accuracy / 400k token throughput in under 2 seconds. Watch the demo and get the link to the docs and repo here: lamini.ai/classifier-age…
English
0
0
2
1.1K
Lamini
Lamini@LaminiAI·
🙌 Our new Enterprise Guide to Fine-Tuning is out! If you can't get above 40-50% accuracy with RAG, fine-tuning might be the answer. Learn the basics of fine-tuning and specific applications and use cases. bit.ly/495Q7c9
Lamini tweet media
English
1
1
2
774
Lamini
Lamini@LaminiAI·
🎉🎉🎉 Excited to announce our new pay-as-you-go offering, Lamini On-Demand. Get $300 in free credit to run your tuning and inference jobs on our high-performance GPU cluster. Happy tuning! lamini.ai/blog/lamini-on…
English
0
1
6
1.1K
#include Stan.h
#include Stan.h@StanTechAddict·
@LaminiAI @GregoryDiamos Certainly one of the best article on LLM inference link to hardware capabilities. May I recommend a part 2 showing performance capabilities with different inference tools (llamacpp, vLLM..) and configuration (CUDA, Vulkan, SYCL, ROCm…)?
English
1
0
2
141
Lamini
Lamini@LaminiAI·
LLM inference frameworks have hit the “memory wall”, which is a hardware imposed speed limit on memory bound code. Is it possible to tear down the memory wall? @GregoryDiamos explains how it works in his new technical blog post. lamini.ai/blog/evaluate-…
English
1
3
13
2.6K
Lamini
Lamini@LaminiAI·
Vertical vs. horizontal AI use cases? GitHub Copilot started vertical and crossed over into horizontal applications. Low latency + accuracy were key! Thanks for the great discussion @gajenkandiah and @Hitachi! youtube.com/watch?v=4Wn-rE…
YouTube video
YouTube
English
0
0
5
688
Lamini retweetledi
AI at Meta
AI at Meta@AIatMeta·
🆕 New course on @DeepLearningAI: Improving Accuracy of LLM Applications ➡️ go.fb.me/zfwvd8 Created in collaboration with DL, Meta & @LaminiAI, this free course covers topics like evaluation frameworks, instruction & memory fine-tuning, LoRA + training data generation.
English
11
86
326
34K
Lamini retweetledi
Andrew Ng
Andrew Ng@AndrewYNg·
Learn a development pattern to systematically improve the accuracy and reliability of LLM applications in our new short course, Improving Accuracy of LLM Applications, built in partnership with @LaminiAI and @Meta, and taught by Lamini’s CEO @realSharonZhou, and Meta’s Senior Director of Partner Engineering, @asangani7. (Disclosure: I am an investor in Lamini.) The path to tuning an LLM application can be complex. In this course, you'll learn a systematic sequence of steps for improving accuracy by reducing hallucinations: - Create an evaluation dataset to measure model accuracy - Add prompt engineering and self-reflection - Fine-tune your model including "memory-tuning" which is a new method of embedding facts in an LLM Using the Llama 3-8B parameter model, you will: - Build a text-to-SQL agent with a custom schema and simulate situations where it hallucinates - Understand the difference between instruction fine-tuning, which gives pre-trained LLMs instructions to follow, and memory fine-tuning - See how Performance-Efficient Fine-tuning (PEFT) techniques like Low-Rank Adaptation (LoRA) reduce training time by 100x and Mixture of Memory Experts (MoME) reduces it even further I appreciate Meta releasing the Llama's family of open models -- this course gives an example of the unique type of work that developers can do with such models. Please sign up here: deeplearning.ai/short-courses/…
English
22
126
627
66.4K