Datawizz AI

27 posts

Datawizz AI banner
Datawizz AI

Datawizz AI

@datawizzai

Datawizz helps companies transition to Specialized Language Models

Katılım Şubat 2025
5 Takip Edilen122 Takipçiler
Datawizz AI
Datawizz AI@datawizzai·
Launching Continuous Learning today. Training and runtime shouldn't be separate worlds. Now they're not. Every online eval, user feedback signal and agent success or failure continuously improve the model. Let you models get smarter as they run. Link in first comment
English
1
1
4
568
Datawizz AI retweetledi
Iddo Gino 🐙
Iddo Gino 🐙@iddogino·
OpenAI just released GPT 5.1 in the API with adjustable reasoning levels—from none to low, medium, and high. But when should you use reasoning, and how much? We used Datawizz to evaluate GPT-5.1 and found that more reasoning isn't always better! ...1/n
English
1
2
4
6.2K
Datawizz AI retweetledi
Iddo Gino 🐙
Iddo Gino 🐙@iddogino·
After taking some time off post-Rapid, I'm excited to share what I’ve been up to since: @datawizzai! We’ve raised a $12.5M Seed led by @humancapital to make AI 10x cheaper, 2x more accurate and 15x faster by transitioning from LLMs to SLMs. AI is eating the world. But unit economics are eating AI. Looking at the fastest growing AI products, they all share two traits - growing fast, and painful inference bills. General-purpose LLMs are just too expensive to run. A big reason for that is we train LLMs to be good at everything - answer any question, be an expert on any topic. The big labs dub this "generalisation", but for real-world applications, it is unnecessary. In reality - many AI applications need models to be experts in one thing - and do that thing extremely well. Your coding model doesn’t need to memorize ancient recipes for Garum sauce. This is where Datawizz comes in - we sit between the AI applications and automatically create smaller (100x-1,000x) specialized models to handle specific aspects of your work. By focusing the model and combining industry-data in the distillation process - we end up with models that beat SOTA LLMs at a fraction of the cost. We created Datawizz to make AI specialized and scalable. We’re early in the journey, but have already been able to save companies 90%+ on their inference bill and speed up their apps by 10x. Excited to build better AI platforms? Join the Datawizz team (link in first comment)
English
10
14
61
21.9K
Datawizz AI
Datawizz AI@datawizzai·
Are OpenAI's newest models hallucinating more than before? Hallucinations have always been one of the biggest issues plaguing AI deployment. It now seems that this problem is getting worse - not better - with newer AI models - especially powerful reasoning models. The reality is hallucinations are not a bug of LLMs per-se - but rather a byproduct of their core structure. LLMs are statistical token prediction models - they are not built to generate "truth". That means these hallucinations must be addressed at the application layer - in how we prompt AI, extract results and perform quickly check. We've put together a list of some approaches we have deployed with our customers to mitigate LLM hallucinations. Check out the link in the first comment!
Datawizz AI tweet media
English
1
1
3
1.3K
Datawizz AI
Datawizz AI@datawizzai·
We built Prompt Debloat to help visualize which tokens (words / parts of words) have the most (and least) impact on the LLM answers. We use a technique called Token Ablation. How does it work? At every step we remove a token, re-run the prompt and check how the model confidence changes (as measured by average output token logprobs). Removing important tokens dramatically changes the confidence. Removing bloat (like a "could you please") doesn't really change the confidence. This is far from a fool-proof approach, but it provides a good first pass on which parts of your prompt matter and which don't. Check it out here -- promptdebloat.datawizz.ai
English
0
1
2
246
Datawizz AI
Datawizz AI@datawizzai·
How much of the average LLM prompt is just bloat that doesn't impact results? More than 20% it turns out! We built a free tool to help visualize redundant tokens in LLM prompts! Link & examples below!
Datawizz AI tweet media
English
1
2
4
379
Datawizz AI
Datawizz AI@datawizzai·
🚨 Big announcements from OpenAI, Anthropic, Google and Meta this week. Multiple new SOTA model drops in just a few days, including GPT-4.1, o3, o4-mini, Gemini 2.5's crazy 2M context window and the Llama 4 family. Checkout our new model cheat-sheet:
Datawizz AI tweet media
English
0
2
12
8.2K
Datawizz AI
Datawizz AI@datawizzai·
9/9 Selecting the best model is going to be harder than just always defaulting to the newest models. You’ll have to spend more time evaluating models around specific use-cases, and often leverage different models for different use cases inside on application.
English
0
0
0
106
Datawizz AI
Datawizz AI@datawizzai·
8/9 Our prediction is that we’ll quickly see other leading labs releasing more specialized models. Think: - Coding Models - Agent Models (function calling) - Vision / Extraction Models - Large Context Models - Human-sounding models (EQ) Etc…
English
1
0
0
134
Datawizz AI
Datawizz AI@datawizzai·
OpenAI changing strategy with GPT-4.1? OpenAI just released their newest flagship model - GPT 4.1. They notably focused this model on coding, positioning GPT 4.1 as a specialized coding model. Is this a new trend of new OpenAI models being more specialized? 🧵 1/9...
Datawizz AI tweet media
English
1
2
3
384
Datawizz AI
Datawizz AI@datawizzai·
@adriankuleszo It was a tough call between these direction - but this variant does look 🔥
English
0
0
2
58
Adrian
Adrian@adriankuleszo·
Another take on @datawizzai branding. Our team loved this one, but they went with the last option we presented. Sometimes, that happens - what you like might not always resonate with your client's vision. This is exactly why we always create multiple strong concepts. Let them choose what aligns with their goals. All the "rejected" options still make for great portfolio pieces. This one was too good not to share! 😁
Adrian tweet media
English
4
0
65
4.7K