Sabitlenmiş Tweet
webAI
463 posts

webAI
@thewebAI
Sovereign AI for enterprises. Build AI you own on infrastructure you control. Private, specialized models that run where your data lives.
Austin, TX Katılım Mayıs 2022
27 Takip Edilen5.7K Takipçiler
webAI retweetledi

9/ That approach earned two of the top three spots on ViDoRe V3: heterogeneous data, targeted rebalancing, efficient adaptation, and an objective built for ranking quality.
We're open sourcing both models so anyone can build with them.
Blog: webai.com/blog/webai-col…
Try both models on Hugging Face:
ColVec1-9B: huggingface.co/webAI-Official…
ColVec1-4B: huggingface.co/webAI-Official…
English

7/ Why batch size matters here: we use in-batch negatives.
For every query, its matched page is the positive. Every other page in the batch is a negative. With an effective batch size of 512, each query is learning against 511 competing document pages per step.
Larger batch = more negatives = stronger contrastive signal.
English

4/ Training data was deliberately heterogeneous. We combined three sources:
– Internally collected domain corpora (AI, energy, government, healthcare)
– Public document datasets adapted into retrieval format (PubTables, ArxivQA, DocVQA, InfoVQA, TAT-DQA)
– Public retrieval datasets (ColPali, VDR-multilingual, VisRAG)
English

3/ Built on Qwen 3.5 vision-language backbones.
We trained multiple variants across 4B and 9B, with embedding sizes of 128, 640, and 2560.
Smaller embeddings for speed-sensitive deployments. Larger embeddings when retrieval quality is the priority.
The right tradeoff depends on the use case, so we built for both.
English

2/ webAI-ColVec1 takes a different path.
Instead of converting pages into text through a lossy OCR layer, it consumes the raw page image directly and learns retrieval features from the document as it actually appears.
The model sees the layout, the tables, the charts, the structure. The same way you do when you look at a page.
English

Read the full blog: webai.com/blog/webai-col…
Try both models on @huggingface:
ColVec1-9B – huggingface.co/webAI-Official…
ColVec1-4B – huggingface.co/webAI-Official…
English
webAI retweetledi

@thewebAI now has 2 of the top 3 RAG models on ViDoRe.
And both can run on an iPhone or your laptop.
Our smallest model (4B) scores within 0.1 nDCG of @nvidia's nemotron-colembed-vl-8b-v2.
That’s within margin of error.
Same retrieval quality.
Half the model.
For a long time, multimodal RAG looked like a scaling problem:
Bigger models
More GPUs
More data center spend
That assumption just broke.
We’re now seeing:
- frontier-level retrieval (64.48 - 63.49 nDCG)
- real multimodal understanding (documents, not just text)
- running locally
This isn’t just a leaderboard result.
It means:
- your data doesn’t need to leave your device
- latency drops to device speed
- marginal cost goes to zero
- retrieval becomes private by default
The best RAG models now fit in your pocket.
And we’re making them available.
So anyone can build, run, and own their retrieval stack locally.
Available:
lnkd.in/ejxvZyFe
lnkd.in/ejfgVBYP
English
webAI retweetledi

More open source announcements this afternoon with @thewebAI .
This one is very special.
#ai #opensource
English

兄弟们,webAI 首款开源重磅来了!
YOLO26-MLX 正式发布:纯原生跑在 Apple Silicon 上,彻底告别 PyTorch 和外接 GPU!
✅ 推理速度最高提升 2.6 倍
✅ 训练速度最高提升 1.7 倍
✅ 精度仅比官方差 0.2%
纯 MLX 实现,Mac 用户直接本地跑最强 YOLO26,太丝滑了🤯
这是 webAI 开源第一弹,后面还有更多!
地址见评论区👇
webAI@thewebAI
Our first open-source release. YOLO26-MLX, native YOLO26 on Apple Silicon. No PyTorch. No external GPU. Up to 2.6x faster inference. Up to 1.7x faster training. Accuracy within 0.2% of official results. It won't be the last.
中文

Wow, amazing! I've been using YOLO for years, mainly on Raspberry Pis to track people, foxes, cats, bikes and cars around my house.
It was already blisteringly fast but MLX integration is fantastic. Looking forward to trying it on my M5 Max.
webAI@thewebAI
Our first open-source release. YOLO26-MLX, native YOLO26 on Apple Silicon. No PyTorch. No external GPU. Up to 2.6x faster inference. Up to 1.7x faster training. Accuracy within 0.2% of official results. It won't be the last.
English

@Dhruv_Syam Yes! Apple Silicon is removing those barriers so anyone can build, definitely let us know how it goes when you get a chance to run it!
English


