TensorChord

141 posts

TensorChord banner
TensorChord

TensorChord

@TensorChord

We build cloud-native AI infrastructure.

Development ➡️ Production Inscrit le Haziran 2022
149 Abonnements904 Abonnés
TensorChord retweeté
Vectorize
Vectorize@Vectorizeio·
With the 0.4.11 release, Hindsight now supports VectorChord (vchord) from @TensorChord, a high-performance, open-source PostgreSQL extension for similarity search
Vectorize tweet media
English
0
4
7
379
TensorChord retweeté
Intelligent Internet
Intelligent Internet@ii_posts·
II-Commons  Infrastructure for shared knowledge.  Transparent. Distributed. Open source.  The foundation for trustworthy AI.
Intelligent Internet tweet media
English
1
21
74
31.8K
TensorChord retweeté
Earth Genome
Earth Genome@EarthGenome·
Building Earth Index as a searchable planet has always been ambitious & we've made sure to use the right tools to make this a reality. We're happy to use @TensorChord's PostgreSQL extension to enable environmental action at this scale. Read more here➡️blog.vectorchord.ai/3-billion-vect…
Earth Genome tweet media
English
0
1
7
568
TensorChord retweeté
Phil Eaton
Phil Eaton@eatonphil·
> First, we’re introducing explicit functions that treat external object storage locations as first-class data citizens. Second, we’re integrating the ability to run LLMs directly within the Postgres platform. Interview with my senior coworker, Torsten. odbms.org/2024/10/on-edb…
Phil Eaton tweet media
English
2
4
23
3.7K
Daniel Kaiser
Daniel Kaiser@spectate_or·
> find new promising vector search extension (pgvecto.rs) for postgres > convert all 20m vectors into its format > learn that the extension has a memory leak > can't convert the vectors back into pgvector format due to OOM > need to drop the column and recompute all embeddings
GIF
English
1
0
3
173
TensorChord retweeté
Ce Gao
Ce Gao@gaocegege·
We're excited to announce the release of pg_bestmatch.rs, a PostgreSQL extension that brings the power of Best Matching 25 Score (BM25) text queries to your database, enhancing your ability to perform efficient and accurate text retrieval. blog.pgvecto.rs/pgbestmatchrs-…
English
1
3
21
2K
TensorChord
TensorChord@TensorChord·
The results are conveniently visualized in the terminal using @textualizeio. The reranking step has dramatically improved quality of the top-k candidates. You can try different queries to see which retrieval method suit your requirements better.
TensorChord tweet media
English
0
3
9
1.6K
TensorChord retweeté
LlamaIndex 🦙
LlamaIndex 🦙@llama_index·
There’s thousands of RAG techniques and tutorials, but which ones perform the best? ARAGOG by Matous Eibich is one of the most comprehensive evaluation surveys on advanced RAG techniques, testing everything from “classic vector database” to reranking (@cohere, LLM) to MMR to @llama_index native advanced techniques (sentence window retrieval, document summary index). The findings 💡: ✅ HyDE and LLM reranking enhance retrieval precision ⚠️ MMR and multi-query techniques didn’t seem to be as effective ✅ Sentence window retrieval, Auto-merging retrieval, and the document summary index (all native @llama_index techniques) offer promising benefits in either retrieval precision and answer similarity! (And also interesting tradeoffs). It’s definitely worth giving the full paper a skim. Check it out: arxiv.org/pdf/2404.01037…
LlamaIndex 🦙 tweet media
English
5
153
645
125.1K
TensorChord retweeté
Simon Willison
Simon Willison@simonw·
TIL about binary vector search... apparently there's a trick where you can take an embedding vector like [0.0051, 0.017, -0.0186, -0.0185...] and turn that into a binary vector just reflecting if each value is > 0 - so [1, 1, -1, -1, ...] and still get useful cosine similarities!
English
65
143
1.6K
271K
TensorChord
TensorChord@TensorChord·
It is important to note that this is closely tied to the embedding model being used. The benchmark we conducted specifically focused on OpenAI's text-embedding-3-large model. Maybe some binary embedding models have better performance. e.g. @cohere @JinaAI_
Simon Willison@simonw

TIL about binary vector search... apparently there's a trick where you can take an embedding vector like [0.0051, 0.017, -0.0186, -0.0185...] and turn that into a binary vector just reflecting if each value is > 0 - so [1, 1, -1, -1, ...] and still get useful cosine similarities!

English
1
2
24
4.8K