Jose Sahad

1.1K posts

Jose Sahad banner
Jose Sahad

Jose Sahad

@JoseSahad

Software engineer, mediocre full stack musician, history geek and literature lover / VP Engineering @TigerDatabase (creators of TimescaleDB)

Berlin, Germany Katılım Mayıs 2010
400 Takip Edilen379 Takipçiler
Jose Sahad retweetledi
Mike Freedman
Mike Freedman@michaelfreedman·
Introducing TigerFS - a filesystem backed by PostgreSQL, and a filesystem interface to PostgreSQL. Idea is simple: Agents don't need fancy APIs or SDKs, they love the file system. ls, cat, find, grep. Pipelined UNIX tools. So let’s make files transactional and concurrent by backing them with a real database. There are two ways to use it: File-first: Write markdown, organize into directories. Writes are atomic, everything is auto-versioned. Any tool that works with files -- Claude Code, Cursor, grep, emacs -- just works. Multi-agent task coordination is just mv'ing files between todo/doing/done directories. Data-first: Mount any Postgres database and explore it with Unix tools. For large databases, chain filters into paths that push down to SQL: .by/customer_id/123/.order/created_at/.last/10/.export/json. Bulk import/export, no SQL needed, and ships with Claude Code skills. Every file is a real PostgreSQL row. Multiple agents and humans read and write concurrently with full ACID guarantees. The filesystem /is/ the API. Mounts via FUSE on Linux and NFS on macOS, no extra dependencies. Point it at an existing Postgres database, or spin up a free one on Tiger Cloud or Ghost. I built this mostly for agent workflows, but curious what else people would use it for. It's early but the core is solid. Feedback welcome. tigerfs.io
English
77
106
1.1K
124.5K
Jose Sahad retweetledi
Tiger Data - Creators of TimescaleDB
Postgres isn't broken. Your workload eliminated the quiet period it was designed around. Here's a pattern that trips up even experienced teams. Write latency develops a rhythm you can't explain by traffic. Autovacuum is always running. Maintenance that used to take minutes now takes hours. Indexes, query plans, configs — everything looks correct. Nothing is misconfigured. The problem is architectural. Postgres maintenance was built around valleys. Batch ETL writes for two hours, the database rests and catches up. Continuous ingestion has no valleys. Every maintenance process runs in direct competition with writes. All day. All night. @mattstratton from @TigerDatabase breaks down exactly where the mechanics stop working. The workloads where this hits hardest — IoT, financial feeds, observability — share one trait: the data source runs on its own schedule, independent of what the database needs. Every quarter optimizing within the wrong architecture is a quarter where migration gets harder. Full breakdown from Matty: tsdb.co/1jjv7yx2
English
0
4
35
1.7K
Jose Sahad retweetledi
Jose Sahad retweetledi
Mike Freedman
Mike Freedman@michaelfreedman·
We just open-sourced pg_textsearch (permissive Postgres license). It's our Postgres extension that brings true BM25 ranking directly into Postgres. If you care about fast, relevance-ranked keyword search without leaving Postgres -- or hybrid retrieval by combining pg_textsearch with pgvector/pgvectorscale -- this is for you. Go 🌟 it on GitHub!
English
22
61
571
37.1K
Jose Sahad retweetledi
Branko
Branko@brankopetric00·
We chose PostgreSQL over MongoDB for our analytics platform. The context: - 50GB of time-series data daily - Complex queries with joins across multiple dimensions - Team had more SQL experience than NoSQL MongoDB seemed obvious for scale, but: - Query complexity made aggregation pipelines unwieldy - Horizontal scaling wasn't needed yet - PostgreSQL's JSON support gave us flexibility - TimescaleDB extension handled time-series perfectly 18 months later: PostgreSQL handles 2TB with sub-second queries. Sometimes boring technology wins.
English
42
41
696
204.2K
Jose Sahad retweetledi
Paul Snively
Paul Snively@JustDeezGuy·
I remain a PostgreSQL maximalist for reasons like this. If you aren’t a FAANG, you almost certainly don’t need a “document DB” or a “hot read cache” in front of PostgreSQL… if you actually use PostgreSQL (i.e. don’t make the other popular mistake and “abstract away from it” because “you might switch databases.”)
Branko@brankopetric00

We chose PostgreSQL over MongoDB for our analytics platform. The context: - 50GB of time-series data daily - Complex queries with joins across multiple dimensions - Team had more SQL experience than NoSQL MongoDB seemed obvious for scale, but: - Query complexity made aggregation pipelines unwieldy - Horizontal scaling wasn't needed yet - PostgreSQL's JSON support gave us flexibility - TimescaleDB extension handled time-series perfectly 18 months later: PostgreSQL handles 2TB with sub-second queries. Sometimes boring technology wins.

English
38
42
842
148.5K
Jose Sahad retweetledi
Ajay Kulkarni
Ajay Kulkarni@acoustik·
Tiger 🤝 AWS Big news this week. We're announcing a Strategic Collaboration Agreement (SCA) with AWS: a deep, multi-year partnership across go-to-market, product, and engineering. Why this matters: Postgres has quietly become the backbone of modern data infrastructure. And together with AWS, we’re doubling down on that future. This collaboration accelerates the next era of Postgres-powered applications, from high-scale systems to devices to AI agents, all tightly integrated with AWS. A big step forward for developers. And just the beginning. Let's go! 🐯🚀 accessnewswire.com/newsroom/en/co… cc @TigerDatabase @awscloud
English
1
8
42
3.9K
Jose Sahad retweetledi
Mike Freedman
Mike Freedman@michaelfreedman·
Today we're announcing a Strategic Collaboration Agreement (SCA) between @TigerDatabase and @awscloud. An SCA is a deep, multi-year partnership at the highest levels, with joint go-to-market, technical collaboration, and shared objectives. It reflects the strength of our existing joint customer base and the scale of the opportunity ahead. Together, we're accelerating modern data infrastructure built on Postgres -- a unified platform for developers, devices, and AI agents, with deep integrations across AWS analytics and AI services. Excited for what's to come! 🚀
English
2
2
15
1.5K
Jose Sahad retweetledi
Tiger Data - Creators of TimescaleDB
☁️ Announcing Tiger Cloud on Azure While Azure Postgres has long embedded TimescaleDB Apache 2 Edition, this lacks advanced capabilities for query speed and cost savings like compression and hypercore columnar storage. Teams split workloads between Postgres and ADX. Dashboards slowed down at scale. Today that changes. Tiger Cloud on Azure brings the full TimescaleDB experience to Azure as a managed service: ✅ Hypercore columnar storage for fast analytics ✅ Incremental continuous aggregates for real-time rollups ✅ Native compression that actually works ✅ Hyperfunctions for advanced time-series operations ✅ Stay 100% in PostgreSQL—no KQL, no pipelines Initial benchmarks between Tiger Cloud vs Azure Database for PostgreSQL show significant improvements for workloads mixing recent and historical data: 225× faster queries, 95% average compression, sub-10ms response times. Available now in: ▪️ East US 2 (Virginia) ▪️ West Europe (Amsterdam) More regions are coming based on demand. Ready to get started? Signup for a free trial at tigerdata.com. Enter your credit card directly (or get invoice billing if needed) via your Tiger Cloud console. Alternatively, if you want to pay through Azure Marketplace, select from two options: Pay-As-You-Go and Annual Commit. Your existing Azure setup stays the same. With Tiger Cloud on Azure, you no longer have to choose between fast analytics and simple architecture. Start your free trial: tsdb.co/az219rs5
English
2
1
8
4.2K
Jose Sahad retweetledi
Mike Freedman
Mike Freedman@michaelfreedman·
Thanks for the post, @AlexMillerDB. Haven’t shared about Fusion before, so here’s the background: Disaggregated object stores are the default for data lakes and lakehouses. But their interaction with analytics formats (e.g., Parquet, Iceberg) often leaves performance suboptimal: erasure coding stripes data into fixed blocks that cut across row-groups or column chunks. As a result, selective queries must reassemble data across nodes before doing useful work, undermining pushdown and degrading tail performance. In Fusion (ASPLOS ’25), my research group @Princeton explored a new design for analytics object stores: – File-format–aware coding (FAC): Co-design erasure coding with analytics layouts so row-groups and column chunks stay local. – Variable-size stripes with bin packing: Align code blocks to file boundaries while adding only ~1.2% overhead. – Cost-based pushdown: Decide when to compute in place vs. ship compressed blocks, based on selectivity and compressibility. Results: On TPC-H, median latency drops 64% and p99 drops 81%; on production SQL workloads, up to 40% and 48%. Fusion is an academic approach for analytics in lakehouse environments. At Tiger (@TimescaleDB), our focus is operational and transactional workloads on Postgres. While this leads to different needs and challenges, we've lately been guided by a similar principle: how can we rethink the storage substrate around the workloads it serves. Finally, proud to note: @ashwiniraina, an author of Fusion, joined TigerData last year and helping drive some big new storage efforts. More, very soon.
Alex Miller@AlexMillerDB

[ASPLOS'25] Fusion: An Analytics Object Store Optimized for Query Pushdown cs.princeton.edu/~mfreed/docs/f… Tightly integrating an Iceberg catalog with an object store means that one could make file-format aware erasure coding decisions, to permit pushing down filters and aggregations.

English
0
8
10
1.6K
Jose Sahad retweetledi
Mike Freedman
Mike Freedman@michaelfreedman·
I’m hiring a PM who will report directly to me (CTO/cofounder) at @TigerDatabase. This is a founder’s PM role: ⚡️ Entrepreneurial, fast-moving 🤖 Deep in AI + infra 🤝 Works directly with partners 🔥 Builds, ships, and makes things happen Remote co, but I want you in SF. Help define the future of Postgres + AI-native infra: tigerdata.com/careers/542068…
English
0
6
28
9.5K
Jose Sahad retweetledi
Tiger Data - Creators of TimescaleDB
“We tried to replicate what TimescaleDB does—behind the scenes—in MongoDB. It sort of worked… but also didn’t.”  — Lead engineer @ Evergen They wired together Kafka, daily backfills, and one-day buckets. All just to make MongoDB usable for time-series. Then they switched to TigerData and got: ✅ Smooth ingestion at scale ✅ Dashboards under 500ms ✅ Built-in compression, retention, tiering ✅ Half the infra complexity They wrote about the whole migration: tsdb.co/zrspspzz Post your worst MongoDB time-series hack below. We’ve probably already fixed it.
Tiger Data - Creators of TimescaleDB tweet media
English
1
5
7
733
Jose Sahad retweetledi
Tobias_Petry.sql
Tobias_Petry.sql@tobias_petry·
My free course about fast analytics with @TimescaleDB is near - after a long time. The design is almost finished and I am beautifying the first two modules now for a release at start of September. So happy the first things will finally be released 😃
Tobias_Petry.sql tweet media
English
7
6
53
4.1K
Jose Sahad retweetledi
Tiger Data - Creators of TimescaleDB
RTABench was designed to reflect real-world analytics workloads. Think: ➡️ Highly selective queries ➡️ Complex joins ➡️ JSONB expressions ➡️ Pre-aggregations and filters you'd actually write If you want to understand how Postgres handles real analytical queries—not synthetic scan-everything tests—RTABench is a great place to start. Really enjoyed this deep dive by Andrei Lepikhov, who walks through RTABench query 0 and the nuances of Postgres performance tuning: 🔗 open.substack.com/pub/danolivo/p… We built RTABench to be open, extensible, and grounded in reality. Excited to see more folks experimenting and sharing what they learn and contribute. 👉 github.com/timescale/rtab…
English
0
2
2
167
Jose Sahad retweetledi
Mike Freedman
Mike Freedman@michaelfreedman·
Having spent many years in academia, surrounded by amazing scientists, it's awesome to learn how @TigerDatabase is enabling Big Physics. @CERN, working with @Siemens, has adopted TimescaleDB to power next-gen SCADA monitoring of the Large Hadron Collider (LHC). 🐯❤️⚛️
Mike Freedman tweet media
English
2
4
10
708