Latest

6/recent/ticker-posts

Header Ads Widget

How S3 is Built ☁️, Fixing Vendor SQL 🦆, Notebooks Without Setup 🐍

Amazon S3 processes hundreds of millions of transactions per second, manages over 500 trillion objects, and operates across hundreds of exabytes ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌  ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ 

TLDR

Together With Aquata

TLDR Data 2026-01-26

Aquata: batteries-included data platform for financial services (Sponsor)

Finance is the OG data-driven industry... but it's often running on a data stack that's held together with duct tape and breaks at the smallest schema change.

With Aquata, you get one cloud-native data platform purpose-built for financial services. It's used by asset managers & data teams at hedge funds, private markets, and capital markets to deliver faster insights - with total confidence in every figure.

  • Hundreds of built-in data models + thousands of attributes.
  • Connectors to custodians, fund admins, and market data providers. 
  • Governed data pipelines for APIs, databases, CSV, and cloud sources.

Try it free for 30 days. Secure, single tenant, no payment required. Start here

📱

Deep Dives

How AWS S3 is built (78 minute podcast)

Amazon S3 processes hundreds of millions of transactions per second, manages over 500 trillion objects, and operates across hundreds of exabytes with 11 nines of durability, achieved through auditor microservices and automated repair systems. Recent architectural advances include a near-total Rust rewrite for core pathways, rigorous formal methods for correctness, and rollout of new primitives like S3 Vectors supporting 20 trillion vectors per bucket with sub-100ms queries. S3's design emphasizes simplicity at scale, crash consistency, proactive defense against correlated failures, and engineering practices where increased scale enhances reliability and performance.
How I Reduced AI Token Costs by 91% with Semantic Tool Selection and Redis (11 minute read)

An enterprise AI platform eliminated the need to include all 70+ tool definitions in every LLM prompt by using multi-component embeddings stored in Redis and selecting only the 2–5 most relevant tools per query. This reduced input tokens by 91.5%, cut total costs by ~49%, boosted precision@3 to 95%, and improved response times by 31%.
High-Risk, High-Scale: Guaranteeing Ad Budget Precision at 1 Million Events/Second (5 minute read)

Flipkart Ads processes over 1 million events per second with a horizontally scalable, stateful stream-processing architecture that prioritizes rapid spend enforcement while ensuring strict budget control. The system employs Flink-managed state for distributed deduplication and key-based idempotency, watermarking to manage temporal skew due to delayed mobile events, and a Lambda architecture that separates sub-second enforcement from batch reconciliation for financial accuracy.
🚀

Opinions & Advice

MCP is Not the Problem, It's your Server: Best Practices for Building MCP Servers (10 minute read)

Many MCP servers fail not because of the protocol itself but due to poor design, treating MCP as mere REST API wrappers instead of thoughtful user interfaces for AI agents. To succeed, focus on high-level outcome-oriented tools, ruthlessly curate tools, use clear prefixed names for discoverability, and paginate large results to prevent context overload.
Path forward for Data Governance: Existence Over Essence (8 minute read)

Traditional compliance and documentation models fail as automation and autonomous agents scale, making embedded, computable governance essential for observable, auditable, and context-sensitive oversight. Data governance must evolve from rigid, essence-based frameworks to a dynamic, socio-technical discipline that enables real-time negotiation of meaning, strategic direction, and continuous accountability within distributed, AI-driven environments.
The Desperate Need For An "Agent Contract" (6 minute read)

The AI Agent Contract is a framework that establishes explicit, enforceable agreements between data producers, engineers, and AI agents to ensure reliable inputs, consistent definitions, and controlled changes, preventing common failures from schema drifts, tool breakage, or mismatched expectations.
💻

Launches & Tools

Your pipelines work great for 3 data sources. Will they work with 30? (Sponsor)

Fivetran helps you manage your data stack, no matter how much it grows. With 700+ prebuilt connectors, you can set up fully managed, self-healing pipelines in just a few steps. Fivetran handles: schema drift and pipeline maintenance. You handle: strategy, architecture, and AI. Try it free today — no credit card needed.
pynb (Tool)

pynb is a macOS tool for running Python notebooks without kernels or environment setup, using plain .py files that are git friendly and scalable to large datasets. It supports SQL alongside Python, works with your existing ChatGPT subscription and agents, and keeps all data local.
Benchmarking 1B Vectors with Low Latency and High Throughput (5 minute read)

ScyllaDB Vector Search achieves industry-leading p99 latencies as low as 1.7 ms and throughput up to 252,000 QPS on billion-scale datasets, demonstrated using the yandex-deep_1b benchmark with 96-dimensional vectors. The architecture co-locates structured and vector data, supporting hybrid queries. Now generally available, upcoming enhancements include native filtering, quantization, and optimized hybrid retrieval.
Feast Joins the PyTorch Ecosystem: Bridging Feature Stores and Deep Learning (5 minute read)

Feast, an open-source feature store, has joined the PyTorch Ecosystem, taming the training-serving skew by ensuring models receive identical feature transformations in both development and production. With unified APIs, point-in-time joins, support for Spark/Snowflake/Flink, OpenTelemetry observability, and RBAC-based governance, Feast empowers teams to maintain data consistency and lineage for large-scale AI deployments.
Sirius (GitHub Repo)

Sirius is a GPU-native SQL engine that plugs into existing query engines (currently DuckDB, with Doris "coming soon") via Substrait, so teams can offload execution to GPUs without rewriting SQL or rebuilding the whole platform. Built on NVIDIA CUDA-X, it reports ~10x TPC-H (on SF=100) speedup at similar on-demand cost. Today, it accelerates a limited operator/type set (joins, group-bys, etc.; common scalar types) and falls back to CPU for unsupported features.
🎁

Miscellaneous

The Joy of SQL - If Properly Implemented (18 minute read)

SQL's frustrations mostly come from poor vendor implementations rather than the language itself. DuckDB shows how better defaults, simpler syntax, and improved tooling can make SQL feel enjoyable and productive.
Optimizing Data Transfer in Distributed AI/ML Training Workloads (15 minute read)

Profiling GPU-to-GPU data transfer in distributed training reveals stark differences in performance between PCIe and NVLink interconnects. NVLink impacts throughput by only 8% versus over 6x slowdown for PCIe. Techniques like gradient compression, optimized memory usage, and parallelized reduction drive substantial performance and cost gains, illustrating the value of regular profiling with NVIDIA Nsight™ Systems for all AI/ML teams.

Quick Links

Small Kafka: Tansu + SQLite on a free t3.micro (8 minute read)

You can run a Kafka-compatible Tansu broker cheaply on AWS free tier using SQLite.
Anthropic Economic Index report: economic primitives (43 minute read)

Anthropic's latest Economic Index reveals that Claude AI usage is increasingly concentrated in high-value software, administrative, and educational tasks, with enterprise API requests now 74% work-related and 52% focused on coding/data workflows.

Love TLDR? Tell your friends and get rewards!

Share your referral link below with friends to get free TLDR swag!
Track your referrals here.

Want to advertise in TLDR? 📰

If your company is interested in reaching an audience of data engineering professionals and decision makers, you may want to advertise with us.

Want to work at TLDR? 💼

Apply here, create your own role or send a friend's resume to jobs@tldr.tech and get $1k if we hire them! TLDR is one of Inc.'s Best Bootstrapped businesses of 2025.

If you have any comments or feedback, just respond to this email!

Thanks for reading,
Joel Van Veluwen, Tzu-Ruey Ching & Remi Turpaud


Manage your subscriptions to our other newsletters on tech, startups, and programming. Or if TLDR Data isn't for you, please unsubscribe.

Post a Comment

0 Comments