
Processing 3B Events/Day: Lessons from Financial Services
A deep dive into how we architected Kafka + Flink pipelines at JPMorgan to handle extreme throughput with exactly-once guarantees.
Loading Portfolio
Senior Data & Backend Engineer at JPMorgan Chase. Crafting high-performance distributed systems and real-time microservices, data pipelines that process petabytes of data daily.

With 8+ years of experience in software engineering, I specialize in designing and building distributed data systems and high-throughput backend services at enterprise scale.
At JPMorgan Chase, I lead the architecture and development of real-time data pipelines processing billions of financial events daily, powering critical risk analytics, trading systems, and regulatory reporting infrastructure.
I'm passionate about system design, data engineering, and building resilient architectures. When I'm not coding, you'll find me writing technical blogs, contributing to open-source, mentoring engineers, or diving deep into distributed systems research.
Lead architect for real-time data pipelines processing 3B+ events/day. Designed event-driven microservices for risk analytics. Reduced data latency from hours to seconds. Manage 15 engineers.
Built core services for DynamoDB handling millions of req/sec. Implemented auto-scaling reducing costs by 40%.
Data Infrastructure team. Built internal ETL frameworks. Optimized MapReduce jobs achieving 60% improvement.
A selection of impactful projects I've designed, built, and shipped at scale.

Event-driven architecture processing 3B+ financial events/day with sub-second P99 latency and exactly-once semantics.

Service mesh managing 200+ microservices with canary deployments and distributed tracing across multi-region.

Unified lakehouse on Databricks + Snowflake serving 500+ analysts. ACID transactions on petabyte-scale. JPM Tech Innovation Award.

ML-powered anomaly detection. 10K+ GitHub stars. Reduced false alarms by 85% across 300+ services.
Thoughts on distributed systems, data engineering, and building at scale.

A deep dive into how we architected Kafka + Flink pipelines at JPMorgan to handle extreme throughput with exactly-once guarantees.

An honest comparison of Apache Iceberg vs Delta Lake based on our real-world migration of 2.5PB of financial data.

Everything I learned about implementing exactly-once processing in production, including the pitfalls that aren't in the docs.

My journey from Senior to Staff Engineer at JPMorgan — the skills gaps, the mindset shift, and what I'd do differently.

We benchmarked both on 100M rows of financial trade data. The results surprised everyone on the team.

What they don't tell you about managing a popular OSS project: issues, PRs, burnout, and how I learned to set boundaries.
The risk analytics team was running batch jobs every 4 hours to calculate VaR and stress tests across the firm's trading portfolio. This created unacceptable gaps in risk visibility. We redesigned the entire pipeline as a real-time stream processing system.
Latency reduction
Events processed daily
Data loss incidents
Infra cost savings
DynamoDB costs were growing unsustainably with traffic. Through intelligent partitioning, adaptive capacity mode, and read-replica optimization, we cut costs dramatically while maintaining sub-10ms P99 reads.
Annual savings
P99 read latency
Availability
Throughput increase
Enterprise data lakehouse that transformed analytics across the firm.
2023Professional certification in distributed systems on AWS.
2022"Processing 3B Events/Day: Lessons from Financial Services."
2023"Real-Time Financial Risk Calculation Using Stream Processing."
2023Contributor to Kafka, Spark, Flink. Maintainer of 3 observability tools.
OngoingVLDB, SIGMOD, and ICDE publications on stream processing and data systems.
2019–2024Prefer email, WhatsApp, or a quick form? I respond within 24 hours.