Architecting and Engineering Data at Scale

Now Streaming

Around six years of architecting and building fault-tolerant, event-driven systems that process millions of transactions across global markets. Currently at Tiger Analytics, developing production solutions for Fortune 500 clients.

Siddhartha Kumar
Open to opportunities
8,000+ concurrent executions·
99.9% reliability·
15+ international markets·
$150K+ annual savings·
85% faster deployments·
10+ enterprise clients served·
8,000+ concurrent executions·
99.9% reliability·
15+ international markets·
$150K+ annual savings·
85% faster deployments·
10+ enterprise clients served·
Scroll to dive in

01 / The Architect

I build the systems that move data — reliably, at scale, under pressure.

My career started on a factory floor at Bosch, where I learned that data isn't abstract. It's a sensor on a manufacturing line, an alert at 3 AM, the difference between $150K saved or lost. That perspective never left.

Six years later, I'm architecting serverless event-driven pipelines that handle 8,000 concurrent executions with sub-second latency. I've migrated databases from PostgreSQL to DynamoDB without losing a transaction, built ML registries that compress 14-day deployments into 2, and engineered platforms that 50+ data scientists rely on daily.

The best infrastructure is the kind nobody notices — until it isn't there. I obsess over the boring, hard parts: idempotency, fault tolerance, auditability, the 99.9th percentile. That's where reliability lives. That's the work I love.

Bengaluru, IN6 Years Experience15+ Markets7 Certifications

~ / siddhartha.kumar — zsh

02 / Impact

Numbers that ship.

Every metric below ran in production. None of them are estimates.

0+

Concurrent Executions

Step Functions orchestrating real-time POS data

0.0%

Message Reliability

Across 15+ international markets, sub-3-min SLA

0%

Faster Deployments

ML model rollout: 14 days → 2 days

$0K+

Annual Savings

Reduced MTBF by 20%, MTTR by 25% at Bosch

03 / Featured Systems

Production-grade infrastructure,
built for the long run.

A selection of systems I've architected, migrated, or rebuilt from scratch. Click any card for the full architecture breakdown.

Production Systems

Mission-critical infrastructure running in production today.

Internal Platforms

Reusable infrastructure that compounds across teams.

Foundations

Earlier work that shaped the systems thinking.

04 / Trajectory

The path so far.

2022Present

Tiger Analytics India Consulting Pvt. Ltd. logo

Data Engineer

Currently here

Tiger Analytics India Consulting Pvt. Ltd.

Bengaluru, India

  • Architected and built POS pipeline at 8,000+ concurrent executions
  • Led PostgreSQL → DynamoDB migration (95% latency cut)
  • Built ML Model Repository for pharma client (85% faster deployments)
  • Engineered multi-tenant portal for 10+ enterprise clients
  • Designed fault-tolerant Lambda functions with 99.9% reliability
  • Automated marketing-mix workflows (90% manual intervention reduced)

20212022

Futurense Technologies Pvt. Ltd. logo

Trainee Data Engineer

Futurense Technologies Pvt. Ltd.

Bengaluru, India

  • Built distributed ETL POCs on AWS EMR + PySpark (500GB+ workloads)
  • Created real-time Tableau dashboards (65% faster reporting)

20192020

Bosch Limited (Robert Bosch India) logo

Graduate Apprentice — Data Analytics

Bosch Limited (Robert Bosch India)

Jaipur, India

  • Digitized executive KPI reporting (70% cycle reduction)
  • IoT root cause analysis: 20% MTBF reduction, $150K+ saved annually
  • Reduced MTTR by 25% via predictive maintenance insights

05 / Tech Arsenal

The tools I trust in production.

Everything below has shipped in real systems — not just listed for keywords.

Languages & Frameworks

Daily-use, production-tested

Python

Expert

SQL

Expert

PySpark

Expert

Pandas

Expert

NumPy

Advanced

Boto3

Expert

Flask

Advanced

REST APIs

Expert

Shell Scripting

Advanced

AWS Cloud

Primary cloud — production architecture

Lambda

Expert

Step Functions

Expert

DynamoDB

Expert

S3

Expert

Athena

Advanced

Glue

Advanced

EMR

Advanced

EC2

Advanced

API Gateway

Advanced

CloudWatch

Expert

EventBridge

Advanced

SNS / SQS

Advanced

IAM

Advanced

Data Platforms

Storage & query engines

Snowflake

Expert

PostgreSQL

Advanced

DynamoDB

Expert

Hive

Proficient

Data Lakes

Advanced

Data Warehouses

Advanced

Distributed Systems

Architecture patterns & frameworks

Apache Spark

Expert

Apache Airflow

Expert

Event-Driven Architecture

Expert

Microservices

Advanced

Message Queues

Advanced

Data Engineering

Patterns & methodologies

ETL / ELT Pipelines

Expert

Dimensional Modeling

Advanced

OLAP

Advanced

SCD Type 2

Advanced

Schema Design

Expert

Partitioning

Expert

Indexing

Advanced

Data Quality

Advanced

DevOps & Tooling

Build, ship, observe

Docker

Advanced

Git

Expert

Azure DevOps

Advanced

CI/CD

Advanced

Terraform

Proficient

Grafana

Advanced

JIRA

Expert

Agile / Scrum

Expert

Analytics & BI

Visualization & analytical platforms

Power BI

Advanced

Tableau

Advanced

Databricks

Expert

Dataiku DSS

Advanced

Proficiency:ExpertAdvancedProficient

06 / Credentials

Validated, not just claimed.

Featured2024

Astronomer Champions Program

Astronomer

Recognized by Astronomer for expertise and contributions to the Apache Airflow community. One of a select group invited annually.

Amazon Web Services logo

AWS Certified Cloud Practitioner

Amazon Web Services

Microsoft logo

Microsoft Azure Fundamentals (AZ-900)

Microsoft

Microsoft logo

Microsoft Azure Data Fundamentals (DP-900)

Microsoft

Microsoft logo

Microsoft Azure AI Fundamentals (AI-900)

Microsoft

Astronomer logo

Astronomer Airflow 3 Fundamentals & DAG Authoring

Astronomer

Databricks logo

Databricks Generative AI Fundamentals

Databricks

Google logo

Google Generative AI Fundamentals

Google

07 / Open Source

Code in the open.

Live activity from my GitHub. The work I'm building, learning, and shipping in public.

08 / Direct Line

Let's build something
that scales.

Working on infrastructure that needs to be fast, fault-tolerant, and audited? That's the kind of problem I solve. Reach out — I read everything.

Open to senior data engineering & architecture roles