Architecting data infrastructure that processes 10M+ transactions daily, delivering $2M+ annual savings, and building the backbone of fraud detection systems in banking.
I'm a Senior Data Engineer based in Bengaluru & Gurgaon, India, with a decade of experience architecting data solutions that power critical banking operations.
I specialize in building end-to-end data pipelines, real-time streaming architectures, and feature engineering systems serving ML teams at scale. My work spans fraud detection, risk management, and regulatory compliance in financial services.
Currently expanding into Generative AI, LLMs, MLOps, and cloud-native architectures to push the boundaries of what data engineering can achieve.
class DataEngineer:
def __init__(self):
self.name = "Krishna Kumar Yadav"
self.role = "Senior Data Engineer"
self.experience = "10+ years"
self.impact = {
"savings": "$2M+ annual",
"optimization": "50% faster",
"scale": "10M+ txns/day"
}
Bank of America
Built a comprehensive data platform processing 10M+ transactions using TigerGraph, Kafka, Spark, Delta Lake, Airflow, and Feature Store. Supplied features to ML team.
Standard Chartered Bank
Orchestrated Kafka streaming pipeline processing 1M+ events/hour with Spark Structured Streaming, Delta Lake, and Databricks.
Wells Fargo
Spearheaded migration to Snowflake data warehouse using dbt, Airflow, PySpark. Transformed legacy systems into modern, scalable architecture.
Have a project? Need data engineering expertise? Let's connect.