Senior Data Engineer with expertise in designing scalable Data Mesh architectures, optimizing low-latency data processing using Apache Spark, Apache Flink, and Apache Kafka, and implementing robust microservices-based distributed systems. Experienced in developing and automating real-time ingestion pipelines with Airflow, DBT, and Trino on AWS. Strong background in data governance (RBAC, PII compliance, Apache Ranger) and performance optimization. Passionate about leveraging open-source technologies to build self-service, high-performance data platforms.
Financial Data Platform – AWS-Based, Regulatory-Compliant
Technologies: Apache Spark, Flink, Kafka, Airflow, Hudi, Trino, Open Metadata, AWS (S3, EMR, Glue), Python, Scala
Enterprise Data Mesh Platform
Technologies: AWS (S3, EMR, Glue), Spark, Airflow, DBT, Apache Ranger, Trino