HELLO, MY NAME IS

Aravindan
Ramesh

Senior Data Engineer

aravindanramesh910@gmail.com​
9659959575​
About Me

Hi, I’m Aravindan—a Senior Data Engineer with 5+ years’ experience solving tough data problems for fintech and enterprise teams.

What I Do

I build and automate scalable data pipelines using tools like PySpark, Kafka, and Airflow.

I create secure, cloud-native solutions on AWS and GCP that make data easy to use and understand.
My work helps companies turn complex raw data into simple, business-driving insights.

How I Work

I love optimizing workflows: for example, I’ve migrated systems from Oracle to the cloud, saving time and ensuring reliability.
I focus on data quality, security, and performance, always with a practical, business-first mindset.
I mentor teams, share ideas, and enjoy collaborating to achieve results that matter.

DATA ENGINEERING

I love making data work smarter. For example, I’ve migrated legacy systems to the cloud, cutting costs and boosting reliability for large-scale analytics.

CLOUD SOLUTIONS

Cloud-first is my thing—I build secure, scalable data platforms on AWS and GCP. My recent open-source data lake project saved teams hours each week and improved access for everyone.

REAL-TIME DATA

Fast data drives smart decisions. I’ve built real-time pipelines with Spark and Kafka—like tracking customer locations live for targeted marketing, all while staying GDPR compliant.

Skills

%
PySpark
%
Apache Airflow
%
AWS & GCP
%
Python
%
SQL, Hive, Spark-SQL

My Experience

Feb 2025 – Present

 M2P Fintech, Chennai

Senior Data Engineer (SDE 2)

Architected an open-source data lake on AWS EKS, integrating Spark, Airflow, Iceberg, Trino, and S3 for scalable, cost-effective analytics.
Automated ETL workflows and metadata management for high-performance data processing using Airflow, Spark, and Hive Metastore.
Enhanced security by implementing IAM Roles for Service Accounts, enabling secure, cloud-native workload authentication.

Apr 2021 – Feb 2025

Tata Consultancy Services, Chennai

Data Engineer

Developed and optimized PySpark apps on Kubernetes, focusing on scalable and reliable analytics.
Automated ETL pipelines with Airflow and migrated critical data from Oracle RDBMS to GCP Cloud.
Led partitioned Hive deployment with advanced compression, improving storage and query speeds.
Integrated Spark and Kafka for real-time processing and managed VDS migration to Dremio for seamless data access.
Recognized for technical excellence and led a 5-member team to deliver high-impact projects.

Portfolio