Design & Development
- Create and maintain ETL / ELT pipelines for large datasets in formats like CSV, JSON, Parquet, and Avro using Hadoop MapReduce and Spark frameworks. +2
Data Modeling
Analyze existing schemas and recommend improvements to E-R and dimensional models. Develop logical and physical data models. +2Performance Optimization
Configure metadata, fine-tune pipeline performance, and optimize job execution. +10Collaboration
Work with cross-functional teams (DevOps, BI, stakeholders) to ensure smooth data workflows and scalable solutions. +10Governance & Quality
Apply best practices for data modeling, governance, and quality controlsSkills Required
Csv, Json, Devops, Bi, Etl