Role Descriptions:
Key Responsibilities
Lead the design and development of end-to-end data pipelines using cloud-native and modern data engineering tools.
Architect scalable data ingestion transformation and storage solutions across structured and unstructured data sources.
Drive implementation of ETLELT frameworks data orchestration and workflow automation.
Collaborate with data architects product owners SMEs and cross-functional teams to gather requirements and define data solutions.
Build and govern data models metadata standards and data quality frameworks.
Optimize data pipelines for reliability performance and cost efficiency.
Lead the adoption of cloud platforms such as AWS Azure or GCP including services like S3 Glue Redshift EMR Databricks or Snowflake.
Oversee code reviews best practices and engineering standards within the team.
Ensure data security compliance and governance alignment (PII encryption audit controls).
Mentor junior data engineers and drive continuous improvement initiatives across the team.
Required Qualifications:
8-12 years of experience in data engineering with at least 3 years leading teams or technical initiatives.
Strong hands-on experience with distributed data processing frameworks (Spark PySpark EMR Databricks).
Proficiency in building scalable pipelines using ETLELT tools and orchestration platforms (Airflow Glue DBT Step Functions).
Strong SQL and programming skills (Python Scala or Java).
Experience with cloud data platforms AWS (preferred) Azure or GCP.
Knowledge of relational NoSQL and analytical databases (Redshift Snowflake BigQuery DynamoDB etc.).
Familiarity with CICD pipelines version control (GitGitHub) and DevOps practices for data engineering.
Strong understanding of data modelling concepts (3NF StarSnowflake schema).
Excellent communication skills and experience working in onshoreoffshore delivery models.
Skills:
Digital : Microsoft AzureDigital : SnowflakeDigital : Databricks
Experience Required:
6-8