Job Role :- Senior / Lead Data Engineer
Location :- Alpharetta GA (Hybrid 3 Days in Office)
Duration :- Fulltime
Job Description :-
We are looking for an experienced Senior/Lead Data Engineer with 8 years of expertise in designing and delivering scalable high-performing data solutions on the Azure ecosystem. The ideal candidate will have deep hands-on experience with Databricks Spark modern data lake house architectures data modelling and both batch and real-time data processing. You will be responsible for driving end-to-end data engineering initiatives influencing architectural decisions and ensuring robust high-quality data pipelines.
Responsibility :-
- Architect design and implement scalable data platforms and pipelines on Azure and Databricks.
- Build and optimize data ingestion transformation and processing workflows across batch and real-time data streams.
- Work extensively with ADLS Delta Lake and Spark (Python) for large-scale data engineering.
- Lead the development of complex ETL/ELT pipelines ensuring high quality reliability and performance.
- Design and implement data models including conceptual logical and physical models for analytics and operational workloads.
- Work with relational and lakehouse systems including PostgreSQL and Delta Lake.
- Define and enforce best practices in data governance data quality security and architecture.
- Collaborate with architects data scientists analysts and business teams to translate requirements into technical solutions.
- Troubleshoot production issues optimize performance and support continuous improvement of the data platform.
- Mentor junior engineers and contribute to building engineering standards and reusable components.
Required Skills :-
- 8 years of hands-on data engineering experience in enterprise environments.
- Strong expertise in Azure services especially Azure Databricks Functions and Azure Data Factory (preferred).
- Advanced proficiency in Apache Spark with Python (PySpark).
- Strong command over SQL query optimization and performance tuning.
- Deep understanding of ETL/ELT methodologies data pipelines and scheduling/orchestration.
- Hands-on experience with Delta Lake (ACID transactions optimization schema evolution).
- Strong experience in data modelling (normalized dimensional lakehouse modelling).
- Experience in both batch processing and real-time/streaming data (Kafka Event Hub or similar).
- Solid understanding of data architecture principles distributed systems and cloud-native design patterns.
- Ability to design end-to-end solutions evaluate trade-offs and recommend best-fit architectures.
- Strong analytical problem-solving and communication skills.
- Ability to collaborate with cross-functional teams and lead technical discussions.
Preferred Skills:
- Experience with CI/CD tools such as Azure DevOps and Git.
- Familiarity with IaC tools (Terraform ARM).
- Exposure to data governance and cataloging tools (Azure Purview).
- Experience supporting machine learning or BI workloads on Databricks.
Skills :- Java Python Spark
Job Role :- Senior / Lead Data Engineer Location :- Alpharetta GA (Hybrid 3 Days in Office) Duration :- Fulltime Job Description :- We are looking for an experienced Senior/Lead Data Engineer with 8 years of expertise in designing and delivering scalable high-performing data solutions on the ...
Job Role :- Senior / Lead Data Engineer
Location :- Alpharetta GA (Hybrid 3 Days in Office)
Duration :- Fulltime
Job Description :-
We are looking for an experienced Senior/Lead Data Engineer with 8 years of expertise in designing and delivering scalable high-performing data solutions on the Azure ecosystem. The ideal candidate will have deep hands-on experience with Databricks Spark modern data lake house architectures data modelling and both batch and real-time data processing. You will be responsible for driving end-to-end data engineering initiatives influencing architectural decisions and ensuring robust high-quality data pipelines.
Responsibility :-
- Architect design and implement scalable data platforms and pipelines on Azure and Databricks.
- Build and optimize data ingestion transformation and processing workflows across batch and real-time data streams.
- Work extensively with ADLS Delta Lake and Spark (Python) for large-scale data engineering.
- Lead the development of complex ETL/ELT pipelines ensuring high quality reliability and performance.
- Design and implement data models including conceptual logical and physical models for analytics and operational workloads.
- Work with relational and lakehouse systems including PostgreSQL and Delta Lake.
- Define and enforce best practices in data governance data quality security and architecture.
- Collaborate with architects data scientists analysts and business teams to translate requirements into technical solutions.
- Troubleshoot production issues optimize performance and support continuous improvement of the data platform.
- Mentor junior engineers and contribute to building engineering standards and reusable components.
Required Skills :-
- 8 years of hands-on data engineering experience in enterprise environments.
- Strong expertise in Azure services especially Azure Databricks Functions and Azure Data Factory (preferred).
- Advanced proficiency in Apache Spark with Python (PySpark).
- Strong command over SQL query optimization and performance tuning.
- Deep understanding of ETL/ELT methodologies data pipelines and scheduling/orchestration.
- Hands-on experience with Delta Lake (ACID transactions optimization schema evolution).
- Strong experience in data modelling (normalized dimensional lakehouse modelling).
- Experience in both batch processing and real-time/streaming data (Kafka Event Hub or similar).
- Solid understanding of data architecture principles distributed systems and cloud-native design patterns.
- Ability to design end-to-end solutions evaluate trade-offs and recommend best-fit architectures.
- Strong analytical problem-solving and communication skills.
- Ability to collaborate with cross-functional teams and lead technical discussions.
Preferred Skills:
- Experience with CI/CD tools such as Azure DevOps and Git.
- Familiarity with IaC tools (Terraform ARM).
- Exposure to data governance and cataloging tools (Azure Purview).
- Experience supporting machine learning or BI workloads on Databricks.
Skills :- Java Python Spark
View more
View less