- Develop optimize and operate ETL/ELT pipelines in Databricks using PySpark SparkSQL and Scala;
- Run migration programmes to Databricks including legacy ETL processes into scalable data pipelines;
- Build robust data pipelines on Databricks for real time and batch processing;
- Configure and manage Databricks clusters and jobs for cost and performance efficiency;
- Apply CI/CD to automate testing and deployment of notebooks jobs and pipelines;
- Drive adoption of Databricks Lakehouse architecture and data engineering practice across our clients.
Qualifications :
Naturally you are keen to discover and learn new things each day and you are confident and rely on your self-worth. But most importantly be the true you. The one and only you. With your personal strengths view of the world and unique personality. You need the following qualifications for the position of Databricks Data Engineer:
- A technical Masters degree preferably in Data Science Big Data Econometrics Physics Mathematics Computer Science Artificial Intelligence or a related field;
- At least 4 years of work experience in data engineering; consulting experience is a plus;
- Strong hands-on experience with Databricks and Apache Spark (PySpark and Spark SQL);
- Deep understanding of Delta Lake and medallion architecture;
- Proficiency in SQL and one or more data engineering related programming languages (Python Scala etc);
- Relevant Databricks certifications (Data Engineer Associate/Professional Spark Developer Associate Data Analyst Associate);
- Excellent command of English and Dutch.
Aanvullende informatie :
As a Databricks Data Engineer you will be a trusted adviser to our clients designing and building reliable ETL/ELT pipelines in Databricks. You will apply the medallion architecture and Delta Lake best practice optimize clusters for performance and cost and embed data quality to deliver reliable data products. You will work closely with cloud platform engineers to ensure data solutions are secure scalable and easy to operate.
Remote Work :
No
Employment Type :
Full-time
Develop optimize and operate ETL/ELT pipelines in Databricks using PySpark SparkSQL and Scala;Run migration programmes to Databricks including legacy ETL processes into scalable data pipelines;Build robust data pipelines on Databricks for real time and batch processing;Configure and manage Databrick...
- Develop optimize and operate ETL/ELT pipelines in Databricks using PySpark SparkSQL and Scala;
- Run migration programmes to Databricks including legacy ETL processes into scalable data pipelines;
- Build robust data pipelines on Databricks for real time and batch processing;
- Configure and manage Databricks clusters and jobs for cost and performance efficiency;
- Apply CI/CD to automate testing and deployment of notebooks jobs and pipelines;
- Drive adoption of Databricks Lakehouse architecture and data engineering practice across our clients.
Qualifications :
Naturally you are keen to discover and learn new things each day and you are confident and rely on your self-worth. But most importantly be the true you. The one and only you. With your personal strengths view of the world and unique personality. You need the following qualifications for the position of Databricks Data Engineer:
- A technical Masters degree preferably in Data Science Big Data Econometrics Physics Mathematics Computer Science Artificial Intelligence or a related field;
- At least 4 years of work experience in data engineering; consulting experience is a plus;
- Strong hands-on experience with Databricks and Apache Spark (PySpark and Spark SQL);
- Deep understanding of Delta Lake and medallion architecture;
- Proficiency in SQL and one or more data engineering related programming languages (Python Scala etc);
- Relevant Databricks certifications (Data Engineer Associate/Professional Spark Developer Associate Data Analyst Associate);
- Excellent command of English and Dutch.
Aanvullende informatie :
As a Databricks Data Engineer you will be a trusted adviser to our clients designing and building reliable ETL/ELT pipelines in Databricks. You will apply the medallion architecture and Delta Lake best practice optimize clusters for performance and cost and embed data quality to deliver reliable data products. You will work closely with cloud platform engineers to ensure data solutions are secure scalable and easy to operate.
Remote Work :
No
Employment Type :
Full-time
View more
View less