Job Overview
We are looking for a highly senior deeply hands on Databricks Architect to lead a large scale Oracle to Databricks migration covering schema migration code conversion and ODI job modernization. The ideal candidate has extensive experience building enterprise-grade data platforms on Databricks has executed at least one greenfield Databricks implementation and is exceptionally strong in PySpark Spark SQL framework development and Databricks Workflows.
Key Responsibilities
- Architect design and implement cloud-native data platforms using Databricks (ingestion transformation consumption).
- Lead the full Oracle Databricks migration including schema translation ETL/ELT logic modernization and ODI job replacement.
- Develop reusable PySpark frameworks data processing patterns and orchestration using Databricks Workflows.
- Build scalable secure and cost optimized Databricks infrastructure and data pipelines.
- Collaborate with business and technical stakeholders to drive data modernization strategy.
- Establish development best practices coding standards CI/CD and DevOps/DataOps patterns.
- Provide technical mentorship and create training plans for engineering teams.
- Contribute to building MLOps and advanced operations frameworks.
Required Qualifications
- 14 years in Data Engineering/Architecture with at least 4 years hands-on Databricks experience delivering end to end cloud data solutions.
- Strong experience migrating from Oracle/on prem systems to Databricks including SQL PL/SQL ETL logic and ODI pipelines.
- Deep hands-on expertise in:
- PySpark Spark SQL Delta Lake Unity Catalog
- Building reusable data frameworks
- Designing high performance batch and streaming pipelines
- Proven experience with greenfield Databricks implementations.
- Strong understanding of cloud-native architectures on AWS and modern data platform concepts.
- Solid knowledge of data warehousing columnar databases and performance optimization.
- Good understanding of Agile/Scrum development processes.
- Bonus: Experience designing Data Products Data Mesh architectures Data Vault or enterprise data governance models.
- Good Understanding of Golde Gate.
Job Overview We are looking for a highly senior deeply hands on Databricks Architect to lead a large scale Oracle to Databricks migration covering schema migration code conversion and ODI job modernization. The ideal candidate has extensive experience building enterprise-grade data platforms o...
Job Overview
We are looking for a highly senior deeply hands on Databricks Architect to lead a large scale Oracle to Databricks migration covering schema migration code conversion and ODI job modernization. The ideal candidate has extensive experience building enterprise-grade data platforms on Databricks has executed at least one greenfield Databricks implementation and is exceptionally strong in PySpark Spark SQL framework development and Databricks Workflows.
Key Responsibilities
- Architect design and implement cloud-native data platforms using Databricks (ingestion transformation consumption).
- Lead the full Oracle Databricks migration including schema translation ETL/ELT logic modernization and ODI job replacement.
- Develop reusable PySpark frameworks data processing patterns and orchestration using Databricks Workflows.
- Build scalable secure and cost optimized Databricks infrastructure and data pipelines.
- Collaborate with business and technical stakeholders to drive data modernization strategy.
- Establish development best practices coding standards CI/CD and DevOps/DataOps patterns.
- Provide technical mentorship and create training plans for engineering teams.
- Contribute to building MLOps and advanced operations frameworks.
Required Qualifications
- 14 years in Data Engineering/Architecture with at least 4 years hands-on Databricks experience delivering end to end cloud data solutions.
- Strong experience migrating from Oracle/on prem systems to Databricks including SQL PL/SQL ETL logic and ODI pipelines.
- Deep hands-on expertise in:
- PySpark Spark SQL Delta Lake Unity Catalog
- Building reusable data frameworks
- Designing high performance batch and streaming pipelines
- Proven experience with greenfield Databricks implementations.
- Strong understanding of cloud-native architectures on AWS and modern data platform concepts.
- Solid knowledge of data warehousing columnar databases and performance optimization.
- Good understanding of Agile/Scrum development processes.
- Bonus: Experience designing Data Products Data Mesh architectures Data Vault or enterprise data governance models.
- Good Understanding of Golde Gate.
View more
View less