Job Title: Oracle Fusion Data Engineer
Job Location: Princeton New Jersey
Job Type: Contract
Job Description:
Data Pipeline Development
Design build and optimize robust ETLELT pipelines to ingest transform and load data from Oracle Fusion applications and other sources into the Databricks Lakehouse This involves using PySpark SQL and Databricks notebooks
Databricks Platform Expertise
Leverage Databricks functionalities such as Delta Lake Unity Catalog and Spark optimization techniques to ensure data quality performance and governance
Oracle Fusion Integration
Develop connectors and integration strategies to extract data from Oracle Fusion modules eg Financials HCM SCM using APIs SQL or other appropriate methods
Data Modelling and Warehousing
Design and implement data models within Databricks potentially following a medallion architecture to support analytical and reporting requirements
Performance Optimization
Tune Spark jobs and optimize data processing within Databricks for efficiency and cost effectiveness
Data Quality and Governance
Implement data quality checks error handling and data validation frameworks to ensure data integrity Adhere to data governance policies and security best practices
Collaboration
Work closely with data architects data scientists business analysts and other stakeholders to understand data requirements and deliver solutions that meet business needs
Automation and CICD
Develop automation scripts and implement CICD pipelines for Databricks workflows and deployments
Troubleshooting and Support
Provide operational support troubleshoot data related issues and perform root cause analysis
Required Skills and Qualifications
Strong proficiency in Databricks
Including PySparkScala Delta Lake Unity Catalog and Databricks notebooks
Experience with Oracle Fusion
Knowledge of Oracle Fusion data structures APIs and data extraction methods
Expertise in SQL
For querying manipulating and optimizing data in both Oracle and Databricks
Cloud Platform Experience
Familiarity with a major cloud provider eg AWS Azure GCP where Databricks is deployed
Data Warehousing and ETLELT Concepts
Solid understanding of data warehousing principles and experience in building and optimizing data pipelines
Problem solving and Analytical Skills
Ability to analyse complex data issues and propose effective solutions
Communication and Collaboration
Strong interpersonal skills to work effectively within cross functional teams