General Responsibilities:
General Skills:
Additional Responsibilities:
Tools used are:
Data design:
Development of ETL strategy and solution for different sets of data modules:
Development and testing of data pipelines with Incremental and Full Load:
Performance review data consistency checks:
End-to-end integrated testing for Full Load and Incremental Load Plan for Go Live Production Deployment:
Go Live Support and Review after Go Live:
Knowledge Transfer to Ministry staff development of documentation on the work completed:
Experience and Skill Set Requirements:
Must Haves:
Skill Set Requirements:
Experience:
Technical Skills:
Design Documentation and Analysis Skills:
- Schema definitions
- Error handling and logging
- ETL Process Documentation
- Job Scheduling and Dependency Management
- Data Quality and Validation Checks
- Performance Optimization and Scalability Plans
- Troubleshooting Guides
- Data Lineage
- Security and Access Control Policies applied within ETL
Communication and Leadership Skills:
Required Skills:
Experience and Skill Set Requirements: Must Haves: Experience with ETL tools such as Microsoft SSIS stored procedures T-SQL Experience in Delta Lake Databricks and Azure Databricks pipelines Strong knowledge of Delta Lake for data management and optimization. Familiarity with Databricks Workflows for scheduling and orchestrating tasks. Python and PySpark Solid understanding of the Medallion Architecture (Bronze Silver Gold) and experience implementing it in production environments. Hands-on experience in real-time data ingestion and streaming architecture using Oracle GoldenGate for CDC SQL Server Oracle Skill Set Requirements: Experience: Experience working with SQL Server T-SQL Oracle PL/SQL development or similar relational databases Experience working with Azure Data Factory Databricks and Python development Experience building data ingestion and change data capture using Oracle Golden Gate Experience in designing developing and implementing ETL pipelines using Databricks and related tools to ingest transform and store large-scale datasets Experience in leveraging Databricks Delta Lake Delta Live Tables and Spark to process structured and unstructured data. Experience working with building databases data warehouses and working with delta and full loads Experience on Data modeling and tools e.g. SAP Power Designer Visio or similar Experience working with SQL Server SSIS or other ETL tools solid knowledge and experience with SQL scripting Experience developing in an Agile environment Understanding data warehouse architecture with a delta lake Ability to analyze design develop test and document ETL pipelines from detailed and high-level specifications and assist in troubleshooting. Ability to utilize SQL to perform DDL tasks and complex queries Good knowledge of database performance optimization techniques Ability to assist in the requirements analysis and subsequent developments Ability to conduct unit testing and assist in test preparations to ensure data integrity Work closely with Designers Business Analysts and other Developers Liaise with Project Managers Quality Assurance Analysts and Business Intelligence Consultants Design and implement technical enhancements of Data Warehouse as required. Technical Skills: Experience in developing and managing ETL pipelines jobs and workflows in Databricks. Deep understanding of Delta Lake for building data lakes and managing ACID transactions schema evolution and data versioning. Experience automating ETL pipelines using Delta Live Tables including handling Change Data Capture (CDC) for incremental data loads. Proficient in structuring data pipelines with the Medallion Architecture to scale data pipelines and ensure data quality. Hands-on experience developing streaming tables in Databricks using Structured Streaming and readStream to handle real-time data. Expertise in integrating CDC tools like GoldenGate or Debezium for processing incremental updates and managing real-time data ingestion. Experience using Unity Catalog to manage data governance access control and ensure compliance. Skilled in managing clusters jobs autoscaling monitoring and performance optimization in Databricks environments. Knowledge of using Databricks Autoloader for efficient batch and real-time data ingestion. Experience with data governance best practices including implementing security policies access control and auditing with Unity Catalog. Proficient in creating and managing Databricks Workflows to orchestrate job dependencies and schedule tasks. Knowledge of FHIR Standards and their application within real-time streaming ecosystems Strong knowledge of Python PySpark and SQL for data manipulation and transformation. Experience integrating Databricks with cloud storage solutions such as Azure Blob Storage AWS S3 or Google Cloud Storage. Familiarity with external orchestration tools like Azure Data Factory Implementing logical and physical data models Design Documentation and Analysis Skills: Demonstrated experience in creating design documentation such as: Schema definitions Error handling and logging ETL Process Documentation Job Scheduling and Dependency Management Data Quality and Validation Checks Performance Optimization and Scalability Plans Troubleshooting Guides Data Lineage Security and Access Control Policies applied within ETL Experience in Fit-Gap analysis system use case reviews requirements reviews coding exercises and reviews. Participate in defect fixing testing support and development activities for ETL Analyze and document solution complexity and interdependencies including providing support for data validation. Strong analytical skills for troubleshooting problem-solving and ensuring data quality. Communication and Leadership Skills: Ability to collaborate effectively with cross-functional teams and communicate complex technical concepts to non-technical stakeholders. Experience working with external vendors Strong problem-solving skills and experience working in an Agile or Scrum environment. Ability to provide technical guidance and support to other team members on Databricks best practices. Must have previous work experience in conducting Knowledge Transfer sessions ensuring the resources will receive the required knowledge to support the system. Must develop documentation and materials as part of a review and knowledge transfer to other members.
IT Services and IT Consulting