drjobs Data Engineer (CPT - Onsite) (3-month Contract to go perm)

Data Engineer (CPT - Onsite) (3-month Contract to go perm)

Employer Active

1 Vacancy
drjobs

Job Alert

You will be updated with latest job alerts via email
Valid email field required
Send jobs
Send me jobs like this
drjobs

Job Alert

You will be updated with latest job alerts via email

Valid email field required
Send jobs
Job Location drjobs

Cape Town - South Africa

Monthly Salary drjobs

Not Disclosed

drjobs

Salary Not Disclosed

Vacancy

1 Vacancy

Job Description

DataFin was established in 1999. We pride ourselves on the fact that we have established relationships with industry leaders and a vast majority of our business is repeat business.

Duties & Responsibilities

ENVIRONMENT:

A leading digital command center based in Cape Town is seeking a skilled Data Engineer to become part of their team. The role involves designing developing and deploying scalable production-ready data pipelines on Google Cloud Platform utilizing services such as Cloud Dataflow Cloud Dataproc and Cloud Functions to ingest process and transform large volumes of both structured and unstructured data. The ideal candidate will have proven experience as a Data Engineer with a strong track record of building data solutions specifically within the Google Cloud ecosystem.

DUTIES:

  • Design & Build Data Pipelines: Design develop and deploy scalable production-grade data pipelines on Google Cloud Platform using services like Cloud Dataflow Cloud Dataproc and Cloud Functions to ingest process and transform large volumes of structured and unstructured data.
  • Architect Data Warehouses & Lakes: Develop and maintain their core data warehouse architecture in BigQuery and their data lake on Google Cloud Storage ensuring data integrity security performance and cost-optimization.
  • Data Modeling & ETL/ELT: Design and implement robust data models and schemas within BigQuery. Build and manage complex ETL/ELT processes using SQL Dataflow and other GCP-native tools to support analytics and reporting.
  • Custom API Integration: Design build and maintain custom RESTful APIs using Python (e.g. Flask FastAPI) and PHP to handle data ingress from and egress to various internal systems bot platforms and third-party services.
  • Integration & Streaming: Develop and maintain data integration solutions to connect various internal and external data sources using Pub/Sub for real-time streaming and Cloud Data Fusion for batch integration.
  • Monitoring & Optimization: Proactively monitor troubleshoot and optimize data pipelines and BigQuery performance. Utilize Cloud Monitoring and Cloud Logging to ensure reliability identify bottlenecks and resolve issues.
  • Data Governance: Implement and manage data quality and governance frameworks within the GCP ecosystem leveraging tools like Dataplex for data discovery metadata management and policy enforcement.
  • Collaboration: Work closely with data scientists analysts and software engineers to understand data requirements translate them into technical specifications and integrate data solutions into production applications.
  • Innovation & Best Practices: Stay current with the latest GCP data services and industry best practices. Champion and implement innovative solutions to continuously improve their data platform.
  • Documentation: Create and maintain comprehensive documentation for data pipelines architectures and processes to ensure knowledge sharing and long-term maintainability.

REQUIREMENTS:

  • Proven Experience: Demonstrated experience as a Data Engineer with a strong portfolio of designing and building data solutions specifically on Google Cloud Platform.
  • Programming Proficiency: Strong proficiency in backend programming with Python and PHP for data processing API development and infrastructure automation. Ability to write clean efficient and maintainable code.
  • SQL Mastery: Expert-level SQL skills with extensive experience writing complex queries optimizing performance and modeling data in BigQuery.
  • Core GCP Services: Hands-on in-depth experience with core GCP data services including BigQuery Cloud Storage Cloud Dataflow and Pub/Sub.
  • Data Architecture: Solid understanding of modern data warehousing concepts (e.g. Kimball Inmon) data lake architectures distributed systems and data governance best practices within a cloud context.
  • Data Integration: Proficiency in various data integration techniques including batch processing real-time streaming and API-based data ingestion on GCP.

Preferred Qualifications:

  • Orchestration: Experience with workflow orchestration tools preferably Cloud Composer (Managed Apache Airflow).
  • Infrastructure as Code (IaC): Familiarity with managing GCP resources using IaC tools like Terraform.
  • Database Knowledge: Experience with relational databases like Cloud SQL (MySQL PostgreSQL) and NoSQL databases.
  • Machine Learning: Familiarity with ML concepts and experience with GCPs ML services such as Vertex AI or BigQuery ML.
  • Data Visualization: Experience building reports and dashboards using Looker or Looker Studio.
  • Security & Compliance: Knowledge of GCP security best practices IAM roles and data compliance standards (e.g. GDPR POPIA).
  • Software Engineering Practices: Understanding of CI/CD principles and experience using tools like Cloud Build and version control systems like Git.

ATTRIBUTES:

  • Problem-Solving: Excellent analytical and problem-solving skills with the ability to work effectively in a fast-paced agile environment.
  • Communication: Strong verbal and written communication skills with the ability to explain complex technical concepts to both technical and non-technical stakeholders.

While we would really like to respond to every application should you not be contacted for this position within 10 working days please consider your application unsuccessful.

COMMENTS:

When applying for jobs ensure that you have the minimum job requirements. OnlySA Citizens will be considered for this role. If you are not in the mentioned location of any of the jobs please note your relocation plans in all applications for jobs and correspondence. Apply here e-mail a Word copy of your CV to and mention the reference number of the job.

Employment Type

Full-Time

Report This Job
Disclaimer: Drjobpro.com is only a platform that connects job seekers and employers. Applicants are advised to conduct their own independent research into the credentials of the prospective employer.We always make certain that our clients do not endorse any request for money payments, thus we advise against sharing any personal or bank-related information with any third party. If you suspect fraud or malpractice, please contact us via contact us page.