Tasks
Kafka Developer
Location:Eindhoven Netherlands (Onsite)
12months
Key Responsibilities
Pipeline Development: Take ownership of the CDC ingestion framework utilizing Kafka connectors (Debezium Iceberg sink S3 sink).
Containerized Infrastructure Management: Deploy and manage Debezium and Kafka Connect workers using Docker containers orchestrating on AWS ECS (Elastic Container Service) and ECR.
Data Lake Integration: Manage data ingestion into AWS S3 utilizing Parquet and Apache Iceberg formats.
Infrastructure as Code: Use Terraform to provision and manage AWS resources supporting the data platform.
CI/CD: Build and maintain deployment pipelines using GitHub and GitHub Actions.
Operational Excellence: Monitor pipeline health troubleshoot connectivity issues and ensure the reliability of the Kafka ecosystem.
Optional: Support and optimize workflow orchestration using Airflow where applicable.
Requirements
Skills Required:
Apache Kafka & Kafka Connect: Multiple years of hands-on experience configuring deploying and managing Kafka Connect clusters in a production environment.
Containerization: Extensive experience with Docker is required. You must be comfortable building images and managing container lifecycles.
AWS Compute: Proven experience running containers on AWS ECS and managing images via AWS ECR.
Tech Stack Summary
Streaming: Apache Kafka Kafka Connect Debezium
Compute/Containerization: AWS ECS AWS ECR Docker
Storage/Format: AWS S3 Apache Iceberg Parquet
DevOps: Terraform GitHub Actions
Languages: Python Bash
Optional Orchestration: Apache Airflow
TasksKafka DeveloperLocation:Eindhoven Netherlands (Onsite)12monthsKey ResponsibilitiesPipeline Development: Take ownership of the CDC ingestion framework utilizing Kafka connectors (Debezium Iceberg sink S3 sink).Containerized Infrastructure Management: Deploy and manage Debezium and Kafka Connect ...
Tasks
Kafka Developer
Location:Eindhoven Netherlands (Onsite)
12months
Key Responsibilities
Pipeline Development: Take ownership of the CDC ingestion framework utilizing Kafka connectors (Debezium Iceberg sink S3 sink).
Containerized Infrastructure Management: Deploy and manage Debezium and Kafka Connect workers using Docker containers orchestrating on AWS ECS (Elastic Container Service) and ECR.
Data Lake Integration: Manage data ingestion into AWS S3 utilizing Parquet and Apache Iceberg formats.
Infrastructure as Code: Use Terraform to provision and manage AWS resources supporting the data platform.
CI/CD: Build and maintain deployment pipelines using GitHub and GitHub Actions.
Operational Excellence: Monitor pipeline health troubleshoot connectivity issues and ensure the reliability of the Kafka ecosystem.
Optional: Support and optimize workflow orchestration using Airflow where applicable.
Requirements
Skills Required:
Apache Kafka & Kafka Connect: Multiple years of hands-on experience configuring deploying and managing Kafka Connect clusters in a production environment.
Containerization: Extensive experience with Docker is required. You must be comfortable building images and managing container lifecycles.
AWS Compute: Proven experience running containers on AWS ECS and managing images via AWS ECR.
Tech Stack Summary
Streaming: Apache Kafka Kafka Connect Debezium
Compute/Containerization: AWS ECS AWS ECR Docker
Storage/Format: AWS S3 Apache Iceberg Parquet
DevOps: Terraform GitHub Actions
Languages: Python Bash
Optional Orchestration: Apache Airflow
View more
View less