Data Engineer

Not Interested
Bookmark
Report This Job

profile Job Location:

Bangalore - India

profile Monthly Salary: Not Disclosed
Posted on: 8 hours ago
Vacancies: 1 Vacancy

Job Summary

Job Summary:

We are looking for an experienced Data Engineer with strong hands-on expertise in designing building and optimizing data pipelines and data processing systems. The ideal candidate should have solid experience in Java Spark Kafka and Google Cloud Platform (GCP). Knowledge of Python is an added advantage.

Key Responsibilities:

  • Design develop and maintain scalable data pipelines and ETL workflows.
  • Work with large complex datasets to extract transform and load data efficiently.
  • Build streaming and batch processing systems using Apache Spark and Kafka.
  • Develop high-performance data ingestion solutions using Java.
  • Create and manage data infrastructure on GCP (BigQuery Dataflow Pub/Sub GCS).
  • Collaborate with data scientists analysts and product teams to ensure reliable data flow.
  • Optimize performance of existing data platforms and pipelines.
  • Implement best practices for data quality governance and security.
  • Monitor and troubleshoot data jobs and cloud services.

Required Skills:

  • Strong experience in Java for data engineering or backend processing.
  • Hands-on experience with Apache Spark (batch streaming).
  • Proficient in Kafka architecture producers/consumers and stream processing.
  • Strong expertise in Google Cloud Platform (GCP) services.
  • Solid understanding of SQL and distributed data systems.
  • Experience with CI/CD version control (Git) and automation tools.
  • Python knowledge is an advantage (optional).
Job Summary: We are looking for an experienced Data Engineer with strong hands-on expertise in designing building and optimizing data pipelines and data processing systems. The ideal candidate should have solid experience in Java Spark Kafka and Google Cloud Platform (GCP). Knowledge of Python is an...
View more view more

Key Skills

  • Apache Hive
  • S3
  • Hadoop
  • Redshift
  • Spark
  • AWS
  • Apache Pig
  • NoSQL
  • Big Data
  • Data Warehouse
  • Kafka
  • Scala