GCP Data Engineer

Not Interested
Bookmark
Report This Job

profile Job Location:

Hartford, CT - USA

profile Monthly Salary: Not Disclosed
Posted on: 30+ days ago
Vacancies: 1 Vacancy

Job Summary

GCP Data Engineer -

Healthcare domain

Design and implement robust data pipelines using Google Cloud Platform (GCP) services such as BigQuery Cloud Storage and Pub/Sub.

Develop and manage workflows using Cloud Composer (Apache Airflow) for efficient scheduling and orchestration.

Write clean efficient scalable code in Python leveraging advanced programming techniques.

Craft complex SQL queries in BigQuery including window functions CTEs and performance tuning strategies.

Build and maintain real-time data processing systems using Apache Kafka.

Model and manage NoSQL databases particularly MongoDB with a focus on scalable schema design.

Utilize Shell scripting and perform Linux system administration tasks to support data infrastructure.

Conduct data profiling and implement validation techniques to ensure data quality and integrity.

Develop and maintain API integration scripts for seamless service automation and data exchange.

Troubleshoot and resolve data-related issues with strong analytical and problem-solving skills.

Create and maintain data flow diagrams to clearly communicate architecture and pipeline logic to stakeholders.

Required Skills:
3 years of proven hands-on experience with GCP data services.

Experience in Agile methodology with preferred experience in SAFe methodology

Experience working in onshore/offshore support model collaborating work with offshore teams.

Strong understanding of distributed systems and data engineering best practices.

Excellent communication and documentation skills.

Ability to work independently and collaboratively in a fast-paced environment.

Bonus Skills:

GCP Google Cloud Professional Data Engineer Certification

Experience working with GitHub RTC automation tools

Experience AI programming IVR Technologies like AYAYA CISCO Chatbot etc

GCP Data Engineer - Healthcare domain Design and implement robust data pipelines using Google Cloud Platform (GCP) services such as BigQuery Cloud Storage and Pub/Sub. Develop and manage workflows using Cloud Composer (Apache Airflow) for efficient scheduling and orchestration. Write clean eff...
View more view more

Key Skills

  • Apache Hive
  • S3
  • Hadoop
  • Redshift
  • Spark
  • AWS
  • Apache Pig
  • NoSQL
  • Big Data
  • Data Warehouse
  • Kafka
  • Scala