GCP Data Engineer -
Healthcare domain
Design and implement robust data pipelines using Google Cloud Platform (GCP) services such as BigQuery Cloud Storage and Pub/Sub.
Develop and manage workflows using Cloud Composer (Apache Airflow) for efficient scheduling and orchestration.
Write clean efficient scalable code in Python leveraging advanced programming techniques.
Craft complex SQL queries in BigQuery including window functions CTEs and performance tuning strategies.
Build and maintain real-time data processing systems using Apache Kafka.
Model and manage NoSQL databases particularly MongoDB with a focus on scalable schema design.
Utilize Shell scripting and perform Linux system administration tasks to support data infrastructure.
Conduct data profiling and implement validation techniques to ensure data quality and integrity.
Develop and maintain API integration scripts for seamless service automation and data exchange.
Troubleshoot and resolve data-related issues with strong analytical and problem-solving skills.
Create and maintain data flow diagrams to clearly communicate architecture and pipeline logic to stakeholders.
Required Skills:
3 years of proven hands-on experience with GCP data services.
Experience in Agile methodology with preferred experience in SAFe methodology
Experience working in onshore/offshore support model collaborating work with offshore teams.
Strong understanding of distributed systems and data engineering best practices.
Excellent communication and documentation skills.
Ability to work independently and collaboratively in a fast-paced environment.
Bonus Skills:
GCP Google Cloud Professional Data Engineer Certification
Experience working with GitHub RTC automation tools
Experience AI programming IVR Technologies like AYAYA CISCO Chatbot etc
GCP Data Engineer - Healthcare domain Design and implement robust data pipelines using Google Cloud Platform (GCP) services such as BigQuery Cloud Storage and Pub/Sub. Develop and manage workflows using Cloud Composer (Apache Airflow) for efficient scheduling and orchestration. Write clean eff...
GCP Data Engineer -
Healthcare domain
Design and implement robust data pipelines using Google Cloud Platform (GCP) services such as BigQuery Cloud Storage and Pub/Sub.
Develop and manage workflows using Cloud Composer (Apache Airflow) for efficient scheduling and orchestration.
Write clean efficient scalable code in Python leveraging advanced programming techniques.
Craft complex SQL queries in BigQuery including window functions CTEs and performance tuning strategies.
Build and maintain real-time data processing systems using Apache Kafka.
Model and manage NoSQL databases particularly MongoDB with a focus on scalable schema design.
Utilize Shell scripting and perform Linux system administration tasks to support data infrastructure.
Conduct data profiling and implement validation techniques to ensure data quality and integrity.
Develop and maintain API integration scripts for seamless service automation and data exchange.
Troubleshoot and resolve data-related issues with strong analytical and problem-solving skills.
Create and maintain data flow diagrams to clearly communicate architecture and pipeline logic to stakeholders.
Required Skills:
3 years of proven hands-on experience with GCP data services.
Experience in Agile methodology with preferred experience in SAFe methodology
Experience working in onshore/offshore support model collaborating work with offshore teams.
Strong understanding of distributed systems and data engineering best practices.
Excellent communication and documentation skills.
Ability to work independently and collaboratively in a fast-paced environment.
Bonus Skills:
GCP Google Cloud Professional Data Engineer Certification
Experience working with GitHub RTC automation tools
Experience AI programming IVR Technologies like AYAYA CISCO Chatbot etc
View more
View less