Job title : GCP Cloud Data Engineer
Location: Phoenix AZ
Hire Type : Contract
Key Responsibilities:
- Design and manage data pipelines using GCP tools (BigQuery Dataflow Dataproc).
- Collaborate with stakeholders to define data requirements and models.
- Implement cloud data storage processing and analytics solutions.
- Optimize Big Data ecosystems (Hadoop HDFS Hive Spark).
- Develop ETL processes and workflows for data transformation.
- Apply advanced Java skills for scalable data processing.
- Troubleshoot data issues and improve performance.
- Automate tasks and ensure system reliability using GCP services.
- Maintain data quality security and compliance standards.
- Stay updated with best practices in cloud infrastructure.
Key Skills & Qualifications:
- Proficient with GCP services (BigQuery Dataflow Dataproc).
- Expertise in Big Data and Hadoop ecosystems (Spark Hive HBase).
- Strong Java programming skills for cloud-based data tasks.
- Excellent problem-solving and analytical abilities.
- Experience with data pipelines workflows and cloud data architecture.
- Familiarity with containerization (Docker Kubernetes) and CI/CD.
- Strong communication and team collaboration skills.