GCP Data Engineer with Python

Saransh Inc

Not Interested
Bookmark
Report This Job

profile Job Location:

Dearborn, MI - USA

profile Monthly Salary: Not Disclosed
Posted on: 13 hours ago
Vacancies: 1 Vacancy

Job Summary

Role: GCP Data Engineer with Python
Location: Dearborn MI (4 days a week onsite)
Job Type: Contract
Experience: Overall 8 to 12 years

Job Summary:
  • The Data Engineer will be responsible for supporting the Credit Global Securitization (GS) teams upskilling initiative by contributing to data engineering efforts across cloud and traditional platforms.
  • This role is intended to accelerate development and delivery.
  • The engineer will work closely with cross-functional teams to build optimize and maintain data pipelines and workflows using GCP Python and ETL tools.

Required Technical Skills:
  • Minimum 3 years of hands-on experience with Google Cloud Platform (GCP) specifically using Astronomer/Composer for orchestration.
  • Strong proficiency in Python for data engineering and automation.
  • Experience with RDBMS technologies such as DB2 and Teradata.
  • Exposure to Big Data ecosystems and distributed data processing.

Nice to have Technical Skills:
  • Prior experience with ETL tools like DataStage or Informatica.

Responsibilities:
  • The Data Engineer will play a key role in the developing and maintaining scalable data pipelines and workflows.
  • The engineer will work with GCP tools like Astronomer/Composer and leverage Python for automation and transformation tasks.
  • The role involves integrating data from RDBMS platforms such as DB2 and Teradata and supporting ETL processes using tools like DataStage or Informatica.
  • The engineer will collaborate with existing team members including Software Analysts and Scrum Masters and will be expected to contribute to knowledge sharing and process improvement.
Specifically:
  • Develop and implement solutions using GCP Python Big Data technologies to enhance data analysis capabilities.
  • Collaborate with cross-functional teams to design and optimize data models in Teradata and DB2 environments.
  • Utilize Python for scripting and automation to streamline geospatial data processing tasks.
  • Integrate and manage data workflows using Cloud Composer to ensure efficient data pipeline operations.
  • Leverage GCP Cloud to deploy scalable applications and services.
Role: GCP Data Engineer with Python Location: Dearborn MI (4 days a week onsite) Job Type: Contract Experience: Overall 8 to 12 years Job Summary: The Data Engineer will be responsible for supporting the Credit Global Securitization (GS) teams upskilling initiative by contributing to data eng...
View more view more

Key Skills

  • Apache Hive
  • S3
  • Hadoop
  • Redshift
  • Spark
  • AWS
  • Apache Pig
  • NoSQL
  • Big Data
  • Data Warehouse
  • Kafka
  • Scala