Overview
The Data Engineer plays a crucial role in the organization being responsible for developing constructing testing and maintaining architectures such as databases and largescale processing systems. They work closely with data scientists analysts and other stakeholders to understand their goals and requirements. Data engineers are instrumental in transforming data into a format that can be easily analyzed to support the organizations decisionmaking.
Key responsibilities
Mandatory skills
AWS Transfer Family
Amazon Glue
AWS Lambda
Amazon MWAA (Airflow) Redshift S3
Skills required
Work with Sr. Data engineer to understand the data pipeline design.
Responsible for data acquisition and data processing. Build and test data pipelines.
Configure data pipelines to connect to data sources and target data warehouses.
Orchestrate data pipelines
Required qualifications
- Bachelors or Masters degree in Computer Science Information Technology or a related field.
- Proven experience as a Data Engineer or in a similar role.
- Proficiency in programming languages such as Python Java or Scala.
- Strong SQL skills with the ability to develop optimize and debug complex SQL queries.
- Experience with data modeling data warehousing and building ETL pipelines.
- Knowledge of big data technologies such as Hadoop Spark or NoSQL databases.
- Handson experience with cloud platforms such as AWS Azure or GCP.
- Experience in creating and maintaining scalable faulttolerant and highperformance data processing solutions.
- Ability to work with stakeholders to understand their needs and present solutions in a clear and concise manner.
- Strong analytical and problemsolving skills with great attention to detail.
- Understanding of data governance best practices and data security principles.
- Excellent communication and teamwork skills to work effectively with crossfunctional teams.
- Familiarity with version control systems and CI/CD pipelines.
- Experience with agile methodologies and the software development life cycle.
- Ability to learn and apply new technologies quickly and independently.
spark,aws,databases,redshift,etl pipelines,amazon glue,aws transfer family,cloud platforms (aws, azure, gcp),data engineering,version control systems,python,data security,data warehousing,airflow,data governance,agile methodologies,glue,s3,amazon mwaa (airflow),ci/cd pipelines,aws lambda,amazon redshift,sql,hadoop,nosql databases,java,stakeholder management,communication skills,teamwork,data modeling,scala,software development life cycle