Overview:
The role of a Data Engineer is crucial to our organization as they are responsible for developing constructing testing and maintaining architectures such as databases and largescale processing systems. Their work ensures that the data is available and accessible for insights and analysis enabling the organization to make datadriven decisions.
Key Responsibilities:
- Develop construct test and maintain scalable data pipelines and datadriven systems.
- Design and implement data infrastructure and perform data modeling.
- Work closely with data scientists and analysts to understand data needs.
- Optimize and improve existing database and data processes.
- Create and maintain optimal data pipeline architecture.
- Identify design and implement internal process improvements including automating manual processes optimizing data delivery and redesigning infrastructure for greater scalability.
- Assemble large complex data sets for analysis.
- Build analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition operational efficiency and other key business performance metrics.
- Implement and maintain data security and integrity.
- Support and work with crossfunctional teams in gathering and maintaining data from different sources.
- Troubleshoot datarelated problems and authorize maintenance or modifications.
- Ensure proper data governance policies are followed by implementing or validating data quality checks and monitoring.
- Collaborate with DevOps software developers and other engineers to optimize systems.
- Documentation of data infrastructure processes and workflows.
Required Qualifications:
- Bachelors degree in Computer Science Information Technology or related field; or equivalent work experience.
- Proven experience as a Data Engineer or similar role.
- Strong understanding of ETL processes and data modeling.
- Proficient in SQL and NoSQL database design and development.
- Experience with programming languages such as Python Java or Scala.
- Indepth knowledge of big data technologies and frameworks such as Hadoop Spark or Hive.
- Experience with cloud platforms such as AWS Azure or Google Cloud.
- Solid understanding of data warehousing and data lake concepts.
- Ability to work with unstructured data and various types of data formats.
- Experience in building and optimizing big data data pipelines architectures and data sets.
- Excellent analytical and problemsolving skills.
- Strong teamwork and communication skills.
- Experience with data governance and data security practices.
- Ability to work in a fastpaced and continuously evolving environment.
- Relevant certifications in data engineering or related fields would be a plus.
etl,sql,python,data modeling,big data,aws,spark,data warehousing,infrastructure,data governance,cloud,design,data infrastructure,database,data security,data