Experience as a Data Engineer with strong track record of designing and implementing data solutions.
Experience in programming language as Python or Java with experience in building data pipelines and workflows.
Experience with cloud data warehousing technologies such as Snowflake and Redshift.
Experience with distributed computing frameworks such as Cloudera Apache Hadoop and Spark.
Experience with Cloud platforms such as AWS Azure or Google Cloud Platform.
Experience with AWS cloud services such as S3 EC2 EMR Glue CloudWatch Athena Lambda.
Experience with containerization and orchestration technologies such as Docker and Kubernetes.
Experience with building CI/CD pipeline using tools such has GitLab and Bitbucket.
Experience with data pipeline orchestration tools such as Airflow and Jenkins.
Knowledge of database concepts data modelling schemas and query languages SQL and Hive.
Knowledge of data visualization and reporting tools such as MicroStrategy Tableau and Power BI.
Knowledge of data quality and monitoring techniques and tools such as Great Expectations or similar.
Knowledge of data governance processes lineage cataloging dictionaries using tools such as DataHub or similar.
Knowledge of streaming data processing and real-time analytics technologies such as Kafka.