Data Engineer with AI Kubernetes
Columbus, NE - USA
Job Summary
Required Skills & Qualifications:
Strong proficiency in Python for data processing and pipeline development
Hands-on experience with Apache Spark (PySpark preferred)
Solid experience with AWS services such as S3 Glue EMR Redshift Athena Lambda
Experience with SQL and relational/non-relational databases
Knowledge of data modeling data warehousing concepts and ETL frameworks
Experience working with large-scale distributed data systems
Familiarity with CI/CD pipelines and version control tools (Git)
Strong problem-solving and communication skills
Preferred / Nice to Have:
Experience with Airflow or other workflow orchestration tools
Knowledge of Kafka Kinesis or streaming data platforms
Experience with Docker/Kubernetes
Exposure to Delta Lake Iceberg or HuD.
Strong proficiency in Python for data processing and pipeline development
Hands-on experience with Apache Spark (PySpark preferred)
Solid experience with AWS services such as S3 Glue EMR Redshift Athena Lambda
Experience with SQL and relational/non-relational databases
Knowledge of data modeling data warehousing concepts and ETL frameworks
Experience working with large-scale distributed data systems
Familiarity with CI/CD pipelines and version control tools (Git)
Strong problem-solving and communication skills
Preferred / Nice to Have:
Experience with Airflow or other workflow orchestration tools
Knowledge of Kafka Kinesis or streaming data platforms
Experience with Docker/Kubernetes
Exposure to Delta Lake Iceberg or HuD.