This is a remote position.
Futuralis is looking for a Data Engineer. An ideal candidate must have five plus years of experience as a Data Engineer and three or more years of handson experience designing and deploying cloud architecture on AWS. This candidate must be comfortable presenting solutions to diverse stakeholders including other engineers data scientists and business leaders. Must be comfortable creating structure for ambiguous engineering problems and be adept at providing scalable solutions.
Requirements
Proficiency in Python programming for data manipulation scripting and automation.
Strong working knowledge of Apache Spark for distributed data processing and analytics.
Handson experience with Amazon Redshift or similar data warehousing technologies.
Advanced proficiency in SQL for querying analyzing and manipulating large datasets.
Experience with Advanced CDK (Cloud Development Kit) for infrastructure as code (IaC) development and deployment is highly desirable.
AWS CLI AWS APIs AWS CloudFormation templates AWS Management Console scripting language.
Integrating external APIs
Building CI/CD pipelines to deploy machine learning models
Big data processing: able to model database schema design and implement scalable data processing workflows
Data modeling data warehousing ETL
Machine Learning: able to productize ML models develop ML endpoints etc.
Familiarity with the full life cycle of software development from design and implementation to testing and deployment
Preferred Experience
Certifications
Proficiency in Python programming for data manipulation, scripting, and automation. Strong working knowledge of Apache Spark for distributed data processing and analytics. Hands-on experience with Amazon Redshift or similar data warehousing technologies. Advanced proficiency in SQL for querying, analyzing, and manipulating large datasets. Experience with Advanced CDK (Cloud Development Kit) for infrastructure as code (IaC) development and deployment is highly desirable. AWS CLI, AWS APIs, AWS CloudFormation templates, AWS Management Console, scripting language. Integrating external APIs Building CI/CD pipelines to deploy machine learning models Big data processing: able to model database schema, design and implement scalable data processing workflows Data modeling, data warehousing, ETL Machine Learning: able to productize ML models, develop ML endpoints, etc. Familiarity with full life cycle of software development, from design and implementation to testing and deployment Preferred Experience Amazon toolset experience - Brazil, Pipelines Certifications AWS Certified Data Analytics - Specialty