Location: LATAM (Remote supporting US-based teams)
Job Type: Contractor (Full-Time Remote)
Project: Data Platform Modernization Healthcare Sector
Time Zone: Aligned with GMT-3 (Argentina)
English Level: B2/C1
Get to Know Us
At Darwoft we build software that drives real change. But were more than just tech were people first. With a remote-first culture and a highly collaborative team spread across LATAM we partner with global companies to co-create reliable scalable and impactful digital products.
Were currently working with a leading US-based healthtech platform in a major transformation of their data pipeline ecosystem migrating legacy SQL logic into modern scalable cloud-based infrastructure using DBT Spark Argo and AWS.
Were Looking For a Senior Data Engineer (DBT Spark Argo)
In this role you will be at the core of a strategic data transformation initiative: converting monolithic SQL Server logic into a modular testable DBT architecture while integrating Spark for performance and Argo for orchestration. You will work with cutting-edge lakehouse formats like Apache Hudi Parquet and Iceberg and enable real-time analytics through ElasticSearch integration.
If youre passionate about modern data engineering and want to work in a data-driven cloud-native healthcare-focused environment this is the role for you.
What Youll Be Doing
-
Translate legacy T-SQL logic into modular scalable DBT models powered by Spark SQL
-
Build reusable and performant data transformation pipelines
-
Develop testing frameworks to ensure data accuracy and integrity in DBT workflows
-
Design and orchestrate workflows using Argo Workflows and CI/CD pipelines with Argo CD
-
Manage mock data and reference datasets (e.g. ICD-10 CPT) ensuring version control and governance
-
Implement efficient storage/query strategies using Apache Hudi Parquet and Iceberg
-
Integrate ElasticSearch for analytics by building APIs and pipelines to support indexing and querying
-
Collaborate with DevOps teams to optimize S3 usage enforce data security and ensure compliance
-
Work in Agile squads and participate in planning estimation and sprint reviews
What You Bring
-
Strong experience with DBT for data modeling testing and deployment
-
Hands-on proficiency in Spark SQL including performance tuning
-
Solid programming skills in Python for automation and data manipulation
-
Familiarity with Jinja templating for building reusable DBT components
-
Practical experience with data lake formats: Apache Hudi Parquet Iceberg
-
Expertise in Argo Workflows and CI/CD integration with Argo CD
-
Deep understanding of AWS S3 data storage performance tuning and cost optimization
-
Strong command of ElasticSearch for indexing structured/unstructured data
-
Knowledge of ICD-10 CPT and other healthcare data standards
-
Ability to work cross-functionally in Agile environments
Nice to Have
-
Experience with Docker Kubernetes and container orchestration
-
Familiarity with cloud-native data tools: AWS Glue Databricks EMR or GCP equivalents
-
Prior work on CI/CD automation for data engineering workflows
-
Knowledge of data compliance standards: HIPAA SOC2 etc.
-
Contributions to open-source projects in DBT Spark or data engineering frameworks
Perks & Benefits
-
Contractor agreement with payment in USD
-
100% remote work
-
Argentinas public holidays
-
English classes
-
Referral program
-
Access to learning platforms
Explore this and other opportunities at:
Location: LATAM (Remote supporting US-based teams) Job Type: Contractor (Full-Time Remote) Project: Data Platform Modernization Healthcare Sector Time Zone: Aligned with GMT-3 (Argentina) English Level: B2/C1 Get to Know Us At Darwoft we build software that drives real change. But were more than jus...
Location: LATAM (Remote supporting US-based teams)
Job Type: Contractor (Full-Time Remote)
Project: Data Platform Modernization Healthcare Sector
Time Zone: Aligned with GMT-3 (Argentina)
English Level: B2/C1
Get to Know Us
At Darwoft we build software that drives real change. But were more than just tech were people first. With a remote-first culture and a highly collaborative team spread across LATAM we partner with global companies to co-create reliable scalable and impactful digital products.
Were currently working with a leading US-based healthtech platform in a major transformation of their data pipeline ecosystem migrating legacy SQL logic into modern scalable cloud-based infrastructure using DBT Spark Argo and AWS.
Were Looking For a Senior Data Engineer (DBT Spark Argo)
In this role you will be at the core of a strategic data transformation initiative: converting monolithic SQL Server logic into a modular testable DBT architecture while integrating Spark for performance and Argo for orchestration. You will work with cutting-edge lakehouse formats like Apache Hudi Parquet and Iceberg and enable real-time analytics through ElasticSearch integration.
If youre passionate about modern data engineering and want to work in a data-driven cloud-native healthcare-focused environment this is the role for you.
What Youll Be Doing
-
Translate legacy T-SQL logic into modular scalable DBT models powered by Spark SQL
-
Build reusable and performant data transformation pipelines
-
Develop testing frameworks to ensure data accuracy and integrity in DBT workflows
-
Design and orchestrate workflows using Argo Workflows and CI/CD pipelines with Argo CD
-
Manage mock data and reference datasets (e.g. ICD-10 CPT) ensuring version control and governance
-
Implement efficient storage/query strategies using Apache Hudi Parquet and Iceberg
-
Integrate ElasticSearch for analytics by building APIs and pipelines to support indexing and querying
-
Collaborate with DevOps teams to optimize S3 usage enforce data security and ensure compliance
-
Work in Agile squads and participate in planning estimation and sprint reviews
What You Bring
-
Strong experience with DBT for data modeling testing and deployment
-
Hands-on proficiency in Spark SQL including performance tuning
-
Solid programming skills in Python for automation and data manipulation
-
Familiarity with Jinja templating for building reusable DBT components
-
Practical experience with data lake formats: Apache Hudi Parquet Iceberg
-
Expertise in Argo Workflows and CI/CD integration with Argo CD
-
Deep understanding of AWS S3 data storage performance tuning and cost optimization
-
Strong command of ElasticSearch for indexing structured/unstructured data
-
Knowledge of ICD-10 CPT and other healthcare data standards
-
Ability to work cross-functionally in Agile environments
Nice to Have
-
Experience with Docker Kubernetes and container orchestration
-
Familiarity with cloud-native data tools: AWS Glue Databricks EMR or GCP equivalents
-
Prior work on CI/CD automation for data engineering workflows
-
Knowledge of data compliance standards: HIPAA SOC2 etc.
-
Contributions to open-source projects in DBT Spark or data engineering frameworks
Perks & Benefits
-
Contractor agreement with payment in USD
-
100% remote work
-
Argentinas public holidays
-
English classes
-
Referral program
-
Access to learning platforms
Explore this and other opportunities at:
View more
View less