Halvik Corp delivers a wide range of services to 13 executive agencies and 15 independent agencies. Halvik is a highly successful WOB business with more than 50 prime contracts and 500 professionals delivering Digital Services Advanced Analytics Artificial Intelligence/Machine Learning Cyber Security and Cutting-Edge Technology across the US Government. Be a part of something special!
Role Summary
The Senior Data Engineer will be a key contributor within the Data Products Delivery Area responsible for designing building and operating scalable secure and high-quality data pipelines that power Business Intelligence & Analytics and AIML products. This role supports a Databricks Data Lakehouse using Medallion Architecture built on Data Fabric principles and operates within a release-based Agile framework.
This role requires strong hands-on development using Python PySpark SQL and APIs and experience integrating diverse enterprise and external data sources.
Key Responsibilities
Data Engineering & Architecture
Design develop and optimize end-to-end data pipelines using Python and PySpark.
Build and maintain data ingestion frameworks leveraging REST APIs streaming APIs and batch interfaces.
Implement Medallion Architecture (Bronze Silver Gold) within a Databricks Lakehouse.
Integrate data from structured semi-structured and unstructured sources including API-based and event-driven sources.
Apply Data Fabric principles including metadata-driven ingestion lineage observability and reusability.
Collaborate with Senior Data Architects on logical and physical data models.
Data Products & Delivery
Deliver data assets as releasable products
Support downstream consumers including:
o BI&A dashboards and visualizations
o AIML feature engineering training and inference pipelines
o Data services and curated datasets exposed via APIs
Decompose requirements into well-defined user stories aligned with MoSCoW prioritization.
Data Quality Governance & Security
Implement data validation reconciliation and quality checks within PySpark pipelines.
Ensure secure handling of PII and sensitive data including encryption masking and access controls.
Partner with CloudOps (DevSecOps FinOps InfraOps) to ensure:
o Secure API access and secrets management
o CI/CD automation for Python and PySpark workloads
o Cost-optimized compute and storage usage
Production Support & Optimization
Support O&M for production data pipelines APIs and analytics products.
Tune Spark jobs SQL queries and API integrations for performance and reliability.
Implement monitoring logging and alerting for data workflows.
Drive automation and refactoring to improve resiliency and scalability.
Collaboration & Leadership
Mentor junior and mid-level data engineers on Python PySpark and API integration best practices.
Work closely with:
o Business Analysts & Data Architects
o Data Scientists & ML Engineers
o BI Developers
o Automation Test Engineers
Contribute to reusable frameworks coding standards and engineering best practices.
Required Skills & Experience
Core Technical Skills
Strong hands-on experience with:
o Python for data engineering and API integration
o PySpark / Apache Spark for large-scale data processing
o SQL (advanced querying and performance tuning)
Proven experience building API-driven data ingestion pipelines (REST JSON OAuth pagination throttling).
Strong experience with Databricks Delta Lake and Lakehouse Architecture.
Experience implementing Medallion Architecture in production environments.
Cloud & DevOps
Experience working in cloud platforms (AWS Azure or GCP).
Familiarity with:
o CI/CD pipelines for Python and Spark workloads
o Infrastructure as Code concepts
o Secure API authentication and secrets management
Cost-conscious engineering mindset (FinOps exposure is a plus).
Data Management & Analytics
Experience supporting BI&A workloads at scale.
Understanding of data modeling for analytics and reporting.
Experience enabling data consumption for AI/ML use cases.
Agile & Delivery
Proven experience working in Agile Scrum / Scrum of Scrums environments.
Comfortable working with release-based delivery and production support.
Nice to Have
Experience exposing data via internal or external APIs.
Familiarity with streaming technologies (Kinesis Firehose Kafka Event Hubs etc.).
Experience with data governance cataloging and lineage tools.
Exposure to feature stores or ML data pipelines.
Soft Skills
Strong analytical and problem-solving skills.
Excellent communication and collaboration abilities.
Ownership mindset with a focus on reliability and quality.
Halvik offers a competitive full benefits package including:
Company-supported medical dental vision life STD and LTD insurance
Benefits include 11 federal holidays and PTO
Eligible employees may receive performance-based incentives in recognition of individual and/or team achievements.
401(k) with company matching
Flexible Spending Accounts for commuter medical and dependent care expenses
Tuition Assistance
Charitable Contribution matching
Halvik Corp is an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to race color religion sex sexual orientation gender identity national origin disability or veteran status.
Halviks pay range for this job level is a general guideline only and not a guarantee of compensation or salary. Additional factors considered in extending an offer include (but are not limited to) responsibilities of the job education experience knowledge skills and abilities as well as internal equity alignment with market data applicable bargaining agreement (if any) or other law.
Required Experience:
Senior IC
Halvik Corp delivers a wide range of services to 13 executive agencies and 15 independent agencies. Halvik is a highly successful WOB business with more than 50 prime contracts and 500 professionals delivering Digital Services Advanced Analytics Artificial Intelligence/Machine Learning Cyber Securit...
Halvik Corp delivers a wide range of services to 13 executive agencies and 15 independent agencies. Halvik is a highly successful WOB business with more than 50 prime contracts and 500 professionals delivering Digital Services Advanced Analytics Artificial Intelligence/Machine Learning Cyber Security and Cutting-Edge Technology across the US Government. Be a part of something special!
Role Summary
The Senior Data Engineer will be a key contributor within the Data Products Delivery Area responsible for designing building and operating scalable secure and high-quality data pipelines that power Business Intelligence & Analytics and AIML products. This role supports a Databricks Data Lakehouse using Medallion Architecture built on Data Fabric principles and operates within a release-based Agile framework.
This role requires strong hands-on development using Python PySpark SQL and APIs and experience integrating diverse enterprise and external data sources.
Key Responsibilities
Data Engineering & Architecture
Design develop and optimize end-to-end data pipelines using Python and PySpark.
Build and maintain data ingestion frameworks leveraging REST APIs streaming APIs and batch interfaces.
Implement Medallion Architecture (Bronze Silver Gold) within a Databricks Lakehouse.
Integrate data from structured semi-structured and unstructured sources including API-based and event-driven sources.
Apply Data Fabric principles including metadata-driven ingestion lineage observability and reusability.
Collaborate with Senior Data Architects on logical and physical data models.
Data Products & Delivery
Deliver data assets as releasable products
Support downstream consumers including:
o BI&A dashboards and visualizations
o AIML feature engineering training and inference pipelines
o Data services and curated datasets exposed via APIs
Decompose requirements into well-defined user stories aligned with MoSCoW prioritization.
Data Quality Governance & Security
Implement data validation reconciliation and quality checks within PySpark pipelines.
Ensure secure handling of PII and sensitive data including encryption masking and access controls.
Partner with CloudOps (DevSecOps FinOps InfraOps) to ensure:
o Secure API access and secrets management
o CI/CD automation for Python and PySpark workloads
o Cost-optimized compute and storage usage
Production Support & Optimization
Support O&M for production data pipelines APIs and analytics products.
Tune Spark jobs SQL queries and API integrations for performance and reliability.
Implement monitoring logging and alerting for data workflows.
Drive automation and refactoring to improve resiliency and scalability.
Collaboration & Leadership
Mentor junior and mid-level data engineers on Python PySpark and API integration best practices.
Work closely with:
o Business Analysts & Data Architects
o Data Scientists & ML Engineers
o BI Developers
o Automation Test Engineers
Contribute to reusable frameworks coding standards and engineering best practices.
Required Skills & Experience
Core Technical Skills
Strong hands-on experience with:
o Python for data engineering and API integration
o PySpark / Apache Spark for large-scale data processing
o SQL (advanced querying and performance tuning)
Proven experience building API-driven data ingestion pipelines (REST JSON OAuth pagination throttling).
Strong experience with Databricks Delta Lake and Lakehouse Architecture.
Experience implementing Medallion Architecture in production environments.
Cloud & DevOps
Experience working in cloud platforms (AWS Azure or GCP).
Familiarity with:
o CI/CD pipelines for Python and Spark workloads
o Infrastructure as Code concepts
o Secure API authentication and secrets management
Cost-conscious engineering mindset (FinOps exposure is a plus).
Data Management & Analytics
Experience supporting BI&A workloads at scale.
Understanding of data modeling for analytics and reporting.
Experience enabling data consumption for AI/ML use cases.
Agile & Delivery
Proven experience working in Agile Scrum / Scrum of Scrums environments.
Comfortable working with release-based delivery and production support.
Nice to Have
Experience exposing data via internal or external APIs.
Familiarity with streaming technologies (Kinesis Firehose Kafka Event Hubs etc.).
Experience with data governance cataloging and lineage tools.
Exposure to feature stores or ML data pipelines.
Soft Skills
Strong analytical and problem-solving skills.
Excellent communication and collaboration abilities.
Ownership mindset with a focus on reliability and quality.
Halvik offers a competitive full benefits package including:
Company-supported medical dental vision life STD and LTD insurance
Benefits include 11 federal holidays and PTO
Eligible employees may receive performance-based incentives in recognition of individual and/or team achievements.
401(k) with company matching
Flexible Spending Accounts for commuter medical and dependent care expenses
Tuition Assistance
Charitable Contribution matching
Halvik Corp is an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to race color religion sex sexual orientation gender identity national origin disability or veteran status.
Halviks pay range for this job level is a general guideline only and not a guarantee of compensation or salary. Additional factors considered in extending an offer include (but are not limited to) responsibilities of the job education experience knowledge skills and abilities as well as internal equity alignment with market data applicable bargaining agreement (if any) or other law.
Required Experience:
Senior IC
View more
View less