Job Title:- and nbsp;Senior Data Engineer
Location:- Houston Texas (On-Site/Hybrid)
Job Type:- Long Term Contract
Must be Local to Houston TX
Only GC USC H4EAD Candidates on W2 or 1099 will be entertained
and nbsp;
Overview:
Delivers the Palantir Foundry exit on a modern Snowflake stack by building reliable performant and testable ELT pipelines; recreates Foundry transformations and rule-based event logic; and ensures historical data extraction reconciliation and cutover readiness.
Years of Experience:
7 years overall; 3 years hands-on with Snowflake.
- Key Responsibilities:Extract historical datasets from Palantir (dataset export parquet) to S3/ADLS and load into Snowflake; implement checksum and reconciliation controls.
- Rebuild Foundry transformations as dbt models and/or Snowflake SQL; implement curated schemas and incremental patterns using Streams and Tasks.
- Implement the batch event/rules engine that evaluates time-series plus reference data on a schedule (e.g. 3060 minutes) and produces auditable event tables.
- Configure orchestration in Airflow running on AKS and where appropriate Snowflake Tasks; monitor alert and document operational runbooks.
- Optimize warehouses queries clustering and caching; manage cost with Resource Monitors and usage telemetry.
- Author automated tests (dbt tests Great Expectations or equivalent) validate parity versus legacy outputs and support UAT and cutover.
- Collaborate with BI/analytics teams (Sigma Power BI) on dataset contracts performance and security requirements.
- Required Qualifications:Strong Snowflake SQL and Python for ELT utilities and data validation.
- Production experience with and nbsp;dbt and nbsp;(models tests macros documentation lineage).
- Orchestration with and nbsp;Airflow and nbsp;(preferably on and nbsp;AKS/Kubernetes) and use of and nbsp;Snowflake Tasks/Streams and nbsp;for incrementals.
- Proficiency with cloud object storage (S3/ADLS) file formats (Parquet/CSV) and bulk/incremental load patterns (Snowpipe External Tables).
- Version control and CI/CD with and nbsp;GitHub/GitLab; environment promotion and release hygiene.
- Data quality and reconciliation fundamentals including checksums row/aggregate parity and schema integrity tests.
- Performance and cost tuning using query profiles micro-partitioning behavior and warehouse sizing policies.
- Preferred Qualifications:Experience migrating from legacy platforms (Palantir Foundry and nbsp;Cloudera/Hive/Spark) and familiarity with and nbsp;Trino/Starburst and nbsp;federation patterns.
- Time-series data handling and rules/pattern detection; exposure to and nbsp;Snowpark and nbsp;or UDFs for complex transforms.
- Familiarity with consumption patterns in and nbsp;Sigma and nbsp;and and nbsp;Power BI and nbsp;(Import DirectQuery composite models RLS/OLS considerations).
- Security and governance in Snowflake (RBAC masking row/column policies) tagging and cost allocation.
- Exposure to containerized workloads on and nbsp;AKS lightweight apps for surfacing data (e.g. and nbsp;Streamlit) and basic observability practices.
Job Title:- and nbsp;Senior Data EngineerLocation:- Houston Texas (On-Site/Hybrid)Job Type:- Long Term ContractMust be Local to Houston TXOnly GC USC H4EAD Candidates on W2 or 1099 will be entertained and nbsp;Overview:Delivers the Palantir Foundry exit on a modern Snowflake stack by building reliab...
Job Title:- and nbsp;Senior Data Engineer
Location:- Houston Texas (On-Site/Hybrid)
Job Type:- Long Term Contract
Must be Local to Houston TX
Only GC USC H4EAD Candidates on W2 or 1099 will be entertained
and nbsp;
Overview:
Delivers the Palantir Foundry exit on a modern Snowflake stack by building reliable performant and testable ELT pipelines; recreates Foundry transformations and rule-based event logic; and ensures historical data extraction reconciliation and cutover readiness.
Years of Experience:
7 years overall; 3 years hands-on with Snowflake.
- Key Responsibilities:Extract historical datasets from Palantir (dataset export parquet) to S3/ADLS and load into Snowflake; implement checksum and reconciliation controls.
- Rebuild Foundry transformations as dbt models and/or Snowflake SQL; implement curated schemas and incremental patterns using Streams and Tasks.
- Implement the batch event/rules engine that evaluates time-series plus reference data on a schedule (e.g. 3060 minutes) and produces auditable event tables.
- Configure orchestration in Airflow running on AKS and where appropriate Snowflake Tasks; monitor alert and document operational runbooks.
- Optimize warehouses queries clustering and caching; manage cost with Resource Monitors and usage telemetry.
- Author automated tests (dbt tests Great Expectations or equivalent) validate parity versus legacy outputs and support UAT and cutover.
- Collaborate with BI/analytics teams (Sigma Power BI) on dataset contracts performance and security requirements.
- Required Qualifications:Strong Snowflake SQL and Python for ELT utilities and data validation.
- Production experience with and nbsp;dbt and nbsp;(models tests macros documentation lineage).
- Orchestration with and nbsp;Airflow and nbsp;(preferably on and nbsp;AKS/Kubernetes) and use of and nbsp;Snowflake Tasks/Streams and nbsp;for incrementals.
- Proficiency with cloud object storage (S3/ADLS) file formats (Parquet/CSV) and bulk/incremental load patterns (Snowpipe External Tables).
- Version control and CI/CD with and nbsp;GitHub/GitLab; environment promotion and release hygiene.
- Data quality and reconciliation fundamentals including checksums row/aggregate parity and schema integrity tests.
- Performance and cost tuning using query profiles micro-partitioning behavior and warehouse sizing policies.
- Preferred Qualifications:Experience migrating from legacy platforms (Palantir Foundry and nbsp;Cloudera/Hive/Spark) and familiarity with and nbsp;Trino/Starburst and nbsp;federation patterns.
- Time-series data handling and rules/pattern detection; exposure to and nbsp;Snowpark and nbsp;or UDFs for complex transforms.
- Familiarity with consumption patterns in and nbsp;Sigma and nbsp;and and nbsp;Power BI and nbsp;(Import DirectQuery composite models RLS/OLS considerations).
- Security and governance in Snowflake (RBAC masking row/column policies) tagging and cost allocation.
- Exposure to containerized workloads on and nbsp;AKS lightweight apps for surfacing data (e.g. and nbsp;Streamlit) and basic observability practices.
View more
View less