The main goal of the DE team in 202425 is to build robust golden data sets to power our business goals of creating more insights based datadriven decisions is key to Plaids culture. To support that we need to scale our data systems while maintaining correct and complete data. We provide tooling and guidance to teams across engineering product and business and help them explore our data quickly and safely to get the data insights they need which ultimately helps Plaid serve our customers more Engineers heavily leverage SQL and Python to build data workflows. We use tools like DBT Airflow Redshift ElasticSearch Atlanta and Retool to orchestrate data pipelines and define workflows. We work with engineers product managers business intelligence data analysts and many other teams to build Plaids data strategy and a datafirst mindset. Our engineering culture is ICdriven we favor bottomup ideation and empowerment of our incredibly talented team. We are looking for engineers who are motivated by creating impact for our consumers and customers growing together as a team shipping the MVP and leaving things better than we found them.
You will be in a high impact role that will directly enable business leaders to make faster and more informed business judgements based on the datasets you build. You will have the opportunity to carve out the ownership and scope of internal datasets and visualizations across Plaid which is a currently unowned area that we intend to take over and build SLAs on. You will have the opportunity to learn best practices and uplevel your technical skills from our strong DE team and from the broader Data Platform team. You will collaborate with and have strong and cross functional partnerships with literally all teams at Plaid from Engineering to Product to Marketing/Finance etc.
Responsibilities
Understanding different aspects of the Plaid product and strategy to inform golden dataset choices design and data usage principles.
Have data quality and performance top of mind while designing datasets
Advocating for adopting industry tools and practices at the right time
Owning core SQL and Python data pipelines that power our data lake and data warehouse
Welldocumented data with defined dataset quality uptime and usefulness.
Qualifications
2 years of dedicated data engineering experience solving complex data pipeline issues at scale.
You have experience building data models and data pipelines on top of large datasets (in the order of 500TB to petabytes)
You value SQL as a flexible and extensible tool and are comfortable with modern SQL data orchestration tools like DBT Mode and Airflow.
Nice to have You have experience working with different performant warehouses and data lakes; Redshift Snowflake Databricks
Nice to have You have experience building and maintaining batch and realtime pipelines using technologies like Spark Kafka.
$163200 $223200 a year
The target base salary for this position ranges from $163200/year to $223200/year in Zone 1. The target base salary will vary based on the jobs location.
Our geographic zones are as follows:
Zone 1 New York City and San Francisco Bay Area
Zone 2 Los Angeles Seattle Washington D.C.
Zone 3 Austin Boston Denver Houston Portland Sacramento San Diego
Zone 4 RaleighDurham and all other US cities
Additional compensation in the form(s) of equity and/or commission are dependent on the position offered. Plaid provides a comprehensive benefit plan including medical dental vision and 401(k). Pay is based on factors such as (but not limited to) scope and responsibilities of the position candidates work experience and skillset and location. Pay and benefits are subject to change at any time consistent with the terms of any applicable compensation or benefit plans.
Disclaimer: Drjobpro.com is only a platform that connects job seekers and employers. Applicants are advised to conduct their own independent research into the credentials of the prospective employer.We always make certain that our clients do not endorse any request for money payments, thus we advise against sharing any personal or bank-related information with any third party. If you suspect fraud or malpractice, please contact us via contact us page.