Position : Senior Data Architect
Location: Santa Clara CA (Onsite)
Job Type: Full Time only
Responsibilities :
USC/GC/GC EAD/H4EAD/H1B
Job Description:
- Build data pipelines to transport data from a data source to the data lake.
- Build data systems and pipelines ensuring that data sources ingestion components validation functions transformation functions and destination are well understood for implementation.
- Develop and implement the new EndtoEnd Data systems for our Planning Logistics and Services initiatives.
- Prepare data for prescriptive and predictive modeling by making sure that the data is complete has been cleansed and has the necessary rules in place.
- Analyze and organize raw operational data including structured and unstructured data.
- Lead discussions with stakeholders and IT to identify and implement the right data strategy given data sources data locations and use cases.
- Interpret trends and patterns by performing complex data analysis.
- Build/develop algorithms prototypes and analytical tools that enable the Ops teams to make critical business decisions.
What we need to see:
- Masters or Bachelors degree in Computer Science or Information System or equivalent experience.
- 8 years of relevant experience including programming knowledge (i.e Python SQL).
- 5 years of relevant experience in big data technologies and cloud platforms (i.e Spark AWS).
- 5 years of relevant experience in data lake technologies (i.e Iceberg Delta Huidi).
- 5 years of experience in development best practices like CICD Unit testing Integration testing
- 5 years of experience grabbing data from source systems like REST APIs other databases using JDBC ODBC SFTP servers.
- Differentiating skill sets:
- 2 years of experience with kubernetes and docker
- Experience in developing required infrastructure for optimal extraction transformation and loading of data from various sources using AWS Azure SQL or other technologies.
- Experience architecting designing developing and maintaining data warehouses/data lakes for complex data ecosystems.
- Experience working with large datasets databases and the software used to analyze the data.
- Expert in data and database management including data pipeline responsibilities in replication and mass ingestion streaming API and application and data integration.
- Strong analytical skills with the ability to collect organize and disseminate significant amounts of information with attention to detail and accuracy.
- Highly independent able to lead key technical decisions influence project roadmap and work effectively with team members.
Ways to stand out from the crowd:
- Knowledge in operational processes in chips boards systems and servers with a view of data landscape.
- Selfstarter positive mindset with integrity and accountability highly motivated driven and highreaching.
- Solid ability to drive continuous improvement of systems and processes.
- A consistent record to work in a fastpaced environment where good interpersonal skills are crucial.