This is a remote position.
We are seeking a Data Engineer (Python and SQL) to join our team.
Responsibilties:
- AI patterns recognition: Youll be developing multiple AI features from user categorization to autofilled descriptions market and conversations sentiment analysis and AI insights to help crypto marketers.
- SDK User Graph improvement and reliability: Were constantly improving our proprietary user graph by matching wallets to social profiles. Your mission will be supporting new crypto wallets and networks automating a system to match more users to their identities and applying data checks to ensure reliability.
- Integrations: As a customer data platform were expanding the number of data sources our customers can import from Web2 and Web3. Youll manage multiple API endpoints and integrate new third-party tools like Mixpanel Amplitude Segment Dune Analytics and DeFi Llama. This involves not only integration work but also data modeling and architectural design.
- Social Data analysis: Youll work with social data APIs like Twitter to analyze Key Opinion Leaders performance and trends.
Requirements
- Proven track record as a Data Engineer delivering complex data solutions.
- Advanced SQL skills and expertise with complex queries.
- Mastery in Python development and strong experience with PySpark.
Extensive proficiency managing cloud services including AWS Redshift RDS Postgres S3 Lambda Kinesis SQS ECS EC2. - Strong competency implementing and supporting various data models such as highly normalized star schema and Data Vault.
- Practical experience with orchestration tools like Airflow Dagster or Prefect.
- Demonstrated proficiency consuming and automating interactions with APIs.
- Hands-on experience creating data pipelines using dbt for various platforms (ideally AWS Redshift and Postgres).
- Experience in SaaS analytics marketing or crypto companies is a plus.
Benefits
- Work Location: Remote
- 5 days working
Proven track record as a Data Engineer delivering complex data solutions. Advanced SQL skills and expertise with complex queries. Mastery in Python development and strong experience with PySpark. Extensive proficiency managing cloud services, including AWS Redshift, RDS Postgres, S3, Lambda, Kinesis, SQS, ECS, EC2. Strong competency implementing and supporting various data models, such as highly normalized, star schema and Data Vault. Practical experience with orchestration tools like Airflow, Dagster or Prefect. Demonstrated proficiency consuming and automating interactions with APIs. Hands-on experience creating data pipelines using dbt for various platforms (ideally AWS Redshift and Postgres). Experience in SaaS analytics, marketing, or crypto companies is a plus.