Employer Active
Job Alert
You will be updated with latest job alerts via emailJob Alert
You will be updated with latest job alerts via emailACG2677JOB
Our client is a fastgrowing firm specialized in fintech who is looking for a qualified candidate to join their firm.
Data Infrastructure Design & Optimization: Take charge of designing managing and refining core data infrastructure components such as data lakes databases pipelines and metadata repositories to ensure consistent and accurate delivery of data insights to business and operational teams.
Crossfunctional Collaboration: Engage closely with data science and product teams to build scalable data models and structures onboard new data sources and partner with fellow engineers to apply modern technologies in data engineering.
Scalable Data Processing: Build and tune both realtime and batch data processing frameworks to support evolving business needs and ensure performance at scale.
Workflow Orchestration: Manage and monitor complex workflows using tools such as Apache Airflow or AWS Batch ensuring smooth data operations.
System Quality & Testing: Implement rigorous testing methodologies to validate the performance reliability and scalability of data pipelines and systems.
Continuous Enhancement: Regularly research and apply new advancements in data engineering to improve overall development workflows and system performance.
Solid grasp of distributed system design and architecture principles.
Handson experience in building and maintaining scalable ETL/ELT pipelines.
Proficient in optimizing Hadoop or Sparkbased data solutions.
Advanced Python skills especially in libraries like PySpark Pandas NumPy/SciPy and Polars.
Familiarity with version control tools such as Git.
Skilled in working with analytical and big data platforms such as Athena Redshift Hive or Vertica.
Strong experience with cloud platforms ideally AWS (Glue EMR S3 EC2 Lambda etc..
Practical experience in container tools such as Docker ECS or Kubernetes.
Working knowledge of CI/CD tools (e.g. Jenkins AWS CodePipeline CircleCI).
Familiarity with JVMbased languages such as Java or Scala is a plus.
Exposure to RDBMS like MySQL/PostgreSQL and NoSQL solutions like Redis or DynamoDB is an advantage.
Experience using enterprise BI platforms such as Power BI Tableau or Looker is preferred.
Handson experience with platforms like AWS SageMaker or Databricks is a strong plus.
Familiarity with observability tools such as the ELK stack or Datadog is a preferred asset.
Awareness of data privacy protocols and security best practices is considered an advantage.
Knowledge of event streaming technologies like Apache Kafka or RabbitMQ is a nice bonus.
Due to the immense number of applications only shortlisted candidates will be contacted.
Full Time