Backend AI & Data Pipeline Engineer

Not Interested
Bookmark
Report This Job

profile Job Location:

Islamabad - Pakistan

profile Monthly Salary: Not Disclosed
Posted on: Yesterday
Vacancies: 1 Vacancy

Job Summary

About the role

We are looking for a Backend AI & Data Pipeline Engineer to own the end-to-end data processing infrastructure that powers Yuzees intelligent course and job matching platform. You will design and maintain scalable event-driven pipelines that process tens of thousands of daily records generate semantic embeddings and feed a growing knowledge graph used for personalised career pathway recommendations.

What youll do

  • Design and maintain three distinct processing pipelines scheduled job ingestion event-driven course processing and a periodic knowledge graph builder each with independent trigger logic and cost controls
  • Generate and manage semantic embeddings via Amazon Bedrock (Titan v2) index them in MongoDB Atlas Vector Search and calibrate similarity thresholds to ensure match accuracy
  • Build and maintain a knowledge graph linking jobs courses skills and industries using FP-Growth association rules and archetype-to-SOC code mapping
  • Build and improve a two-stage discovery and matching API on AWS Lambda vector retrieval first then deep eligibility scoring with LLM re-ranking
  • Right-size Fargate Spot instances and design resumable processing loops that tolerate interruption keeping infrastructure costs under control as data volume scales
  • Maintain and improve daily job scrapers across multiple sources and build institution data scrapers with robust HTML cleaning pipelines

What were looking for

  • 1 years of backend engineering experience focused on data pipelines ML infrastructure or search systems
  • Hands-on experience with AWS serverless and container services Lambda ECS Fargate EventBridge and Step Functions
  • Strong Python skills Pandas async processing bulk database operations and text cleaning
  • Familiarity with vector databases and semantic similarity search; MongoDB Atlas Vector Search experience is a strong plus
  • Cost-conscious infrastructure mindset you think in per-record compute costs free tiers Spot resilience and right-sizing
  • Ability to document and communicate complex architecture clearly to both technical and non-technical stakeholders

Nice to have

  • Experience with knowledge graphs or association rule mining (FP-Growth Apriori)
  • Experience using LLMs for re-ranking or eligibility assessment on top of vector retrieval results
  • Background in edtech jobtech or recommendation/matching systems

Qualifications :

Degree or existing proven experience 


Additional Information :

Benefits 

  • You can work from home for the whole internship period
  • A reference letter can be requested upon completion of internship
  • A bit of flexibility with working time aside from the usual 9am to 6pm (Ex. 8am to 5pm / 7:30am to 4:30pm)
  • The possibility of retainment for part-time or Full-time work post-internship based on your performance even if you are not based in Malaysia

Remote Work :

Yes


Employment Type :

Intern

About the roleWe are looking for a Backend AI & Data Pipeline Engineer to own the end-to-end data processing infrastructure that powers Yuzees intelligent course and job matching platform. You will design and maintain scalable event-driven pipelines that process tens of thousands of daily records ge...
View more view more

Key Skills


About Company

SEEKA Technologies (Not Seeka Limited) is a project under its parent organization called Fresh Futures Australia which is an education consultant based in both Australia and Malaysia. It will be a huge platform that utilises A.I. to help match students and job seekers to the right opp ... View more

View Profile View Profile