At a Glance
- Tasks: Design and enhance scalable ML platforms for cutting-edge AI projects.
- Company: Join talabat, a leading on-demand food and delivery app with a vibrant culture.
- Benefits: Enjoy competitive salary, health benefits, and opportunities for remote work.
- Why this job: Make a real impact in the AI space while working with innovative technologies.
- Qualifications: 3+ years in ML platform engineering and strong Python skills required.
- Other info: Be part of a dynamic team with excellent career growth opportunities.
The predicted salary is between 54000 - 84000 ÂŁ per year.
Since launching in Kuwait in 2004, talabat, the leading on-demand food and Q-commerce app for everyday deliveries, has been offering convenience and reliability to its customers. talabat’s local roots run deep, offering a real understanding of the needs of the communities we serve in eight countries across the region. We harness innovative technology and knowledge to simplify everyday life for our customers, optimize operations for our restaurants and local shops, and provide our riders with reliable earning opportunities daily. Here at talabat, we are building a high‑performance culture through an engaged workforce and growing talent density. We’re all about keeping it real and making a difference. Our 6,000+ strong talabaty are on an awesome mission to spread positive vibes. We are proud to be a multi‑great place to work award winner.
As the leading delivery platform in the region, we have a unique responsibility and opportunity to positively impact millions of customers, restaurant partners, and riders. To achieve our mission, we must scale and continuously evolve our machine learning capabilities, including cutting‑edge Generative AI (genAI) initiatives. This demands robust, efficient, and scalable ML platforms that empower our teams to rapidly develop, deploy, and operate intelligent systems. As an ML Platform Engineer, your mission is to design, build, and enhance the infrastructure and tooling that accelerates the development, deployment, and monitoring of traditional ML and genAI models at scale. You’ll collaborate closely with data scientists, ML engineers, genAI specialists, and product teams to deliver seamless ML workflows—from experimentation to production serving—ensuring operational excellence across our ML and genAI systems.
Responsibilities
- Design, build, and maintain scalable, reusable, and reliable ML platforms and tooling that support the entire ML lifecycle, including data ingestion, model training, evaluation, deployment, and monitoring for both traditional and generative AI models.
- Develop standardized ML workflows and templates using MLflow and other platforms, enabling rapid experimentation and deployment cycles.
- Implement robust CI/CD pipelines, Docker containerization, model registries, and experiment tracking to support reproducibility, scalability, and governance in ML and genAI.
- Collaborate closely with genAI experts to integrate and optimize genAI technologies, including transformers, embeddings, vector databases (e.g., Pinecone, Redis, Weaviate), and real‑time retrieval‑augmented generation (RAG) systems.
- Automate and streamline ML and genAI model training, inference, deployment, and versioning workflows, ensuring consistency, reliability, and adherence to industry best practices.
- Ensure reliability, observability, and scalability of production ML and genAI workloads by implementing comprehensive monitoring, alerting, and continuous performance evaluation.
- Integrate infrastructure components such as real‑time model serving frameworks (e.g., TensorFlow Serving, NVIDIA Triton, Seldon), Kubernetes orchestration, and cloud solutions (AWS/GCP) for robust production environments.
- Drive infrastructure optimization for generative AI use‑cases, including efficient inference techniques (batching, caching, quantization), fine‑tuning, prompt management, and model updates at scale.
- Partner with data engineering, product, infrastructure, and genAI teams to align ML platform initiatives with broader company goals, infrastructure strategy, and innovation roadmap.
- Contribute actively to internal documentation, onboarding, and training programs, promoting platform adoption and continuous improvement.
Technical Experience
- Strong software engineering background with experience in building distributed systems or platforms designed for machine learning and AI workloads.
- Expert‑level proficiency in Python and familiarity with ML frameworks (TensorFlow, PyTorch), infrastructure tooling (MLflow, Kubeflow, Ray), and popular APIs (Hugging Face, OpenAI, LangChain).
- Experience implementing modern MLOps practices, including model lifecycle management, CI/CD, Docker, Kubernetes, model registries, and infrastructure‑as‑code tools (Terraform, Helm).
- Demonstrated experience working with cloud infrastructure, ideally AWS or GCP, including Kubernetes clusters (GKE/EKS), serverless architectures, and managed ML services (e.g., Vertex AI, SageMaker).
- Proven experience with generative AI technologies: transformers, embeddings, prompt engineering strategies, fine‑tuning vs. prompt‑tuning, vector databases, and retrieval‑augmented generation (RAG) systems.
- Experience designing and maintaining real‑time inference pipelines, including integrations with feature stores, streaming data platforms (Kafka, Kinesis), and observability platforms.
- Familiarity with SQL and data warehouse modeling; capable of managing complex data queries, joins, aggregations, and transformations.
- Solid understanding of ML monitoring, including identifying model drift, decay, latency optimization, cost management, and scaling API‑based genAI applications efficiently.
Qualifications
- Bachelor’s degree in Computer Science, Engineering, or a related field; advanced degree is a plus.
- 3+ years of experience in ML platform engineering, ML infrastructure, generative AI, or closely related roles.
- Proven track record of successfully building and operating ML infrastructure at scale, ideally supporting generative AI use‑cases and complex inference scenarios.
- Strategic mindset with strong problem‑solving skills and effective technical decision‑making abilities.
- Excellent communication and collaboration skills, comfortable working cross‑functionally across diverse teams and stakeholders.
- Strong sense of ownership, accountability, pragmatism, and proactive bias for action.
Sr. Engineer, ML Platform (Based in Dubai) employer: talabat
Contact Detail:
talabat Recruiting Team
StudySmarter Expert Advice 🤫
We think this is how you could land Sr. Engineer, ML Platform (Based in Dubai)
✨Tip Number 1
Network like a pro! Reach out to people in the industry, especially those at talabat. A friendly chat can open doors that applications alone can't.
✨Tip Number 2
Show off your skills! If you have a portfolio or projects related to ML platforms, share them during interviews or networking events. It’s a great way to demonstrate your expertise.
✨Tip Number 3
Prepare for technical interviews by brushing up on relevant ML concepts and tools. Practice coding challenges and be ready to discuss your past experiences in detail.
✨Tip Number 4
Don’t forget to apply through our website! It’s the best way to ensure your application gets noticed and shows your genuine interest in joining the talabat team.
Some tips for your application 🫡
Tailor Your CV: Make sure your CV is tailored to the role of Sr. Engineer, ML Platform. Highlight your experience with machine learning platforms and any relevant projects you've worked on. We want to see how your skills align with what we're looking for!
Craft a Compelling Cover Letter: Your cover letter is your chance to shine! Use it to explain why you're passionate about the role and how you can contribute to our mission at talabat. Keep it engaging and personal – we love to see your personality come through!
Showcase Your Technical Skills: Don’t forget to highlight your technical expertise in Python, ML frameworks, and cloud infrastructure. We’re keen on seeing how you’ve applied these skills in real-world scenarios, so be specific about your achievements!
Apply Through Our Website: We encourage you to apply directly through our website. It’s the best way to ensure your application gets into the right hands. Plus, it shows us that you’re genuinely interested in joining our awesome team!
How to prepare for a job interview at talabat
✨Know Your Tech Inside Out
Make sure you’re well-versed in the technologies mentioned in the job description, like Python, TensorFlow, and Kubernetes. Brush up on your knowledge of ML frameworks and generative AI technologies, as you might be asked to discuss how you've used them in past projects.
✨Showcase Your Problem-Solving Skills
Prepare to share specific examples of challenges you've faced in ML platform engineering. Use the STAR method (Situation, Task, Action, Result) to structure your answers, highlighting how you approached problems and what solutions you implemented.
✨Collaborate Like a Pro
Since the role involves working closely with various teams, be ready to discuss your experience in cross-functional collaboration. Share instances where you’ve successfully partnered with data scientists or product teams to deliver projects, emphasising your communication skills.
✨Ask Insightful Questions
Prepare thoughtful questions about the company’s ML initiatives and future projects. This shows your genuine interest in the role and helps you gauge if the company culture aligns with your values, especially their focus on innovation and community impact.