DevOps Specialist

DevOps Specialist

Peterborough Full-Time 43200 - 72000 Β£ / year (est.) Home office possible
O

At a Glance

  • Tasks: Design and maintain robust infrastructure and data pipelines for AI projects.
  • Company: Join a pioneering AI company aiming to revolutionise the spicy content industry.
  • Benefits: Enjoy a competitive salary, equity, and a flexible remote work culture.
  • Why this job: Be part of a fast-paced team working on cutting-edge AI technology with high impact.
  • Qualifications: 5+ years in DevOps or Data Engineering, strong Python skills, and cloud expertise required.
  • Other info: Opportunity for professional growth and exposure to innovative AI/ML technologies.

The predicted salary is between 43200 - 72000 Β£ per year.

Location: Remote

Type: Full-time

Experience Level: Senior

Industry: Generative AI / Artificial Intelligence / Machine Learning

Reports To: Head of Engineering / CTO

About Us

Ready to join a cutting edge AI company? We’re on a mission to become the OpenAI of the spicy content industry, building a full-spectrum ecosystem of revolutionary AI infrastructure and products. Our platform, OhChat, features digital twins of real-world personalities and original AI characters, enabling users to interact with lifelike AI-generated characters through text, voice, and images, with a roadmap that includes agentic superModels, API integrations, and video capabilities.

Role Overview

We are looking for a Senior DevOps Specialist with a strong python and data engineering background to support our R&D and tech teams by designing, building, and maintaining robust infrastructure and data pipelines across AWS and GCP. You will be instrumental in ensuring our systems are scalable, observable, cost-effective, and secure. This role is hands-on, cross-functional, and central to our product and research success.

Key Responsibilities

  • DevOps & Infrastructure
    • Design, implement, and maintain infrastructure on AWS and Google Cloud Platform (GCP) to support high-performance computing workloads and scalable services.
    • Collaborate with R&D teams to provision and manage compute environments for model training and experimentation.
    • Maintain / monitor systems, implement observability solutions (e.g., logging, metrics, tracing), and proactively resolve infrastructure issues.
    • Manage CI/CD pipelines for rapid, reliable deployment of services and models.
    • Ensure high availability, disaster recovery, and robust security practices across environments.
  • Data Engineering
    • Build and maintain data processing pipelines for model training, experimentation, and analytics.
    • Work closely with machine learning engineers and researchers to understand data requirements and workflows.
    • Design and implement solutions for data ingestion, transformation, and storage using tools such as Scrappy, Playwright, agentic workflows (e.g. crawl4ai) or equivalent.
    • Optimize and benchmark AI training / inference / data workflows to ensure high performance, scalability, cost and an exceptional customer experience.
    • Maintain data quality, lineage, and compliance across multiple environments.

Key Requirements

  • 5+ years of experience in DevOps, Site Reliability Engineering, or Data Engineering roles.
  • Deep expertise with AWS and GCP, including services like EC2, S3, Lambda, IAM, GKE, BigQuery, and more.
  • Strong proficiency in infrastructure-as-code tools (e.g., Terraform, Pulumi, CloudFormation).
  • Extensive hands-on experience with Docker, Kubernetes, and CI/CD tools such as GitHub Actions, Bitbucket Pipelines, or Jenkins, with a strong ability to optimize CI/CD workflows as well as AI training and inference pipelines for performance and reliability.
  • Exceptional programming skills in Python. You are expected to write clean, efficient, and production-ready code. You should be highly proficient with modern Python programming paradigms and tooling.
  • Proficiency in data-centric programming and scripting languages (e.g., Python, SQL, Bash).
  • Proven experience designing and maintaining scalable ETL/ELT pipelines.
  • Focused, sharp, and results-oriented: You are decisive, work with a high degree of autonomy, and consistently deliver high-quality results. You are quick to understand and solve the core of a problem and know how to summarize it efficiently for stakeholders.
  • Effective communicator and concise in reporting: You should be able to communicate technical insights in a clear and actionable manner, both verbally and in written form. Your reports should be precise, insightful, and aligned with business objectives.

Nice to Have

  • Experience supporting AI/ML model training infrastructure (e.g., GPU orchestration, model serving) for both Diffusion- and LLM pipelines.
  • Familiarity with data lake architectures and tools like Delta Lake, LakeFS, or Databricks.
  • Knowledge of security and compliance best practices (e.g., SOC2, ISO 27001).
  • Exposure to MLOps platforms or frameworks (e.g., MLflow, Kubeflow, Vertex AI).

What We Offer

  • Competitive salary + equity
  • Flexible work environment and remote-friendly culture
  • Opportunities to work on cutting-edge AI/ML technology
  • Fast-paced environment with high impact and visibility
  • Professional growth support and resources

DevOps Specialist employer: OhChat

Join a pioneering AI company that is redefining the spicy content industry with innovative technology and a commitment to employee growth. As a remote DevOps Specialist, you'll enjoy a flexible work environment, competitive salary, and the chance to work on groundbreaking AI/ML projects while collaborating with talented professionals in a supportive culture that values autonomy and high-quality results.
O

Contact Detail:

OhChat Recruiting Team

StudySmarter Expert Advice 🀫

We think this is how you could land DevOps Specialist

✨Tip Number 1

Familiarise yourself with the specific tools and technologies mentioned in the job description, such as AWS, GCP, Docker, and Kubernetes. Having hands-on experience or projects that showcase your skills with these platforms can set you apart from other candidates.

✨Tip Number 2

Network with professionals in the AI and DevOps fields. Attend relevant meetups, webinars, or online forums where you can connect with current employees or industry experts. This can provide valuable insights into the company culture and potentially lead to referrals.

✨Tip Number 3

Prepare to discuss your past experiences in detail, especially those related to building and maintaining infrastructure and data pipelines. Be ready to share specific examples of challenges you've faced and how you overcame them, as this demonstrates your problem-solving abilities.

✨Tip Number 4

Showcase your communication skills by being clear and concise when discussing technical topics. Practice explaining complex concepts in simple terms, as effective communication is key in cross-functional roles like this one.

We think you need these skills to ace DevOps Specialist

AWS Expertise
GCP Expertise
Python Programming
Data Engineering
CI/CD Pipeline Management
Infrastructure as Code (Terraform, Pulumi, CloudFormation)
Docker and Kubernetes
ETL/ELT Pipeline Design
Observability Solutions Implementation
Data Quality and Compliance Management
Effective Communication Skills
Problem-Solving Skills
Scalability and Performance Optimisation
Security Best Practices Knowledge
Collaboration with R&D Teams

Some tips for your application 🫑

Tailor Your CV: Make sure your CV highlights relevant experience in DevOps, data engineering, and cloud platforms like AWS and GCP. Use specific examples that demonstrate your expertise with tools such as Docker, Kubernetes, and CI/CD pipelines.

Craft a Compelling Cover Letter: In your cover letter, express your enthusiasm for the role and the company’s mission. Mention how your skills align with their needs, particularly in building scalable infrastructure and data pipelines. Be sure to include any experience you have with AI/ML technologies.

Showcase Your Technical Skills: When detailing your technical skills, focus on your proficiency in Python and any relevant programming languages. Highlight your experience with infrastructure-as-code tools and your ability to write clean, efficient code. This will demonstrate your readiness for the hands-on nature of the role.

Prepare for Technical Questions: Anticipate technical questions related to DevOps practices, data engineering workflows, and cloud services. Be ready to discuss your past projects and how you approached challenges, especially those involving scalability and performance optimisation.

How to prepare for a job interview at OhChat

✨Showcase Your Technical Skills

Be prepared to discuss your experience with AWS and GCP in detail. Highlight specific projects where you've implemented infrastructure-as-code tools like Terraform or CloudFormation, and be ready to explain how you optimised CI/CD workflows.

✨Demonstrate Problem-Solving Abilities

Expect to face scenario-based questions that assess your ability to troubleshoot and resolve infrastructure issues. Use examples from your past experiences to illustrate how you approached and solved complex problems.

✨Communicate Clearly and Concisely

As effective communication is key, practice explaining technical concepts in a straightforward manner. Be ready to summarise your insights and findings in a way that aligns with business objectives, as this will be crucial for collaboration with R&D teams.

✨Prepare for Data Engineering Discussions

Since the role involves data pipelines, brush up on your knowledge of ETL/ELT processes and data quality management. Be ready to discuss your experience with data-centric programming languages and any relevant tools you've used for data ingestion and transformation.

O
Similar positions in other companies
UK’s top job board for Gen Z
discover-jobs-cta
Discover now
>