At a Glance
- Tasks: Design and maintain robust infrastructure and data pipelines for AI projects.
- Company: Join a pioneering AI company aiming to revolutionise the spicy content industry.
- Benefits: Enjoy a competitive salary, equity, and a flexible remote work culture.
- Why this job: Be part of a fast-paced team working on cutting-edge AI technology with high impact.
- Qualifications: 5+ years in DevOps or Data Engineering, strong Python skills, and cloud expertise required.
- Other info: Opportunity for professional growth and exposure to innovative AI/ML technologies.
The predicted salary is between 43200 - 72000 £ per year.
Location: Remote
Type: Full-time
Experience Level: Senior
Industry: Generative AI / Artificial Intelligence / Machine Learning
Reports To: Head of Engineering / CTO
About Us
Ready to join a cutting edge AI company? We’re on a mission to become the OpenAI of the spicy content industry, building a full-spectrum ecosystem of revolutionary AI infrastructure and products. Our platform, OhChat, features digital twins of real-world personalities and original AI characters, enabling users to interact with lifelike AI-generated characters through text, voice, and images, with a roadmap that includes agentic superModels, API integrations, and video capabilities.
Role Overview
We are looking for a Senior DevOps Specialist with a strong python and data engineering background to support our R&D and tech teams by designing, building, and maintaining robust infrastructure and data pipelines across AWS and GCP. You will be instrumental in ensuring our systems are scalable, observable, cost-effective, and secure. This role is hands-on, cross-functional, and central to our product and research success.
Key Responsibilities
- DevOps & Infrastructure
- Design, implement, and maintain infrastructure on AWS and Google Cloud Platform (GCP) to support high-performance computing workloads and scalable services.
- Collaborate with R&D teams to provision and manage compute environments for model training and experimentation.
- Maintain / monitor systems, implement observability solutions (e.g., logging, metrics, tracing), and proactively resolve infrastructure issues.
- Manage CI/CD pipelines for rapid, reliable deployment of services and models.
- Ensure high availability, disaster recovery, and robust security practices across environments.
- Build and maintain data processing pipelines for model training, experimentation, and analytics.
- Work closely with machine learning engineers and researchers to understand data requirements and workflows.
- Design and implement solutions for data ingestion, transformation, and storage using tools such as Scrappy, Playwright, agentic workflows (e.g. crawl4ai) or equivalent.
- Optimize and benchmark AI training / inference / data workflows to ensure high performance, scalability, cost and an exceptional customer experience.
- Maintain data quality, lineage, and compliance across multiple environments.
Key Requirements
- 5+ years of experience in DevOps, Site Reliability Engineering, or Data Engineering roles.
- Deep expertise with AWS and GCP, including services like EC2, S3, Lambda, IAM, GKE, BigQuery, and more.
- Strong proficiency in infrastructure-as-code tools (e.g., Terraform, Pulumi, CloudFormation).
- Extensive hands-on experience with Docker, Kubernetes, and CI/CD tools such as GitHub Actions, Bitbucket Pipelines, or Jenkins, with a strong ability to optimize CI/CD workflows as well as AI training and inference pipelines for performance and reliability.
- Exceptional programming skills in Python. You are expected to write clean, efficient, and production-ready code. You should be highly proficient with modern Python programming paradigms and tooling.
- Proficiency in data-centric programming and scripting languages (e.g., Python, SQL, Bash).
- Proven experience designing and maintaining scalable ETL/ELT pipelines.
- Focused, sharp, and results-oriented: You are decisive, work with a high degree of autonomy, and consistently deliver high-quality results. You are quick to understand and solve the core of a problem and know how to summarize it efficiently for stakeholders.
- Effective communicator and concise in reporting: You should be able to communicate technical insights in a clear and actionable manner, both verbally and in written form. Your reports should be precise, insightful, and aligned with business objectives.
Nice to Have
- Experience supporting AI/ML model training infrastructure (e.g., GPU orchestration, model serving) for both Diffusion- and LLM pipelines.
- Familiarity with data lake architectures and tools like Delta Lake, LakeFS, or Databricks.
- Knowledge of security and compliance best practices (e.g., SOC2, ISO 27001).
- Exposure to MLOps platforms or frameworks (e.g., MLflow, Kubeflow, Vertex AI).
What We Offer
- Competitive salary + equity
- Flexible work environment and remote-friendly culture
- Opportunities to work on cutting-edge AI/ML technology
- Fast-paced environment with high impact and visibility
- Professional growth support and resources
Locations
DevOps Specialist employer: OhChat
Contact Detail:
OhChat Recruiting Team
StudySmarter Expert Advice 🤫
We think this is how you could land DevOps Specialist
✨Tip Number 1
Familiarise yourself with the specific tools and technologies mentioned in the job description, such as AWS, GCP, Docker, and Kubernetes. Having hands-on experience or projects that showcase your skills with these platforms can set you apart from other candidates.
✨Tip Number 2
Network with professionals in the AI and DevOps fields. Attend relevant meetups, webinars, or online forums where you can connect with current employees or industry experts. This can provide valuable insights into the company culture and potentially lead to referrals.
✨Tip Number 3
Prepare to discuss your past experiences in detail, especially those related to building and maintaining infrastructure and data pipelines. Be ready to share specific examples of challenges you've faced and how you overcame them, as this demonstrates your problem-solving abilities.
✨Tip Number 4
Showcase your communication skills by being clear and concise when discussing technical topics. Practice explaining complex concepts in simple terms, as effective communication is crucial for collaborating with cross-functional teams in this role.
We think you need these skills to ace DevOps Specialist
Some tips for your application 🫡
Tailor Your CV: Make sure your CV highlights relevant experience in DevOps, data engineering, and cloud platforms like AWS and GCP. Use specific examples that demonstrate your expertise with tools such as Docker, Kubernetes, and CI/CD pipelines.
Craft a Compelling Cover Letter: In your cover letter, express your enthusiasm for the role and the company’s mission. Mention how your skills align with their needs, particularly in building scalable infrastructure and data pipelines. Be sure to include any experience you have with AI/ML technologies.
Showcase Your Technical Skills: When detailing your technical skills, focus on your proficiency in Python and any relevant programming languages. Highlight your experience with infrastructure-as-code tools and your ability to write clean, efficient code. Consider including links to any projects or repositories that showcase your work.
Prepare for Technical Questions: Anticipate technical questions related to DevOps practices, data engineering workflows, and cloud services. Be ready to discuss your problem-solving approach and provide examples of how you've optimised CI/CD workflows or managed infrastructure issues in previous roles.
How to prepare for a job interview at OhChat
✨Showcase Your Technical Skills
Be prepared to discuss your experience with AWS and GCP in detail. Highlight specific projects where you've implemented infrastructure-as-code tools like Terraform or CloudFormation, and be ready to explain how you optimised CI/CD workflows.
✨Demonstrate Problem-Solving Abilities
During the interview, expect to face technical challenges or scenarios. Use these opportunities to showcase your analytical skills and how you approach problem-solving, especially in high-pressure situations related to DevOps and data engineering.
✨Communicate Clearly and Concisely
Effective communication is key. Practice explaining complex technical concepts in simple terms, as you'll need to convey insights to both technical and non-technical stakeholders. Prepare examples of how you've done this in past roles.
✨Align with Company Values
Research the company's mission and values, particularly their focus on AI and innovation. Be ready to discuss how your personal values align with theirs and how you can contribute to their goal of becoming a leader in the spicy content industry.