At a Glance
- Tasks: Join a team to build cutting-edge data platforms for AI and analytics.
- Company: Goldman Sachs, a leader in financial services with a focus on innovation.
- Benefits: Competitive salary, diverse work environment, and opportunities for growth.
- Other info: Collaborative culture with a focus on continuous learning and development.
- Why this job: Make an impact by developing reliable data solutions in a fast-paced setting.
- Qualifications: Degree in a relevant field and strong programming skills in Python or Java.
The predicted salary is between 55000 - 65000 ÂŁ per year.
Join a team building the data foundations that support the firm’s AI and analytics capabilities. This role sits within the engineering effort to develop a modern Lakehouse and AI data platform that enables reliable, well‑governed and high‑performing data use across the firm.
At Goldman Sachs, engineering teams are positioned at the centre of the business, building scalable systems, solving complex technical problems and turning data into action. In data engineering roles, the emphasis is on designing, building and maintaining large‑scale data platforms, delivering production pipelines, improving reliability and quality, and partnering closely with users of the platform. This is a delivery‑focused role for engineers who want to build robust data assets in production, work with modern data technologies, and grow over time within the firm. You will contribute to the data models, pipelines and platform capabilities that underpin analytics, operational decision‑making and emerging AI use cases, and may also help extend platform tooling where additional functionality is needed.
As a Data Engineer in the Lakehouse and AI Data Platform team, you will design, build, test and support data pipelines and curated datasets on the firm’s modern data platform. You will work across ingestion, transformation, modelling, optimisation and data quality, helping to deliver data products that are reliable, scalable and fit for purpose. Where there are gaps in platform functionality, you may also contribute to shared tooling or framework components that improve how the platform is used and operated.
Key Responsibilities- Pipeline Engineering: Build, enhance and support batch and streaming data pipelines on the Lakehouse and AI data platform. Refactor or modernise existing data flows where needed to improve reliability, performance and maintainability. Build reusable tooling to improve delivery, consistency and operational support. Ensure data pipelines are production‑ready, well‑tested and operationally supportable.
- Data Modelling and Curation: Develop raw, refined and curated datasets that support analytics, reporting and AI use cases. Apply sound data modelling principles to represent business entities, relationships and historical change accurately. Work with consumers to shape data products that are usable, well documented and aligned to business needs.
- Data Quality and Reconciliation: Implement controls to validate completeness, accuracy and consistency of data across pipelines and datasets. Use reconciliation approaches to build confidence in production outputs and investigate breaks where they arise. Contribute to clear standards for testing, monitoring and issue resolution. Contribute to practical improvements in testing, monitoring or reconciliation tooling where this strengthens platform reliability and day‑to‑day delivery.
- Delivery and Partnership: Work closely with engineers, platform teams and data consumers to deliver agreed outcomes to time and quality expectations. Communicate clearly on progress, risks, dependencies and design choices, including where delivery would benefit from improvements to shared platform tooling. For more senior candidates, take a broader role in technical leadership, task breakdown and support for junior engineers.
- Bachelor’s or master’s degree in a relevant discipline, or equivalent practical experience, with evidence of strong quantitative skills or data engineering expertise.
- Strong hands‑on programming experience in Python or Java.
- Good working knowledge of SQL, including troubleshooting, optimisation and data analysis.
- Ability to learn new tools, internal platforms and delivery workflows quickly.
- Familiarity with software engineering fundamentals, including version control, testing, release discipline and CI/CD practices.
- Understanding of temporal data modelling, including handling of historical state and change over time.
- Knowledge of schema design, schema evolution and data compatibility considerations.
- Understanding of partitioning, clustering and other techniques used to improve data performance at scale.
- Ability to make sensible design choices across normalised and denormalised models, and between natural and surrogate keys.
- Practical approach to data quality, reconciliation and root‑cause analysis.
- Experience building or supporting production data pipelines in a collaborative engineering environment.
- Experience working with distributed data processing frameworks such as Apache Spark.
- Working knowledge of common data formats such as JSON, Avro and Parquet.
- Stronger ownership of technical design across multiple datasets or pipeline domains.
- Experience guiding implementation standards, code quality and engineering practices within a team.
- Ability to lead delivery for a workstream, manage dependencies and support less experienced engineers.
The role will involve working with a modern and evolving data stack. Candidates are not expected to have deep expertise in every tool from day one but should bring relevant experience and the ability to work across comparable technologies.
Examples of Technologies In Scope- Data processing and logic: ANSI SQL, Apache Spark, Kafka
- Data formats: JSON, Avro, Parquet
- Platforms and storage: Snowflake, Apache Iceberg, Databricks, Hadoop ecosystem technologies, Sybase IQ
- Engineering and deployment: CI/CD tooling, containerised or Kubernetes‑based deployment approaches where relevant
We are looking for engineers who can deliver well‑structured, reliable solutions in production and who take ownership of the quality of what they build. The role suits candidates who are technically strong, pragmatic and comfortable working in a fast‑paced environment where data platforms support important business outcomes. It will also suit candidates who are willing to contribute to shared tooling or platform components that make the wider engineering environment more effective.
Stronger Candidates Will Typically Demonstrate- Sound judgement in technical trade‑offs
- Attention to detail in data correctness and testing
- A clear and structured approach to problem solving
- Willingness to work closely with stakeholders and partner teams
- An ability to identify when delivery problems would be better solved through reusable tooling or platform improvements
- An interest in developing long‑term expertise within the firm
Data Engineering - Data, Lakehouse and AI Data Platform Engineer - Analyst/Associate - London employer: Goldman Sachs
Contact Detail:
Goldman Sachs Recruiting Team
StudySmarter Expert Advice 🤫
We think this is how you could land Data Engineering - Data, Lakehouse and AI Data Platform Engineer - Analyst/Associate - London
✨Tip Number 1
Network like a pro! Reach out to current employees at Goldman Sachs or in the data engineering field. A friendly chat can give you insider info and maybe even a referral, which can really boost your chances.
✨Tip Number 2
Show off your skills! Create a portfolio showcasing your data projects, especially those involving Python, SQL, or Apache Spark. This gives you something tangible to discuss during interviews and shows you're serious about your craft.
✨Tip Number 3
Prepare for technical interviews by brushing up on your coding skills and understanding data modelling principles. Practice common data engineering problems and be ready to explain your thought process clearly.
✨Tip Number 4
Don’t forget to apply through our website! It’s the best way to ensure your application gets seen. Plus, it shows you’re genuinely interested in joining the team at Goldman Sachs.
We think you need these skills to ace Data Engineering - Data, Lakehouse and AI Data Platform Engineer - Analyst/Associate - London
Some tips for your application 🫡
Tailor Your CV: Make sure your CV is tailored to the Data Engineering role. Highlight your experience with data pipelines, programming skills in Python or Java, and any relevant projects that showcase your ability to work with modern data technologies.
Craft a Compelling Cover Letter: Your cover letter is your chance to shine! Use it to explain why you're passionate about data engineering and how your skills align with the responsibilities outlined in the job description. Be sure to mention your interest in building robust data assets.
Showcase Your Technical Skills: Don’t forget to highlight your technical skills! Mention your experience with SQL, Apache Spark, and any other relevant tools. If you've worked on data quality or reconciliation, make sure to include that too—it’s a big plus!
Apply Through Our Website: We encourage you to apply through our website for the best chance of being noticed. It’s straightforward and ensures your application goes directly to the right team. Plus, we love seeing candidates who take the initiative!
How to prepare for a job interview at Goldman Sachs
✨Know Your Data Engineering Fundamentals
Before the interview, brush up on your understanding of data modelling, pipeline engineering, and data quality principles. Be ready to discuss how you’ve applied these concepts in past projects, especially with technologies like SQL, Apache Spark, and Kafka.
✨Showcase Your Problem-Solving Skills
Prepare to share specific examples of how you've tackled complex technical challenges. Use the STAR method (Situation, Task, Action, Result) to structure your responses, highlighting your analytical thinking and decision-making process.
✨Familiarise Yourself with the Tech Stack
Make sure you’re comfortable discussing the technologies mentioned in the job description, such as Snowflake, Databricks, and CI/CD practices. If you have experience with similar tools, be ready to draw parallels and explain how that knowledge can transfer.
✨Communicate Clearly and Collaboratively
During the interview, demonstrate your ability to communicate effectively with both technical and non-technical stakeholders. Practice explaining complex concepts in simple terms, and be prepared to discuss how you’ve partnered with teams to deliver successful outcomes.