At a Glance
- Tasks: Design and maintain scalable data pipelines using Azure Synapse Analytics and PySpark.
- Company: Sembcorp is a leading energy solutions provider focused on sustainable urban development.
- Benefits: Enjoy a collaborative culture, opportunities for growth, and the chance to impact Asia's energy transition.
- Why this job: Join a mission-driven team dedicated to driving energy transition and making a real-world impact.
- Qualifications: Bachelor’s degree in Computer Science or related field; 3-5 years of Azure data engineering experience required.
- Other info: Work in a supportive environment that values collaboration, accountability, and innovation.
The predicted salary is between 36000 - 60000 £ per year.
About Sembcorp
Sembcorp is a leading energy and urban solutions provider headquartered in Singapore. Led by its purpose to drive energy transition, Sembcorp delivers sustainable energy solutions and urban developments by leveraging its sector expertise and global track record.
Purpose and Scope
We are seeking a highly skilled and self-driven Azure Data Engineer with expertise in PySpark, Python, and modern Azure data services including Synapse Analytics and Azure Data Explorer. The ideal candidate will design, develop, and maintain scalable data pipelines and architectures, enabling effective data management, analytics, and governance.
Key Roles and Responsibilities
- Design, develop, and maintain scalable and efficient data pipelines (both batch and real-time streaming) using modern data engineering tools.
- Build and manage data lakes, data warehouses, and data marts using Azure Data Services.
- Integrate data from various sources including APIs, structured/unstructured files, IoT devices, and real-time streams.
- Develop and optimize ETL/ELT workflows using tools such as Azure Data Factory, Databricks, and Apache Spark.
- Implement real-time data ingestion and processing using Azure Stream Analytics, Event Hubs, or Kafka.
- Ensure data quality, availability, and security across the entire data lifecycle.
- Collaborate with analysts, data scientists, and engineering teams to deliver business-aligned data solutions.
- Contribute to data governance efforts and ensure compliance with data privacy standards.
- Establish and manage source system connectivity (on-prem, APIs, sensors, etc.).
- Handle deployment and migration of data pipeline artifacts between environments using Azure DevOps.
- Design, develop, and troubleshoot PySpark scripts and orchestration pipelines.
- Perform data integration using database joins and other transformations aligned with project requirements.
- Any assigned ad-hoc duties.
Requirements:
- Bachelor’s Degree in Computer Science, Engineering, or related field.
- 3–5 years of experience in Azure-based data engineering, PySpark, and Big Data technologies.
- Strong hands-on experience with Azure Synapse Analytics for pipeline orchestration and data handling.
- Expertise in SQL, data warehousing, data marts, and ingestion using PySpark and Python.
- Solid experience building and maintaining cloud-based ETL/ELT pipelines, especially with Azure Data Factory or Synapse.
- Familiarity with cloud data environments such as Azure and optionally AWS.
- Experience with Azure DevOps for CI/CD and artifact deployment.
- Excellent communication, problem-solving, and interpersonal skills.
Good to Have:
- 1–2 years of experience working with Azure Data Explorer (including row-level security and access controls).
- Experience with Azure Purview for metadata management, data lineage, governance, and discovery.
- Ability to work independently and take full ownership of assignments.
- Proactive in identifying and resolving blockers and escalating when needed.
- Exposure to real-time processing with tools like Azure Stream Analytics or Kafka.
Our Culture at Sembcorp
At Sembcorp, our culture is shaped by a strong set of shared behaviours that guide the way we work and uphold our commitment to driving the energy transition. We foster an institution-first mindset, where the success of Sembcorp takes precedence over individual interests. Collaboration is at the heart of what we do, as we work seamlessly across markets, businesses, and functions to achieve our goals together. Accountability is a core principle, ensuring that we take ownership of our commitments and deliver on them with integrity and excellence. These values define who we are and create a workplace where our people can thrive while making a meaningful impact on driving energy transition. Join us in making a real impact!
Data Engineer (Azure Synapse Analytics, PySpark) employer: SEMBCORP UTILITIES PTE LTD
Contact Detail:
SEMBCORP UTILITIES PTE LTD Recruiting Team
StudySmarter Expert Advice 🤫
We think this is how you could land Data Engineer (Azure Synapse Analytics, PySpark)
✨Tip Number 1
Familiarise yourself with Azure Synapse Analytics and PySpark by working on personal projects or contributing to open-source initiatives. This hands-on experience will not only enhance your skills but also give you practical examples to discuss during interviews.
✨Tip Number 2
Network with professionals in the data engineering field, especially those who work with Azure technologies. Attend meetups, webinars, or online forums to connect with others and gain insights into the industry, which can help you stand out as a candidate.
✨Tip Number 3
Stay updated on the latest trends and advancements in data engineering and cloud technologies. Follow relevant blogs, podcasts, and social media channels to ensure you can speak knowledgeably about current practices and tools during your interview.
✨Tip Number 4
Prepare for technical interviews by practising coding challenges related to data pipelines and ETL processes. Use platforms like LeetCode or HackerRank to sharpen your problem-solving skills, particularly in Python and SQL, which are crucial for this role.
We think you need these skills to ace Data Engineer (Azure Synapse Analytics, PySpark)
Some tips for your application 🫡
Understand the Role: Before applying, make sure to thoroughly understand the responsibilities and requirements of the Data Engineer position at Sembcorp. Familiarise yourself with Azure Synapse Analytics, PySpark, and other relevant technologies mentioned in the job description.
Tailor Your CV: Customise your CV to highlight your experience with Azure-based data engineering, particularly focusing on your skills in PySpark, SQL, and data pipeline development. Use specific examples from your past work that align with the key roles and responsibilities outlined in the job description.
Craft a Compelling Cover Letter: Write a cover letter that not only showcases your technical skills but also reflects your understanding of Sembcorp's mission to drive energy transition. Mention how your background and values align with their culture of collaboration and accountability.
Proofread and Edit: Before submitting your application, carefully proofread your CV and cover letter for any grammatical errors or typos. A polished application demonstrates attention to detail and professionalism, which are crucial in a technical role like this.
How to prepare for a job interview at SEMBCORP UTILITIES PTE LTD
✨Showcase Your Technical Skills
Be prepared to discuss your experience with Azure Synapse Analytics, PySpark, and other relevant technologies. Bring examples of projects where you've designed and maintained data pipelines, and be ready to explain the challenges you faced and how you overcame them.
✨Understand Sembcorp's Mission
Familiarise yourself with Sembcorp's commitment to energy transition and sustainable solutions. Be ready to discuss how your skills as a Data Engineer can contribute to their goals and align with their values of collaboration and accountability.
✨Prepare for Problem-Solving Questions
Expect technical questions that assess your problem-solving abilities. Practice explaining your thought process when tackling data integration issues or optimising ETL workflows, as this will demonstrate your analytical skills and approach to real-world challenges.
✨Emphasise Team Collaboration
Sembcorp values teamwork, so be sure to highlight your experience working with cross-functional teams. Share examples of how you've collaborated with analysts, data scientists, and engineers to deliver successful data solutions, showcasing your interpersonal skills.