Job Search and Career Advice Platform

Aktiviere Job-Benachrichtigungen per E-Mail!

Graduate Data Engineer CV-Library Graduate Data Engineer

Scienceabode

Marlow

Hybrid

EUR 45.000 - 55.000

Vollzeit

Heute
Sei unter den ersten Bewerbenden

Erstelle in nur wenigen Minuten einen maßgeschneiderten Lebenslauf

Überzeuge Recruiter und verdiene mehr Geld. Mehr erfahren

Zusammenfassung

A leading pharmaceutical company in Marlow is seeking a Graduate Data Engineer to build and maintain scalable data pipelines in Palantir Foundry. This role involves collaborating with Product, Engineering, and GTM teams to transform raw data into reliable datasets for advanced reporting and analytics. The ideal candidate will have a degree in a relevant field and up to 2 years of experience, along with solid skills in Python/PySpark and familiarity with analytics tools. This position offers a hybrid working environment with expectations of in-office attendance.

Qualifikationen

  • Degree in Computer Science, Engineering, Mathematics, or similar.
  • Up to 2 years of experience in building data pipelines.
  • Ability to write clear and reliable Python/PySpark code.

Aufgaben

  • Build and maintain data pipelines in Palantir Foundry.
  • Prepare and optimize data pipelines for machine learning.
  • Develop dashboards and reports in Foundry.

Kenntnisse

Python/PySpark coding
Familiarity with analytics tools (pandas, numpy, matplotlib)
Understanding of big data frameworks (Spark)
Cloud service knowledge (Palantir, AWS, Azure, Google Cloud)
Data model expertise

Ausbildung

Degree in Computer Science, Engineering, Mathematics or similar
Master's degree in relevant fields (AI/ML, Data Systems)
Jobbeschreibung

Role Title: Graduate Data Engineer

Contract: 12 months

Location: Marlow (hybrid)

SRG are working with a leading pharmaceutical company based in Marlow. Our client develops and manufactures an impressive portfolio of aesthetics brands and products. Our client is committed to driving innovation and providing high-quality products and services.

Role Overview

As a Graduate Data Engineer, you will build and maintain scalable data pipelines in Palantir Foundry for advanced reporting and analytics while collaborating with cross-functional teams as part of the BTS Data & Analytics team. You will work closely with key stakeholders in Engineering, Product, GTM, and other groups to help build scalable data solutions that support key metrics, reporting, and insights. You will assist in ensuring teams have access to reliable, accurate data as our company grows. You will have the opportunity to support projects that enable self‑serve insights, helping teams make data‑driven decisions, while learning from experienced team members and developing your technical and business skills.

Key Responsibilities
  • Build and maintain data pipelines, leveraging PySpark and/or TypeScript within Foundry, to transform raw data into reliable, usable datasets.
  • Assist in preparing and optimizing data pipelines to support machine learning and AI model development, ensuring datasets are clean, well‑structured, and readily usable by Data Science teams.
  • Support the integration and management of feature engineering processes and model outputs into Foundry's data ecosystem, helping enable scalable deployment and monitoring of AI/ML solutions.
  • Gather and translate stakeholder requirements for key data models and reporting, focusing on Palantir Foundry workflows and tools.
  • Develop and refine dashboards and reports in Foundry to visualize key metrics and insights.
  • Collaborate with Product, Engineering, and GTM teams to align data architecture and solutions, supporting scalable, self‑serve analytics across the organization.
  • Have some prompt engineering experience with large language models, including writing and evaluating complex multi‑step prompts.
  • Continuously develop your understanding of the company's data landscape, including Palantir Foundry's ontology‑driven approach and best practices for data management.
About you

You have a degree in Computer Science, Engineering, Mathematics, or similar, or have similar work experience.

Having up to 2 years of experience building data pipelines at work or through internships is helpful.

You can write clear and reliable Python/PySpark code.

You are familiar with popular analytics tools (like pandas, numpy, matplotlib), big data frameworks (like Spark), and cloud services (like Palantir, AWS, Azure, or Google Cloud).

You have a deep understanding of data models, relational and non‑relational databases, and how they are used to organize, store, and retrieve data efficiently for analytics and machine learning.

Knowing about software engineering methods, including DevOps, DataOps, or MLOps, is also a plus.

Ideal Candidate
  • Master's degree in engineering (e.g., AI/ML, Data Systems, Computer Science, Mathematics, Biotechnology, Physics), or minimum 2 years of relevant technology experience.
  • Experience with Generative AI (GenAI) and agentic systems will be considered a strong plus.
  • Have a proactive and adaptable mindset: willing to take initiative, learn new skills, and contribute to different aspects of a project as needed to drive solutions from start to finish.
  • Show a strong ability to thrive in situations of ambiguity, taking initiative to create clarity for yourself and the team, and proactively driving progress even when details are uncertain or evolving.
Other details

Hybrid working policy: Currently, our client expects all staff to be in their Marlow-based office at least 3 days a week from Jan 2026.

No visa sponsorship. ILR/Citizenship required.

Recruitment Business: Guidant, Carbon60, Lorien & SRG – The Impellam Group Portfolio are acting as an Employment Business in relation to this vacancy.

Hol dir deinen kostenlosen, vertraulichen Lebenslauf-Check.
eine PDF-, DOC-, DOCX-, ODT- oder PAGES-Datei bis zu 5 MB per Drag & Drop ablegen.