Data Engineer (GCP)
Hello, let’s meet!
Who We Are
While Xebia is a global tech company, our journey in CEE started with two Polish companies – PGS Software, known for world-class cloud and software solutions, and GetInData, a pioneer in Big Data. Today, we’re a team of 1,000+ experts delivering top-notch work across cloud, data, and software. And we’re just getting started.
What We Do
We work on projects that matter – and that make a difference. From fintech and e-commerce to aviation, logistics, media, and fashion, we help our clients build scalable platforms, data and AI solutions, and cutting-edge applications to shape the future of tech. Our clients include McLaren, Aviva, Deloitte, Spotify, Disney, ING, UPS, Tesco, Truecaller, AllSaints, Volotea, Schmitz Cargobull, Allegro, InPost, and many, many more.
We value smart tech, real ownership, and continuous growth. We use modern, open-source stacks, and we’re proud to be trusted partners of Databricks, dbt, Snowflake, Azure, GCP, and AWS. Fun fact: we were the first AWS Premier Partner in Poland!
Beyond Projects
What makes Xebia special? Our community. We support tech communities, organize meetups (Software Talks, Data Tech Talks), and have a culture that actively support your growth via Guilds, Labs, and personal development budgets — for both tech and soft skills. It’s not just a job. It’s a place to grow.
What sets us apart?
Our mindset. Our vibe. Our people. And while that’s hard to capture in text – come visit us and see for yourself.
You will be:
- designing, delivering, and maintaining large scale, production grade data platforms,
- designing and implementing cloud native architectures, preferably on Google Cloud Platform,
- designing and developing scalable ETL solutions handling diverse data formats,
- building and managing complex data workflows using Apache Airflow,
- developing and maintaining high volume batch and streaming data processing pipelines using Apache Spark,
- work with large scale data processing workloads, including billions of events per day,
- implementing Infrastructure as Code using Terraform,
- building and maintaining automated CI CD pipelines,
- collaborating on production grade data and ML platforms.
Your profile:
- at least 5 years of professional experience in Data Engineering,
- proven experience delivering and maintaining production data platforms,
- strong experience with Google Cloud Platform and ability to design cloud native architectures,
- expert level knowledge of Apache Airflow for workflow orchestration,
- strong programming skills in Python and SQL,
- hands on experience with Apache Spark for batch and streaming processing,
- experience working with large scale data processing systems,
- experience handling multiple data formats
- practical experience with Infrastructure as Code using Terraform,
- eperience building automated release and deployment pipelines,
- hands-on experience with GitHub Actions, Docker, and Kubernetes,
- openness to on-site workshops max 1-3 per quarter in Warszawa and Poznań.
Work from the European Union region and a work permit are required.
Nice to have:
- experience in productionizing and deploying Machine Learning models at scale,
- experience with Vertex AI,
- experience with scaling high performance data processing systems,
- experience with building scalable REST APIs for Machine Learning inference,
- strong focus on software development best practices,
- experience with Test Driven Development,
- ability to write comprehensive unit and integration tests for Spark and Airflow,
- professional GCP Data Engineer certification.
Recruitment Process:
CV review – HR call – Interview – Client Interview I - Client interview II – Decision
Apply for this job
*
indicates a required field
.png?1773750017)