We are seeking a Data Engineer to join our Data Solutions – Data Engineering team. As a Data Engineer, you will design, develop, and maintain automated data pipelines to ensure reliable and scalable data flows. You will work with modern tools such as AWS, Snowflake, DBT, Airflow, and Python, while following best practices in CI/CD and version control.
Key Highlights
Key Responsibilities
Technical Skills Required
Benefits & Perks
Nice to Have
Job Description
This position is open only to candidates who are legally residing in Poland and can work with us through a registered business entity in Poland (e.g., sole proprietorship/JDG or limited liability company/sp. z o.o.).
As part of onboarding, we kindly ask new joiners to visit our Warsaw office on the first day for a short introduction, identity verification, and equipment pick-up.
Please note that we are unable to consider applications from candidates residing outside Poland.
Location: Warsaw, Poland
Employment type: B2B contract
Work model: 100% remote
Business travel: Occasional, up to once per quarter (e.g., onboarding sessions or workshops)
Seniority level: Mid-level
About Us
Sanoma Learning is the leading European learning company, serving over 20 million students in 11 countries. We offer printed and digital learning materials as well as digital learning and teaching platforms for primary, secondary, and vocational education. The development of our methodologies is based on deep teacher and student insight and really understanding their needs. By combining our educational technologies and pedagogical expertise, we create learning products and services with the highest learning impact. In our Technology organization, you will join the largest cross-cultural community of Sanoma Learning and contribute to the digital transformation and future of education in Europe.
Project Description
The Data Solutions – Data Engineering team at Sanoma builds and maintains the central data platform that powers analytics, reporting, and data-driven products across the company.
As part of this team, you will help design and develop automated data pipelines, create reliable data models, and ensure that high-quality datasets are available for analysts, data scientists, and business teams. You will work with modern tools such as AWS, Snowflake, DBT, Airflow, and Python, while following best practices in CI/CD and version control.
This is a great opportunity for a data engineer who wants to grow in a modern cloud environment, work with large-scale data, and contribute to impactful projects that support Sanoma Learning.
Role Responsibilities
- Design, develop, and maintain automated data pipelines to ensure reliable and scalable data flows
- Build and optimize data models to support analytics, reporting, and AI-driven use cases
- Write clean, efficient, and reusable code in Python and SQL
- Work with cloud-based data technologies such as AWS, Snowflake, and DBT to deliver high-quality solutions
- Use orchestration tools (e.g., Airflow) to schedule and monitor data workflows
- Apply best practices in CI/CD, version control (Git), and testing to maintain robust data processes
- Collaborate with analysts, data scientists, and business stakeholders to understand data needs and deliver solutions
- Support data integration projects, including ingesting data from enterprise systems like SAP or Salesforce
- Troubleshoot, debug, and resolve issues in data pipelines and workflows
Interested in remote work opportunities in Data Science? Discover Data Science Remote Jobs featuring exclusive positions from top companies that offer flexible work arrangements.
- Bachelor’s degree or equivalent training or work experience
- Demonstrable experience building automated data pipelines and creating proper data models
- Strong Python and SQL skills;
- Experience with Python, AWS, Snowflake, DBT, CI/CD, version control (Git), and data pipeline orchestration tools (e.g., Airflow)
- Strong problem-solving, debugging, and analytical skills
Browse our curated collection of remote jobs across all categories and industries, featuring positions from top companies worldwide.
- Experience working with data from enterprise systems (e.g., SAP, Salesforce) in a data warehouse environment
- Familiarity with designing and optimizing ETL/ELT pipelines for large-scale business applications
- Exposure to data integration tools, APIs, or connectors used with systems like SAP or Salesforce
- B2B contract for an indefinite period
- Work-life balance and a supportive, informal atmosphere
- Opportunities for professional development and skill growth
- Work on modern data platforms (e.g., Snowflake, Databricks, cloud environments)
- Build and maintain data pipelines supporting AI-driven solutions in education
- Gain hands-on experience with modern data stack (DBT, CI/CD, orchestration tools)
- Collaborate with Data Engineers, Data Scientists, and Product teams
- Learn best practices in data modeling, data quality, and scalable data processing
- At Sanoma, you’ll work in a result-oriented, flexible environment focused on collaboration
- Be part of an international team working on data solutions across European markets
- Contribute to projects that have real impact on digital education
Similar Jobs
Explore other opportunities that match your interests
Mid-Senior Data Engineer (Azure Databricks)
Miratech
peakdata