Senior Data Engineer (GCP, Databricks, BigQuery, Airflow, dbt)

Leadtech Group • United Kingdom
Remote
Apply
AI Summary

Leadtech Group seeks a Senior Data Engineer to design and optimize data infrastructure on Databricks and GCP. Responsibilities include architecting scalable ETL/ELT pipelines using BigQuery, Airflow, dbt, and ensuring data quality with Great Expectations. Requires 3+ years of data engineering experience, with a focus on cloud-native solutions and GCP stack.

Key Highlights
Design, develop, and optimize data infrastructure on Databricks and GCP.
Architect scalable ETL/ELT pipelines using BigQuery, Airflow, dbt, and ensure data quality.
Implement CI/CD best practices and event-driven architectures for microservices.
Key Responsibilities
Define and implement the overall data architecture on GCP, including data warehousing in BigQuery/Databricks, data lake patterns in Google Cloud Storage, and Data Mart (Data Mach) solutions.
Integrate Terraform for Infrastructure as Code to provision and manage cloud resources efficiently.
Establish both batch and real-time data processing frameworks to ensure reliability, scalability, and cost efficiency.
Design, build, and optimize ETL/ELT pipelines using Apache Airflow for workflow orchestration.
Implement dbt (Data Build Tool) transformations to maintain version-controlled data models in BigQuery, ensuring consistency and reliability across the data pipeline.
Use Google Dataflow (based on Apache Beam) and Pub/Sub for large-scale streaming/batch data processing and ingestion.
Automate job scheduling and data transformations to deliver timely insights for analytics, machine learning, and reporting.
Implement event-driven or asynchronous data workflows between microservices.
Employ Docker and Kubernetes (K8s) for containerization and orchestration, enabling flexible and efficient microservices-based data workflows.
Implement CI/CD pipelines for streamlined development, testing, and deployment of data engineering components.
Enforce data quality standards using Great Expectations or similar frameworks, defining and validating expectations for critical datasets.
Define and uphold metadata management, data lineage, and auditing standards to ensure trustworthy datasets.
Implement security best practices, including encryption at rest and in transit, Identity and Access Management (IAM), and compliance with GDPR or CCPA where applicable.
Collaborate with Data Science, Analytics, and Product teams to ensure the data infrastructure supports advanced analytics, including machine learning initiatives.
Maintain Data Mart (Data Mach) environments that cater to specific business domains, optimizing access and performance for key stakeholders.
Technical Skills Required
Databricks BigQuery Google Cloud Storage Apache Airflow dbt Dataflow Pub/Sub Great Expectations Terraform Docker Kubernetes Python Java Scala RabbitMQ Prometheus Grafana Datadog New Relic Kafka Spark Streaming Vertex AI
Benefits & Perks
Flexible career path
Personalized internal training
Annual budget for external learning opportunities
Flexible schedule with flextime
Option of working full remote or from Barcelona office
Free Friday afternoons
35-hour workweek in July and August
Competitive salary
Full-time permanent contract
Top-tier private health insurance (including dental and psychological services)
25 days of vacation plus your birthday off
Flexible vacation options
Free coffee, fresh fruit, snacks
Game room
Rooftop terrace
Ticket restaurant
Nursery vouchers
Nice to Have
Familiarity with end-to-end ML workflows and model deployment on GCP (e.g., Vertex AI)
Experience with Prometheus, Grafana, Datadog, or New Relic for system health and performance monitoring
Advanced knowledge of compliance frameworks such as HIPAA, SOC 2, or relevant regulations
Additional proficiency in Kafka, Spark Streaming, or other streaming solutions
GCP-specific certifications (e.g., Google Professional Data Engineer)

Job Description


We are looking for a Senior Data Engineer to design, develop, and optimize our data infrastructure on Databricks. You will architect scalable pipelines using BigQuery, Google Cloud Storage, Apache Airflow, dbt, Dataflow, and Pub/Sub, ensuring high availability and performance across our ETL/ELT processes. You will leverage Great Expectations to enforce data quality standards. The role also involves building our Data Mart (Data Mach) environment and implementing CI/CD best practices.

A successful candidate has extensive knowledge of cloud-native data solutions, strong proficiency with ETL/ELT frameworks (including dbt), and a passion for building robust, cost-effective pipelines.

Key Responsibilities

Data Architecture & Strategy

  • Define and implement the overall data architecture on GCP, including data warehousing in BigQuery/Databricks, data lake patterns in Google Cloud Storage, and Data Mart (Data Mach) solutions
  • Integrate Terraform for Infrastructure as Code to provision and manage cloud resources efficiently
  • Establish both batch and real-time data processing frameworks to ensure reliability, scalability, and cost efficiency

Pipeline Development & Orchestration

  • Design, build, and optimize ETL/ELT pipelines using Apache Airflow for workflow orchestration
  • Implement dbt (Data Build Tool) transformations to maintain version-controlled data models in BigQuery, ensuring consistency and reliability across the data pipeline
  • Use Google Dataflow (based on Apache Beam) and Pub/Sub for large-scale streaming/batch data processing and ingestion
  • Automate job scheduling and data transformations to deliver timely insights for analytics, machine learning, and reporting

Event-Driven & Microservices Architecture

  • Implement event-driven or asynchronous data workflows between microservices
  • Employ Docker and Kubernetes (K8s) for containerization and orchestration, enabling flexible and efficient microservices-based data workflows
  • Implement CI/CD pipelines for streamlined development, testing, and deployment of data engineering components

Data Quality, Governance & Security

  • Enforce data quality standards using Great Expectations or similar frameworks, defining and validating expectations for critical datasets
  • Define and uphold metadata management, data lineage, and auditing standards to ensure trustworthy datasets
  • Implement security best practices, including encryption at rest and in transit, Identity and Access Management (IAM), and compliance with GDPR or CCPA where applicable

BI & Analytics Enablement

  • Collaborate with Data Science, Analytics, and Product teams to ensure the data infrastructure supports advanced analytics, including machine learning initiatives
  • Maintain Data Mart (Data Mach) environments that cater to specific business domains, optimizing access and performance for key stakeholders

Requirements

Experience

  • 3+ years of professional experience in data engineering, with at least 1 year in mobile data

Technical Expertise with GCP Stack

  • Proven track record building and maintaining BigQuery environments and Google Cloud Storage based data lakes
  • Deep knowledge of Apache Airflow for scheduling/orchestration and ETL/ELT design
  • Experience implementing dbt for data transformations, RabbitMQ for event-driven workflows, and Pub/Sub + Dataflow for streaming/batch data pipelines
  • Familiarity with designing and implementing Data Mart (Data Mach) solutions, as well as using Terraform for IaC

Programming & Containerization

  • Strong coding capabilities in Python, Java, or Scala, plus scripting for automation
  • Experience with Docker and Kubernetes (K8s) for containerizing data-related services
  • Hands-on with CI/CD pipelines and DevOps tools (e.g., Terraform, Ansible, Jenkins, GitLab CI) to manage infrastructure and deployments

Data Quality & Governance

  • Proficiency in Great Expectations (or similar) to define and enforce data quality standards
  • Expertise in designing systems for data lineage, metadata management, and compliance (GDPR, CCPA)
  • Strong understanding of OLTP (Online Transaction Processing) and OLAP (Online Analytical Processing) systems

Communication

  • Excellent communication skills for both technical and non-technical audiences
  • High level of organization, self-motivation, and problem-solving aptitude

Preferred Skills :

  • Machine Learning (ML) Integration: Familiarity with end-to-end ML workflows and model deployment on GCP (e.g., Vertex AI)
  • Advanced Observability: Experience with Prometheus, Grafana, Datadog, or New Relic for system health and performance monitoring
  • Security & Compliance: Advanced knowledge of compliance frameworks such as HIPAA, SOC 2, or relevant regulations
  • Real-Time Data Architectures: Additional proficiency in Kafka, Spark Streaming, or other streaming solutions
  • Certifications: GCP-specific certifications (e.g., Google Professional Data Engineer) are highly desirable

Will be a plus

  • Machine Learning (ML) Integration: Familiarity with end-to-end ML workflows and model deployment on GCP (e.g., Vertex AI)
  • Advanced Observability: Experience with Prometheus, Grafana, Datadog, or New Relic for system health and performance monitoring
  • Security & Compliance: Advanced knowledge of compliance frameworks such as HIPAA, SOC 2, or relevant regulations
  • Real-Time Data Architectures: Additional proficiency in Kafka, Spark Streaming, or other streaming solutions
  • Certifications: GCP-specific certifications (e.g., Google Professional Data Engineer) are highly desirable

Benefits

Why should you join us?

Growth and career development

  • At Leadtech, we prioritize your growth. Enjoy a flexible career path with personalized internal training and an annual budget for external learning opportunities

Work-Life balance

  • Benefit from a flexible schedule with flextime (7 - 9:30 a.m. start, 3:30 - 6 p.m. end) and the option of working full remote or from our Barcelona office. Enjoy free Friday afternoons with a 7-hour workday, plus a 35-hour workweek in July and August so you can savor summer!

Comprehensive benefits

  • Competitive salary, full-time permanent contract, and top-tier private health insurance (including dental and psychological services)
  • 25 days of vacation plus your birthday off, with flexible vacation options—no blackout days!

Unique Perks

  • If you wish to come, in our office in Barcelona you'll find it coplete with free coffee, fresh fruit, snacks, a game room, and a rooftop terrace with stunning Mediterranean views
  • Additional benefits include ticket restaurant and nursery vouchers, paid directly from your gross salary

Join us in an environment where you're free to innovate, learn, and grow alongside passionate professionals. At Leadtech, you'll tackle exciting challenges and be part of a vibrant team dedicated to delivering exceptional user experiences

Equal Employment Opportunity Employer:

Leadtech is an Equal Employment Opportunity (EEO) Employer, which means we encourage applications from people with different backgrounds, interests, and personal circumstances. Our team welcomes applicants regardless of their race, gender, age, religion, nationality, sexual orientation, and/or disabilities. All we need is your high energy, skills, and willingness to be a part of a great project!

Location

You'll have the flexibility to choose whether you'd like to come to the office every day, from time to time, or work fully remote. We want you to find the best combination for you.

If you prefer to be surrounded with amazing people, our exceptional office is in Barcelona's Blue Building, located right on the city's seafront. Besides our stunning views, you'll enjoy our office perks such as free fruit, snacks, and coffee and you'll also be able to take part in our Mario Kart and table tennis competitions.

The personal data you provide will be processed in order to manage your candidacy for the corporate selection processes that fit your profile. If you wish, you can exercise your rights of access, rectification or cancellation by writing to our address (Avenida Litoral, 12-14, 5ta planta. Barcelona. 08005) or to the email address protecciondedatos@LeadTech.com, attaching to your request a document that can validate your identity.

Similar Jobs

Explore other opportunities that match your interests

Visa Sponsorship Relocation Remote
Job Type Full-time
Experience Level Mid-Senior level

sire

United Kingdom

Senior Data Engineer

Data Science
•
2d ago
Visa Sponsorship Relocation Remote
Job Type Contract
Experience Level Mid-Senior level

Principle

United Kingdom
Visa Sponsorship Relocation Remote
Job Type Contract
Experience Level Mid-Senior level

Leap29

United Kingdom

Subscribe our newsletter

New Things Will Always Update Regularly