This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
We are looking for a Data Pipeline Engineer to design, build, and operate scalable, reliable data pipelines for enterprise Data platforms. The candidate must have strong working knowledge, and this is a hands-on individual contributor role.
Job Responsibility:
Build and maintain data transformation pipelines using Dbt/Spark
Develop and optimize large-scale/CPU intensive data processing using Apache Spark/Dremio
Orchestrate workflows using Airflow and/or Dagster
Implement data quality checks, testing, and monitoring for pipelines
Support schema evolution, backfills, and incremental processing
Ensure pipelines meet SLAs for freshness, reliability, and performance
Expertise/working knowledge in Dremio (semantic layer, virtual datasets, Reflections)
Requirements:
Strong hands-on experience with dbt
Strong hands-on experience with Apache Spark
Experience with Dremio/Trino or similar lakehouse query engines
Experience with Airflow and/or Dagster
Understanding of data catalogs and lineage (e.g., OpenLineage, DataHub, Apache Polaris, openlineage)