This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
The Data Engineer will support the design, development, and optimization of modern data pipelines and cloud-based data platforms. This role involves building scalable ETL/ELT processes, integrating structured and unstructured data sources, and enabling secure, reliable data delivery for analytics, machine learning, and operational systems. This position plays a key role in federal data modernization efforts, which require high-performance data access, governance, and real-time processing capabilities.
Job Responsibility:
Design and implement robust ETL/ELT data pipelines using tools such as Apache Airflow, Spark, dbt, Kafka, or Databricks
Develop and optimize workflows that ingest, transform, and publish data from diverse sources into data lakes and data warehouses
Support batch and real-time data integration using scalable streaming frameworks and distributed processing engines
Manage infrastructure-as-code deployments and collaborate with DevOps teams to ensure data pipeline reliability, scalability, and security
Work closely with data scientists, analysts, and product teams to deliver data that supports BI dashboards, ML pipelines, and reporting
Apply best practices in data governance, schema versioning, and metadata management
Ensure systems are compliant with applicable data privacy regulations and security frameworks, including FISMA and Section 508
Document pipeline workflows, technical specifications, and data lineage to support auditability and transparency
Requirements:
Bachelor’s degree in Computer Science, Engineering, or related field
Minimum 3 years of experience in data engineering or backend data systems
Experience building and maintaining data pipelines using tools such as Apache Airflow, Spark, Kafka, Databricks, dbt, or Flink
Strong proficiency in SQL and one or more programming languages such as Python or Scala
Experience working with cloud-based data infrastructure (e.g., AWS, Azure, or GCP), including data lakes and data warehouses such as Delta Lake, Snowflake, BigQuery, or Redshift
Familiarity with containerization (Docker, Kubernetes), CI/CD workflows, and data version control systems such as DVC
Understanding of modern data architecture patterns including Lakehouse, event streaming, and batch processing models
Proven ability to lead data pipeline projects, mentor junior engineers, and collaborate across engineering and analytics teams
Nice to have:
Experience with real-time data processing and streaming analytics use cases
Knowledge of data cataloging, lineage tracking, and metadata management tools
Familiarity with data privacy and compliance frameworks (e.g., GDPR, HIPAA, IRS Publication 1075)
Experience collaborating with analytics, business intelligence, or ML engineering teams in production environments
Exposure to observability tools and practices to monitor pipeline health and data quality
Prior experience in a federal data engineering environment or regulated enterprise setting
What we offer:
Generous medical, dental, and vision plans
Opportunity to work in different sectors
Flexibility to balance quality work and personal lives