This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
We are seeking a Senior Data Engineer / Data Platform Architect to play a critical role in the delivery of our enterprise-wide Data Mesh and Hybrid Cloud architecture. This role combines hands-on engineering with architectural ownership across core platform components—including data pipelines, storage layers, metadata systems, and federated access patterns. The ideal candidate brings deep technical expertise, an architect’s mindset, and a proactive approach to building resilient, scalable, and cloud-native data platforms. You will work closely with the Data Mesh and Cloud Architect Lead to design, implement, and operationalize enterprise-grade data products and services across Controls Technology and the broader Functions Technology organization.
Job Responsibility:
Co-architect modular, reusable, and secure data platform components supporting Data Mesh principles
Design optimal data storage, processing, and retrieval strategies across hybrid cloud environments, with a focus on Apache Iceberg-based architectures
Integrate federated querying frameworks using Starburst and Stargate, ensuring high-performance access across data sources
Build and optimize complex batch and streaming pipelines for diverse data domains, with an emphasis on performance, fault-tolerance, and scalability
Enable schema evolution and transactional consistency in Iceberg tables, leveraging catalog services and metadata stores (e.g., Hive MetaStore)
Drive the adoption and implementation of Apache Iceberg, Iceberg Catalogs, Hive MetaStore, and Starburst to unify analytics across cloud and on-prem data assets
Operationalize data product templates, ingestion frameworks, and reusable patterns for use across multiple data domains
Architect and deploy data ingestion and processing components that span AWS cloud (S3, Lambda, Glue, Redshift, Athena) and on-prem environments
Implement data movement, governance, and monitoring frameworks that operate transparently across hybrid infrastructure
Embed data quality, lineage, and metadata capture into the platform by default, ensuring compliance and traceability
Enforce role-based access controls, encryption standards, and audit mechanisms aligned with corporate policies
Guide junior engineers and developers through technical reviews, architecture guidance, and solution design
Establish coding standards, testing patterns, CI/CD pipelines, and promote a culture of engineering excellence
Requirements:
Expert in Python, SQL, and scripting for automation and data engineering
Familiarity with Scala or Java is a plus
Strong experience with Apache Spark, Kafka, Flink, or other distributed data processing engines
Advanced knowledge of Apache Iceberg, including partitioning strategies, schema evolution, compaction, and ACID support
Hands-on with AWS cloud services—S3, Glue, Lambda, Redshift, Athena, and EMR
Experience architecting solutions across hybrid environments, integrating cloud-native and legacy systems
Proficient with Hive MetaStore, Iceberg Catalogs, and metadata management
Strong implementation experience with Starburst and Stargate for federated SQL access across disparate systems
Working knowledge of Terraform, CloudFormation, Docker, GitHub Actions, or similar CI/CD and IaC tools
Familiarity with data lineage, cataloging, RBAC, encryption, and compliance standards (GDPR, CCPA, etc.)
Architectural Thinking – Ability to break down complex data platform needs into modular, scalable components
Leadership & Mentorship – Demonstrated experience in guiding junior team members, setting technical direction, and reviewing solutions
Business Orientation – Understands data’s role in business processes and can balance technical rigor with practical implementation
Collaboration – Comfortable working across multiple teams, geographies, and functions, including architects, engineers, product owners, and business users
Problem Solving – Strong analytical skills and a solution-oriented mindset for complex data challenges
13+ years of experience in data engineering, big data platforms, and cloud-based infrastructure
Bachelor’s or Master’s degree in Computer Science, Data Engineering, or related technical discipline
Proven track record in building production-grade data platforms using Iceberg, Spark, and AWS services
Deep understanding of federated architectures, hybrid integration patterns, and metadata-driven pipelines
Welcome to CrawlJobs.com – Your Global Job Discovery Platform
At CrawlJobs.com, we simplify finding your next career opportunity by bringing job listings directly to you from all corners of the web. Using cutting-edge AI and web-crawling technologies, we gather and curate job offers from various sources across the globe, ensuring you have access to the most up-to-date job listings in one place.
We use cookies to enhance your experience, analyze traffic, and serve personalized content. By clicking “Accept”, you agree to the use of cookies.