This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
Are you a visionary architect and administrator with a passion for transforming data ecosystems? We are seeking a senior individual contributor to become the technical authority for our client's enterprise data platform. In this pivotal role, you will not just maintain; you will design, implement, and continuously optimize critical data infrastructure, directly impacting how insights are generated and utilized across the organization. This is an exciting opportunity to drive innovation, enhance efficiency, and ensure the scalability and reliability of a multi-platform data environment, working closely with data engineering, analytics, and infrastructure teams to build robust, high-performing solutions.
Job Responsibility:
Platform Architecture & Governance: Architect and govern the enterprise platform environment, including workspace topology, data catalog structure, and access control frameworks
Data Pipeline Design: Design scalable, performant data pipeline patterns utilizing modern data processing tools and structured streaming
Infrastructure & Operations: Administer and troubleshoot underlying compute environments supporting the platform, including init scripts and cluster lifecycle management
Cost Management & Optimization: Own consumption tracking and reporting, proactively identifying optimization opportunities across jobs, interactive clusters, and data warehouses
Security & Compliance: Design and implement data governance frameworks within the data catalog, including lineage, tagging, and access auditing
Automation & AI Integration: Design and implement end-to-end automation frameworks for platform operations, including cluster lifecycle management, job scheduling, alerting, and self-healing workflows
Strategic Evaluation & Support: Serve as the primary technical escalation point for platform issues, contribute to sprint planning, and manage change requests and incidents
Requirements:
7+ years of experience in data engineering or data platform roles
minimum of 4 years of hands-on implementation experience with enterprise data platforms
Demonstrated expertise with platform capabilities including data catalogs, data lake technologies, workflow orchestration, and SQL warehouses
Strong Unix/Linux proficiency, including shell scripting, process management, file system operations, cron scheduling, and environment configuration
Proficiency in Python and PySpark for distributed data processing, pipeline development, and platform automation
Experience with leading cloud infrastructure providers (AWS, Azure, or GCP), including compute, storage, networking, and identity/access management
Demonstrated ability to design for scale, cost efficiency, and operational reliability in an enterprise data environment
Demonstrated experience designing automation frameworks for data platform operations, including job orchestration, monitoring, alerting, and pipeline self-healing
Familiarity with AI/ML concepts and tooling within enterprise data ecosystems, including experiment tracking, automated machine learning, and model serving
Experience with relational database environments, including SQL development, schema design, and integration patterns for data extraction and pipeline sourcing
Proficiency in Git-based version control, including branching strategies, pull request workflows, repository management, and CI/CD pipeline integration for data platform code
Experience working within ITSM and project delivery frameworks such as ServiceNow and Jira
Strong written and verbal communication skills, with the ability to convey complex architectural concepts to both technical and non-technical audiences
Nice to have:
Hands-on experience with MLflow experiment tracking, model registry, and deployment patterns within enterprise data platforms
Exposure to generative AI frameworks (e.g., LangChain, LlamaIndex) or experience building LLM-integrated data pipelines and retrieval-augmented generation (RAG) workflows
Experience with workflow automation tools such as Apache Airflow or comparable orchestration platforms at enterprise scale
Experience integrating enterprise data platforms with leading ETL/ELT platforms, including Fivetran or Ab Initio
hands-on Ab Initio development or administration experience is a strong plus
Familiarity with enterprise data governance frameworks and catalog tools (e.g., Collibra, Alation, or advanced data catalog features)
Experience supporting enterprise data platforms in regulated industries (e.g., financial services, insurance) with associated audit and compliance requirements
Working knowledge of Infrastructure-as-Code tooling (e.g., Terraform, Ansible) for platform provisioning and configuration management
Background in disaster recovery design and resiliency planning for cloud-hosted data platforms