This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
Are you passionate about building robust data infrastructure and enabling innovation through engineering excellence? As our Data Engineer, your goal is to own and evolve the foundation of our data infrastructure. You will be central in ensuring data reliability, scalability, and accessibility across our lakehouse and transactional systems. This role is ideal for someone who thrives at the intersection of engineering and innovation, ensuring our data platforms are robust today while enabling the products of tomorrow.
Job Responsibility:
Design, build, and operate reliable ETL and ELT pipelines in Python and SQL
Manage ingestion into Bronze, standardization and quality in Silver, and curated serving in Gold layers of our Medallion architecture
Maintain ingestion from transactional MySQL systems into Vaniam Core to keep production data flows seamless
Implement observability, data quality checks, and lineage tracking to ensure trust in all downstream datasets
Develop schemas, tables, and views optimized for analytics, APIs, and product use cases
Apply and enforce best practices for security, privacy, compliance, and access control, ensuring data integrity across sensitive healthcare domains
Maintain clear and consistent documentation for datasets, pipelines, and operating procedures
Lead the integration of third-party datasets, client-provided sources, and new product-generated data into Vaniam Core
Partner with product and innovation teams to build repeatable processes for onboarding new data streams
Ensure harmonization, normalization, and governance across varied data types (scientific, engagement, operational)
Collaborate with the innovation team to prototype and productionize analytics, predictive features, and decision-support tools
Support dashboards, APIs, and services that activate insights for internal stakeholders and clients
Work closely with Data Science and AI colleagues to ensure engineered pipelines meet modeling and deployment requirements
Monitor job execution, storage, and cluster performance, ensuring cost efficiency and uptime
Troubleshoot and resolve data issues, proactively addressing bottlenecks
Conduct code reviews, enforce standards, and contribute to CI/CD practices for data pipelines
Requirements:
5+ years of professional experience in data engineering, ETL, or related roles
Strong proficiency in Python and SQL for data engineering
Hands-on experience building and maintaining pipelines in a lakehouse or modern data platform
Practical understanding of Medallion architectures and layered data design
Familiarity with modern data stack tools, including: Spark or PySpark
Workflow orchestration (Airflow, dbt, or similar)
Testing and observability frameworks
Containers (Docker) and Git-based version control
Excellent communication skills, problem-solving mindset, and a collaborative approach
Nice to have:
Experience with Databricks and the Microsoft Azure ecosystem
Expertise with Delta Lake formats, metadata management, and data catalogs
Familiarity with healthcare, scientific, or engagement data domains
Experience exposing analytics through APIs or lightweight microservices
What we offer:
100% remote environment with opportunities for local meet-ups
Positive, diverse, and supportive culture
Passionate about serving clients focused on Cancer and Blood diseases
Investment in you with opportunities for professional growth and personal development through Vaniam Group University
Health benefits – medical, dental, vision
Generous parental leave benefit
Focused on your financial future with a 401(k) Plan and company match
Work-Life Balance and Flexibility
Flexible Time Off policy for rest and relaxation
Volunteer Time Off for community involvement
Emphasis on Personal Wellness
Virtual workout classes
Discounts on tickets, events, hotels, child care, groceries, etc.
Welcome to CrawlJobs.com – Your Global Job Discovery Platform
At CrawlJobs.com, we simplify finding your next career opportunity by bringing job listings directly to you from all corners of the web. Using cutting-edge AI and web-crawling technologies, we gather and curate job offers from various sources across the globe, ensuring you have access to the most up-to-date job listings in one place.
We use cookies to enhance your experience, analyze traffic, and serve personalized content. By clicking “Accept”, you agree to the use of cookies.