This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
Join Entalpic's Platform team and contribute to the core software foundations of our data stack while learning how data and software engineering practices enable advanced AI and scientific workflows. Your mission will be to help design, implement, and maintain a robust, scalable, and maintainable data infrastructure and internal data platform.
Job Responsibility:
Data Engineering Implementation: Help build and maintain data pipelines that process and integrate multimodal data including simulation results, scientific text, and experimental measurements
Data Storage and Organization: Contribute to the implementation and organization of data storage solutions, ensuring data is well-structured, accessible, and documented
Automation and Tooling: Write scripts to automate data ingestion, cleaning, and transformation tasks under the guidance of senior engineers
Data Visualization and Analytics: Support the creation of basic data visualizations and analytical dashboards to help teams explore datasets, monitor data quality, and gain insights from experimental, simulation, and textual data
Applied ML Research: Prototype end-to-end pipelines using state-of-the-art LLMs for chemical entity extraction and explore ranking methods to improve data quality and retrieval for downstream AI systems
Data Quality and Documentation: Assist with data validation, basic lineage tracking, and documentation to improve data reliability and reproducibility
Infrastructure Collaboration: Work with data engineers, data scientists, and DevOps to understand how data systems are deployed and scaled on cloud platforms such as AWS or GCP
Cross-team Collaboration: Interact with scientists and engineers to understand data needs and help translate them into practical data solutions
Requirements:
Currently pursuing or recently completed a Bachelor’s or Master’s degree in Computer Science, Data Engineering, Software Engineering, or a related field
Strong interest in data engineering, data systems, and infrastructure for machine learning or scientific applications
Proficiency in Python and familiarity with at least one other programming language is a plus
Foundational knowledge of SQL and relational databases, with curiosity about NoSQL systems
Understanding of core data concepts such as ETL, data modeling, and data pipelines
Must be comfortable working in English in a technical environment
Motivated, curious, and eager to learn in a fast-paced startup setting
Nice to have:
Exposure to cloud platforms such as AWS or GCP through coursework, projects, or internships
Interest in machine learning, AI systems, or scientific computing
Familiarity with Git and collaborative software development workflows
Familiarity with containerization tools such as Docker
Interest in materials science, chemistry, or physics is a strong plus but not required
What we offer:
Competitive salary + full reimbursement of your transport card
Flexible work environment: Remote-Friendly with a minimum of 3 days in Paris offices per week (Station F)
1 paid day off per month
Professional development opportunities: access to conferences and internal learning sessions