This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
As a Research Engineer in Data Infrastructure, you will design and implement a “data engine” that uploads the data collected by the robot fleet, makes this data easy to query and train on. Your work ensures high‑quality data pipelines are built and maintained, enabling rapid model development, large‑scale annotation, and smooth integration between on‑robot, on‑premise, and cloud systems.
Job Responsibility:
Optimize operational efficiency of data collection on the NEO fleet
Design triggers on the robot to determine if and when data should be uploaded
Automate ETL pipelines so fleet‑wide data is easily queryable and available for training
Work with external dataset providers to prepare diverse multi-modal pre-training datasets
Build frontend tools for visualizing and automating labeling of very large datasets
Develop machine learning models to automatically label and organize datasets
Requirements:
Strong experience in building data pipelines and ETL systems
Ability to design and implement systems that collect, upload, and manage data from robotic fleets
Familiarity with architectures combining on‑robot components, on‑premises clusters, and cloud systems
Experience with data labeling tools or building tooling for dataset visualization and annotation
Skills in creating or applying machine learning models for dataset organization / automated labeling