CrawlJobs Logo

Research Engineer, Data Infrastructure

1x.tech Logo

1X Technologies

Location Icon

Location:
United States , Palo Alto

Category Icon

Job Type Icon

Contract Type:
Employment contract

Salary Icon

Salary:

180000.00 - 250000.00 USD / Year

Job Description:

As a Research Engineer in Data Infrastructure, you will design and implement a “data engine” that uploads the data collected by the robot fleet, makes this data easy to query and train on. Your work ensures high‑quality data pipelines are built and maintained, enabling rapid model development, large‑scale annotation, and smooth integration between on‑robot, on‑premise, and cloud systems.

Job Responsibility:

  • Optimize operational efficiency of data collection on the NEO fleet
  • Design triggers on the robot to determine if and when data should be uploaded
  • Automate ETL pipelines so fleet‑wide data is easily queryable and available for training
  • Work with external dataset providers to prepare diverse multi-modal pre-training datasets
  • Build frontend tools for visualizing and automating labeling of very large datasets
  • Develop machine learning models to automatically label and organize datasets

Requirements:

  • Strong experience in building data pipelines and ETL systems
  • Ability to design and implement systems that collect, upload, and manage data from robotic fleets
  • Familiarity with architectures combining on‑robot components, on‑premises clusters, and cloud systems
  • Experience with data labeling tools or building tooling for dataset visualization and annotation
  • Skills in creating or applying machine learning models for dataset organization / automated labeling
What we offer:
  • Health, dental, and vision insurance
  • 401(k) with company match
  • Paid time off and holidays

Additional Information:

Job Posted:
December 01, 2025

Employment Type:
Fulltime
Work Type:
On-site work
Job Link Share:

Looking for more opportunities? Search for other job offers that match your skills and interests.

Briefcase Icon

Similar Jobs for Research Engineer, Data Infrastructure

AI Research Engineer, Data Infrastructure

As a Research Engineer in Infrastructure, you will design and implement a robust...
Location
Location
United States , Palo Alto
Salary
Salary:
180000.00 - 250000.00 USD / Year
1x.tech Logo
1X Technologies
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Strong experience in building data pipelines and ETL systems
  • Ability to design and implement systems for data collection and management from robotic fleets
  • Familiarity with architectures that span on-robot components, on-premise clusters, and cloud infrastructure
  • Experience with data labeling tools or building dataset visualization and annotation tooling
  • Proficiency in creating or applying machine learning models for dataset organization and automated labeling
Job Responsibility
Job Responsibility
  • Optimize operational efficiency of data collection across the NEO robot fleet
  • Design intelligent triggers to determine when and what data should be uploaded from the robots
  • Automate ETL pipelines to make fleet-wide data easily queryable and training-ready
  • Collaborate with external dataset providers to prepare diverse multi-modal pre-training datasets
  • Build frontend tools for visualizing and automating the labeling of large datasets
  • Develop machine learning models for automatic dataset labeling and organization
What we offer
What we offer
  • Equity
  • Health, dental, and vision insurance
  • 401(k) with company match
  • Paid time off and holidays
  • Fulltime
Read More
Arrow Right

Software Engineer, Data Infrastructure

The Data Infrastructure team at Figma builds and operates the foundational platf...
Location
Location
United States , San Francisco; New York
Salary
Salary:
149000.00 - 350000.00 USD / Year
figma.com Logo
Figma
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 5+ years of Software Engineering experience, specifically in backend or infrastructure engineering
  • Experience designing and building distributed data infrastructure at scale
  • Strong expertise in batch and streaming data processing technologies such as Spark, Flink, Kafka, or Airflow/Dagster
  • A proven track record of impact-driven problem-solving in a fast-paced environment
  • A strong sense of engineering excellence, with a focus on high-quality, reliable, and performant systems
  • Excellent technical communication skills, with experience working across both technical and non-technical counterparts
  • Experience mentoring and supporting engineers, fostering a culture of learning and technical excellence
Job Responsibility
Job Responsibility
  • Design and build large-scale distributed data systems that power analytics, AI/ML, and business intelligence
  • Develop batch and streaming solutions to ensure data is reliable, efficient, and scalable across the company
  • Manage data ingestion, movement, and processing through core platforms like Snowflake, our ML Datalake, and real-time streaming systems
  • Improve data reliability, consistency, and performance, ensuring high-quality data for engineering, research, and business stakeholders
  • Collaborate with AI researchers, data scientists, product engineers, and business teams to understand data needs and build scalable solutions
  • Drive technical decisions and best practices for data ingestion, orchestration, processing, and storage
What we offer
What we offer
  • equity
  • health, dental & vision
  • retirement with company contribution
  • parental leave & reproductive or family planning support
  • mental health & wellness benefits
  • generous PTO
  • company recharge days
  • a learning & development stipend
  • a work from home stipend
  • cell phone reimbursement
  • Fulltime
Read More
Arrow Right

AI Research Infrastructure Engineer

Block is scaling Customer Insights into an AI-powered insights accelerator that ...
Location
Location
United States , Bay Area
Salary
Salary:
168300.00 - 297000.00 USD / Year
block.xyz Logo
Block
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 7+ years of experience in research, automation implementation, analytics, or related technical fields with hands-on workflow optimization experience
  • 3+ years implementing AI/ML solutions, with experience in automation, LLM integration, or applied AI/analytics workflows
  • Hands-on technical skills in programming languages (Python, R, SQL) for automation development, API/MCP integrations, cloud platforms, and research data pipeline creation
  • Experience with research and analytic platforms and tools (Qualtrics, Snowflake, etc) or transferable experience with analytics and automation platforms
  • Strong technical communication and translation skills with ability to make complex AI/ML concepts, data architecture decisions, and automation workflows accessible and actionable for researchers, product managers, and business stakeholders
  • Proven ability to build stakeholder confidence and alignment during technology transformation
  • Strong project management skills with ability to coordinate multiple complex automation initiatives, manage competing priorities, and deliver measurable operational efficiency gains (reduced cycle times, improved quality outcomes, increased research capacity)
Job Responsibility
Job Responsibility
  • Design, build, and deploy AI agents and agentic workflows that automate research operations from study design through insights delivery, using LLMs, prompt engineering, MCP (Model Context Protocol) integrations, and workflow orchestration integrated with existing research and analytics tech stack
  • Design, build, and maintain automated data pipelines that ingest, transform, and unify research data from diverse sources (surveys, transcripts, analytics, behavioral logs) into AI-ready repositories with RAG capabilities for instant insight access via tools like Goose
  • Architect ETL/ELT frameworks using Python, SQL or equivalent tools to ensure data consistency, traceability, and scalability
  • Develop data models and schemas for research metadata, participant data, and AI-generated insights to support efficient querying and analysis
  • Design and prototype research automation systems using AI/ML techniques, partnering with design & engineering teams to productionize solutions
  • Partner with engineering, design, and platform teams to integrate research automation systems with Block's tech stack (i.e. Goose, GitHub, etc.) and establish governance frameworks for quality, ethics, and compliance
  • Mentor team members on AI agent development, agentic system design, and research automation best practices to build organizational capabilities in intelligent automation
What we offer
What we offer
  • Remote work
  • medical insurance
  • flexible time off
  • retirement savings plans
  • modern family planning
  • Fulltime
Read More
Arrow Right

Cloud Technical Architect / Data DevOps Engineer

The role involves designing, implementing, and optimizing scalable Big Data and ...
Location
Location
United Kingdom , Bristol
Salary
Salary:
Not provided
https://www.hpe.com/ Logo
Hewlett Packard Enterprise
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • An organised and methodical approach
  • Excellent time keeping and task prioritisation skills
  • An ability to provide clear and concise updates
  • An ability to convey technical concepts to all levels of audience
  • Data engineering skills – ETL/ELT
  • Technical implementation skills – application of industry best practices & designs patterns
  • Technical advisory skills – experience in researching technological products / services with the intent to provide advice on system improvements
  • Experience of working in hybrid environments with both classical and DevOps
  • Excellent written & spoken English skills
  • Excellent knowledge of Linux operating system administration and implementation
Job Responsibility
Job Responsibility
  • Detailed development and implementation of scalable clustered Big Data solutions, with a specific focus on automated dynamic scaling, self-healing systems
  • Participating in the full lifecycle of data solution development, from requirements engineering through to continuous optimisation engineering and all the typical activities in between
  • Providing technical thought-leadership and advisory on technologies and processes at the core of the data domain, as well as data domain adjacent technologies
  • Engaging and collaborating with both internal and external teams and be a confident participant as well as a leader
  • Assisting with solution improvement activities driven either by the project or service
  • Support the design and development of new capabilities, preparing solution options, investigating technology, designing and running proof of concepts, providing assessments, advice and solution options, providing high level and low level design documentation
  • Cloud Engineering capability to leverage Public Cloud platform using automated build processes deployed using Infrastructure as Code
  • Provide technical challenge and assurance throughout development and delivery of work
  • Develop re-useable common solutions and patterns to reduce development lead times, improve commonality and lowering Total Cost of Ownership
  • Work independently and/or within a team using a DevOps way of working
What we offer
What we offer
  • Extensive social benefits
  • Flexible working hours
  • Competitive salary
  • Shared values
  • Equal opportunities
  • Work-life balance
  • Evolving career opportunities
  • Comprehensive suite of benefits that supports physical, financial and emotional wellbeing
  • Fulltime
Read More
Arrow Right

Data Engineer

Barbaricum is seeking a Data Engineer to provide support an emerging capability ...
Location
Location
United States , Omaha
Salary
Salary:
Not provided
barbaricum.com Logo
Barbaricum
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Active DoD Top Secret/SCI clearance required
  • 8+ years of demonstrated experience in software engineering
  • Bachelor’s degree in computer science or a related field
  • 8+ years of experience working with AWS big data technologies (S3, EC2) and demonstrate experience in distributed data processing, Data Modeling, ETL Development, and/or Data Warehousing
  • Demonstrated mid-level knowledge of software engineering best practices across the development lifecycle
  • 3+ years of experience using analytical concepts and statistical techniques
  • 8+ years of demonstrated experience across Mathematics, Applied Mathematics, Statistics, Applied Statistics, Machine Learning, Data Science, Operations Research, or Computer Science especially around software engineering and/or designing/implementing machine learning, data mining, advanced analytical algorithms, programming, data science, advanced statistical analysis, artificial intelligence
Job Responsibility
Job Responsibility
  • Design, implement, and operate data management systems for intelligence needs
  • Use Python to automate data workflows
  • Design algorithms databases, and pipelines to access, and optimize data retrieval, storage, use, integration and management by different data regimes and digital systems
  • Work with data users to determine, create, and populate optimal data architectures, structures, and systems
  • and plan, design, and optimize data throughput and query performance
  • Participate in the selection of backend database technologies (e.g. SQL, NoSQL, etc.), its configuration and utilization, and the optimization of the full data pipeline infrastructure to support the actual content, volume, ETL, and periodicity of data to support the intended kinds of queries and analysis to match expected responsiveness
  • Assist and advise the Government with developing, constructing, and maintaining data architectures
  • Research, study, and present technical information, in the form of briefings or written papers, on relevant data engineering methodologies and technologies of interest to or as requested by the Government
  • Align data architecture, acquisition, and processes with intelligence and analytic requirements
  • Prepare data for predictive and prescriptive modeling deploying analytics programs, machine learning and statistical methods to find hidden patterns, discover tasks and processes which can be automated and make recommendations to streamline data processes and visualizations
Read More
Arrow Right

Data Engineer

As a Data Engineer, you’ll build and refine the pipelines, data models, and serv...
Location
Location
United States , Redmond
Salary
Salary:
155000.00 - 175000.00 USD / Year
2a.consulting Logo
2A Consulting
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Proven ability to design and build end-to-end data systems, from ingestion through cleaning, structuring, storage, and serving
  • Experience building and shipping data products that deliver practical value
  • Demonstrated impact using AI models in data workflows (applied use, not ML research)
  • 5+ years of software or data engineering experience, including at least 2 years of hands-on work with data pipelines
  • Comfortable defining architecture and starting systems from scratch, working independently in a small cross-functional team
  • Proficiency in Python, SQL, or similar languages used in data engineering workflows
Job Responsibility
Job Responsibility
  • Build and maintain core data pipelines
  • Build and maintain end-to-end ingestion pipelines for documents, datasets, code repositories, videos, transcripts, and internal knowledge sources
  • Clean, normalize, structure, and store data in formats that support both web applications and AI-driven use cases
  • Use “out of the box” Microsoft tools—such as Fabric, Azure services, Cosmos DB, or Copilot Studio—to create reliable, maintainable systems
  • Enrich and model research data
  • Use AI models to transform unstructured content into structured metadata and durable knowledge assets
  • Design the architecture and foundational data systems, establishing the patterns and infrastructure for a new, scalable environment
  • Develop and refine embeddings, vector indexes, and retrieval components to support semantic search and grounding scenarios
  • Build backend and data services
  • Build data services, APIs, and backend components that power internal applications and agent-supported workflows
What we offer
What we offer
  • Flexible time-off plan
  • 100% employer-paid medical, dental, and vision insurance
  • Employer-paid life insurance for those enrolled in medical coverage
  • 401(k) plan with company match
  • Fertility, surrogacy, and adoption benefits
  • Fitness and caregiver benefits
  • Employee Assistance Program
  • 100% employer-paid short- and long-term disability coverage
  • Fulltime
Read More
Arrow Right

Senior Data Engineer

As a senior member of our engineering team, you will take ownership of critical ...
Location
Location
Poland
Salary
Salary:
Not provided
userlane.com Logo
Userlane GmbH
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Minimum of 5 years of hands-on experience in designing and developing data processing systems
  • Experience being part of a team of software engineers and helping establish processes from scratch
  • Familiarity with DBMS like ClickHouse or a different SQL-based OLAP database
  • Experience with various data engineering tools like Airflow, Kafka, dbt
  • Experience building and maintaining applications with the following languages: Python, Golang, Typescript
  • Knowledge of container technologies like Docker and Kubernetes
  • Experience with CI/CD pipelines and automated testing
  • Ability to solve problems and balance structure with creativity
  • Ability to operate independently and apply strategic thinking with technical depth
  • Willingness to share information and skills with the team
Job Responsibility
Job Responsibility
  • Shape and maintain our various data and backend components - DBs, APIs and services
  • Understand business requirements and analyze their impact on the design of our software services and tools
  • Identify architectural changes needed in our infrastructure to support a smooth process of adding new features
  • Research, propose, and deliver changes to our software architecture to address our engineering and product requirements
  • Design, develop, and maintain a solid and stable RESTful API based on industry standards and best practices
  • Collaborate with internal and external teams to deliver software that fits the overall ecosystem of our products
  • Stay up to date with the new trends and technologies that enable us to work smarter, not harder
What we offer
What we offer
  • Team & Culture: A high-performance culture with great leadership and a fun, engaged, motivated, and diverse team with people from over 20 countries
  • Market: Userlane is among the global leaders in the rapidly growing Digital Adoption industry
  • Growth: We take you and your development seriously. You can expect weekly 121s, a personalised skills assessment and development plan, on the job coaching and a budget for events and training
  • Compensation: Significant financial upside with an attractive and incentivising package on B2B basis
  • Fulltime
Read More
Arrow Right

AI Research Engineer

We're seeking a Research Engineer to conduct innovative research in key AI areas...
Location
Location
United Kingdom
Salary
Salary:
Not provided
prolific.com Logo
Prolific
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 5+ years of engineering experience with significant AI/ML focus
  • Demonstrated research experience through publications, open-source contributions, or impactful projects
  • Strong engineering fundamentals and experience implementing AI systems in production environments
  • Deep knowledge of LLM evaluation methodologies, alignment techniques, and model optimization approaches
  • Experience with model fine-tuning, adapters, quantization, and distillation frameworks
  • Self-motivation and ability to define and pursue research directions independently
  • Excellent understanding of current challenges in AI safety, reliability, and alignment
  • Strong communication skills and ability to explain complex research concepts clearly
  • Passion for staying current with the rapidly evolving AI research landscape
Job Responsibility
Job Responsibility
  • Lead independent research projects in AI evaluation methodologies, alignment techniques, and synthetic data generation
  • Design and implement novel evaluation frameworks for LLMs and agent systems that are grounded in human data
  • Contribute to the academic AI community through publications and open-source contributions
  • Stay at the forefront of AI research and pioneer innovative approaches to tackle pressing open challenges in the field
  • Design and conduct rigorous experiments to study AI models and systems with sound methodological approaches
  • Develop scalable frameworks for systematic evaluation of model behaviours and capabilities
  • Create tools and frameworks that transform research insights into practical applications
  • Build infrastructure to support large-scale research experiments when needed
  • Apply knowledge of model fine-tuning, optimization techniques, distillation, and other ML engineering practices to support research goals
  • Work closely with ML engineers, data scientists, and product teams to translate research insights into practical applications
What we offer
What we offer
  • competitive salary
  • benefits
  • remote working
  • impactful, mission-driven culture
Read More
Arrow Right