CrawlJobs Logo

Senior Data Engineer (LLM)

sunscrapers.com Logo

Sunscrapers sp. z o.o.

Location Icon

Location:
Poland , Warsaw

Category Icon

Job Type Icon

Contract Type:
Not provided

Salary Icon

Salary:

Not provided

Job Description:

Sunscrapers is a technology consultancy that empowers finance and healthcare leaders to succeed by leveraging cutting-edge software, data, and AI. We combine world-class engineering, deep industry expertise, and proprietary know-how to deliver innovative, high-impact solutions. Specializing in software engineering, DevOps, data engineering, and data science, we design and build AI-powered data platforms and web applications tailored to each client’s unique needs. Trusted by over 60 clients across the US, UK, and beyond, we consistently maintain a 4.9/5 client satisfaction rating, with partnerships averaging five years or more. The project: We are carrying out the project for our client, an American private equity and investment management fund - listed on the Forbes 500 list - based in New York. We support them in the area of the infrastructure and data platform, and very recently we also build and experiment with Gen AI applications. The client operates very widely in the world of finance, loans, investments and real estate. As a Senior Data Engineer with Frontend Focus you'll design and implement core systems that enable data science, data visualization, and agent-based applications at companies that implement data-driven decision processes to create a competitive advantage. You'll build a data platform and internal agent-based tooling for data and business teams, including data pipeline orchestrator, data warehouses, and authenticated frontend interfaces.

Job Responsibility:

  • Developing PoCs using latest technologies, experimenting with third party integrations
  • Owning and evolving Next.js/React frontend with authenticated flows and secure session handling
  • Delivering production grade applications once PoCs are validated
  • Creating solutions that enable data scientists and business analysts to be self-sufficient as much as possible
  • Designing and implementing secure, scalable access patterns (OAuth2/OIDC, authorization boundaries)
  • Finding new ways how to leverage Gen AI applications and underlying vector and graph data storages
  • Contributing across the stack including FastAPI backend services and agent-driven workflows
  • Contributing data technology stacks including data warehouses and ETL pipelines
  • Building data flows for fetching, aggregation and data modeling using batch and streaming pipelines
  • Documenting design decisions before implementation

Requirements:

  • At least 5+ years of professional experience in data-related or full-stack engineering role
  • Undergraduate or graduate degree in Computer Science, Engineering, Mathematics, or similar
  • Strong proficiency in Next.js / React for frontend development
  • Expertise in Python and SQL languages
  • Experience with data warehouses (Snowflake) and different database technologies (RDBMS, vector, graphs)
  • Proven experience building secure, scalable web applications
  • Expertise in AWS stack and services, proficiency in using Docker
  • Experience with infrastructure-as-code tools, like Terraform
  • Excellent command in spoken and written English, at least C1
  • Creative problem-solving skills and excellent technical documentation
  • Ability to work with both Windows and Unix-like operating systems

Nice to have:

  • Experience with FastAPI or other Python web frameworks
  • Knowledge of OAuth2/OIDC flows, PKCE, token/session handling
  • Experience with integrating LLMs (OpenAI but also others, maybe open source)
  • Agent frameworks experience (OpenAI Agents, Agno, Strands) or MCP-style integrations
  • Understanding of LLMs fine tuning, embedding and vector semantic searching
  • Experience with Pinecone or Neo4j
  • Observability stacks experience (OpenTelemetry or similar)
  • Experience in building ETL processes and data pipelines with platforms like Airflow
  • AWS ECS, VPCs, and secure networking patterns
  • Proficiency in statistics and machine learning with Python libraries
  • Experience in working with repository manager, for example Jfrog Artifactory
What we offer:
  • Working alongside a talented team of software engineers who are changing the image of Poland abroad
  • Culture of teamwork, professional development and knowledge sharing
  • Flexible working hours and remote work possibility
  • Comfortable office in central Warsaw, equipped with all the necessary tools for conquering the universe (Macbook Pro/Dell, external screen, ergonomic chairs)

Additional Information:

Job Posted:
February 18, 2026

Work Type:
Remote work
Job Link Share:

Looking for more opportunities? Search for other job offers that match your skills and interests.

Briefcase Icon

Similar Jobs for Senior Data Engineer (LLM)

Data Engineer Senior

We are looking for a highly skilled professional to lead the industrialisation o...
Location
Location
Portugal , Lisbon
Salary
Salary:
Not provided
https://www.inetum.com Logo
Inetum
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Minimum of 5 years’ experience in MLOps, data engineering, or DevOps with a focus on ML/DL/LLM/AI agents in production environments
  • Strong proficiency in Python
  • Hands-on experience with CI/CD tools such as GitLab, Docker, Kubernetes, Jenkins
  • Solid understanding of ML, DL, and LLM models
  • Experience with ML lifecycle tools such as MLflow or DVC
  • Good understanding of model lifecycle, data traceability, and governance frameworks
  • Experience with on-premise and hybrid infrastructures
  • Excellent communication skills and ability to collaborate with remote teams
  • Proactive mindset, technical rigour, and engineering mentality
  • Willingness to learn, document, and standardise best practices
Job Responsibility
Job Responsibility
  • Analyse, monitor, and optimise ML models, tracking their performance
  • Design and implement CI/CD pipelines for ML models and data flows
  • Containerise and deploy models via APIs, batch processes, and streaming
  • Manage model versioning and traceability
  • Ensure continuous improvement and adaptation of AI use cases and ML models
  • Set up monitoring and alerting for model performance
  • Establish incident response protocols in collaboration with IT
  • Maintain dashboards and automated reports on model health
  • Implement validation frameworks for data and models (e.g., Great Expectations, unit tests, stress tests), in collaboration with Group Governance
  • Contribute to documentation and apply technical best practices
What we offer
What we offer
  • Work in a constantly evolving environment
  • Contribute to digital impact
  • Opportunity for growth and development
  • Fulltime
Read More
Arrow Right

Senior Data Engineer

We're seeking an experienced Senior Data Engineer to help shape the future of he...
Location
Location
Germany , Berlin
Salary
Salary:
Not provided
audibene.de Logo
Audibene GmbH
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 5+ years of hands on experience with complex ETL processes, data modeling, and large scale data systems
  • Production experience with modern cloud data warehouses (Snowflake, BigQuery, Redshift) on AWS, GCP, or Azure
  • Proficiency in building and optimizing data transformations and pipelines in python
  • Experience with columnar storage, MPP databases, and distributed data processing architectures
  • Ability to translate complex technical concepts for diverse audiences, from engineers to business stakeholders
  • Experience with semantic layers, data catalogs, or metadata management systems
  • Familiarity with modern analytical databases like Snowflake, BigQuery, ClickHouse, DuckDB, or similar systems
  • Experience with streaming technologies like Kafka, Pulsar, Redpanda, or Kinesis
Job Responsibility
Job Responsibility
  • Design and build robust, high performance data pipelines using our modern stack (Airflow, Snowflake, Pulsar, Kubernetes) that feed directly into our semantic layer and data catalog
  • Create data products optimized for consumption by AI agents and LLMs where data quality, context, and semantic richness are crucial
  • Structure and transform data to be inherently machine readable, with rich metadata and clear lineage that powers intelligent applications
  • Take responsibility from raw data ingestion through to semantic modeling, ensuring data is not just accurate but contextually rich and agent ready
  • Champion best practices in building LLM consumable data products, optimize for both human and machine consumers, and help evolve our dbt transformation layer
  • Built data products for AI/LLM consumption, not just analytics dashboards
What we offer
What we offer
  • Work 4 days a week from our office (Berlin/Mainz) with a passionate team, and 1 day a week from home
  • Regularly join on- and offline team events, company off-sites, and the annual audibene Wandertag
  • Cost of the Deutschland-Ticket covered
  • Access to over 50,000 gyms and wellness facilities through Urban Sports Club
  • Support for personal development with a wide range of programs, trainings, and coaching opportunities
  • Dog-friendly office
  • Fulltime
Read More
Arrow Right

Senior Data Engineer

Radix is building the most trusted data and analytics platform in multifamily. J...
Location
Location
United States , Scottsdale
Salary
Salary:
Not provided
radix.com Logo
Radix (AZ)
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 5+ years of experience in data engineering or backend systems, with 2+ years collaborating closely with analytical or scientific practitioners
  • Experience designing at least one end-to-end data application (visualization, model pipeline, etc.) for production use
  • Strong understanding of data modeling, batch processing, and streaming systems
  • Hands-on experience with SQL/NoSQL databases, cloud storage, file-based datasets, and Infrastructure-as-code
  • Experience building or operating data pipelines on AWS cloud services (Lambda, S3, RDS, ECS)
  • Understanding of AI/LLM integration and prompt engineering fundamentals
  • Proficiency with Git/GitHub, and familiarity with Spark and Kubernetes
  • Strong problem-solving skills, ownership, and ability to identify root causes of technical debt and recurring problems
  • Demonstrated ability to translate Product/Science requirements into technical plans and hold teams accountable to delivery timelines
  • Undergraduate degree in computer science, computer engineering, software engineering, or equivalent
Job Responsibility
Job Responsibility
  • Build scalable ETL/ELT pipelines for ingesting structured and unstructured data (Excel, JSON, PDFs, APIs)
  • Design and maintain data pipelines using SQL, Python, Node.js, or TypeScript
  • Work with distributed compute systems (Spark, Kubernetes), message queues, and streaming data
  • Manage and optimize data storage in MongoDB, PostgreSQL, Redis, Snowflake, and S3
  • Develop clean, standardized data schemas and event-driven transformations
  • Integrate AI-assisted parsers, mappers, and LLM-supported transformations
  • Collaborate with backend, analytics, product, and AI teams to break down requirements into well-defined engineering problems
  • Implement monitoring, data validation, and reliability checks (DQ rules, freshness, duplication)
  • Own production readiness, including on-call responsibilities and incident follow-ups
  • Mentor engineers, conduct thorough code reviews, and introduce patterns that raise the team's technical bar
What we offer
What we offer
  • Medical, dental and vision coverage designed to support your wellbeing
  • Pre-IPO Equity
  • Performance Bonus
  • Learn From the Best
  • Build Category-Defining Products
Read More
Arrow Right

Senior ML Data Engineer

As a Senior Data Engineer, you will play a pivotal role in our AI/ML workstream,...
Location
Location
Poland , Warsaw
Salary
Salary:
Not provided
awin.com Logo
Awin Global
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor or Master’s degree in data science, data engineering, Computer Science with focus on math and statistics / Master’s degree is preferred
  • At least 5 years experience as AI/ML data engineer undertaking above task and accountabilities
  • Strong foundation in computer science principes and statistical methods
  • Strong experience with cloud technology (AWS or Azure)
  • Strong experience with creation of data ingestion pipeline and ET process
  • Strong knowledge of big data tool such as Spark, Databricks and Python
  • Strong understanding of common machine learning techniques and frameworks (e.g. mlflow)
  • Strong knowledge of Natural language processing (NPL) concepts
  • Strong knowledge of scrum practices and agile mindset
  • Strong Analytical and Problem-Solving Skills with attention to data quality and accuracy
Job Responsibility
Job Responsibility
  • Design and maintain scalable data pipelines and storage systems for both agentic and traditional ML workloads
  • Productionise LLM- and agent-based workflows, ensuring reliability, observability, and performance
  • Build and maintain feature stores, vector/embedding stores, and core data assets for ML
  • Develop and manage end-to-end traditional ML pipelines: data prep, training, validation, deployment, and monitoring
  • Implement data quality checks, drift detection, and automated retraining processes
  • Optimise cost, latency, and performance across all AI/ML infrastructure
  • Collaborate with data scientists and engineers to deliver production-ready ML and AI systems
  • Ensure AI/ML systems meet governance, security, and compliance requirements
  • Mentor teams and drive innovation across both agentic and classical ML engineering practices
  • Participate in team meetings and contribute to project planning and strategy discussions
What we offer
What we offer
  • Flexi-Week and Work-Life Balance: We prioritise your mental health and well-being, offering you a flexible four-day Flexi-Week at full pay and with no reduction to your annual holiday allowance. We also offer a variety of different paid special leaves as well as volunteer days
  • Remote Working Allowance: You will receive a monthly allowance to cover part of your running costs. In addition, we will support you in setting up your remote workspace appropriately
  • Pension: Awin offers access to an additional pension insurance to all employees in Germany
  • Flexi-Office: We offer an international culture and flexibility through our Flexi-Office and hybrid/remote work possibilities to work across Awin regions
  • Development: We’ve built our extensive training suite Awin Academy to cover a wide range of skills that nurture you professionally and personally, with trainings conveniently packaged together to support your overall development
  • Appreciation: Thank and reward colleagues by sending them a voucher through our peer-to-peer program
Read More
Arrow Right

Senior LLM Backend Engineer

We are looking for a Senior Backend Engineer with a strong focus on Large Langua...
Location
Location
Spain
Salary
Salary:
Not provided
bark.com Logo
Bark
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Extensive production experience with Python in backend engineering
  • Proven experience integrating LLMs into applications via APIs or SDKs
  • Strong experience building and maintaining APIs for LLM-based features
  • Strong experience building and maintaining event-driven workflows
  • Strong experience building and maintaining business logic that consumes AI outputs
  • Strong experience building and maintaining integrations with 3rd party AI/ML platforms
  • Solid SQL and NoSQL experience (especially in AI data pipelines)
  • Production experience with Docker, ideally with Kubernetes or AWS Fargate/ECS/EKS
  • Experience deploying and maintaining AI services in cloud environments
  • Strong organisational skills and ability to deliver in a fast-paced, product-focused environment
Job Responsibility
Job Responsibility
  • Work with product managers to understand user needs and translate them into AI-powered functionality
  • Design and build APIs, services, and workflows that integrate LLMs (both proprietary and open-source)
  • Implement prompt engineering, RAG pipelines, and model fine-tuning where required
  • Optimise AI inference performance, scalability, and cost-effectiveness
  • Ensure AI features meet high standards for security, reliability, and maintainability
  • Collaborate with other engineers to integrate AI features seamlessly into the wider system
  • Stay on top of emerging LLM technologies and best practices, running experiments and sharing knowledge across the team
What we offer
What we offer
  • Fully remote working
  • Personal annual L&D Budgets with 600€ to spend on your development
  • Being at the forefront of an industry with new and exciting problems to solve
  • Fulltime
Read More
Arrow Right

Senior ML Data Engineer

As a Senior Data Engineer, you will play a pivotal role in our AI/ML workstream,...
Location
Location
Salary
Salary:
Not provided
awin.com Logo
Awin Global
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Bachelor or Master’s degree in data science, data engineering, Computer Science with focus on math and statistics / Master’s degree is preferred
  • At least 5 years experience as AI/ML data engineer undertaking above task and accountabilities
  • Strong foundation in computer science principes and statistical methods
  • Strong experience with cloud technology (AWS or Azure)
  • Strong experience with creation of data ingestion pipeline and ET process
  • Strong knowledge of big data tool such as Spark, Databricks and Python
  • Strong understanding of common machine learning techniques and frameworks (e.g. mlflow)
  • Strong knowledge of Natural language processing (NPL) concepts
  • Strong knowledge of scrum practices and agile mindset
Job Responsibility
Job Responsibility
  • Design and maintain scalable data pipelines and storage systems for both agentic and traditional ML workloads
  • Productionise LLM- and agent-based workflows, ensuring reliability, observability, and performance
  • Build and maintain feature stores, vector/embedding stores, and core data assets for ML
  • Develop and manage end-to-end traditional ML pipelines: data prep, training, validation, deployment, and monitoring
  • Implement data quality checks, drift detection, and automated retraining processes
  • Optimise cost, latency, and performance across all AI/ML infrastructure
  • Collaborate with data scientists and engineers to deliver production-ready ML and AI systems
  • Ensure AI/ML systems meet governance, security, and compliance requirements
  • Mentor teams and drive innovation across both agentic and classical ML engineering practices
  • Participate in team meetings and contribute to project planning and strategy discussions
What we offer
What we offer
  • Flexi-Week and Work-Life Balance: We prioritise your mental health and well-being, offering you a flexible four-day Flexi-Week at full pay and with no reduction to your annual holiday allowance. We also offer a variety of different paid special leaves as well as volunteer days
  • Remote Working Allowance: You will receive a monthly allowance to cover part of your running costs. In addition, we will support you in setting up your remote workspace appropriately
  • Pension: Awin offers access to an additional pension insurance to all employees in Germany
  • Flexi-Office: We offer an international culture and flexibility through our Flexi-Office and hybrid/remote work possibilities to work across Awin regions
  • Development: We’ve built our extensive training suite Awin Academy to cover a wide range of skills that nurture you professionally and personally, with trainings conveniently packaged together to support your overall development
  • Appreciation: Thank and reward colleagues by sending them a voucher through our peer-to-peer program
Read More
Arrow Right

LLM - Senior Staff Engineer - Python + Machine Learning

AquSag is seeking a hands-on Machine Learning Senior Staff Engineer to lead cros...
Location
Location
Salary
Salary:
40.00 - 60.00 USD / Hour
aqusag.com Logo
AquSag Technologies
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • 9+ yrs of strong background in Machine Learning, NLP, and modern deep learning architectures (Transformers, LLMs)
  • Hands-on experience with frameworks such as PyTorch, TensorFlow, Hugging Face, or DeepSpeed
  • Hands-on experience in Docker for Production deployment
  • Proven experience managing teams delivering ML/LLM models in production environments
  • Knowledge of distributed training, GPU/TPU optimization, and cloud platforms (AWS, GCP, Azure)
  • Familiarity with MLOps tools like MLflow, Kubeflow, or Vertex AI for scalable ML pipelines
  • Excellent leadership, communication, and cross-functional collaboration skills
  • Bachelor’s or Master’s in Computer Science, Engineering, or related field (PhD preferred)
  • Overlap of 6 hours with PST time zone is mandatory
  • Commitments Required: 8 hours per day with overlap of 6 hours with PST
Job Responsibility
Job Responsibility
  • Lead and mentor a cross-functional team of ML engineers, data scientists, and MLOps professionals
  • Oversee the full lifecycle of LLM and ML projects — from data collection to training, evaluation, and deployment
  • Collaborate with Research, Product, and Infrastructure teams to define goals, milestones, and success metrics
  • Provide technical direction on large-scale model training, fine-tuning, and distributed systems design
  • Implement best practices in MLOps, model governance, experiment tracking, and CI/CD for ML
  • Manage compute resources, budgets, and ensure compliance with data security and responsible AI standards
  • Communicate progress, risks, and results to stakeholders and executives effectively
  • Fulltime
Read More
Arrow Right

Senior Principal Machine Learning Engineer - LLM Post-Training and Optimization

Atlassian is seeking a highly skilled and experienced Senior Principle Machine L...
Location
Location
United States , Mountain View
Salary
Salary:
243100.00 - 407200.00 USD / Year
https://www.atlassian.com Logo
Atlassian
Expiration Date
Until further notice
Flip Icon
Requirements
Requirements
  • Ph.D. or Master’s degree in Computer Science, Machine Learning, Artificial Intelligence, or a related field
  • 8+ years of experience in machine learning, with a focus on large-scale model development and optimization
  • Deep expertise in LLM and transformer architectures (e.g., GPT, BERT, T5)
  • Strong proficiency in Python and ML frameworks such as PyTorch, JAX, or TensorFlow
  • Experience with distributed training techniques and large-scale data processing pipelines
  • Proven track record of deploying machine learning models in production environments
  • Familiarity with model optimization techniques, including quantization, pruning, and knowledge distillation
  • Strong problem-solving skills and ability to work in a fast-paced, collaborative environment
  • Excellent communication skills and ability to translate technical concepts for diverse audiences
Job Responsibility
Job Responsibility
  • Lead the fine-tuning and post-training optimization of large language models (LLMs) for diverse applications
  • Develop and implement techniques for model compression, quantization, pruning, and knowledge distillation to optimize performance and reduce computational costs
  • Conduct research on advanced techniques in transfer learning, reinforcement learning, and prompt engineering for LLMs
  • Design and execute rigorous benchmarking and evaluation frameworks to assess model performance across multiple dimensions
  • Collaborate with infrastructure teams to optimize LLM deployment pipelines, ensuring scalability and efficiency in production environments
  • Stay at the forefront of advancements in LLM technologies, sharing insights, driving innovation within the team, and leading agile development
  • Mentoring other team members, facilitating within/across team workshops, fostering a culture of technical excellence and continuous learning
What we offer
What we offer
  • health coverage
  • paid volunteer days
  • wellness resources
  • Fulltime
Read More
Arrow Right