This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
We are looking for an AI Engineer to maintain and enhance the AI-driven backbone of the Sootra platform. This role involves ensuring production stability of LLM/VLM pipelines, optimizing model interactions, maintaining APIs and queues, and building feedback loops that continuously improve AI outputs.
Job Responsibility:
Maintain and optimize LLM- and VLM-powered services for content generation, compliance scoring, and campaign testing
Manage and scale Flask/FastAPI microservices, ensuring high uptime and low latency
Maintain Dramatiq queues for async AI workflows, campaign generation, and pipeline orchestration
Deploy, monitor, and debug Uvicorn/Gunicorn-based hosting in production environments
Integrate with OpenRouter and equivalent LLM routing tools to balance cost, latency, and quality
Design and refine prompt engineering strategies for reliability, context-awareness, and compliance
Build and maintain feedback pipelines for AI model evaluation (human-in-the-loop scoring, automated quality checks, reinforcement)
Expose and maintain REST APIs for AI services, ensuring secure, versioned endpoints
Collaborate with backend/frontend teams to keep microservice architecture aligned and maintainable
Track token consumption, latency, and error rates to ensure production-grade performance
Requirements:
Strong in Python, with experience in production-grade codebases
Flask (for APIs), FastAPI (optional), Uvicorn/Gunicorn for async hosting
Dramatiq (or Celery/RQ equivalent) for background jobs
Hands-on with LLMs and VLMs, including prompt engineering, fine-tuning, and evaluation
Familiar with OpenRouter or equivalent LLM/VLM routing & fallback tools
Experience designing and maintaining microservice architectures
Strong experience with REST API design (auth, rate limiting, documentation)