This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
About the Role: We are a fast-growing technology company leveraging large language models and AI to power next-generation applications. Our teams deliver high-impact solutions across diverse domains while maintaining a culture of curiosity, collaboration, and continuous learning. We’re seeking a skilled Prompt / AI Engineer to design, implement, and optimize natural-language prompts, chains, and retrieval-augmented workflows. You’ll partner with engineering, product, and data science teams to ensure seamless integration of LLM capabilities into our services.
Job Responsibility:
Prompt Design & Optimization: Craft and iterate prompt templates for a variety of use cases (summarization, classification, Q&A)
Implement few-shot examples, chain-of-thought techniques, and dynamic parameterization
Benchmark, profile, and refine prompts to balance accuracy, latency, and cost
Workflow & Chain Development: Build multi-step LLM pipelines using frameworks like LangChain, LlamaIndex, or equivalent
Integrate retrieval-augmented generation (RAG) with vector stores and document loaders
Develop agentic workflows that orchestrate API calls, data enrichment, and user interactions
Model Integration & Testing: Integrate with public and private LLM services (e.g., OpenAI, Anthropic, self-hosted models)
Automate performance tests, error-handling routines, and hallucination checks
Monitor usage metrics, implement rate-limits, and optimize token budgets
Collaboration & Documentation: Work closely with backend/frontend engineers to expose LLM endpoints
Create and maintain a prompt library, testing suite, and best-practice guidelines
Conduct reviews, knowledge transfers, and training sessions on prompt engineering
Requirements:
3+ years of experience in AI/ML, NLP, or software engineering with a focus on prompt or LLM integration
Strong proficiency in Python and familiarity with LLM client libraries
Hands-on with at least one orchestration framework (LangChain, LlamaIndex, Dify, etc.)
Experience working with vector databases (Pinecone, Weaviate, Qdrant) or embedding services
Solid understanding of prompt-engineering techniques, chain-of-thought, and RAG concepts
Ability to benchmark, profile, and optimize model calls for performance and cost
Excellent written communication for clear documentation and prompt examples
Nice to have:
Experience fine-tuning or prompt-tuning domain-specific models
Familiarity with self-hosted LLM deployments (Hugging Face, OpenAI local runtimes)
Background in automated evaluation metrics (ROUGE, BLEU, custom tests)
Contributions to open-source prompt libraries or agent frameworks
Exposure to CI/CD for AI services and observability tools
What we offer:
Impactful Projects: Shape core LLM-driven capabilities across diverse applications
Innovative Environment: Work with cutting-edge models, RAG pipelines, and agent frameworks
Growth & Learning: Access to mentorship, training budgets, and conference support