This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
At Together.ai, we are building state-of-the-art infrastructure to enable efficient and scalable inference for large language models (LLMs). Our mission is to optimize inference frameworks, algorithms, and infrastructure, pushing the boundaries of performance, scalability, and cost-efficiency. We are seeking an Inference Frameworks and Optimization Engineer to design, develop, and optimize distributed inference engines that support multimodal and language models at scale. This role will focus on low-latency, high-throughput inference, GPU/accelerator optimizations, and software-hardware co-design, ensuring efficient large-scale deployment of LLMs and vision models.
Job Responsibility:
Design and develop fault-tolerant, high-concurrency distributed inference engine for text, image, and multimodal generation models
Implement and optimize distributed inference strategies, including Mixture of Experts (MoE) parallelism, tensor parallelism, pipeline parallelism for high-performance serving
Apply CUDA graph optimizations, TensorRT/TRT-LLM graph optimizations, and PyTorch-based compilation (torch.compile), and speculative decoding to enhance efficiency and scalability
Collaborate with hardware teams on performance bottleneck analysis, co-optimize inference performance for GPUs, TPUs, or custom accelerators
Work closely with AI researchers and infrastructure engineers to develop efficient model execution plans and optimize E2E model serving pipelines
Requirements:
3+ years of experience in deep learning inference frameworks, distributed systems, or high-performance computing
Familiar with at least one LLM inference frameworks (e.g., TensorRT-LLM, vLLM, SGLang, TGI(Text Generation Inference))
Background knowledge and experience in at least one of the following: GPU programming (CUDA/Triton/TensorRT), compiler, model quantization, and GPU cluster scheduling
Deep understanding of KV cache systems like Mooncake, PagedAttention, or custom in-house variants
Proficient in Python and C++/CUDA for high-performance deep learning inference
Deep understanding of Transformer architectures and LLM/VLM/Diffusion model optimization
Knowledge of inference optimization, such as workload scheduling, CUDA graph, compiled, efficient kernels
Strong analytical problem-solving skills with a performance-driven mindset
Excellent collaboration and communication skills across teams
Nice to have:
Experience in developing software systems for large-scale data center networks with RDMA/RoCE
Familiar with distributed filesystem(e.g., 3FS, HDFS, Ceph)
Familiar with open source distributed scheduling/orchestration frameworks, such as Kubernetes (K8S)
Contributions to open-source deep learning inference projects