CrawlJobs Logo

AI Operations Platform Consultant

clearbridgetech.com Logo

ClearBridge Technology Group

Location Icon

Location:
United States, Charlotte

Category Icon
Category:
IT - Software Development

Job Type Icon

Contract Type:
Not provided

Salary Icon

Salary:

75.00 - 110.00 USD / Hour

Job Responsibility:

  • Deploying, managing, operating, and troubleshooting containerized services at scale on Kubernetes for mission-critical applications (OpenShift)
  • Deploying, configuring, and tuning LLMs using TensorRT-LLM and Triton Inference server.
  • Managing MLOps/LLMOps pipelines, using TensorRT-LLM and Triton Inference server to deploy inference services in production
  • Setup and operation of AI inference service monitoring for performance and availability.
  • Experience deploying and troubleshooting LLM models on a containerized platform, monitoring, load balancing, etc.
  • Operation and support of MLOps/LLMOps pipelines, using TensorRT-LLM and Triton Inference server to deploy inference services in production
  • Deploying and troubleshooting LLM models on a containerized platform, monitoring, load balancing, etc.
  • Standard processes for operation of a mission critical system – incident management, change management, event management, etc.
  • Managing scalable infrastructure for deploying and managing LLMs
  • Deploying models in production environments, including containerization, microservices, and API design
  • Triton Inference Server, including its architecture, configuration, and deployment.
  • Model Optimization techniques using Triton with TRTLLM
  • Model optimization techniques, including pruning, quantization, and knowledge distillation

Requirements:

  • Ability to pass an in-depth background check
  • Ability to work onsite out of either Jersey City, NJ or Charlotte, NC 3 days per week
  • Experience deploying, managing, operating, and troubleshooting containerized services at scale on Kubernetes for mission-critical applications (OpenShift)
  • Experience with deploying, configuring, and tuning LLMs using TensorRT-LLM and Triton Inference server.
  • Experience deploying and troubleshooting LLM models on a containerized platform, monitoring, load balancing, etc.
  • Experience with standard processes for operation of a mission critical system – incident management, change management, event management, etc.
What we offer:

excellent benefits and compensation packages

Additional Information:

Job Posted:
December 07, 2025

Work Type:
Hybrid work
Job Link Share:
Welcome to CrawlJobs.com
Your Global Job Discovery Platform
At CrawlJobs.com, we simplify finding your next career opportunity by bringing job listings directly to you from all corners of the web. Using cutting-edge AI and web-crawling technologies, we gather and curate job offers from various sources across the globe, ensuring you have access to the most up-to-date job listings in one place.