Demo Platform LLM optimization and inference leveraging
Links
Red Hat

LLM optimization and inferencing

    • 1. vLLM & Performance Tuning
      • Securing vLLM Endpoints
      • Troubleshooting
      • Configuration
    • 2. LLM Compressor, Model Quantization and Sparsification
      • LLM Compressor
      • Model Optimization - Deep dive
      • Quantization in Practice - Lab
      • Quantization Pipeline - Lab
    • 3. LLM evaluation with GuideLLM
      • Load Testing
    • 4. RH Inference Server on Multiple Platforms
      • RHEL
      • OpenShift
      • OpenShift AI
      • Ubuntu
    • LLM Compressor: Executive Guide
      • Customer Qualification
      • Business Value
      • Model Selection
      • Managing Accuracy
      • Deployment Framework
      • Competitive Positioning
      • Team Guidance
      • Implementation
      • Common Objections
    • llm-d Technical Overview
      • What It Does
      • Architecture
      • Technical Requirements
        • Infrastructure
        • Prerequisites
        • Deployment
      • Key Features
    • LLM Compressor: Model Comparison Examples
  • Dev Mode
    • Inference Server on Multiple Platforms
    • openshift-icons
    • Attributes Page
  • LLM optimization and inferencing
    • master
  • LLM optimization and inferencing
  • Dev Mode
  • openshift-icons
Inference Server on Multiple Platforms Attributes Page

Powered by

Demo Platform