Red Hat AI Inference Server 3.2
Welcome
Release notes
Highlights of what is new and what has changed with this Red Hat AI Inference Server release
Getting started
Getting started with Red Hat AI Inference Server
Deploying Red Hat AI Inference Server in a disconnected environment
Deploy Red Hat AI Inference Server in a disconnected environment using OpenShift Container Platform and a disconnected mirror image registry
Deploying Red Hat AI Inference Server in OpenShift Container Platform
Deploy Red Hat AI Inference Server in OpenShift Container Platform clusters that have supported AI accelerators installed
Supported product and hardware configurations
Supported hardware and software configurations for deploying Red Hat AI software
Validated models
Red Hat AI validated models
Inference Operations
Inference serving language models in OCI-compliant model containers
Inferencing OCI-compliant models in Red Hat AI Inference Server
vLLM server arguments
Server arguments for running Red Hat AI Inference Server
Red Hat AI Model Optimization Toolkit
Compressing large language models with the LLM Compressor library
Related Products
Red Hat OpenShift AI
Find release notes and product documentation for using the OpenShift AI platform and its integrated MLOps capabilities to manage predictive and generative AI models at scale
Red Hat Enterprise Linux AI
Switch to the Red Hat Enterprise Linux AI documentation
Additional Resources
Product life cycle
Understand the product life cycle to plan deployments and support applications using the product
Red Hat AI learning hub
Explore a curated collection of learning resources designed to help you accomplish key tasks with Red Hat AI products and services