# runner **Repository Path**: thuanghai/runner ## Basic Information - **Project Name**: runner - **Description**: clone from github - **Primary Language**: Docker - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2026-02-19 - **Last Updated**: 2026-03-07 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # GPUStack Runner This repository serves as the Docker image pack center for GPUStack Runner. It provides a collection of Dockerfiles to build images for various inference services across different accelerated backends. ## Agenda - [Onboard Services](#onboard-services) - [Directory Structure](#directory-structure) - [Dockerfile Convention](#dockerfile-convention) - [Docker Image Naming Convention](#docker-image-naming-convention) - [Integration Process](#integration-process) ## Onboard Services > [!TIP] > - The list below shows the accelerated backends and inference services available in the latest release. For support of backends or services not shown here, please refer to previous release tags. > - Deprecated inference service versions in the latest release are marked with ~~strikethrough~~ formatting. They may still be available in previous releases, and not recommended for new deployments. > - Polished inference service versions in the latest release are marked with **bold** formatting. If they are using in your deployment, it is recommended to pull the latest images and upgrade. The following table lists the supported accelerated backends and their corresponding inference services with versions. ### Ascend CANN > [!WARNING] > - The Atlas 300I series is currently experimental in vLLM, only supporting eager mode and float16 data type. And there are some known issues for running vLLM, you can refer to vllm-ascend [#3316](https://github.com/vllm-project/vllm-ascend/issues/3316) and [#2795](https://github.com/vllm-project/vllm-ascend/issues/2795). | CANN Version
(Variant) | MindIE | vLLM | SGLang | |------------------------------|-----------|-------------------------------------------|--------------------------------| | 8.5 (A3/910C) | `2.3.0` | `0.15.0`(rc), `0.14.1`(rc),
`0.13.0` | `0.5.9`, `0.5.8.post1` | | 8.5 (910B) | `2.3.0` | `0.15.0`(rc), `0.14.1`(rc),
`0.13.0` | `0.5.9`, `0.5.8.post1` | | 8.5 (310P) | `2.3.0` | `0.15.0`(rc), `0.14.1`(rc) | | | 8.3 (A3/910C) | `2.2.rc1` | `0.12.0`(rc), `0.11.0` | `0.5.7`, `0.5.6.post2` | | 8.3 (910B) | `2.2.rc1` | `0.12.0`(rc), `0.11.0` | `0.5.7`, `0.5.6.post2` | | 8.3 (310P) | `2.2.rc1` | | | | 8.2 (A3/910C) | `2.1.rc2` | `0.10.2`(rc) | ~~`0.5.2`~~, ~~`0.5.1.post3`~~ | | 8.2 (910B) | `2.1.rc2` | `0.10.2`(rc), `0.10.0`(rc),
`0.9.1` | ~~`0.5.2`~~, ~~`0.5.1.post3`~~ | | 8.2 (310P) | `2.1.rc2` | `0.10.0`(rc), `0.9.1` | | ### Iluvatar CoreX | CoreX Version
(Variant) | vLLM | |-------------------------------|---------| | 4.2 | `0.8.3` | ### NVIDIA CUDA > [!NOTE] > - CUDA 12.9 supports Compute Capabilities: `7.5 8.0+PTX 8.9 9.0 10.0 10.3 12.0 12.1+PTX`. > - CUDA 12.8 supports Compute Capabilities: `7.5 8.0+PTX 8.9 9.0 10.0+PTX 12.0+PTX`. > - CUDA 12.6/12.4 supports Compute Capabilities: `7.5 8.0+PTX 8.9 9.0+PTX`. | CUDA Version
(Variant) | vLLM | SGLang | VoxBox | |------------------------------|-------------------------------------------------------------------------------------|-------------------------------------------------------------------------|----------| | 12.9 | `0.16.0`, `0.15.1`,
`0.14.1`, `0.13.0`,
`0.12.0`, `0.11.2` | `0.5.9`, `0.5.8.post1`,
`0.5.7`, `0.5.6.post2` | | | 12.8 | `0.16.0`, `0.15.1`,
`0.14.1`, `0.13.0`,
`0.12.0`, `0.11.2`,
`0.10.2` | `0.5.9`, `0.5.8.post1`,
`0.5.7`, `0.5.6.post2`,
`0.5.5.post3` | `0.0.21` | | 12.6 | `0.15.1`, `0.14.1`,
`0.13.0`, `0.12.0`,
`0.11.2`, `0.10.2` | | `0.0.21` | ### Hygon DTK | DTK Version
(Variant) | vLLM | |-----------------------------|----------------------------| | 25.04 | `0.11.0`, `0.9.2`, `0.8.5` | ### T-Head HGGC | HGGC Version
(Variant) | vLLM | SGLang | |------------------------------|--------------------|------------------| | 12.3 | `0.12.0`, `0.11.1` | `0.5.6`, `0.5.5` | ### MetaX MACA | MACA Version
(Variant) | vLLM | SGLang | |------------------------------|----------|---------| | 3.3 | `0.11.2` | `0.5.6` | | 3.2 | `0.10.2` | | | 3.0 | `0.9.1` | | ### MThreads MUSA | MUSA Version
(Variant) | vLLM | SGLang | |------------------------------|---------|---------| | 4.3.2 | | `0.5.7` | | 4.1.0 | `0.9.2` | | ### AMD ROCm > [!NOTE] > - ROCm 7.0 supports LLVM targets: `gfx908 gfx90a gfx942 gfx950 gfx1030 gfx1100 gfx1101 gfx1200 gfx1201 gfx1150 gfx1151`. > - ROCm 6.4 supports LLVM targets: `gfx908 gfx90a gfx942 gfx1030 gfx1100`. > [!WARNING] > - ROCm 7.0 vLLM `0.11.2` are reusing the official ROCm 6.4 PyTorch 2.9 wheel package rather than a ROCm 7.0 specific PyTorch build. Although supports ROCm 7.0 in vLLM `0.11.2`, `gfx1150/gfx1151` are not supported yet. > - ROCm 6.4 vLLM `0.13.0` supports `gfx903 gfx90a gfx942` only. > - ROCm 6.4 SGLang supports `gfx942` only. > - ROCm 7.0 SGLang supports `gfx950` only. | ROCm Version
(Variant) | vLLM | SGLang | |------------------------------|-------------------------------------------------------------------------------------|-----------------------------------------------------------| | 7.0 | `0.16.0`, `0.15.1`,
`0.14.1`, `0.13.0`,
`0.12.0`, `0.11.2` | `0.5.9`, `0.5.8.post1`,
`0.5.7`, `0.5.6.post2` | | 6.4 | `0.16.0`, `0.15.1`,
`0.14.1`, `0.13.0`,
`0.12.0`, `0.11.2`,
`0.10.2` | `0.5.8.post1`, `0.5.7`,
`0.5.6.post2`, `0.5.5.post3` | ## Directory Structure The pack skeleton is organized by backend: ```text pack ├── {BACKEND 1} │ └── Dockerfile ├── {BACKEND 2} │ └── Dockerfile ├── {BACKEND 3} │ └── Dockerfile ├── ... │ └── Dockerfile └── {BACKEND N} └── Dockerfile ``` ## Dockerfile Convention Each Dockerfile follows these conventions: - Begin with comments describing the package logic in steps and usage of build arguments (`ARG`s). - Use `ARG` for all required and optional build arguments. If a required argument is unused, mark it as `(PLACEHOLDER)`. - Use heredoc syntax for `RUN` commands to improve readability. ### Example Dockerfile Structure ```dockerfile # Describe package logic and ARG usage. # ARG PYTHON_VERSION=... # REQUIRED ARG CMAKE_MAX_JOBS=... # REQUIRED ARG {OTHERS} # OPTIONAL ARG {BACKEND}_VERSION=... # REQUIRED ARG {BACKEND}_VERSION_EXTRA=... # OPTIONAL ARG {BACKEND}_ARCHS=... # REQUIRED ARG {BACKEND}_{OTHERS}=... # OPTIONAL ARG {SERVICE}_BASE_IMAGE=... # REQUIRED ARG {SERVICE}_VERSION=... # REQUIRED ARG {SERVICE}_{OTHERS}=... # OPTIONAL ARG {SERVICE}_{FRAMEWORK}_VERSION=... # REQUIRED ARG {SERVICE}_{FRAMEWORK}_{OTHERS}=... # OPTIONAL # Stage Bake Runtime FROM {BACKEND DEVEL IMAGE} AS runtime SHELL ["/bin/bash", "-eo", "pipefail", "-c"] ARG TARGETPLATFORM ARG TARGETOS ARG TARGETARCH ARG ... RUN <