# model-cli **Repository Path**: mirrors_docker/model-cli ## Basic Information - **Project Name**: model-cli - **Description**: Docker Model Runner CLI Plugin - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2025-03-29 - **Last Updated**: 2026-01-03 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # Docker Model CLI ## ⚠️ This Repository Has Moved This repository has been consolidated into [model-runner](https://github.com/docker/model-runner/tree/main/cmd/cli). All future development, issues, and pull requests should be directed there. Please visit the new repository for the latest updates and to contribute. --- A powerful command-line interface for managing, running, packaging, and deploying AI/ML models using Docker. This CLI lets you install and control the Docker Model Runner, interact with models, manage model artifacts, and integrate with OpenAI and other backends—all from your terminal. ## Features - **Install Model Runner**: Easily set up the Docker Model Runner for local or cloud environments with GPU support. - **Run Models**: Execute models with prompts or in interactive chat mode, supporting multiline input and OpenAI-style backends. - **List Models**: View all models available locally or via OpenAI, with options for JSON and quiet output. - **Package Models**: Convert GGUF files into Docker model OCI artifacts and push them to registries, including license and context size options. - **Configure Models**: Set runtime flags and context sizes for models. - **Logs & Status**: Stream logs and check the status of the Model Runner and individual models. - **Tag, Pull, Push, Remove, Unload**: Full lifecycle management for model artifacts. - **Compose & Desktop Integration**: Advanced orchestration and desktop support for model backends. ## Building 1. **Clone the repo:** ```bash git clone https://github.com/docker/model-cli.git cd model-cli ``` 2. **Build the CLI:** ```bash make build ``` 3. **Install Model Runner:** ```bash ./model install-runner ``` Use `--gpu cuda` for GPU support, or `--gpu auto` for automatic detection. ## Usage Run `./model --help` to see all commands and options. ### Common Commands - `model install-runner` — Install the Docker Model Runner - `model run MODEL [PROMPT]` — Run a model with a prompt or enter chat mode - `model list` — List available models - `model package --gguf --push ` — Package and push a model - `model logs` — View logs - `model status` — Check runner status - `model configure MODEL [flags]` — Configure model runtime - `model unload MODEL` — Unload a model - `model tag SOURCE TARGET` — Tag a model - `model pull MODEL` — Pull a model - `model push MODEL` — Push a model - `model rm MODEL` — Remove a model ## Example: Interactive Chat ```bash ./model run llama.cpp "What is the capital of France?" ``` Or enter chat mode: ```bash ./model run llama.cpp Interactive chat mode started. Type '/bye' to exit. > """ Tell me a joke. """ ``` ## Advanced - **Packaging:** Add licenses and set context size when packaging models for distribution. ## Development - **Run unit tests:** ```bash make unit-tests ``` - **Generate docs:** ```bash make docs ``` ## License [Apache 2.0](LICENSE)