# hello-genai **Repository Path**: mirrors_docker/hello-genai ## Basic Information - **Project Name**: hello-genai - **Description**: Very simple GenAI application to try the Docker Model Runner - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2025-03-29 - **Last Updated**: 2025-12-27 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # hello-genai A simple chatbot web application built in Go, Python and Node.js that connects to a local LLM service (llama.cpp) to provide AI-powered responses. ## Environment Variables The application uses the following environment variables defined in the `.env` file: - `LLM_BASE_URL`: The base URL of the LLM API - `LLM_MODEL_NAME`: The model name to use To change these settings, simply edit the `.env` file in the root directory of the project. ## Quick Start 1. Clone the repository: ```bash git clone https://github.com/docker/hello-genai cd hello-genai ``` 2. Start the application using Docker Compose: ```bash docker compose up ``` 3. Open your browser and visit the following links: http://localhost:8080 for the GenAI Application in Go http://localhost:8081 for the GenAI Application in Python http://localhost:8082 for the GenAI Application in Node http://localhost:8083 for the GenAI Application in Rust ## Requirements - macOS (recent version) - Either: - Docker and Docker Compose (preferred) - Go 1.21 or later - Local LLM server If you're using a different LLM server configuration, you may need to modify the`.env` file.