# crabml **Repository Path**: kinouta/crabml ## Basic Information - **Project Name**: crabml - **Description**: 中专github的项目,学习AI - **Primary Language**: Rust - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2024-01-29 - **Last Updated**: 2024-02-24 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # crabml crabml is an ongoing experiment that aims to reimplement GGML using Rust. Currently it can inference a 3B Q8_0 quantized Llama model at a dog slow speed. Its design goals are: - focus on inference only. - limit tensor operators to the bare minimum required for LLM inference. - fast enough inferencing on cheap hardwares. - `mmap()` from day one. - prioritize SIMD ahead of GPU. ## Build ``` RUSTFLAGS="-C target-feature=+neon" cargo build --release ./target/release/crabml-cli -m ./testdata/open-llama-3b-q8_0.gguf "captain america" --steps 100 -t 0.8 -p 1.0 ```