# useful-transformers **Repository Path**: ssyyw/useful-transformers ## Basic Information - **Project Name**: useful-transformers - **Description**: 镜像 https://github.com/usefulsensors/useful-transformers.git 用于本地加速 - **Primary Language**: Unknown - **License**: GPL-3.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2026-03-22 - **Last Updated**: 2026-03-22 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # Useful Transformers Useful Transformers is a library for efficient inference of Transformer models. The focus is on low cost, low energy processors to run inference at the edge. The initial implementation is aimed at running OpenAI's [Whisper](https://github.com/openai/whisper) speech-to-text model efficiently on the [RK3588](https://www.rock-chips.com/a/en/products/RK35_Series/2022/0926/1660.html) processors' based single-board computers. The tiny.en Whisper model runs transcribes speech at 30x real-time speeds, and 2x better than best [known](https://github.com/guillaumekln/faster-whisper) implementation. ## Getting started The easiest way to try out Whisper transcription is to install the [release](https://github.com/usefulsensors/useful-transformers/releases/download/0.1_rk3588/useful_transformers-0.1-cp310-cp310-linux_aarch64.whl) wheel package. # Preferably inside a virtual environment $ python -m pip install https://github.com/usefulsensors/useful-transformers/releases/download/0.1_rk3588/useful_transformers-0.1-cp310-cp310-linux_aarch64.whl Try transcribing a wav file. $ taskset -c 4-7 python -m useful_transformers.transcribe_wav If you don't have a wav file handy, running the above command will transcribe an example provided in the package. $ taskset -c 4-7 python -m useful_transformers.transcribe_wav Ever tried, ever failed. No matter, try again. Fail again. Fail better. ## Performance ![Performance comparison](https://github.com/usefulsensors/useful-transformers/blob/main/examples/whisper/assets/perf-comparison.png) The plot shows `useful-transformers` Whisper `tiny.en` model's inference times across the examples with varying durations. `useful-transformer` is 2x faster than `faster-whisper`'s int8 implementation. `useful-transformer` uses FP16 matrix multiplication on the NPU available in the RK3588 processor. The majority of benefit comes from the large matrix multiplications (of sizes `1500x384x384` for the tiny.en model) in the encoder. ## TODO - [x] Whisper tiny.en - [x] Whisper base.en - [ ] Larger Whisper models - [ ] Use int8 matmuls from the librknnrt - [ ] Use int4 matmuls (request Rockhip for int4 matmul kernels) - [ ] Use asynchronous kernel launches (request Rockchip for better APIs in general) - [ ] Decode with timestamps ## Contributors * Nat Jeffries (@njeffrie) * Manjunath Kudlur (@keveman) * Guy Nicholson (@guynich) * James Wang (@JamesUseful) * Pete Warden (@petewarden) * Ali Zartash (@aliz64)