# TensorFlowASR **Repository Path**: oo/TensorFlowASR ## Basic Information - **Project Name**: TensorFlowASR - **Description**: No description available - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2026-01-23 - **Last Updated**: 2026-01-23 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README

TensorFlowASR :zap:

GitHub python tensorflow PyPI

Almost State-of-the-art Automatic Speech Recognition in Tensorflow 2

TensorFlowASR implements some automatic speech recognition architectures such as DeepSpeech2, Jasper, RNN Transducer, ContextNet, Conformer, etc. These models can be converted to TFLite to reduce memory and computation for deployment :smile:

## What's New? ## Table of Contents - [What's New?](#whats-new) - [Table of Contents](#table-of-contents) - [:yum: Supported Models](#yum-supported-models) - [Baselines](#baselines) - [Publications](#publications) - [Installation](#installation) - [Training \& Testing Tutorial](#training--testing-tutorial) - [Features Extraction](#features-extraction) - [Augmentations](#augmentations) - [TFLite Convertion](#tflite-convertion) - [Pretrained Models](#pretrained-models) - [Corpus Sources](#corpus-sources) - [English](#english) - [Vietnamese](#vietnamese) - [How to contribute](#how-to-contribute) - [References \& Credits](#references--credits) - [Contact](#contact) ## :yum: Supported Models ### Baselines - **Transducer Models** (End2end models using RNNT Loss for training, currently supported Conformer, ContextNet, Streaming Transducer) - **CTCModel** (End2end models using CTC Loss for training, currently supported DeepSpeech2, Jasper) ### Publications - **Conformer Transducer** (Reference: [https://arxiv.org/abs/2005.08100](https://arxiv.org/abs/2005.08100)) See [examples/models/transducer/conformer](./examples/models/transducer/conformer) - **Streaming Conformer** (Reference: [http://arxiv.org/abs/2010.11395](http://arxiv.org/abs/2010.11395)) See [examples/models/transducer/conformer](./examples/models/transducer/conformer) - **ContextNet** (Reference: [http://arxiv.org/abs/2005.03191](http://arxiv.org/abs/2005.03191)) See [examples/models/transducer/contextnet](./examples/models/transducer/contextnet) - **RNN Transducer** (Reference: [https://arxiv.org/abs/1811.06621](https://arxiv.org/abs/1811.06621)) See [examples/models/transducer/rnnt](./examples/models/transducer/rnnt) - **Deep Speech 2** (Reference: [https://arxiv.org/abs/1512.02595](https://arxiv.org/abs/1512.02595)) See [examples/models/ctc/deepspeech2](./examples/models/ctc/deepspeech2) - **Jasper** (Reference: [https://arxiv.org/abs/1904.03288](https://arxiv.org/abs/1904.03288)) See [examples/models/ctc/jasper](./examples/models/ctc/jasper) ## Installation For training and testing, you should use `git clone` for installing necessary packages from other authors (`ctc_decoders`, `rnnt_loss`, etc.) **NOTE ONLY FOR APPLE SILICON**: TensorFlowASR requires python >= 3.12 See the `requirements.[extra].txt` files for extra dependencies ```bash git clone https://github.com/TensorSpeech/TensorFlowASR.git cd TensorFlowASR ./setup.sh [apple|tpu|gpu] [dev] ``` **Running in a container** ```bash docker-compose up -d ``` ## Training & Testing Tutorial - For training, please read [tutorial_training](./docs/tutorials/training.md) - For testing, please read [tutorial_testing](./docs/tutorials/testing.md) **FYI**: Keras builtin training uses **infinite dataset**, which avoids the potential last partial batch. See [examples](./examples/) for some predefined ASR models and results ## Features Extraction See [features_extraction](./tensorflow_asr/features/README.md) ## Augmentations See [augmentations](./tensorflow_asr/augmentations/README.md) ## TFLite Convertion After converting to tflite, the tflite model is like a function that transforms directly from an **audio signal** to **text and tokens** See [tflite_convertion](./docs/tutorials/tflite.md) ## Pretrained Models See the results on each example folder, e.g. [./examples/models//transducer/conformer/results/sentencepiece/README.md](./examples/models//transducer/conformer/results/sentencepiece/README.md) ## Corpus Sources ### English | **Name** | **Source** | **Hours** | | :----------- | :----------------------------------------------------------------- | :-------- | | LibriSpeech | [LibriSpeech](http://www.openslr.org/12) | 970h | | Common Voice | [https://commonvoice.mozilla.org](https://commonvoice.mozilla.org) | 1932h | ### Vietnamese | **Name** | **Source** | **Hours** | | :------------------------------------- | :------------------------------------------------------------------------------------------------------------------- | :-------- | | Vivos | [https://ailab.hcmus.edu.vn/vivos](https://www.kaggle.com/datasets/kynthesis/vivos-vietnamese-speech-corpus-for-asr) | 15h | | InfoRe Technology 1 | [InfoRe1 (passwd: BroughtToYouByInfoRe)](https://files.huylenguyen.com/datasets/infore/25hours.zip) | 25h | | InfoRe Technology 2 (used in VLSP2019) | [InfoRe2 (passwd: BroughtToYouByInfoRe)](https://files.huylenguyen.com/datasets/infore/audiobooks.zip) | 415h | | VietBud500 | [https://huggingface.co/datasets/linhtran92/viet_bud500](https://huggingface.co/datasets/linhtran92/viet_bud500) | 500h | ## How to contribute 1. Fork the project 2. [Install for development](#installing-for-development) 3. Create a branch 4. Make a pull request to this repo ## References & Credits 1. [NVIDIA OpenSeq2Seq Toolkit](https://github.com/NVIDIA/OpenSeq2Seq) 2. [https://github.com/noahchalifour/warp-transducer](https://github.com/noahchalifour/warp-transducer) 3. [Sequence Transduction with Recurrent Neural Network](https://arxiv.org/abs/1211.3711) 4. [End-to-End Speech Processing Toolkit in PyTorch](https://github.com/espnet/espnet) 5. [https://github.com/iankur/ContextNet](https://github.com/iankur/ContextNet) ## Contact Huy Le Nguyen Email: nlhuy.cs.16@gmail.com