Skip to content

asai95/speech-recognition-api

Repository files navigation

PyPI - Version PyPI - Python Version PyPI - License PyPI - Downloads codecov

Speech Recognition API

Simple but extensible API for Speech Recognition.

Installation

From pip:

pip install speech_recognition_api[all]

From git:

git clone https://github.com/asai95/speech-recognition-api.git
pip install -r requirements.txt

Usage

Simple dev server:

python -m speech_recognition_api

Gunicorn:

gunicorn "speech_recognition_api:create_app()" -k uvicorn.workers.UvicornWorker -w 1 -b 127.0.0.1:8888

Celery worker:

celery -A speech_recognition_api.extra.celery_bus worker

Huey worker:

huey_consumer speech_recognition_api.extra.huey_bus.huey

Description

This project is aimed to simplify building and deploying of applications that require Speech Recognition functionality.

It is designed to work as a microservice, so it does not handle stuff like auth and rate limits.

However, it is also designed to be extensible in 3 major areas:

  • Models
  • File Storages
  • Message Busses

There are two types of APIs available.

Synchronous API

This API is designed for simple workloads, where the machine that runs the server is capable of running a model. You probably want to limit the payload size for these routes.

Routes:

POST /sync/v1/transcribe

Accepts an audio file. File type depends on the model that is being used.

Returns an object with transcription. Response model.

Asynchronous API

This API is designed to process files asynchronously, i.e. to create tasks and process them on separate workers. Typical client flow here is as follows:

  • Create a task and receive task id
  • Use this task id to periodically check if it is completed.

Routes:

POST /async/v1/transcribe

Accepts an audio file. File type depends on the model that is being used.

Returns an object with async task id. Response model.

GET /async/v1/transcribe/{task_id}

Returns an object with status and a transcription (if transcription is available). Response model.

Async API also requires a worker to run the actual work.

Configuring

Configuration is done by .env file or env variables (they take preference).

The main variables required for the API and worker to run are:

  • MODEL - model class path (it will do the actual audio-to-text conversion)
  • STORAGE - storage class path (in Async API it will be responsible for uploading/downloading files)
  • MESSAGE_BUS - message bus class path (in Async API it will be responsible for sending tasks to remoted workers and getting the result back from them)

These classes will be imported only when used for the fist time.

Each class may require its own variables. Please refer to config.py of the specific module to get the config reference.

Built-in classes:

Models:

Storages:

Message Busses:

Extending

It is easy to extend the API by adding models, storages and message busses.

To do that, one can just create a class that implements an interface:

Then just add a path to the class to the config file and that's it!

I suggest to distribute new modules through PyPI, so other people could reuse them.

About

Simple but extensible API for Speech Recognition.

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages