ASR Inference Service for AI App Store
To build the docker container, run
make build
To push the image to a registry, first build the image, then run
docker tag asr-inference-service:1.0.0 <REGISTRY>/<REPO>/asr-inference-service:1.0.0
If not logged in to the registry, run
docker login -u <USERNAME> -p <PASSWORD> <REGISTRY>
Then, push the tagged image to a registry
docker push <REGISTRY>/<REPO>/asr-inference-service:1.0.0
To run the Gradio application locally, run the following
make dev
First, make sure your image is pushed to the registry.
Check out the AI App Store documentation for full details, but in general:
- Create/edit a model card
- Pass the docker image URI (e.g
<REGISTRY>/<REPO>/asr-inference-service:1.0.0
) when creating/editing the inference service
There are other potential deployment options, including:
- Google Cloud Run
- AWS Fargate
- Red Hat Openshift Serverless