riffusion-inference/README.md

57 lines
1.4 KiB
Markdown
Raw Normal View History

2022-11-25 14:20:30 -07:00
# Riffusion Inference Server
2022-11-25 17:30:11 -07:00
Model inference backend for the Riffusion app.
* a diffusers pipeline that performs prompt interpolation combined with image conditioning
* a module for (approximately) converting between spectrograms and waveforms
* a flask server to provide model inference via API to the next.js app
## Install
Tested with Python 3.9 and diffusers 0.9.0
```
conda create --name riffusion-inference python=3.9
conda activate riffusion-inference
python -m pip install -r requirements.txt
```
## Run
Start the Flask server:
```
python -m riffusion.server --port 3013 --host 127.0.0.1 --checkpoint /path/to/diffusers_checkpoint
```
The model endpoint is now available at `http://127.0.0.1:3013/run_inference` via POST request.
Example input (see [InferenceInput](https://github.com/hmartiro/riffusion-inference/blob/main/riffusion/datatypes.py#L28) for the API):
```
{
alpha: 0.75,
num_inference_steps: 50,
seed_image_id: 0,
start: {
prompt: "church bells on sunday",
seed: 42,
denoising: 0.75,
guidance: 7.0,
},
end: {
prompt: "jazz with piano",
seed: 123,
denoising: 0.75,
guidance: 7.0,
},
}
```
Example output (see [InferenceOutput](https://github.com/hmartiro/riffusion-inference/blob/main/riffusion/datatypes.py#L54) for the API):
```
{
image: "< base64 encoded PNG >",
audio: "< base64 encoded MP3 clip >",,
}
```