I am happy to share the project that I have done with @Sayak_Paul. In this project, our main focus is to serve Stable Diffusion in TF Serving (V1 and V2 as well). To this end, we made three
decoder with the pre/post operations included.
In the project repository, you can find out the following information:
- how each
- Docker images of TF Serving for each
- how to deploy each TF Serving on GKE(Google Kubernetes Engine)
- how to run inference by interacting with three TF Serving
Additionally, you could learn some more information about different target environments:
- Hugging Face Inference Endpoint
- FastAPI on GKE
I hope this project is helpful for some of you!