Submitted by OkOkPlayer t3_zdfrnw in MachineLearning
I have a fine-tuned Stable Diffusion Model and would like to host it to make it publicly available. Both options (GPU, CPU) seem to be problematic.
- I can't find a "cheap" GPU hosting platform. AWS etc. are all > 200$ per month and they have no serverless option (only found banana.dev which seems to have relatively limited flexibility)
- CPU seems to be too slow for inference
I am currently running the model on my notebook CPU with 35s/it which is way too slow. Is it possible to host a Stable Diffusion on CPU with close to real-time responses (< 60s for ~100 inference steps) or is there a "cheap" GPU hosting platform I couldn't find yet?
MarkusDeNeutoy t1_iz1dpfj wrote
You may like https://replicate.com/ .