Submitted by TrueBlueDreamin t3_zvzix2 in MachineLearning

Hey folks,

I built TuneMyAI to make it incredibly simple for developers to finetune and deploy Stable Diffusion models to production so they can focus on building great products.

As an app developer myself, I spent a while trying to figure out how to go beyond local GPUs and notebooks and setup our own infra using Kubernetes. In summary, we wanted to make it really simple for anyone to build applications on top of Stable Diffusion without worrying about all the MLOps overhead.

Our API allows you to finetune your Stable Diffusion models for your specific data sets. We handle everything from storage, finetuning, model deployment & inference and integrate with HuggingFace as well. We're working on a bunch of new features including hosted WebUIs, support for additional models like Whisper and more.

Would love for y'all to check us out and share any feedback. You can learn more on ProductHunt.

Thanks & Happy Holidays!

179

Comments

You must log in or register to comment.

monkecoder t1_j1s9tw4 wrote

How much do you guys charge for finetuning? Also how long does it take and what GPUs are you training on?

12

TrueBlueDreamin OP t1_j1sb5mf wrote

Finetuning takes ~15 mins at 1500 steps on about 10-15 images. We train on A100s and it costs about $2.50 per finetune. We're running a holiday promo though right now where you can finetune for $0.99.

Excited for you to try it out! Let me know if you have any feedback :)

19

Bodesterine555 t1_j1slqq5 wrote

Great to hear! How does the quality compare to Astria? I’ve found that Replicate’s API, for example, is significantly lower quality that Astria’s.

6

xKraazY t1_j1ss7op wrote

Interesting. 2.50 for 1500 is quite expensive though. Astria is about 3$ bulk or 5$ one off with about 100 free images. Do you guys add any “secret” sauce for fine tuning models?

14

TrueBlueDreamin OP t1_j1sswpp wrote

So just to compare we do $2.50 for one-off fine tunes + inference for 100 images costs 0.5c so call it $2.51.

By the way, you can choose how many steps you want to train for, would still come around to $2.50 as long as it's within reason (eg: <10k steps).

For bulk pricing + volume, we can definitely do better than $3, I mean our one-off is cheaper than that. DM me with details on your workloads and we can figure out bulk pricing that beats any other vendor.

8

manganime1 t1_j1tc27q wrote

Interesting! Will definitely look into this if I develop any apps in the future.

Are you using AWS for training, or others?

4

_rundown_ t1_j1th53n wrote

I’m interested to beta test. Will you all be supporting LLMs as well? I hear stability has something cooking for those, very interested in that space and running some tests with GPT2 right now.

3

somethingclassy t1_j1tna1y wrote

Great concept. What version(s) of Stable Diffusion does it work with?

3

dalcowboiz t1_j1to4o5 wrote

Does it work with sparse models from the sparse zoo?

1

JClub t1_j1u36tt wrote

Another guy making money out of dreambooth training when you can do it for free on Google Colab...

10

Illustrious_Row_9971 t1_j1uxh50 wrote

Since you are using huggingface for hosting models, you can also create a organization with a webui there and embed it in your website

for example this is stabiity ai org: https://huggingface.co/stabilityai

this is a web demo: https://huggingface.co/spaces/stabilityai/stable-diffusion

you can duplicate and embed it:

&lt;script
type="module"
src="https://gradio.s3-us-west-2.amazonaws.com/3.11.0/gradio.js"
\&gt;&lt;/script&gt;

&lt;gradio-app src="https://stabilityai-stable-diffusion.hf.space"&gt;&lt;/gradio-app&gt;

1

QuerulousPanda t1_j1uzydb wrote

This may be a dumb question, but what products is stable diffusion useful for? It's an awesome technology but does it have any kind of widespread applicability?

1

Evoke_App t1_j1v9hoi wrote

Yes. Just look at all the AI art apps popping up. Logo makers too. The AI avatar apps killing it like AI Avatars or Lensa.

We'll have a stable diffusion and dreambooth API up soon on our site as well. So I hope it makes SD app dev's lives that much easier

1

TrueBlueDreamin OP t1_j1van9a wrote

Yes, will be adding support for additional models starting with Whisper, GPT2 and Stability's LLM once it comes out :) Let me know if there's any custom models you'd like hosted and we'll get that up and running. Eventually will be able to do that self serve.

5

TrueBlueDreamin OP t1_j1vb0ft wrote

There's a ton of Colabs out there for folks to use, keep in mind you're still paying Google however much a month for compute credits. I would recommend Colab for folks looking to experiment in a notebook. What we're offering is an API ready for production grade apps.

Hope you try us out :)

2

_rundown_ t1_j1vyb2x wrote

Looking into GPT-NeoX, seems like the right direction if you're looking at supporting Stability's LLM (Stability is providing EleutherAI cloud services, I'd guess the forthcoming LLM will be a collaboration).

Anyway we can get that up and running? Would love to test that on your API.

3

mac4281 t1_j1w1bd4 wrote

This is a great response. I feel like I’m one of many who have no interest in experimenting with a colab. I just want an api that I can start building on quickly. I’ll head over and check it out later today!

2

xKraazY t1_j1waw4m wrote

I was just wondering if you guys do anything different while tuning. I would use your service if tunes yielded similiar result to astria. Right now I have my own infra setup but can't get it fine tuned quite like it.

2

xKraazY t1_j1xxsb7 wrote

Just bought a credit and trying it out :) Btw do you use the huggingface diffusers repo? Or another one? (Also running the api, and how what should the file array include? Path to images or public image url?)

1

JanssonsFrestelse t1_j1y2npq wrote

Also can you supply your own regularization images, or do you have some selection (w. recommendations for e.g. fine-tuning on a person) to choose from? Training the text-encoder as well I assume? What about jointly learning different concepts when fine-tuning on an object/person?

1

TrueBlueDreamin OP t1_j1y4zjh wrote

We can support regularization with your own class images if you'd want, however it's recommended to use model generated regularization images for prior preservation. You don't want to introduce bias into the model with curated images.

We train the text encoder as well, correct.

You should be able to train multiple concepts/subjects although there is an unsolved problem regarding bleeding when used in the same prompt. Shoot me a DM and we can probably figure something out!

1

Impressive_Maximum32 t1_j1ybvzl wrote

Sounds great! I d be happy to switch over from astria, but i am missing two things:
- Astria has a "super_resolution" and a "face_correct" flag that work pretty good

2

JanssonsFrestelse t1_j1ydl3c wrote

Curated images would be generated by the model being trained using the same prompt for reg images as for the subject training images (found via clip interrogation, swapping out e. g. "a woman" to my subject's token). Not a big deal though, if you can train the 768x768 model I'll try it out. Can't run it locally and colabs for the 768 model have been unreliable. Might write my own later on if the model trained by you shows good quality.

Edit: probably not much use having the exact same prompt, but I'm thinking something similar to the clip classification of the image(s) + the general style/concept you want to learn. Or do you see some issues with the method I've described?

1

JClub t1_j1yp8uf wrote

I understand your point. My point is that these guys are making you pay for it when this API that you want could definitely be for free.

Without open source they would not be able to make you pay for any of this, so using open source tools to make paid APIs just doesn't go along with me.

1

Unreal_777 t1_j1z1gf3 wrote

Hello u/TrueBlueDreamin,

Would this website (of yours) allow a website (that I created) to let users send their images to my website, then my website would send them to your website and generate a training , then my website would receive back the new model? then produce some images to the end user? If yes, then everytime that happends you get payed right? (I am still new)

1

Paid-Not-Payed-Bot t1_j1z1h6r wrote

> you get paid right? (I

FTFY.

Although payed exists (the reason why autocorrection didn't help you), it is only correct in:

  • Nautical context, when it means to paint a surface, or to cover with something like tar or resin in order to make it waterproof or corrosion-resistant. The deck is yet to be payed.

  • Payed out when letting strings, cables or ropes out, by slacking them. The rope is payed out! You can pull now.

Unfortunately, I was unable to find nautical or rope-related words in your comment.

Beep, boop, I'm a bot

1

TrueBlueDreamin OP t1_j206nhy wrote

Hey Unreal, yes you can use our API to create your own app/website that does exactly that. You can choose to export the model after it's done finetuning and run inference (creating images) yourself if you'd like, we do also have an option for you to run inference through an API. Shoot me a DM and I can help you get setup

2

mxby7e t1_j20q0hg wrote

Does your api support the ability to train a style instead of a subject?

1

TrueBlueDreamin OP t1_j210gk8 wrote

Ah, sorry about that. Our docs are currently a little hard to parse, reworking them soon.

If you go to the inference params, theres fields called facetool and facetool_strength for face correction.

Documenting the upscaling endpoint as we speak :) New docs should be live EOD tomorrow. In the meantime, feel free to hop in the Discord to stay posted on when it's live.

2

TrueBlueDreamin OP t1_j210wo4 wrote

It's currently optimized for tuning on subjects but we're launching finetuning for styles in the coming week since a lot of folks have asked for it. Feel free to hop in the Discord to be notified when we launch it, in any case, I'll circle back here once it's live

2

mxby7e t1_j215ewa wrote

Looking forward to it! I’ve been doing style training on the lastBen colab but I have a lot of artists that want a more accessible way to build and use models based on their own styles

2