Submitted by TrueBlueDreamin t3_zvzix2 in MachineLearning

Hey folks,

I built TuneMyAI to make it incredibly simple for developers to finetune and deploy Stable Diffusion models to production so they can focus on building great products.

As an app developer myself, I spent a while trying to figure out how to go beyond local GPUs and notebooks and setup our own infra using Kubernetes. In summary, we wanted to make it really simple for anyone to build applications on top of Stable Diffusion without worrying about all the MLOps overhead.

Our API allows you to finetune your Stable Diffusion models for your specific data sets. We handle everything from storage, finetuning, model deployment & inference and integrate with HuggingFace as well. We're working on a bunch of new features including hosted WebUIs, support for additional models like Whisper and more.

Would love for y'all to check us out and share any feedback. You can learn more on ProductHunt.

Thanks & Happy Holidays!

179

Comments

You must log in or register to comment.

xKraazY t1_j1ss7op wrote

Interesting. 2.50 for 1500 is quite expensive though. Astria is about 3$ bulk or 5$ one off with about 100 free images. Do you guys add any “secret” sauce for fine tuning models?

14

TrueBlueDreamin OP t1_j1sswpp wrote

So just to compare we do $2.50 for one-off fine tunes + inference for 100 images costs 0.5c so call it $2.51.

By the way, you can choose how many steps you want to train for, would still come around to $2.50 as long as it's within reason (eg: <10k steps).

For bulk pricing + volume, we can definitely do better than $3, I mean our one-off is cheaper than that. DM me with details on your workloads and we can figure out bulk pricing that beats any other vendor.

8

xKraazY t1_j1t5rt3 wrote

Do we get to configure the specific fine tuning?

2

TrueBlueDreamin OP t1_j1t7f04 wrote

Yes you do, what parameters you thinking specifically?

5

xKraazY t1_j1waw4m wrote

I was just wondering if you guys do anything different while tuning. I would use your service if tunes yielded similiar result to astria. Right now I have my own infra setup but can't get it fine tuned quite like it.

2

TrueBlueDreamin OP t1_j1wbo3j wrote

We do indeed have a few optimizations in place to make sure renders are high quality. Take a look at some of our renders

2

xKraazY t1_j1xxsb7 wrote

Just bought a credit and trying it out :) Btw do you use the huggingface diffusers repo? Or another one? (Also running the api, and how what should the file array include? Path to images or public image url?)

1

monkecoder t1_j1s9tw4 wrote

How much do you guys charge for finetuning? Also how long does it take and what GPUs are you training on?

12

TrueBlueDreamin OP t1_j1sb5mf wrote

Finetuning takes ~15 mins at 1500 steps on about 10-15 images. We train on A100s and it costs about $2.50 per finetune. We're running a holiday promo though right now where you can finetune for $0.99.

Excited for you to try it out! Let me know if you have any feedback :)

19

3deal t1_j1thrua wrote

Can we download the model ?

12

rowleboat t1_j1sbj15 wrote

How much does inference cost? Does that run on A100s too?

10

TrueBlueDreamin OP t1_j1sdqat wrote

Yeap, you can run inference on upto 100 images at a time for 0.5c

7

JClub t1_j1u36tt wrote

Another guy making money out of dreambooth training when you can do it for free on Google Colab...

10

the_magic_gardener t1_j1ue4x5 wrote

Is it possible to do dreambooth with less than 16 GB of GPU? I figured you couldn't do attention slicing for fine tuning the way you can for just running the model.

3

hopbel t1_j1yn3y6 wrote

You can do full finetuning with 12gb already (kohya_ss)

1

TrueBlueDreamin OP t1_j1vb0ft wrote

There's a ton of Colabs out there for folks to use, keep in mind you're still paying Google however much a month for compute credits. I would recommend Colab for folks looking to experiment in a notebook. What we're offering is an API ready for production grade apps.

Hope you try us out :)

2

mac4281 t1_j1w1bd4 wrote

This is a great response. I feel like I’m one of many who have no interest in experimenting with a colab. I just want an api that I can start building on quickly. I’ll head over and check it out later today!

2

TrueBlueDreamin OP t1_j1w5t5a wrote

Excited for you to give us a try, feel free to hop in the Discord too. I'm always around to answer Q's!

3

JClub t1_j1yp8uf wrote

I understand your point. My point is that these guys are making you pay for it when this API that you want could definitely be for free.

Without open source they would not be able to make you pay for any of this, so using open source tools to make paid APIs just doesn't go along with me.

1

Bodesterine555 t1_j1slqq5 wrote

Great to hear! How does the quality compare to Astria? I’ve found that Replicate’s API, for example, is significantly lower quality that Astria’s.

6

TrueBlueDreamin OP t1_j1ssh25 wrote

We can benchmark it, our renders are pretty high quality. I'll pick a celeb and train it against both and post results sometime tomorrow.

9

Bodesterine555 t1_j1t06xc wrote

Thanks, this would be hugely helpful! I’m actually trying to make an app using an API, I’d love to Ty this out!

2

TrueBlueDreamin OP t1_j1subhb wrote

In the meantime, here's an example of some of the renders. This ofcourse is highly dependent on the prompts + inference config (samplers, steps, seed etc...)

7

mudman13 t1_j1tru8u wrote

No Emma Watson or Taylor Swift. I declare it a failure.

5

mearco t1_j1u49f5 wrote

Bottom right?

8

mudman13 t1_j1u9s9a wrote

If it is then I stand corrected.....said the man in the orthopedic Shoes

3

manganime1 t1_j1tc27q wrote

Interesting! Will definitely look into this if I develop any apps in the future.

Are you using AWS for training, or others?

4

TrueBlueDreamin OP t1_j1tdmxw wrote

Excited to have you try us out, let me know if I can answer any questions or help you get started. We don't use AWS, we've got our own GPUs :)

6

_rundown_ t1_j1th53n wrote

I’m interested to beta test. Will you all be supporting LLMs as well? I hear stability has something cooking for those, very interested in that space and running some tests with GPT2 right now.

3

TrueBlueDreamin OP t1_j1van9a wrote

Yes, will be adding support for additional models starting with Whisper, GPT2 and Stability's LLM once it comes out :) Let me know if there's any custom models you'd like hosted and we'll get that up and running. Eventually will be able to do that self serve.

5

_rundown_ t1_j1vyb2x wrote

Looking into GPT-NeoX, seems like the right direction if you're looking at supporting Stability's LLM (Stability is providing EleutherAI cloud services, I'd guess the forthcoming LLM will be a collaboration).

Anyway we can get that up and running? Would love to test that on your API.

3

somethingclassy t1_j1tna1y wrote

Great concept. What version(s) of Stable Diffusion does it work with?

3

TrueBlueDreamin OP t1_j1vai7n wrote

You can finetune on all versions, 1.4, 1.5 and the newer SD 2.x models :)

4

JanssonsFrestelse t1_j1y2df0 wrote

Fine-tuning the SD 2.1 768x768 resolution model as well or just the 2.1-base 512x512 model?

1

JanssonsFrestelse t1_j1y2npq wrote

Also can you supply your own regularization images, or do you have some selection (w. recommendations for e.g. fine-tuning on a person) to choose from? Training the text-encoder as well I assume? What about jointly learning different concepts when fine-tuning on an object/person?

1

TrueBlueDreamin OP t1_j1y4zjh wrote

We can support regularization with your own class images if you'd want, however it's recommended to use model generated regularization images for prior preservation. You don't want to introduce bias into the model with curated images.

We train the text encoder as well, correct.

You should be able to train multiple concepts/subjects although there is an unsolved problem regarding bleeding when used in the same prompt. Shoot me a DM and we can probably figure something out!

1

JanssonsFrestelse t1_j1ydl3c wrote

Curated images would be generated by the model being trained using the same prompt for reg images as for the subject training images (found via clip interrogation, swapping out e. g. "a woman" to my subject's token). Not a big deal though, if you can train the 768x768 model I'll try it out. Can't run it locally and colabs for the 768 model have been unreliable. Might write my own later on if the model trained by you shows good quality.

Edit: probably not much use having the exact same prompt, but I'm thinking something similar to the clip classification of the image(s) + the general style/concept you want to learn. Or do you see some issues with the method I've described?

1

Cultural_Contract512 t1_j1ssm92 wrote

No UI, just api calls, is that correct?

2

TrueBlueDreamin OP t1_j1st61c wrote

Hosted WebUI's is on the roadmap and will be launching soon! We have a few users beta testing it right now. Let me know if you'd like to join the beta.

5

pablonaj t1_j1to3cs wrote

I'd be interested in trying the beta!

1

Illustrious_Row_9971 t1_j1uxh50 wrote

Since you are using huggingface for hosting models, you can also create a organization with a webui there and embed it in your website

for example this is stabiity ai org: https://huggingface.co/stabilityai

this is a web demo: https://huggingface.co/spaces/stabilityai/stable-diffusion

you can duplicate and embed it:

&lt;script
type="module"
src="https://gradio.s3-us-west-2.amazonaws.com/3.11.0/gradio.js"
\&gt;&lt;/script&gt;

&lt;gradio-app src="https://stabilityai-stable-diffusion.hf.space"&gt;&lt;/gradio-app&gt;

1

TrueBlueDreamin OP t1_j1w99bh wrote

We host the models on our end, our huggingface integration allows you to pull in any model from there for finetuning and push finetuned models to your private repos.

3

Impressive_Maximum32 t1_j1ybvzl wrote

Sounds great! I d be happy to switch over from astria, but i am missing two things:
- Astria has a "super_resolution" and a "face_correct" flag that work pretty good

2

TrueBlueDreamin OP t1_j206vle wrote

We also support "face_correct" which is just face restoration using GFPGAN/CodeFormers and "super_resolution" which is what's known as upscaling :)

Cheers! Excited to have you try us out

2

Impressive_Maximum32 t1_j20qby5 wrote

thanks for the reply!
I wasnt able to find these in your API-Documentation! Shouldnt that be under inference?

1

TrueBlueDreamin OP t1_j210gk8 wrote

Ah, sorry about that. Our docs are currently a little hard to parse, reworking them soon.

If you go to the inference params, theres fields called facetool and facetool_strength for face correction.

Documenting the upscaling endpoint as we speak :) New docs should be live EOD tomorrow. In the meantime, feel free to hop in the Discord to stay posted on when it's live.

2

dalcowboiz t1_j1to4o5 wrote

Does it work with sparse models from the sparse zoo?

1

TrueBlueDreamin OP t1_j1wcrph wrote

We can setup any model from there for you to finetune + run inference on. Shoot me a DM and I'll help you get started

1

QuerulousPanda t1_j1uzydb wrote

This may be a dumb question, but what products is stable diffusion useful for? It's an awesome technology but does it have any kind of widespread applicability?

1

Evoke_App t1_j1v9hoi wrote

Yes. Just look at all the AI art apps popping up. Logo makers too. The AI avatar apps killing it like AI Avatars or Lensa.

We'll have a stable diffusion and dreambooth API up soon on our site as well. So I hope it makes SD app dev's lives that much easier

1

Unreal_777 t1_j1z1gf3 wrote

Hello u/TrueBlueDreamin,

Would this website (of yours) allow a website (that I created) to let users send their images to my website, then my website would send them to your website and generate a training , then my website would receive back the new model? then produce some images to the end user? If yes, then everytime that happends you get payed right? (I am still new)

1

TrueBlueDreamin OP t1_j206nhy wrote

Hey Unreal, yes you can use our API to create your own app/website that does exactly that. You can choose to export the model after it's done finetuning and run inference (creating images) yourself if you'd like, we do also have an option for you to run inference through an API. Shoot me a DM and I can help you get setup

2

Paid-Not-Payed-Bot t1_j1z1h6r wrote

> you get paid right? (I

FTFY.

Although payed exists (the reason why autocorrection didn't help you), it is only correct in:

  • Nautical context, when it means to paint a surface, or to cover with something like tar or resin in order to make it waterproof or corrosion-resistant. The deck is yet to be payed.

  • Payed out when letting strings, cables or ropes out, by slacking them. The rope is payed out! You can pull now.

Unfortunately, I was unable to find nautical or rope-related words in your comment.

Beep, boop, I'm a bot

1

mxby7e t1_j20q0hg wrote

Does your api support the ability to train a style instead of a subject?

1

TrueBlueDreamin OP t1_j210wo4 wrote

It's currently optimized for tuning on subjects but we're launching finetuning for styles in the coming week since a lot of folks have asked for it. Feel free to hop in the Discord to be notified when we launch it, in any case, I'll circle back here once it's live

2

mxby7e t1_j215ewa wrote

Looking forward to it! I’ve been doing style training on the lastBen colab but I have a lot of artists that want a more accessible way to build and use models based on their own styles

2