Submitted by bikeskata t3_10r7k0h in MachineLearning
2blazen t1_j6ykrcq wrote
Reply to comment by arhetorical in [N] OpenAI starts selling subscriptions to its ChatGPT bot by bikeskata
I've been using the GPT3 API for around 0.4c per request with 0 down time. With my current usage this sums up to around 10c a day, 3usd per month. I don't see how 20usd is reasonable
CowardlyVelociraptor t1_j7025ho wrote
You're paying a premium for the nice UI
2blazen t1_j70vh2g wrote
Might be just me, but I really hate how the reply is returned in the UI. Even if the subscription will solve the random interruptions during generation, the word-by-word printing kills me, I'd rather wait a bit but receive my answer in one piece
danielbln t1_j7c9mpc wrote
I much prefer to see the tokens as they are generated, it's much better UX as you can abort the generation if you feel it's not going in the right direction. All my GPT3 integrations use stream:true and display every word as it comes in.
arhetorical t1_j70ndxc wrote
Isn't ChatGPT more advanced than the davinci models available through the API? In any case, the point is that if you use it for work, $20 is negligible compared to the time you'll save.
2blazen t1_j70ux9o wrote
I thought so too, but haven't actually notice any difference, other than how the davinci models don't have the extensive content filters.
>if you use it for work, $20 is negligible
If my company pays for it, sure, otherwise I'll always prefer the request-based pricing with a nice API that I can just call from my terminal
Viewing a single comment thread. View all comments