Viewing a single comment thread. View all comments

2blazen t1_j6ykrcq wrote

I've been using the GPT3 API for around 0.4c per request with 0 down time. With my current usage this sums up to around 10c a day, 3usd per month. I don't see how 20usd is reasonable

1

CowardlyVelociraptor t1_j7025ho wrote

You're paying a premium for the nice UI

1

2blazen t1_j70vh2g wrote

Might be just me, but I really hate how the reply is returned in the UI. Even if the subscription will solve the random interruptions during generation, the word-by-word printing kills me, I'd rather wait a bit but receive my answer in one piece

1

danielbln t1_j7c9mpc wrote

I much prefer to see the tokens as they are generated, it's much better UX as you can abort the generation if you feel it's not going in the right direction. All my GPT3 integrations use stream:true and display every word as it comes in.

1

arhetorical t1_j70ndxc wrote

Isn't ChatGPT more advanced than the davinci models available through the API? In any case, the point is that if you use it for work, $20 is negligible compared to the time you'll save.

1

2blazen t1_j70ux9o wrote

I thought so too, but haven't actually notice any difference, other than how the davinci models don't have the extensive content filters.

>if you use it for work, $20 is negligible

If my company pays for it, sure, otherwise I'll always prefer the request-based pricing with a nice API that I can just call from my terminal

1