Viewing a single comment thread. View all comments

Ghostglitch07 t1_izy5qmb wrote

It's weird because of how quick it is to claim it is unable to do things. In their attempt to make it safer they severely limited it's useability.They drilled the boilerplate text of "as a large language model trained by OpenAI I can't..." So hard that it throws it out far too often.

9

LetMeGuessYourAlts t1_j035ugy wrote

And if you carry a similar prompt over to the playground and run it on a davinci-003 model it will still attempt to answer your question without just giving up like that, so it's likely outside the model itself producing that response and then just having the model complete the error message. I was wondering if perhaps if confidence was low if it just defaults to an "I'm sorry..." and then let's the model produce the error.

1