Submitted by [deleted] t3_11u4u6b in MachineLearning
Disastrous_Elk_6375 t1_jco0q62 wrote
> can I expect better outputs than what Stanford Alpaca achieved?
I think better is a bit subjective. They do note that the answers are generally shorter than ChatGPT, because they used text-davinci-003. Using gpt-3.5-turbo would get your answers closer to ChatGPT, but it could also "grab" that boring monotone "firstly... secondly... in conclusion" that often gives it away.
[deleted] OP t1_jcok6zs wrote
Yeah, it does that. I can modify ChatGPT behavior through the system message, which should change the personality and response type in the final data. I could maybe start training it with examples on how it should act when the system message is present.
Example:
### System:{Act as a best friend}
### Instruction:{hi}### Input:{noinput}### Response: Hey there! What's up? How's your day going?
I could feed the model with thousands of examples like this which would result in complete personality change if the system message is present
Viewing a single comment thread. View all comments