Submitted by buggaby t3_11qgasm in MachineLearning
MysteryInc152 t1_jc3hxpq wrote
Reply to comment by buggaby in [D] Are modern generative AI models on a path to significantly improved truthfulness? by buggaby
Yup. Decided to go over it properly.
If you compare all the instruct tuned models on there. Greater size equals Greater truthfulness. From Ada to Babbage to Curie to Claude to Davinci-002/003.
https://crfm.stanford.edu/helm/latest/?group=core_scenarios
So it does seem once again that scale will be in part the issue
buggaby OP t1_jc3ifnw wrote
Informative. Thanks. I'm a complexity scientist with training in some ML approaches, but not in transformers or other RL approaches. I'll review this (though not as fast as a LLM can...)
buggaby OP t1_jc3jw39 wrote
How do you find the model size? All those you listed appear to be based on GPT-3 or 3.5 which, according to my searching, are both 175B parameters. It looks to me like they are different only in the kind and amount of fine-tuning. What am I missing?
MysteryInc152 t1_jc3kb0x wrote
MysteryInc152 t1_jc3klp8 wrote
Claude is the informal name for Anthropic-LM v4-s3 (52B)
MysteryInc152 t1_jc3kufz wrote
Finally the instruct versions are prepended with "text-"
Viewing a single comment thread. View all comments