Submitted by RadioFreeAmerika t3_122ilav in singularity
ArcticWinterZzZ t1_jdt0plo wrote
Reply to comment by zero_for_effort in Why is maths so hard for LLMs? by RadioFreeAmerika
GPT-4 always takes the same amount of time to output a token. However, multiplication has been proven to take more time than GPT-4 has available. Therefore, an LLM like GPT-4 cannot possibly "grow" the requisite structures required to actually calculate multiplication "instantly". There are probably quite a few more problems like this, which is why chain-of-thought prompting can be so powerful.
zero_for_effort t1_jdt4nz1 wrote
This is the explanation I found easiest to understand, cheers.
Viewing a single comment thread. View all comments