Submitted by terserterseness t3_10fxryj in MachineLearning
All the examples from langchain and on huggingface create memory by pasting the entire history in every prompt. This seems to violate the max input prompt length pretty quickly. And it’s expensive. Does chatgpt use something revolutionary? It forgets everything when you create a new session so it ‘feels’ it’s using the convo as memory as well.
But then the question; how do they get past prompt limits? Chunking doesn’t help as it still doesn’t get context in that case between prompts. Maybe they ask the same question with different chunks many times and then ask for a final result?
Apologies if this was answered somewhere, I cannot find it at all and all examples use the same kind of history memory.
DaLameLama t1_j4zhqqj wrote
Does ChatGPT actually get past the token limit? Codex supports ~8000 tokens. You might underestimate how much this is. Has anyone tested the limits?
Unfortunately, OpenAI aren't serious about publishing technical reports anymore.