Viewing a single comment thread. View all comments

2lazy2buy t1_jcaary6 wrote

How is one achieving long context lengths for LLM? Chatgpt has a length 32k? Is the transformer decoder "just" that big?

2