Submitted by imgonnarelph t3_11wqmga in MachineLearning
Unlucky_Excitement_2 t1_jdavhcr wrote
Reply to comment by KerfuffleV2 in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
Bro what are you talking about LOL. Its context length he's discussing. There are multiple ways[all of which I'm expertimenting with] ->
- flash attention
- strided context window
- finetuning on a dataset with longer sequences
KerfuffleV2 t1_jdbrkc1 wrote
Uh, did you reply to the wrong person or something? Your post doesn't have anything to do with either mine or the parent.
Viewing a single comment thread. View all comments