Submitted by ChristmasInOct t3_11ium8l in deeplearning
ChristmasInOct OP t1_jb2cwwf wrote
Reply to comment by karyo in LLaMA model parallelization and server configuration by ChristmasInOct
Thanks for the response. Do you recall where you read the "only 200 people" bit? I'll take a look around for it as well; seems like the context could have found itself surrounded by interesting conversation.
P2P is not so much of a limitation so long as you can fit the entire model / pipeline into a single cards VRAM though, correct?
So for example, if you have a 7B Param model at FP16 and its around 14GB, presumably you should be safe with 24GB VRAM?
Thanks again for your time.
Viewing a single comment thread. View all comments