SpaceCockatoo
SpaceCockatoo t1_jblj2so wrote
Reply to comment by ortegaalfredo in [R] Created a Discord server with LLaMA 13B by ortegaalfredo
4bit quant already out
SpaceCockatoo t1_j12q9me wrote
I too would like to know if this is even theoretically possible
SpaceCockatoo t1_jbz2mns wrote
Reply to [P] vanilla-llama an hackable plain-pytorch implementation of LLaMA that can be run on any system (if you have enough resources) by poppear
Any plans to implement 4/8-bit quantization?