Getting GGUF quantization

#2
by hbacard - opened

@TheBloke : I really like what you are doing ! I would like to know how you actually converted-quantized this model [which script from llama.cpp you used]. I am coming from academia and want to understand what's going on under the hood :).

Thanks a lot !

Sign up or log in to comment