Closed as not planned
Description
The correct command for quantizing the model is not reflected accurately in lines 206 and 207 of README.md (Prepare Data & Run). To perform model quantization, you should use the following command instead.
# quantize the model to 4-bits (using method 2 = q4_0)
./quantize ./models/7B/ggml-model-f16.bin ./models/7B/ggml-model-q4_0.bin 2
Metadata
Metadata
Assignees
Labels
No labels