Quantized model at q4_k_4 gives below error: llama_model_load: error loading model: done_getting_tensors: wrong number of tensors; expected 292, got 291 Maybe related to this: https://github.com/abetlen/llama-cpp-python/issues/1646
Quantized model at q4_k_4 gives below error:
llama_model_load: error loading model: done_getting_tensors: wrong number of tensors; expected 292, got 291
Maybe related to this:
abetlen/llama-cpp-python#1646