-
Notifications
You must be signed in to change notification settings - Fork 13.3k
Closed
Labels
buildCompilation issuesCompilation issues
Description
I try the following with the latest master (6b2cb63)
python convert-pth-to-ggml.py models/13B/ 1
./quantize ./models/13B/ggml-model-f16.bin ./models/13B/ggml-model-q4_0.bin 2
./quantize ./models/13B/ggml-model-f16.bin.1 ./models/13B/ggml-model-q4_0.bin.1 2
ls models/13B/
checklist.chk consolidated.00.pth consolidated.01.pth ggml-model-f16.bin ggml-model-f16.bin.1 ggml-model-q4_0.bin ggml-model-q4_0.bin.1 params.json
./main -m ./models/13B/ggml-model-q4_0.bin -t 8 -n 128
main: seed = 1678568386
llama_model_load: loading model from './models/13B/ggml-model-q4_0.bin' - please wait ...
llama_model_load: n_vocab = 32000
llama_model_load: n_ctx = 512
llama_model_load: n_embd = 5120
llama_model_load: n_mult = 256
llama_model_load: n_head = 40
llama_model_load: n_layer = 40
llama_model_load: n_rot = 128
llama_model_load: f16 = 2
llama_model_load: n_ff = 13824
llama_model_load: ggml ctx size = 8559.49 MB
llama_model_load: memory_size = 800.00 MB, n_mem = 20480
llama_model_load: tensor 'tok_embeddings.weight' has wrong size in model file
main: failed to load model from './models/13B/ggml-model-q4_0.bin'
llama_model_load: ⏎
What would tensor 'tok_embeddings.weight' has wrong size in model file mean?
Metadata
Metadata
Assignees
Labels
buildCompilation issuesCompilation issues