LoRA integration with llama-export-lora.exe results in blank/gibberish output #10911
Unanswered
AkiMatsushita
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm experiencing an issue when integrating a LoRA model with a base model using
llama-export-lora.exe
. After merging the LoRA into the base model's gguf file, the resulting model either produces blank output or gibberish (e.g., "---------") regardless of the prompt used.Steps to reproduce:
peft
library with the following settings:convert_lora_to_gguf.py
with the--outtype f16
option.Command:
python convert_lora_to_gguf.py --base ./base_model --outfile ./lora_model.gguf --outtype f16 ./lora_model
llama-export-lora.exe
.Command:
llama-export-lora.exe -m base_model.gguf --lora lora_model.gguf -o merged_model.gguf
Command:
./main -m merged_model.gguf -p "Test prompt"
Expected behavior:
The merged model should generate text based on the given prompt, incorporating the knowledge learned by the LoRA.
Actual behavior:
The merged model produces either blank output or gibberish (e.g., "---------"). This happens even with prompts unrelated to the training data.
Environment:
Additional context:
transformers
andpeft
.Thank you for your time and assistance.
Beta Was this translation helpful? Give feedback.
All reactions