Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

deepseek微调模型无法转换为rkllm #212

Open
jianfeng-xiyue opened this issue Mar 6, 2025 · 3 comments
Open

deepseek微调模型无法转换为rkllm #212

jianfeng-xiyue opened this issue Mar 6, 2025 · 3 comments

Comments

@jianfeng-xiyue
Copy link

我用unsloth 微调了1.5B模型,生成的文件目录如下图,

Image

将export_deepseek_1.5b_rkllm.py中mdol_lora改为了新路径,ret = llm.load_huggingface(model=modelpath, model_lora = modelpath, device='cpu')但转换时报错。

ython3 export_deepseek_1.5b_rkllm.py 
INFO: rkllm-toolkit version: 1.1.4
The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
Unused kwargs: ['_load_in_4bit', '_load_in_8bit', 'quant_method']. These kwargs are not used in <class 'transformers.utils.quantization_config.BitsAndBytesConfig'>.
ERROR: Catch exception when loading model: bitsandbytes
Load model failed!
@waydong
Copy link
Collaborator

waydong commented Mar 7, 2025

你好,方便提供下除权重以外的配置文件吗?

@jianfeng-xiyue
Copy link
Author

deep-1.5B.zip
请查收附件

你好,方便提供下除权重以外的配置文件吗?

@waydong
Copy link
Collaborator

waydong commented Mar 7, 2025

请先确认下huggingface模型能通过如下接口正常推理哈,model = AutoModelForCausalLM.from_pretrained(modelpath, trust_remote_code=True),这边测试都加载不起来。

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants