-
Notifications
You must be signed in to change notification settings - Fork 335
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
chatglm-ggml_q4_0.bin GGML_ASSERT ggml-metal.m:1453: false #184
Comments
产生报错的环境是 mac + metal 吗? |
同样遇到了这个问题。 M1 pro 16GB,chatglm3-f16使用mps后端就导致了这个。 系统是Sonoma 14.2. |
执行一下 |
cmake在终端的输出:
顺便编译警告里有个这个不知道有没有关系:
|
这个warning应该没关系,方便再发下执行指令和执行log吗 |
看 cmake 编译log其实主要想看 编译ggml那段的参数,你发的这个没有 |
@Weaxs 我删除了build目录重新cmake了一遍,获取了更完整的输出更新在上面了,但好像仍然没有ggml相关的参数。 执行指令和log跟楼主的一样,但我已经删掉模型了所以不方便重新运行了...磁盘空间伤不起 但我觉得原因只是单纯的爆内存了。我用llama.cpp运行qwen-14B-Q4_K,上下文长度较短时一切正常,但设到1000左右就会发生一模一样的错误。之前运行chatglm3-f16时,从敲回车到GGML_ASSERT输出之间,大概有十几秒的等待,期间内存压力也是先攀升到顶后瞬间跌落,肯定是爆内存了。 Update:通过修改代码,输出 修改方法: MTLCommandBufferDescriptor* descriptor = [[MTLCommandBufferDescriptor alloc] init];
descriptor.errorOptions = MTLCommandBufferErrorOptionEncoderExecutionStatus;
ctx->command_buffers[i] = [ctx->queue commandBufferWithDescriptor:descriptor];
[descriptor release]; 并且在报错行之前插入: NSError*error = [ctx->command_buffers[i] error];
if(error && ([ctx->command_buffers[i] errorOptions] &
MTLCommandBufferErrorOptionEncoderExecutionStatus)) {
GGML_METAL_LOG_INFO("%s", error.localizedDescription.UTF8String);
} 即可看到错误信息。 |
./build/bin/main -m ../GGUF_Models/chatglm-ggml_q4_0.bin -l 256 -p "你好"
GGML_ASSERT: /Users/apple/PycharmProjects/NLPProject/chatglm.cpp/third_party/ggml/src/ggml-metal.m:1453: false
The text was updated successfully, but these errors were encountered: