Skip to content

Commit b9d7b2b

Browse files
committed
follow the prefix parameter and avoid hardcoded max number
1 parent 77c679f commit b9d7b2b

File tree

1 file changed

+8
-3
lines changed

1 file changed

+8
-3
lines changed

qwen_image.hpp

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -493,11 +493,16 @@ namespace Qwen {
493493
: GGMLRunner(backend, offload_params_to_cpu) {
494494
qwen_image_params.flash_attn = flash_attn;
495495

496-
int model_layers = 60;
497-
int num_layers = 1;
496+
int model_layers = qwen_image_params.num_layers;
497+
int num_layers = 1;
498+
std::string layer_prefix = "transformer_blocks.";
499+
if (prefix.size() > 0) {
500+
layer_prefix = prefix + "." + layer_prefix;
501+
}
498502
for (int layer = model_layers; layer > num_layers; layer--) {
503+
std::string layer_name = layer_prefix + std::to_string(layer-1) + ".attn.add_k_proj.bias";
499504
for (auto pair : tensor_types) {
500-
if (pair.first.find("model.diffusion_model.transformer_blocks." + std::to_string(layer-1) + ".attn.add_k_proj.bias") != std::string::npos) {
505+
if (pair.first.find(layer_name) != std::string::npos) {
501506
num_layers = layer;
502507
break;
503508
}

0 commit comments

Comments
 (0)