@@ -3812,7 +3812,7 @@ struct server_context {
38123812                            if  (slot.n_past  > 0  && slot.n_past  < (int ) slot.prompt .tokens .size ()) {
38133813                                const  auto  pos_min = llama_memory_seq_pos_min (llama_get_memory (ctx), slot.id );
38143814                                if  (pos_min == -1 ) {
3815-                                     SLT_ERR (slot, " n_past = %d, cache_tokens .size() = %d, seq_id = %d, pos_min = %d\n " n_past , (int ) slot.prompt .tokens .size (), slot.id , pos_min);
3815+                                     SLT_ERR (slot, " n_past = %d, slot.prompt.tokens .size() = %d, seq_id = %d, pos_min = %d\n " n_past , (int ) slot.prompt .tokens .size (), slot.id , pos_min);
38163816                                    GGML_ABORT (" pos_min == -1, but n_past > 0 - should not happen: https://github.com/ggml-org/llama.cpp/pull/13833#discussion_r2116181237" 
38173817                                }
38183818
@@ -3860,7 +3860,7 @@ struct server_context {
38603860                                }
38613861
38623862                                if  (pos_min > pos_min_thold) {
3863-                                     SLT_WRN (slot, " n_past = %d, cache_tokens .size() = %d, seq_id = %d, pos_min = %d, n_swa = %d\n " n_past , (int ) slot.prompt .tokens .size (), slot.id , pos_min, n_swa);
3863+                                     SLT_WRN (slot, " n_past = %d, slot.prompt.tokens .size() = %d, seq_id = %d, pos_min = %d, n_swa = %d\n " n_past , (int ) slot.prompt .tokens .size (), slot.id , pos_min, n_swa);
38643864
38653865                                    //  search for a context checkpoint
38663866                                    const  auto  it = std::find_if (
@@ -4028,7 +4028,7 @@ struct server_context {
40284028                        }
40294029                    }
40304030
4031-                     //  SLT_INF(slot, "new cache_tokens : %s\n", slot.cache_tokens .str().c_str());
4031+                     //  SLT_INF(slot, "new slot.prompt.tokens : %s\n", slot.slot.prompt.tokens .str().c_str());
40324032
40334033                    SLT_INF (slot, " prompt processing progress, n_past = %d, n_tokens = %d, progress = %f\n " n_past , batch.n_tokens , (float ) slot.n_past  / slot.n_prompt_tokens ());
40344034
0 commit comments