From 34516a3557d7c2aec1fe9509a861c6fe27cbe2a6 Mon Sep 17 00:00:00 2001 From: Keval Morabia <28916987+kevalmorabia97@users.noreply.github.com> Date: Sat, 28 Feb 2026 05:08:26 +0530 Subject: [PATCH] Fix lint and import error in perf script llama3_llm_finetune.py (#2592) Signed-off-by: Keval Morabia <28916987+kevalmorabia97@users.noreply.github.com> --- .../configs/llama/llama3_llm_finetune.py | 24 +++++++++---------- 1 file changed, 12 insertions(+), 12 deletions(-) diff --git a/scripts/performance/configs/llama/llama3_llm_finetune.py b/scripts/performance/configs/llama/llama3_llm_finetune.py index 39f458ad70..b6f8c5c94e 100644 --- a/scripts/performance/configs/llama/llama3_llm_finetune.py +++ b/scripts/performance/configs/llama/llama3_llm_finetune.py @@ -274,12 +274,12 @@ def llama3_70b_lora_config_b300(precision: str = "bf16", config_variant: str = " ) precision_config = get_precision_config(precision) - cfg = llama3_70b_finetune_config( - peft="lora", - precision_config=precision_config, - packed_sequence=True, - seq_length=4096, - ) + cfg = llama3_70b_peft_config(peft_scheme="lora") + cfg.mixed_precision = precision_config + seq_length = 4096 + cfg.model.seq_length = seq_length + cfg.dataset.seq_length = seq_length + cfg.dataset.packed_sequence_specs.packed_sequence_size = seq_length set_llama3_common_peft_configs(cfg) set_workload_base_configs(cfg, base_cfg) # Enable pad_cu_seqlens for CUDA graphs compatibility with packed sequences. @@ -307,12 +307,12 @@ def llama3_70b_lora_config_b200(precision: str = "bf16", config_variant: str = " ) precision_config = get_precision_config(precision) - cfg = llama3_70b_finetune_config( - peft="lora", - precision_config=precision_config, - packed_sequence=True, - seq_length=4096, - ) + cfg = llama3_70b_peft_config(peft_scheme="lora") + cfg.mixed_precision = precision_config + seq_length = 4096 + cfg.model.seq_length = seq_length + cfg.dataset.seq_length = seq_length + cfg.dataset.packed_sequence_specs.packed_sequence_size = seq_length set_llama3_common_peft_configs(cfg) set_workload_base_configs(cfg, base_cfg) # Enable pad_cu_seqlens for CUDA graphs compatibility with packed sequences.