diff --git a/examples/model_configs/transformers_model.yaml b/examples/model_configs/transformers_model.yaml index 44e095dd3..377e348b3 100644 --- a/examples/model_configs/transformers_model.yaml +++ b/examples/model_configs/transformers_model.yaml @@ -3,10 +3,10 @@ model: model_args: "pretrained=HuggingFaceTB/SmolLM-1.7B,revision=main" # pretrained=model_name,trust_remote_code=boolean,revision=revision_to_use,model_parallel=True ... dtype: "bfloat16" compile: true + multichoice_continuations_start_space: null # If true/false, will force multiple choice continuations to start/not start with a space. If none, will do nothing merged_weights: # Ignore this section if you are not using PEFT models delta_weights: false # set to True of your model should be merged with a base model, also need to provide the base model name adapter_weights: false # set to True of your model has been trained with peft, also need to provide the base model name base_model: null # path to the base_model generation: - multichoice_continuations_start_space: null # If true/false, will force multiple choice continuations to start/not start with a space. If none, will do nothing temperature: 0.5 diff --git a/src/lighteval/main_accelerate.py b/src/lighteval/main_accelerate.py index d8d69f30f..3764cacdb 100644 --- a/src/lighteval/main_accelerate.py +++ b/src/lighteval/main_accelerate.py @@ -168,7 +168,7 @@ def accelerate( # noqa C901 args_dict["accelerator"] = accelerator args_dict["quantization_config"] = quantization_config args_dict["batch_size"] = override_batch_size - args_dict["multichoice_continuations_start_space"] = config["generation"][ + args_dict["multichoice_continuations_start_space"] = config["base_params"][ "multichoice_continuations_start_space" ] args_dict["use_chat_template"] = use_chat_template