Skip to content

Commit 452e031

Browse files
sadra-barikbinNathanHBclefourrier
authored
Tiny improvements to endpoint_model.py, base_model.py,... (#219)
--------- Co-authored-by: Nathan Habib <[email protected]> Co-authored-by: Clémentine Fourrier <[email protected]>
1 parent 5a28b22 commit 452e031

File tree

3 files changed

+12
-10
lines changed

3 files changed

+12
-10
lines changed

src/lighteval/models/endpoints/endpoint_model.py

Lines changed: 8 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -510,7 +510,7 @@ def greedy_until(
510510

511511
for _, _ in tqdm(
512512
dataset.splits_start_end_iterator(),
513-
total=self.DATASET_SPLITS,
513+
total=dataset.num_dataset_splits,
514514
desc="Splits",
515515
position=0,
516516
disable=self.disable_tqdm,
@@ -532,12 +532,15 @@ def greedy_until(
532532
responses = asyncio.run(self._async_process_batch_generate(batch))
533533
else:
534534
responses = self._process_batch_generate(batch)
535-
for response in responses:
535+
for i, response in enumerate(responses):
536536
results.append(
537537
GenerativeResponse(
538538
result=response.generated_text,
539539
logits=[item.logprob for item in response.details.prefill] if returns_logits else None,
540-
truncated_tokens_count=-1,
540+
generated_tokens=[token.id for token in response.details.tokens],
541+
truncated_tokens_count=max(
542+
len(self.tokenizer.encode(batch[i].context)) - self.max_length, 0
543+
),
541544
padded_tokens_count=-1,
542545
)
543546
)
@@ -556,7 +559,7 @@ def loglikelihood(
556559

557560
for _, _ in tqdm(
558561
dataset.splits_start_end_iterator(),
559-
total=self.DATASET_SPLITS,
562+
total=dataset.num_dataset_splits,
560563
desc="Splits",
561564
position=0,
562565
disable=self.disable_tqdm,
@@ -607,7 +610,7 @@ def loglikelihood_rolling(
607610

608611
for _, _ in tqdm(
609612
dataset.splits_start_end_iterator(),
610-
total=self.DATASET_SPLITS,
613+
total=dataset.num_dataset_splits,
611614
desc="Splits",
612615
position=0,
613616
disable=self.disable_tqdm,

src/lighteval/models/model_loader.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -131,7 +131,9 @@ def load_openai_model(config: OpenAIModelConfig, env_config: EnvConfig):
131131
return model
132132

133133

134-
def load_model_with_inference_endpoints(config: InferenceEndpointModelConfig, env_config: EnvConfig):
134+
def load_model_with_inference_endpoints(
135+
config: Union[InferenceEndpointModelConfig, ServerlessEndpointModelConfig], env_config: EnvConfig
136+
):
135137
logger.info("Spin up model using inference endpoint.")
136138
model = InferenceEndpointModel(config=config, env_config=env_config)
137139
return model

src/lighteval/models/transformers/transformers_model.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -880,10 +880,7 @@ def greedy_until(
880880
input_ids=tokenized["input_ids"],
881881
input_lengths=[len(item == 1) for item in tokenized["attention_mask"]],
882882
input_mask=tokenized["attention_mask"],
883-
truncated=[
884-
len(c) - tokenized["input_ids"].shape[1] if len(c) > tokenized["input_ids"].shape[1] else 0
885-
for c in context
886-
],
883+
truncated=[max(len(c) - tokenized["input_ids"].shape[1], 0) for c in context],
887884
padded=[sum(mask == 0) for mask in tokenized["attention_mask"]],
888885
)
889886

0 commit comments

Comments
 (0)