Skip to content

run_mlm example is missing block_size and torch_dtype args (present in run_clm) #29697

@galtay

Description

@galtay

System Info

  • transformers version: 4.39.0.dev0
  • Platform: Linux-6.5.0-25-generic-x86_64-with-glibc2.35
  • Python version: 3.11.8
  • Huggingface_hub version: 0.21.4
  • Safetensors version: 0.4.2
  • Accelerate version: 0.28.0
  • Accelerate config: not found
  • PyTorch version (GPU?): 2.2.1+cu121 (True)
  • Tensorflow version (GPU?): not installed (NA)
  • Flax version (CPU?/GPU?/TPU?): not installed (NA)
  • Jax version: not installed
  • JaxLib version: not installed
  • Using GPU in script?: no
  • Using distributed or parallel set-up in script?: no

Who can help?

@ArthurZucker and @younesbelkada

Information

  • The official example scripts
  • My own modified scripts

Tasks

  • An officially supported task in the examples folder (such as GLUE/SQuAD, ...)
  • My own task or dataset (give details below)

Reproduction

inspection of args in https://github.com/huggingface/transformers/blob/main/examples/pytorch/language-modeling/run_mlm.py
vs args in https://github.com/huggingface/transformers/blob/main/examples/pytorch/language-modeling/run_clm.py

Expected behavior

the ability to use torch_dtype and block_size args in the run_mlm.py script as they are used in the run_clm.py script.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions