-
Notifications
You must be signed in to change notification settings - Fork 31.9k
Remove all hf-internal-testing checkpoints that can be removed #21199
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
The documentation is not available anymore as the PR was closed or merged. |
ydshieh
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for working on this. I believe there are 2 places to correct, and one for you to decide (reformer)
|
|
||
| _CHECKPOINT_FOR_DOC = "allenai/longformer-base-4096" | ||
| _CONFIG_FOR_DOC = "LongformerConfig" | ||
| _TOKENIZER_FOR_DOC = "LongformerTokenizer" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I believe you intended to change the file modeling_longformer.py instead of this TF file. As the TF_xxx_SAMPLE still have something like (at this moment)
from transformers import {processor_class}, {model_class}There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ah yes, thanks!
|
|
||
| Returns: | ||
|
|
||
| <Tip warning={true}> |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think this could use _CHECKPOINT_FOR_DOC but remove the parts that check the expected output values, and we don't need this warning?
At least, this is what you have done for some models that used tiny-random-xxx checkpoints.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
(Hmm, it raises a question: when we use real checkpoints which don't contain head weights, so far we don't include the output checks, but we don't do any warning. The warning introduced in your last PR is only used for 2 models so far where no small enough real checkpoints exist).
Would like to hear from you regarding this - as I remember one major reason is to avoid user confusion regarding the results they get.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Here the checkpoint fails because it has is_decoder=True and we need False to use the MLM head.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
OK, thanks for explaining :-)
|
|
||
| # General docstring | ||
| _CONFIG_FOR_DOC = "UniSpeechSatConfig" | ||
| _PROCESSOR_FOR_DOC = "Wav2Vec2Processor" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Not able to comment at the exact line, but for UniSpeechSatForPreTraining, I think you missed to change Wav2Vec2FeatureExtractor (you changed this for another model file UniSpeechForPreTraining)
ydshieh
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM + thank you again!
What does this PR do?
This PR continues the work on docstrings and removes all checkpoints from the hf-internal-testing org where they can be removed.