From 63218bde8924520e543d96e58230143254b25486 Mon Sep 17 00:00:00 2001 From: Stas Bekman Date: Thu, 11 Aug 2022 19:16:07 -0700 Subject: [PATCH] [doc] fix anchors the manual anchors end up being duplicated with automatically added anchors and no longer work. --- docs/source/en/glossary.mdx | 13 ++++++------- 1 file changed, 6 insertions(+), 7 deletions(-) diff --git a/docs/source/en/glossary.mdx b/docs/source/en/glossary.mdx index b6cb2259d67d..a61eb86eaab4 100644 --- a/docs/source/en/glossary.mdx +++ b/docs/source/en/glossary.mdx @@ -44,7 +44,7 @@ specific language governing permissions and limitations under the License. Every model is different yet bears similarities with the others. Therefore most models use the same inputs, which are detailed here alongside usage examples. - + ### Input IDs @@ -113,7 +113,7 @@ we will see because this is the way a [`BertModel`] is going to expect its inputs. - + ### Attention mask @@ -171,7 +171,7 @@ in the dictionary returned by the tokenizer under the key "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ``` - + ### Token Type IDs @@ -224,7 +224,7 @@ second sequence, corresponding to the "question", has all its tokens represented Some models, like [`XLNetModel`] use an additional token represented by a `2`. - + ### Position IDs @@ -238,7 +238,7 @@ absolute positional embeddings. Absolute positional embeddings are selected in the range `[0, config.max_position_embeddings - 1]`. Some models use other types of positional embeddings, such as sinusoidal position embeddings or relative position embeddings. - + ### Labels @@ -266,7 +266,7 @@ These labels are different according to the model head, for example: The base models (e.g., [`BertModel`]) do not accept labels, as these are the base transformer models, simply outputting features. - + ### Decoder input IDs @@ -279,7 +279,6 @@ such models, passing the `labels` is the preferred way to handle training. Please check each model's docs to see how they handle these input IDs for sequence to sequence training. - ### Feed Forward Chunking