Skip to content

Conversation

@HaoboGu
Copy link
Contributor

@HaoboGu HaoboGu commented Jul 11, 2022

This PR fixes #1013.
Tokenizer can be created from bytes in memory using added Tokenizers::from_bytes.

Signed-off-by: HaoboGu haobogu@outlook.com

Signed-off-by: HaoboGu <haobogu@outlook.com>
Copy link
Contributor

@Narsil Narsil left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, left a nit (feel free to ignore it if you don't agree).

@HuggingFaceDocBuilderDev
Copy link

HuggingFaceDocBuilderDev commented Jul 15, 2022

The documentation is not available anymore as the PR was closed or merged.

@Narsil
Copy link
Contributor

Narsil commented Jul 15, 2022

We can consider the tests wokring, this is linked to clippy update.

@Narsil Narsil merged commit 3564f24 into huggingface:main Jul 18, 2022
Narsil pushed a commit that referenced this pull request Aug 23, 2022
Signed-off-by: HaoboGu <haobogu@outlook.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Load pre-trained tokenizer from memory

3 participants