From fc5954392e9260cf5b59eb3bc7193f24af53914a Mon Sep 17 00:00:00 2001 From: Matt Watson Date: Fri, 12 Jan 2024 15:03:18 -0800 Subject: [PATCH] Update llama conversion script for new kaggle format --- keras_nlp/models/__init__.py | 1 + .../convert_llama_checkpoints.py | 266 +++++++++++------- 2 files changed, 165 insertions(+), 102 deletions(-) diff --git a/keras_nlp/models/__init__.py b/keras_nlp/models/__init__.py index ab04d8eae0..b42652efd0 100644 --- a/keras_nlp/models/__init__.py +++ b/keras_nlp/models/__init__.py @@ -90,6 +90,7 @@ ) from keras_nlp.models.gpt_neo_x.gpt_neo_x_tokenizer import GPTNeoXTokenizer from keras_nlp.models.llama.llama_backbone import LlamaBackbone +from keras_nlp.models.llama.llama_tokenizer import LlamaTokenizer from keras_nlp.models.mistral.mistral_backbone import MistralBackbone from keras_nlp.models.opt.opt_backbone import OPTBackbone from keras_nlp.models.opt.opt_causal_lm import OPTCausalLM diff --git a/tools/checkpoint_conversion/convert_llama_checkpoints.py b/tools/checkpoint_conversion/convert_llama_checkpoints.py index 5eb3973f36..f71e536c65 100644 --- a/tools/checkpoint_conversion/convert_llama_checkpoints.py +++ b/tools/checkpoint_conversion/convert_llama_checkpoints.py @@ -11,131 +11,193 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. -import os - -import torch -from transformers import AutoModel +""" +Llama weight conversion script. + +To run, install the CPU only development environment and huggingface libraries: +``` +pip install -r requirements.txt +pip install transformers huggingface-cli +``` + +Login to Huggingface: +``` +huggingface-cli login +``` + +Finally run this script to convert, validate and upload weights. +``` +python tools/checkpoint_conversion/convert_llama_checkpoints.py \ + --preset tiny_llama_1b_chat_en +``` +""" -from keras_nlp.models.llama.llama_backbone import LlamaBackbone +import os os.environ["KERAS_BACKEND"] = "torch" +os.environ["CUDA_VISIBLE_DEVICES"] = "-1" -# from huggingface_hub import login -# llama weights as of now are on request access -# login(token='