diff --git a/gallery/index.yaml b/gallery/index.yaml index 384fa21dda07..acd91884749a 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -1,4 +1,37 @@ --- +- name: "onerec-8b" + url: "github:mudler/LocalAI/gallery/virtual.yaml@master" + urls: + - https://huggingface.co/mradermacher/OneRec-8B-GGUF + description: | + The model `mradermacher/OneRec-8B-GGUF` is a quantized version of the base model `OpenOneRec/OneRec-8B`, a large language model designed for tasks like recommendations or content generation. It is optimized for efficiency with various quantization schemes (e.g., Q2_K, Q4_K, Q8_0) and available in multiple sizes (3.5–9.0 GB). The model uses the GGUF format and is licensed under Apache-2.0. Key features include: + + - **Base Model**: `OpenOneRec/OneRec-8B` (a pre-trained language model for recommendations). + - **Quantization**: Supports multiple quantized variants (Q2_K, Q3_K, Q4_K, etc.), with the best quality for `Q4_K_S` and `Q8_0`. + - **Sizes**: Available in sizes ranging from 3.5 GB (Q2_K) to 9.0 GB (Q8_0), with faster speeds for lower-bit quantized versions. + - **Usage**: Compatible with GGUF files, suitable for deployment in applications requiring efficient model inference. + - **Licence**: Apache-2.0, available at [https://huggingface.co/OpenOneRec/OneRec-8B/blob/main/LICENSE](https://huggingface.co/OpenOneRec/OneRec-8B/blob/main/LICENSE). + + For detailed specifications, refer to the [model page](https://hf.tst.eu/model#OneRec-8B-GGUF). + overrides: + parameters: + model: llama-cpp/models/OneRec-8B.Q4_K_M.gguf + name: OneRec-8B-GGUF + backend: llama-cpp + template: + use_tokenizer_template: true + known_usecases: + - chat + function: + grammar: + disable: true + description: Imported from https://huggingface.co/mradermacher/OneRec-8B-GGUF + options: + - use_jinja:true + files: + - filename: llama-cpp/models/OneRec-8B.Q4_K_M.gguf + sha256: f19217971ee5a7a909c9217a79d09fb573380f5018e25dcb32693139e59b434f + uri: https://huggingface.co/mradermacher/OneRec-8B-GGUF/resolve/main/OneRec-8B.Q4_K_M.gguf - name: "minimax-m2.1-i1" url: "github:mudler/LocalAI/gallery/virtual.yaml@master" urls: