-
Shanghai Jiao Tong University
Pinned Loading
-
intel/neural-compressor
intel/neural-compressor PublicSOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
-
EleutherAI/lm-evaluation-harness
EleutherAI/lm-evaluation-harness PublicA framework for few-shot evaluation of language models.
-
intel/intel-extension-for-transformers
intel/intel-extension-for-transformers Public archive⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡
-
optimum
optimum PublicForked from huggingface/optimum
🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools
Python
-
optimum-intel
optimum-intel PublicForked from huggingface/optimum-intel
Accelerate inference of 🤗 Transformers with Intel optimization tools
Jupyter Notebook
-
transformers
transformers PublicForked from huggingface/transformers
🤗 Transformers: State-of-the-art Natural Language Processing for Pytorch, TensorFlow, and JAX.
Python
If the problem persists, check the GitHub status page or contact support.