- Blog published on Intel Developer News: Intel AI Solutions Support the New Llama 3.2 Models (Sep 2024)
- Blog published on Intel information: 解决方案为最新Meta Llama 3.1模型提供加速 (July 2024)
- Blog published on Intel Developer News: [Intel AI Solutions Boost LLMs: Unleashing the Power of Meta* Llama 3.1] (https://www.intel.com/content/www/us/en/developer/articles/technical/intel-ai-solutions-support-meta-llama-3-1-launch.html) (July 2024)
- Blog published on digit.in: AI hallucination in LLM and beyond: Will it ever be fixed? (July 2024)
- Blog published on Medium: Accelerating Qwen2 Models with Intel Extension for Transformers (June 2024)
- Blog published on Huggingface: Building Cost-Efficient Enterprise RAG applications with Intel Gaudi 2 and Intel Xeon (May 2024)
- Blog published on Intel Developer News: Efficient Natural Language Embedding Models with Intel® Extension for Transformers (May 2024)
- Blog published on Intel NewsRoom: Intel Welcomes Open Platform for Enterprise AI (April 2024)
- Blog published on Intel Developer News: Weight-only Quantization in LLM Inference Using Intel® Extension for PyTorch* on Intel® Core™ Ultra Processors for AI PCs (Apr 2024)
- Blog published on Techcrunch: Intel and others commit to building open generative AI tools for the enterprise (Apr 2024)
- Video on YouTube: Intel Vision Keynotes 2024 (Apr 2024)
- Blog published on Vectara: Do Smaller Models Hallucinate More? (Apr 2024)
- Blog published on Intel NewsRoom: Intel Gaudi 2 Remains Only Benchmarked Alternative to NV H100 for GenAI Performance (March 2024)
- Blog published on Intel Developer News: Use the neural-chat-7b Model for Advanced Fraud Detection: An AI-Driven Approach in Cybersecurity (March 2024)
- CES 2024: CES 2024 Great Minds Keynote: Bringing the Limitless Potential of AI Everywhere: Intel Hybrid Copilot demo (Jan 2024)
- Blog published on Hugging Face: 2023, year of open LLMs (Dec 2023)
- Blog published on Medium: Connect an AI agent with your API: Intel Neural-Chat 7b LLM can replace Open AI Function Calling (Dec 2023)
- Blog published on 360 EEA (A News Platform about AI and LLMs): Intel neural-chat-7b-v3-1 (Dec 2023)
- Apple Podcasts from Papers Read on AI: Efficient LLM Inference on CPUs (Dec 2023)
- Blog published on Medium: Supercharging NLP Inference: Leveraging INT8 BGE-1.5 Models on Intel CPUs for Ultra-Low Latency (Nov 2023)
- NeurIPS'2023 on Efficient Natural Language and Speech Processing: Efficient LLM Inference on CPUs (Nov 2023)
- NeurIPS'2023 on Diffusion Models: Effective Quantization for Diffusion Models on CPUs (Nov 2023)
- Blog published on datalearner: Analysis of the top ten popular open source LLM of HuggingFace in the fourth week of November 2023 - the explosion of multi-modal large models and small-scale models (Nov 2023)
- Blog published on zaker: With this toolkit, the inference performance of large models can be accelerated by 40 times (Nov 2023)
- Blog published on geeky-gadgets: [New Intel Neural-Chat 7B LLM tops Hugging Face leaderboard beating original Mistral 7B] (https://www.geeky-gadgets.com/intel-neural-chat-7b-llm/) (Nov 2023)
- Blog published on Huggingface: Intel Neural-Chat 7b: Fine-Tuning on Gaudi2 for Top LLM Performance (Nov 2023)
- Video on YouTube: Neural Chat 7B v3-1 Installation on Windows - Step by Step (Nov 2023)
- Video on YouTube: Intel's Neural-Chat 7b: Most Powerful 7B Model! Beats GPT-4!? (Nov 2023)
- Blog published on marktechpost: Intel Researchers Propose a New Artificial Intelligence Approach to Deploy LLMs on CPUs More Efficiently (Nov 2023)
- Blog published on VMware: AI without GPUs: A Technical Brief for VMware Private AI with Intel (Nov 2023)
- News releases on VMware: VMware Collaborates with Intel to Unlock Private AI Everywhere (Nov 2023)
- Video on YouTube:Build Your Own ChatBot with Neural Chat | Intel Software (Oct 2023)
- Blog published on Medium: Layer-wise Low-bit Weight Only Quantization on a Laptop (Oct 2023)
- Blog published on Medium: Intel-Optimized Llama.CPP in Intel Extension for Transformers (Oct 2023)
- Blog published on Medium: Reduce the Carbon Footprint of Large Language Models (Oct 2023)
- Blog on GOVINDH Tech: Neural Chat vs. Competitors: A Detailed Guide (Sep 2023)
- Blog published on Medium: Empower Applications with Optimized LLMs: Performance, Cost, and Beyond (Sep 2023)
- Blog published on Medium: NeuralChat: Simplifying Supervised Instruction Fine-tuning and Reinforcement Aligning for Chatbots (Sep 2023)
- Keynote: Intel Innovation 2023 Livestream - Day2 (Sep 2023)
- Hands-On Lab at Intel Innovation 2023: Demystifying Generative AI: Develop and Optimize Your Own Talking Chatbot (Sep 2023)
- Blog published on Medium: NeuralChat: A Customizable Chatbot Framework (Sep 2023)
- Blog published on Medium: Faster Stable Diffusion Inference with Intel Extension for Transformers (July 2023)
- Blog published on Intel Developer News: The Moat Is Trust, Or Maybe Just Responsible AI (July 2023)
- Blog published on Intel Developer News: Create Your Own Custom Chatbot (July 2023)
- Blog published on Intel Developer News: Accelerate Llama 2 with Intel AI Hardware and Software Optimizations (July 2023)
- Arxiv: An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs (June 2023)
- Blog published on Medium: Simplify Your Custom Chatbot Deployment (June 2023)
- Blog published on Medium: Create Your Own Custom Chatbot (April 2023)
- Blog of Tech-Innovation Artificial-Intelligence(AI): Intel® Xeon® Processors Are Still the Only CPU With MLPerf Results, Raising the Bar By 5x - Intel Communities (April 2023)
- Blog published on Medium: MLefficiency — Optimizing transformer models for efficiency (Dec 2022)
- NeurIPS'2022: Fast Distilbert on CPUs (Nov 2022)
- NeurIPS'2022: QuaLA-MiniLM: a Quantized Length Adaptive MiniLM (Nov 2022)
- Blog published by Cohere: Top NLP Papers—November 2022 (Nov 2022)
- Blog published by Alibaba: Deep learning inference optimization for Address Purification (Aug 2022)
- NeurIPS'2021: Prune Once for All: Sparse Pre-Trained Language Models (Nov 2021)