From 7bb86cf593e07f5c5bc6ed058ae82a42feb88a43 Mon Sep 17 00:00:00 2001 From: "Huang, Tai" Date: Wed, 4 Sep 2024 15:39:13 +0800 Subject: [PATCH 1/2] add recent publications Signed-off-by: Huang, Tai --- README.md | 1 + docs/source/publication_list.md | 4 +++- 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index fa82961dd75..16e16a2a29c 100644 --- a/README.md +++ b/README.md @@ -181,6 +181,7 @@ Intel Neural Compressor will convert the model format from auto-gptq to hpu form > From 3.0 release, we recommend to use 3.X API. Compression techniques during training such as QAT, Pruning, Distillation only available in [2.X API](https://github.com/intel/neural-compressor/blob/master/docs/source/2x_user_guide.md) currently. ## Selected Publications/Events +* Blog on Medium: [Quantization on Intel Gaudi Series AI Accelerators](https://medium.com/intel-analytics-software/intel-neural-compressor-v3-0-a-quantization-tool-across-intel-hardware-9856adee6f11) (Aug 2024) * Blog by Intel: [Neural Compressor: Boosting AI Model Efficiency](https://community.intel.com/t5/Blogs/Tech-Innovation/Artificial-Intelligence-AI/Neural-Compressor-Boosting-AI-Model-Efficiency/post/1604740) (June 2024) * Blog by Intel: [Optimization of Intel AI Solutions for Alibaba Cloud’s Qwen2 Large Language Models](https://www.intel.com/content/www/us/en/developer/articles/technical/intel-ai-solutions-accelerate-alibaba-qwen2-llms.html) (June 2024) * Blog by Intel: [Accelerate Meta* Llama 3 with Intel AI Solutions](https://www.intel.com/content/www/us/en/developer/articles/technical/accelerate-meta-llama3-with-intel-ai-solutions.html) (Apr 2024) diff --git a/docs/source/publication_list.md b/docs/source/publication_list.md index db558a39195..61d2027a564 100644 --- a/docs/source/publication_list.md +++ b/docs/source/publication_list.md @@ -1,6 +1,8 @@ Full Publications/Events (82) ========== -## 2024 (3) +## 2024 (5) +* Blog on Medium: [Quantization on Intel Gaudi Series AI Accelerators](https://medium.com/intel-analytics-software/intel-neural-compressor-v3-0-a-quantization-tool-across-intel-hardware-9856adee6f11) (Aug 2024) +* Blog on Medium: [Accelerating Qwen2 Models with Intel Extension for Transformers](https://medium.com/intel-analytics-software/accelerating-qwen2-models-with-intel-extension-for-transformers-99403de82f68) (June 2024) * Blog by Intel: [Neural Compressor: Boosting AI Model Efficiency](https://community.intel.com/t5/Blogs/Tech-Innovation/Artificial-Intelligence-AI/Neural-Compressor-Boosting-AI-Model-Efficiency/post/1604740) (June 2024) * Blog by Intel: [Optimization of Intel AI Solutions for Alibaba Cloud’s Qwen2 Large Language Models](https://www.intel.com/content/www/us/en/developer/articles/technical/intel-ai-solutions-accelerate-alibaba-qwen2-llms.html) (June 2024) * Blog by Intel: [Accelerate Meta* Llama 3 with Intel AI Solutions](https://www.intel.com/content/www/us/en/developer/articles/technical/accelerate-meta-llama3-with-intel-ai-solutions.html) (Apr 2024) From 1a71d89c833b11b2544dcc386ed37e864d8f823b Mon Sep 17 00:00:00 2001 From: "Huang, Tai" Date: Sat, 7 Sep 2024 22:19:22 +0800 Subject: [PATCH 2/2] update total count Signed-off-by: Huang, Tai --- docs/source/publication_list.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/source/publication_list.md b/docs/source/publication_list.md index 61d2027a564..65d12714289 100644 --- a/docs/source/publication_list.md +++ b/docs/source/publication_list.md @@ -1,4 +1,4 @@ -Full Publications/Events (82) +Full Publications/Events (84) ========== ## 2024 (5) * Blog on Medium: [Quantization on Intel Gaudi Series AI Accelerators](https://medium.com/intel-analytics-software/intel-neural-compressor-v3-0-a-quantization-tool-across-intel-hardware-9856adee6f11) (Aug 2024)