diff --git a/python/llm/example/GPU/HuggingFace/Advanced-Quantizations/AWQ/README.md b/python/llm/example/GPU/HuggingFace/Advanced-Quantizations/AWQ/README.md index cf281a8fdd1..b67bbe981f2 100644 --- a/python/llm/example/GPU/HuggingFace/Advanced-Quantizations/AWQ/README.md +++ b/python/llm/example/GPU/HuggingFace/Advanced-Quantizations/AWQ/README.md @@ -37,15 +37,10 @@ conda create -n llm python=3.11 conda activate llm # below command will install intel_extension_for_pytorch==2.1.10+xpu as default pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ -pip install transformers==4.35.0 pip install autoawq==0.1.8 --no-deps pip install accelerate==0.25.0 pip install einops ``` -**Note: For Mixtral model, please use transformers 4.36.0:** -```bash -pip install transformers==4.36.0 -``` ### 2. Configures OneAPI environment variables diff --git a/python/llm/example/GPU/HuggingFace/Advanced-Quantizations/GPTQ/README.md b/python/llm/example/GPU/HuggingFace/Advanced-Quantizations/GPTQ/README.md index a8040d31a5a..344b6db3dbc 100644 --- a/python/llm/example/GPU/HuggingFace/Advanced-Quantizations/GPTQ/README.md +++ b/python/llm/example/GPU/HuggingFace/Advanced-Quantizations/GPTQ/README.md @@ -13,7 +13,6 @@ conda create -n llm python=3.11 conda activate llm # below command will install intel_extension_for_pytorch==2.1.10+xpu as default pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ -pip install transformers==4.37.0 pip install auto_gptq==0.7.1 pip install optimum==1.14.0 ```