Skip to content

Commit ee12fa8

Browse files
xin3hexinhe3
andauthored
replace XPU with Intel GPU (#2198)
* replace XPU with Intel GPU Signed-off-by: Xin He <[email protected]> --------- Signed-off-by: Xin He <[email protected]> Co-authored-by: Xin He <[email protected]>
1 parent 88a9fd3 commit ee12fa8

File tree

7 files changed

+11
-11
lines changed

7 files changed

+11
-11
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ support AMD CPU, ARM CPU, and NVidia GPU through ONNX Runtime with limited testi
3535
Choose the necessary framework dependencies to install based on your deploy environment.
3636
### Install Framework
3737
* [Install intel_extension_for_pytorch for CPU](https://intel.github.io/intel-extension-for-pytorch/cpu/latest/)
38-
* [Install intel_extension_for_pytorch for XPU](https://intel.github.io/intel-extension-for-pytorch/xpu/latest/)
38+
* [Install intel_extension_for_pytorch for Intel GPU](https://intel.github.io/intel-extension-for-pytorch/xpu/latest/)
3939
* [Use Docker Image with torch installed for HPU](https://docs.habana.ai/en/latest/Installation_Guide/Bare_Metal_Fresh_OS.html#bare-metal-fresh-os-single-click)
4040
**Note**: There is a version mapping between Intel Neural Compressor and Gaudi Software Stack, please refer to this [table](./docs/source/3x/gaudi_version_map.md) and make sure to use a matched combination.
4141
* [Install torch for other platform](https://pytorch.org/get-started/locally)

docs/source/3x/PyTorch.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -267,7 +267,7 @@ Deep Learning</a></td>
267267

268268
3. How to specify an accelerator?
269269

270-
> Neural Compressor provides automatic accelerator detection, including HPU, XPU, CUDA, and CPU.
270+
> Neural Compressor provides automatic accelerator detection, including HPU, Intel GPU, CUDA, and CPU.
271271

272272
> The automatically detected accelerator may not be suitable for some special cases, such as poor performance, memory limitations. In such situations, users can override the detected accelerator by setting the environment variable `INC_TARGET_DEVICE`.
273273

docs/source/faq.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -33,12 +33,12 @@ torch._C._LinAlgError: linalg.cholesky: The factorization could not be completed
3333
Try increasing `percdamp` (percent of the average Hessian diagonal to use for dampening),
3434
or increasing `nsamples` (the number of calibration samples).
3535
#### Issue 7:
36-
If you run GPTQ quantization with transformers-like API on xpu device, then you may encounter the following error:
36+
If you run GPTQ quantization with transformers-like API on Intel GPU device, then you may encounter the following error:
3737
```shell
3838
[ERROR][modeling_auto.py:128] index 133 is out of bounds for dimension 0 with size 128
3939
[ERROR][modeling_auto.py:129] Saved low bit model loading failed, please check your model.
4040
HINT:
41-
XPU device does not support `g_idx` for GPTQ quantization now. Please stay tuned.
41+
Intel GPU device does not support `g_idx` for GPTQ quantization now. Please stay tuned.
4242
You can set desc_act=False.
4343
```
4444
#### Issue 8:

docs/source/quantization.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -459,7 +459,7 @@ Intel(R) Neural Compressor support multi-framework: PyTorch, Tensorflow and ONNX
459459
<td align="left">IPEX</td>
460460
<td align="left">OneDNN</td>
461461
<td align="left">"ipex"</td>
462-
<td align="left">cpu | xpu</td>
462+
<td align="left">cpu | Intel GPU</td>
463463
</tr>
464464
<tr>
465465
<td rowspan="5" align="left">ONNX Runtime</td>
@@ -524,7 +524,7 @@ conf = PostTrainingQuantConfig()
524524
```python
525525
# run with IPEX on CPU
526526
conf = PostTrainingQuantConfig(backend="ipex")
527-
# run with IPEX on XPU
527+
# run with IPEX on Intel GPU
528528
conf = PostTrainingQuantConfig(backend="ipex", device="xpu")
529529
```
530530
```python
@@ -543,4 +543,4 @@ conf = PostTrainingQuantConfig(backend="itex", device="gpu")
543543
## Examples
544544

545545
User could refer to [examples](https://github.com/intel/neural-compressor/blob/master/examples/README.md) on how to quantize a new model.
546-
If user wants to quantize an onnx model with npu, please refer to this [example](../../examples/onnxrt/image_recognition/onnx_model_zoo/shufflenet/quantization/ptq_static/README.md). If user wants to quantize a pytorch model with xpu, please refer to this [example](../../examples/pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/ipex/README.md).
546+
If user wants to quantize an onnx model with npu, please refer to this [example](../../examples/onnxrt/image_recognition/onnx_model_zoo/shufflenet/quantization/ptq_static/README.md). If user wants to quantize a pytorch model with Intel GPU, please refer to this [example](../../examples/pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/ipex/README.md).

examples/3.x_api/pytorch/image_recognition/torchvision_models/quantization/static_quant/ipex/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ Please refer to [intel/intel-extension-for-pytorch(github.com)](https://github.c
2525
python -m pip install intel_extension_for_pytorch -f https://software.intel.com/ipex-whl-stable
2626
```
2727

28-
### Install IPEX XPU
28+
### Install IPEX Intel GPU
2929
Please build an IPEX docker container according to the [official guide](https://intel.github.io/intel-extension-for-pytorch/index.html#installation?platform=gpu&version=v2.1.30%2bxpu&os=linux%2fwsl2&package=docker).
3030

3131
You can run a simple sanity test to double confirm if the correct version is installed, and if the software stack can get correct hardware information onboard your system. The command should return PyTorch and IPEX versions installed, as well as GPU card(s) information detected.
@@ -84,7 +84,7 @@ bash run_quant.sh --input_model=resnext101_32x16d_wsl --dataset_location=/path/t
8484
bash run_benchmark.sh --input_model=resnext101_32x16d_wsl --dataset_location=/path/to/imagenet --mode=performance/accuracy --int8=true/false
8585
```
8686

87-
# Run with XPU
87+
# Run with Intel GPU
8888

8989
> Note: All torchvision model names can be passed as long as they are included in `torchvision.models`, below are some examples.
9090

examples/3.x_api/pytorch/nlp/huggingface_models/question-answering/quantization/static_quant/ipex/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ python run_qa.py \
2929
--output_dir ./savedresult
3030
```
3131

32-
## 2. Quantization with XPU
32+
## 2. Quantization with Intel GPU
3333
### 2.1 Environment Setting
3434
Please build an IPEX docker container according to the [official guide](https://intel.github.io/intel-extension-for-pytorch/index.html#installation?platform=gpu&version=v2.1.30%2bxpu&os=linux%2fwsl2&package=docker).
3535

examples/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -496,7 +496,7 @@ Intel® Neural Compressor validated examples with multiple compression technique
496496
<td>bert-large-uncased-whole-word-masking-finetuned-squad</td>
497497
<td>Natural Language Processing</td>
498498
<td>Post-Training Static Quantization</td>
499-
<td><a href="./pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/fx">fx</a> / <a href="./pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/ipex">ipex(xpu)</a></td>
499+
<td><a href="./pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/fx">fx</a> / <a href="./pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/ipex">ipex(Intel GPU)</a></td>
500500
</tr>
501501
<tr>
502502
<td>distilbert-base-uncased-distilled-squad</td>

0 commit comments

Comments
 (0)