Skip to content

Commit

Permalink
update doc entry for IPEX xpu example (#1516)
Browse files Browse the repository at this point in the history
Signed-off-by: Cheng, Zixuan <zixuan.cheng@intel.com>
  • Loading branch information
violetch24 authored Jan 5, 2024
1 parent c097b7c commit d69c552
Show file tree
Hide file tree
Showing 2 changed files with 5 additions and 3 deletions.
6 changes: 4 additions & 2 deletions docs/source/quantization.md
Original file line number Diff line number Diff line change
Expand Up @@ -466,7 +466,7 @@ Intel(R) Neural Compressor support multi-framework: PyTorch, Tensorflow, ONNX Ru
<td align="left">IPEX</td>
<td align="left">OneDNN</td>
<td align="left">"ipex"</td>
<td align="left">cpu | gpu</td>
<td align="left">cpu | xpu</td>
</tr>
<tr>
<td rowspan="5" align="left">ONNX Runtime</td>
Expand Down Expand Up @@ -538,6 +538,8 @@ conf = PostTrainingQuantConfig()
```python
# run with IPEX on CPU
conf = PostTrainingQuantConfig(backend="ipex")
# run with IPEX on XPU
conf = PostTrainingQuantConfig(backend="ipex", device="xpu")
```
```python
# run with ONNXRT CUDAExecutionProvider on GPU
Expand All @@ -555,4 +557,4 @@ conf = PostTrainingQuantConfig(backend="itex", device="gpu")
## Examples

User could refer to [examples](https://github.com/intel/neural-compressor/blob/master/examples/README.md) on how to quantize a new model.
If user wants to quantize an onnx model with npu, please refer to this [example](../../examples/onnxrt/image_recognition/onnx_model_zoo/shufflenet/quantization/ptq_static/README.md).
If user wants to quantize an onnx model with npu, please refer to this [example](../../examples/onnxrt/image_recognition/onnx_model_zoo/shufflenet/quantization/ptq_static/README.md). If user wants to quantize a pytorch model with xpu, please refer to this [example](../../examples/pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/ipex/README.md).
2 changes: 1 addition & 1 deletion examples/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -531,7 +531,7 @@ Intel® Neural Compressor validated examples with multiple compression technique
<td>bert-large-uncased-whole-word-masking-finetuned-squad</td>
<td>Natural Language Processing</td>
<td>Post-Training Static Quantization</td>
<td><a href="./pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/fx">fx</a> / <a href="./pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/ipex">ipex</a></td>
<td><a href="./pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/fx">fx</a> / <a href="./pytorch/nlp/huggingface_models/question-answering/quantization/ptq_static/ipex">ipex(xpu)</a></td>
</tr>
<tr>
<td>distilbert-base-uncased-distilled-squad</td>
Expand Down

0 comments on commit d69c552

Please sign in to comment.