Skip to content

Commit 500d7b1

Browse files
authored
Add link to quantization tutorial (#218)
Signed-off-by: Ashwini Khade <askhade@microsoft.com>
1 parent 030a7d2 commit 500d7b1

1 file changed

Lines changed: 3 additions & 0 deletions

File tree

README.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -71,6 +71,9 @@ Once you have an ONNX model, it can be scored with a variety of tools.
7171
* [Inference Bert Model for High Performance with ONNX Runtime on AzureML](https://github.com/microsoft/onnxruntime/blob/master/onnxruntime/python/tools/transformers/notebooks/Inference_Bert_with_OnnxRuntime_on_AzureML.ipynb)
7272
* [Various Samples: Inferencing ONNX models using ONNX Runtime (Python, C#, C, Java, etc)](https://github.com/microsoft/onnxruntime/tree/master/samples)
7373

74+
### ONNX Quantization
75+
* [HuggingFace Bert Quantization with ONNX Runtime](https://github.com/microsoft/onnxruntime/blob/master/onnxruntime/python/tools/quantization/notebooks/Bert-GLUE_OnnxRuntime_quantization.ipynb)
76+
7477
### Serving
7578
* [Serving ONNX models with AI-Serving](https://github.com/autodeployai/ai-serving/blob/master/examples/AIServingMnistOnnxModel.ipynb)
7679
* [Serving ONNX models with Cortex](https://towardsdatascience.com/how-to-deploy-onnx-models-in-production-60bd6abfd3ae)

0 commit comments

Comments
 (0)