From 12e5c97c78b8ec9a8b38aeaafe74cde0980be871 Mon Sep 17 00:00:00 2001 From: Wang Xin Date: Mon, 19 Feb 2024 14:21:51 +0800 Subject: [PATCH] fix doc style (#61688) --- python/paddle/quantization/quantize.py | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git a/python/paddle/quantization/quantize.py b/python/paddle/quantization/quantize.py index b7887ffc46e1c..425b7eab8de32 100644 --- a/python/paddle/quantization/quantize.py +++ b/python/paddle/quantization/quantize.py @@ -28,8 +28,9 @@ class Quantization(metaclass=abc.ABCMeta): r""" Abstract class used to prepares a copy of the model for quantization calibration or quantization-aware training. + Args: - config(QuantConfig) - Quantization configuration + config(QuantConfig): Quantization configuration """ def __init__(self, config: QuantConfig): @@ -43,10 +44,11 @@ def quantize(self, model: Layer, inplace=False): def convert(self, model: Layer, inplace=False, remain_weight=False): r"""Convert the quantization model to ONNX style. And the converted model can be saved as inference model by calling paddle.jit.save. + Args: - model(Layer) - The quantized model to be converted. - inplace(bool, optional) - Whether to modify the model in-place, default is False. - remain_weight(bool, optional) - Whether to remain weights in floats, default is False. + model(Layer): The quantized model to be converted. + inplace(bool, optional): Whether to modify the model in-place, default is False. + remain_weight(bool, optional): Whether to remain weights in floats, default is False. Return: The converted model