You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/source/tutorials/qat_finetune.rst
-5Lines changed: 0 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -168,11 +168,6 @@ modifications accordingly:
168
168
fake_quant_after_n_steps: 1000
169
169
memory_efficient_fsdp_wrap: False
170
170
171
-
.. note::
172
-
173
-
QAT in torchtune is currently not compatible with `memory_efficient_fsdp_wrap <https://pytorch.org/torchtune/stable/generated/torchtune.utils.get_full_finetune_fsdp_wrap_policy.html#torchtune.utils.get_full_finetune_fsdp_wrap_policy>`_.
174
-
This is a known issue and will be fixed in a future torchtune version.
175
-
176
171
Empirically, we observed that disabling fake quantization for the first N steps
177
172
led to better results, presumably because doing so allows the weights to stabilize
178
173
before we start introducing quantization noise to the fine-tuning process.
0 commit comments