-
Notifications
You must be signed in to change notification settings - Fork 27k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Command-line interface Document Bug #39
Comments
Thanks! |
stevezheng23
added a commit
to stevezheng23/transformers
that referenced
this issue
Mar 24, 2020
add rational labeling multi-task support for mt-coqa runner
1 task
ocavue
pushed a commit
to ocavue/transformers
that referenced
this issue
Sep 13, 2023
ZYC-ModelCloud
pushed a commit
to ZYC-ModelCloud/transformers
that referenced
this issue
Nov 14, 2024
…exists Check that model_save_name exists before trying to load it, to avoid confusing checkpoint error
ZYC-ModelCloud
pushed a commit
to ZYC-ModelCloud/transformers
that referenced
this issue
Nov 14, 2024
* Fix model loader code is using bad default of float16 * Try to fix from_quantized error * set dtype as fp16 in inference mode
ZYC-ModelCloud
pushed a commit
to ZYC-ModelCloud/transformers
that referenced
this issue
Nov 14, 2024
* support bitblas. * Add use_bitblas flag to load_model_tokenizer function and main function * Fix bitblas backend initialization and matrix multiplication * Remove print statement and save quantized model * Fix bitblas backend initialization and matrix multiplication * BitBLAS Support * revert example * cleanup * add apache * clean * cleanup * clean hip * msg * fix package name * monkey patch auto_detect_nvidia_target() * Fix safe_save code that was mistakenly removed * pass GPU ID * fix wrong FORMAT state set at end of conversion + remove compat and replace print with logger.info * bitblas will be compiled by default * add SUPPORTED_BITS and SUPPORTED_GROUP_SIZES to BaseQuantLinear so we can consolidate validation code * Remove impossible code * refractor structure and make sure params are passed in the same order * fix pg509-232 not mapped to a100 * resolve commit issues * resolve commit issues * refract * resolve commit * lock pci id order for tests * fix issue: convert_to_marlin() missing 2 required positional arguments: 'sym' and 'desc_act' * resolve test * Fix if statement * accelerate.utils.modeling.load_checkpoint_in_model() is no longer called when use_bitblas is True * add refractor todo lists * If we use marlin or bitblas to load the quantized model, the model is already a converted model, and we no longer need to call load_checkpoint_in_model() * Now supports saving quant models in BitBLAS format * cleanup * add test_q4_bitblas.py * Fix unpacked_zeros calculation error * add FORMAT.BITBLAS * cleanup * BITBLAS Only supported when desc_act is False. --------- Co-authored-by: LeiWang <leiwang1999@outlook.com> Co-authored-by: diegomontoya <xing@fictionpress.com> Co-authored-by: Qubitium <Qubitium@modelcloud.ai>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
There is a bug in README.md about Command-line interface:
export BERT_BASE_DIR=chinese_L-12_H-768_A-12
Wrong:
Right:
The text was updated successfully, but these errors were encountered: