Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Command-line interface Document Bug #39

Closed
delldu opened this issue Nov 19, 2018 · 1 comment
Closed

Command-line interface Document Bug #39

delldu opened this issue Nov 19, 2018 · 1 comment

Comments

@delldu
Copy link

delldu commented Nov 19, 2018

There is a bug in README.md about Command-line interface:
export BERT_BASE_DIR=chinese_L-12_H-768_A-12

Wrong:

pytorch_pretrained_bert convert_tf_checkpoint_to_pytorch \
   --tf_checkpoint_path $BERT_BASE_DIR/bert_model.ckpt.index \
   --bert_config_file $BERT_BASE_DIR/bert_config.json \
   --pytorch_dump_path $BERT_BASE_DIR/pytorch_model.bin

Right:

pytorch_pretrained_bert convert_tf_checkpoint_to_pytorch \
  $BERT_BASE_DIR/bert_model.ckpt.index \
  $BERT_BASE_DIR/bert_config.json \
  $BERT_BASE_DIR/pytorch_model.bin
@thomwolf
Copy link
Member

Thanks!

stevezheng23 added a commit to stevezheng23/transformers that referenced this issue Mar 24, 2020
add rational labeling multi-task support for mt-coqa runner
ocavue pushed a commit to ocavue/transformers that referenced this issue Sep 13, 2023
ZYC-ModelCloud pushed a commit to ZYC-ModelCloud/transformers that referenced this issue Nov 14, 2024
…exists

Check that model_save_name exists before trying to load it, to avoid confusing checkpoint error
ZYC-ModelCloud pushed a commit to ZYC-ModelCloud/transformers that referenced this issue Nov 14, 2024
* Fix model loader code is using bad default of float16

* Try to fix from_quantized error

* set dtype as fp16 in inference mode
ZYC-ModelCloud pushed a commit to ZYC-ModelCloud/transformers that referenced this issue Nov 14, 2024
* support bitblas.

* Add use_bitblas flag to load_model_tokenizer function and main function

* Fix bitblas backend initialization and matrix multiplication

* Remove print statement and save quantized model

* Fix bitblas backend initialization and matrix multiplication

* BitBLAS Support

* revert example

* cleanup

* add apache

* clean

* cleanup

* clean hip

* msg

* fix package name

* monkey patch auto_detect_nvidia_target()

* Fix safe_save code that was mistakenly removed

* pass GPU ID

* fix wrong FORMAT state set at end of conversion + remove compat and replace print with logger.info

* bitblas will be compiled by default

* add SUPPORTED_BITS and SUPPORTED_GROUP_SIZES to BaseQuantLinear so we can consolidate validation code

* Remove impossible code

* refractor structure and make sure params are passed in the same order

* fix pg509-232 not mapped to a100

* resolve commit issues

* resolve commit issues

* refract

* resolve commit

* lock pci id order for tests

* fix issue: convert_to_marlin() missing 2 required positional arguments: 'sym' and 'desc_act'

* resolve test

* Fix if statement

* accelerate.utils.modeling.load_checkpoint_in_model() is no longer called when use_bitblas is True

* add refractor todo lists

* If we use marlin or bitblas to load the quantized model, the model is already a converted model, and we no longer need to call load_checkpoint_in_model()

* Now supports saving quant models in BitBLAS format

* cleanup

* add test_q4_bitblas.py

* Fix unpacked_zeros calculation error

* add FORMAT.BITBLAS

* cleanup

* BITBLAS Only supported when desc_act is False.

---------

Co-authored-by: LeiWang <leiwang1999@outlook.com>
Co-authored-by: diegomontoya <xing@fictionpress.com>
Co-authored-by: Qubitium <Qubitium@modelcloud.ai>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants