You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@LeiWang1999 Happy Chinese New Years! When you have time, can you point to us or help us find the cause of bitblas v0.1.0 issue. Thanks!
For GPTQModel v1.8.0-dev release I am trying to migrate gptqmodel's locked support for bitblas from working 0.0.1-dev13 to 0.1.0 but running into multiple issues.
It appears since bitblas version >= 0.0.1-dev14 bitblas kernel is no longer compatible and throws shape mistmatch errors.
I tried to migrate the code to bitblas kernel and pack code to bitblas 0.1.0 in PR and failed #1184 (now partially reverted back to 0.0.1-dev13 state) with Segmentation fault core dumped at the self.bitblas_matmul.lib.call(). 0.1.0 no longer has call_lib but we see there is a lib.call but trying to use it just seg faults.
fromgptqmodelimportGPTQModel, BACKENDmodel=GPTQModel.load(
"ModelCloud/Qwen2.5-0.5B-Instruct-ci-test-bitblas",
backend=BACKEND.BITBLAS, # change to BACKEND.TORCH or MARLIN for normal generate
)
print(model.tokenizer.decode(model.generate("What is the capital of United States?")[0]))
p.s. GPTQModel was recently merged into HF Transformers main so the next Transformers non-patch release should have wide-support of optional bitblas support for all gptq models with backend toggle!
The text was updated successfully, but these errors were encountered:
@LeiWang1999 Happy Chinese New Years! When you have time, can you point to us or help us find the cause of bitblas v0.1.0 issue. Thanks!
For GPTQModel v1.8.0-dev release I am trying to migrate gptqmodel's locked support for bitblas from working
0.0.1-dev13
to0.1.0
but running into multiple issues.It appears since bitblas version >=
0.0.1-dev14
bitblas kernel is no longer compatible and throwsshape
mistmatch errors.I tried to migrate the code to bitblas kernel and pack code to bitblas 0.1.0 in PR and failed #1184 (now partially reverted back to 0.0.1-dev13 state) with
Segmentation fault
core dumped at theself.bitblas_matmul.lib.call()
. 0.1.0 no longer hascall_lib
but we see there is alib.call
but trying to use it just seg faults.Please use the following code to replicate:
shape
exceptionsp.s. GPTQModel was recently merged into HF Transformers
main
so the next Transformers non-patch release should have wide-support of optional bitblas support for all gptq models withbackend
toggle!The text was updated successfully, but these errors were encountered: