-
Notifications
You must be signed in to change notification settings - Fork 412
Issues: SJTU-IPADS/PowerInfer
Meta: Implementing hybrid inference across key desktop platforms
#92
opened Dec 27, 2023 by
hodlen
Open
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Error: the provided PTX was compiled with an unsupported toolchain
bug-unconfirmed
Unconfirmed bugs
#229
opened Nov 3, 2024 by
jiangzizi
3 tasks done
about the use of OPT model
question
Further information is requested
#228
opened Oct 21, 2024 by
bobzhang208
add new model in power-infer2
question
Further information is requested
#227
opened Oct 21, 2024 by
Francis235
Qualcomm chips support
question
Further information is requested
#226
opened Oct 21, 2024 by
Francis235
Question about the perplexity
question
Further information is requested
#225
opened Oct 13, 2024 by
eljrte
请问我该如何获得opt模型相关的weight文件?
question
Further information is requested
#223
opened Sep 23, 2024 by
a1bc2def6g
统计predictor的overhead
question
Further information is requested
#220
opened Sep 16, 2024 by
guanchenl
3 tasks done
Help! Want a toy example to run matmul with q40 weight by cuda kernel
question
Further information is requested
#219
opened Sep 11, 2024 by
Eutenacity
CUDA toolkit version?
question
Further information is requested
#218
opened Sep 6, 2024 by
shujiehan
Am i doing something wrong?
question
Further information is requested
#216
opened Aug 28, 2024 by
RealMrCactus
3 tasks done
Some question about Fig4.
question
Further information is requested
#213
opened Jul 23, 2024 by
rhmaaa
我要如何获得预测文件呢
question
Further information is requested
#211
opened Jul 15, 2024 by
LDLINGLINGLING
3 tasks
Feature request : Support for PHI3 mini
enhancement
New feature or request
#210
opened Jul 14, 2024 by
raymond-infinitecode
3 tasks
请问powerinfer能否兼容llama.cpp的模型呢
question
Further information is requested
#209
opened Jul 5, 2024 by
mailonghua
the output for Q4_gguf is strange again!!
bug-unconfirmed
Unconfirmed bugs
#208
opened Jul 4, 2024 by
milktea888
About powerinfer-2
enhancement
New feature or request
#207
opened Jul 2, 2024 by
Ther-nullptr
3 tasks done
Where is the TurboSparse-Mixtral mlp_predictor?
question
Further information is requested
#203
opened Jun 27, 2024 by
MatthewCroughan
How to convert ProSparse-LLaMA-2-13B model to .gguf?
question
Further information is requested
#201
opened Jun 23, 2024 by
Graysonicc
3 tasks done
Source for v2 (mobile inference engine)
question
Further information is requested
#194
opened Jun 12, 2024 by
peeteeman
Need quite a long time to load the model
question
Further information is requested
#188
opened May 21, 2024 by
meicale
Previous Next
ProTip!
Updated in the last three days: updated:>2024-11-10.