-
Notifications
You must be signed in to change notification settings - Fork 608
Issues: THUDM/GLM-130B
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Question] Can I finetune GLM-130B with SAT framework?
#55
opened Jan 8, 2023 by
smeyerhot
updated Jan 8, 2023
TASK_NAME/validation.jsonl used for tuning hyperparameters for zero-shot testing?
#64
opened Jan 12, 2023 by
tomyoung903
updated Jan 12, 2023
will the GLM generation add some common text generation penalties like GPT-3?
#65
opened Jan 13, 2023 by
cyente
updated Jan 13, 2023
Can i have only the encoder of the model ?
#66
opened Jan 13, 2023 by
MohamedAliRashad
updated Jan 19, 2023
zero-shot or few-shot for summarization task question.
#84
opened Feb 14, 2023 by
siyuanxue
updated Feb 14, 2023
will you release DCU Ascend train and inference code?
#86
opened Feb 16, 2023 by
clockfly
updated Feb 16, 2023
Why is there a backward method in the W8A16Linear class?
#90
opened Feb 21, 2023 by
Ant0082
updated Feb 21, 2023
How to eval the glm-130b loss? I got a loss 6 on wudao corpus
#91
opened Feb 23, 2023 by
Syno8
updated Feb 23, 2023
GLM-130B model evaluation on the 4 x RTX 3090 GPU machine
#94
opened Mar 1, 2023 by
Tomas0413
updated Mar 1, 2023
[Disscussion] Can we align GLM-130B to human like chatgpt?
#43
opened Dec 10, 2022 by
AnShengqiang
updated Mar 10, 2023
单机离线状态下无法运行,报错[errno 11001]getaddrinfo failed
#103
opened Mar 24, 2023 by
gsxy456
updated Mar 24, 2023
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.