Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

损失函数的值不下降,一直维持在3到4 #118

Open
MichealZhangxa opened this issue Sep 24, 2024 · 4 comments
Open

损失函数的值不下降,一直维持在3到4 #118

MichealZhangxa opened this issue Sep 24, 2024 · 4 comments

Comments

@MichealZhangxa
Copy link

你好,我在你的代码的基础上训练,前期的损失一直在3到4之间正常吗,训练的也是llava的预训练数据集

@shiym2000
Copy link
Collaborator

您好,请问可以看一下您的实验脚本/配置吗?

@shiym2000
Copy link
Collaborator

Please provide your training script and any modified code (if applicable) so that we can better identify the issues in the code.

@shiym2000
Copy link
Collaborator

You might try using the two encoders separately to train the whole model, ensuring that each individual encoder can properly train the whole model.

@shiym2000
Copy link
Collaborator

And you can adjust the model training length by modifying the num_train_epochs parameter.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants