Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

the costs of fine-tune #5

Open
TheCutestCat opened this issue Apr 3, 2023 · 5 comments
Open

the costs of fine-tune #5

TheCutestCat opened this issue Apr 3, 2023 · 5 comments

Comments

@TheCutestCat
Copy link

Your idea is really interesting, but I'm more worried about the fine-tune program spending too much money.

  1. Can you share the approximate cost of your previous work?
  2. As well as GPT-4 has been released and it shows very strong thinking power, perhaps GPT-4 can further improve the prediction performance of the model. Do you have a plan to use the GPT-4 ?
@asoyemi1
Copy link

asoyemi1 commented Apr 6, 2023

To piggyback on the above question, is there a way to save and load a fined-tuned for later inference? This could save some cost on having to retrain the model from scratch.

@kjappelbaum
Copy link
Owner

Sorry for not following up on this issue.

  1. It depends. A large chunk was for the davinci runs that we did not even include in the main text, and there are things we have done very often. We report around 4k for the full work.
  2. GPT-4 can currently not be fine-tuned. But we have seen some promising results with few-shot inference (e.g. check my Twitter)

Yes, you can load a fine-tuned model for later inference, you can save the model name from the output the Tuner gives you.

@TheCutestCat
Copy link
Author

@kjappelbaum Thanks for your response. Large language model can be so powerful. Perhaps we can develop a open-source model that based on LLaMA(Meta's "Chatgpt") and all researchers can participate. This could be very interesting🤣

@kjappelbaum
Copy link
Owner

kjappelbaum commented Apr 8, 2023

I'll soon upload a revised model of our paper; we also have some results on consumer hardware in there :)

@bmiles
Copy link
Contributor

bmiles commented Apr 19, 2023

From my experience I did a finetune with 4000 rows of data (943,748 trained tokens) and the training cost with openai was $0.38

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants