-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running finetune_mind2web.py #11
Comments
MODEL_PATH is the path (or model name) of the base language model. CHECKPOINT_PATH is the path for the lora module. In finetuning, we didn't try using memory to construct few-shot prompts because of the context limit of the base model and those few-shot examples are also in the training data. |
Thanks for the quick reply. It would be very helpful if you could guide me on where to find the Lora module and the base language model. |
In the finetuning code, the --lora_dir is just the location to save the trained lora. If you download the LM locally, use --cache_dir to specify the location. Otherwise, it will download automatically into huggingface cache folder. |
Thanks, I understand now! I'm looking to simulate the output of the model as actions on a website and wondered if you've done something similar before. I've looked into frameworks that could support this but haven't found anything suitable. Do you have any suggestions on how to approach this? |
In this command:
python finetune_mind2web.py --data_dir path/to/Mind2Web/data --base_model codellama/CodeLlama-7b-Instruct-hf --cache_dir <MODEL_PATH> --lora_dir <CHECKPOINT_PATH> --top_k_elements 20
I am not able to understand what path should be given for MODEL_PATH and CHECKPOINT_PATH.
Also, for the finetuning part, in the README file, there were only two options:
Is there any reason for not using the -no_memory option?.
The text was updated successfully, but these errors were encountered: