Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

docs: bit misspell comments model adapter default template name conversation #2594

Merged
merged 3 commits into from
Oct 21, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion fastchat/conversation.py
Original file line number Diff line number Diff line change
Expand Up @@ -762,7 +762,7 @@ def get_conv_template(name: str) -> Conversation:
)
)

# ChagGPT default template
# ChangGPT default template
register_conv_template(
Conversation(
name="polyglot_changgpt",
Expand Down
8 changes: 4 additions & 4 deletions fastchat/model/model_adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -572,7 +572,7 @@ def get_default_conv_template(self, model_path: str) -> Conversation:


class VicunaAdapter(BaseModelAdapter):
"Model adapater for Vicuna models (e.g., lmsys/vicuna-7b-v1.3)" ""
"Model adapter for Vicuna models (e.g., lmsys/vicuna-7b-v1.3)" ""

use_fast_tokenizer = False

Expand Down Expand Up @@ -605,7 +605,7 @@ def raise_warning_for_old_weights(self, model):
"current fastchat.\nYou can try one of the following methods:\n"
"1. Upgrade your weights to the new Vicuna-v1.3: https://github.com/lm-sys/FastChat#vicuna-weights.\n"
"2. Use the old conversation template by `python3 -m fastchat.serve.cli --model-path /path/to/vicuna-v0 --conv-template one_shot`\n"
"3. Downgrade fschat to fschat==0.1.10 (Not recommonded).\n"
"3. Downgrade fschat to fschat==0.1.10 (Not recommended).\n"
)


Expand Down Expand Up @@ -641,7 +641,7 @@ def load_model(self, model_path: str, from_pretrained_kwargs: dict):


class LongChatAdapter(BaseModelAdapter):
"Model adapater for LongChat models (e.g., lmsys/longchat-7b-16k)."
"Model adapter for LongChat models (e.g., lmsys/longchat-7b-16k)."

use_fast_tokenizer = False

Expand Down Expand Up @@ -1345,7 +1345,7 @@ def get_default_conv_template(self, model_path: str) -> Conversation:


class OpenOrcaAdapter(BaseModelAdapter):
"""Model adapater for Open-Orca models wich may use different prompt templates
"""Model adapter for Open-Orca models which may use different prompt templates
- (e.g. Open-Orca/OpenOrcaxOpenChat-Preview2-13B, Open-Orca/Mistral-7B-OpenOrca)
- `OpenOrcaxOpenChat-Preview2-13B` uses their "OpenChat Llama2 V1" prompt template.
- [Open-Orca/OpenOrcaxOpenChat-Preview2-13B #Prompt Template](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B#prompt-template)
Expand Down