-
Notifications
You must be signed in to change notification settings - Fork 252
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support Llama3 8b/70b #256
Conversation
7404baf
to
1d037e3
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
awesome, thanks for adding this!
left a couple minor grammar nits and one nit on naming, but overall looks fantastic.
Thank you for adding llama3 support so fast. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good! Can't wait to try running them. Had some inline comments.
train_configs/llama3_8b.toml
Outdated
[model] | ||
name = "llama3" | ||
flavor = "8B" | ||
tokenizer_path = "./torchtitan/datasets/tokenizer/original/tokenizer.model" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Putting llama2 tokenizer.model under torchtitan/datasets/tokenizer/
and llama3 tokenizer.model under /torchtitan/datasets/tokenizer/original/
is confusing, especially because they share the same file name. Can we organize them better?
requirements.txt
Outdated
datasets | ||
tomli >= 1.1.0 ; python_version < "3.11" | ||
tensorboard | ||
sentencepiece | ||
tiktoken==0.4.0 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I was asked for pip install blobfile
when tiktoken is first time running
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is it still the case for tiktokenizer version > 0.5.2? Or we have to install the blob file?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
that's a good question... I haven't tried
on mast it requires an additional pkg chardet
, which was not asked on local devgpu
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
hmmm I guess I'll try to remove the version pin for now
This PR adds support for Llama3 8b/70b, mainly it: - add tiktonizer, add instructions to download tokenizer - add options for the llama model to support Llama3 - add Llama3 8b/70b configs Have to remove integration test first, will add it back later once we figure out the details
@gnadathur I had to make |
This PR adds support for Llama3 8b/70b, mainly it: - add tiktonizer, add instructions to download tokenizer - add options for the llama model to support Llama3 - add Llama3 8b/70b configs
This PR adds support for Llama3 8b/70b, mainly it: - add tiktonizer, add instructions to download tokenizer - add options for the llama model to support Llama3 - add Llama3 8b/70b configs
This PR adds support for Llama3 8b/70b, mainly it: