Skip to content

Commit

Permalink
add window_size to nemo_config.
Browse files Browse the repository at this point in the history
Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
  • Loading branch information
akoumpa committed Dec 19, 2023
1 parent 2b98ee2 commit 4eb7ce4
Showing 1 changed file with 2 additions and 0 deletions.
2 changes: 2 additions & 0 deletions scripts/nlp_language_modeling/convert_mistral_7b_to_nemo.py
Original file line number Diff line number Diff line change
Expand Up @@ -98,12 +98,14 @@ def load_config(mistral_config, tokenizer_path):
nemo_config = OmegaConf.load(
os.path.join(os.path.dirname(__file__), '../../examples/nlp/language_modeling/conf/megatron_llama_config.yaml')
).model
# akoumparouli: verify this.
nemo_config.encoder_seq_length = mistral_config['sliding_window']
nemo_config.num_layers = int(mistral_config['n_layers'])
nemo_config.hidden_size = mistral_config['dim']
nemo_config.ffn_hidden_size = mistral_config['hidden_dim']
nemo_config.num_attention_heads = mistral_config['n_heads']
nemo_config.max_position_embeddings = 32_768
nemo_config.window_size = mistral_config['sliding_window']
nemo_config.init_method_std = 0.02
# RMSNorm's epsilon.
nemo_config.layernorm_epsilon = mistral_config['norm_eps']
Expand Down

0 comments on commit 4eb7ce4

Please sign in to comment.