You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When constructing the n-gram lexicon for memory, did you use test dataset? I notice in run.sh, in --eval_data_path the value is the test dataset. But from the code, I assume constructing the lexicon uses only train and eval data, and test set is used for --do_test option.
Moreover, if the test set is used to construct lexicon, some of the words and their statistical features are known to the model.
It doesn't feel right for experiments.
The text was updated successfully, but these errors were encountered:
When constructing the n-gram lexicon for memory, did you use test dataset? I notice in run.sh, in --eval_data_path the value is the test dataset. But from the code, I assume constructing the lexicon uses only train and eval data, and test set is used for --do_test option.
Moreover, if the test set is used to construct lexicon, some of the words and their statistical features are known to the model.
It doesn't feel right for experiments.
The text was updated successfully, but these errors were encountered: