Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some question #3

Open
lolsd opened this issue Mar 6, 2022 · 1 comment
Open

Some question #3

lolsd opened this issue Mar 6, 2022 · 1 comment

Comments

@lolsd
Copy link

lolsd commented Mar 6, 2022

When I ran your code on Github, I found some bugs:

you used "loss, sample_size, logging_output = super().valid_step(sample, model, criterion)" in fairseq/tasks/translation.py. The main code in "super().valid_step()" is "loss, sample_size, logging_output = criterion(model, sample)" and "criterion" is Class LabelSmoothedCrossEntropyCriterion. In Class LabelSmoothedCrossEntropyCriterion, you called function "model.mix()" which need parameter "pair_sample", but you provided the parameter "sample" in the valid_step.

When I fixed this bug, although I got 28.1 on WMT14EN-DE, I got 35.1 on iWSLt14DE-en instead of 36.2.

@Zhaoyi-Li21
Copy link

hello, have you noticed that the description on sampling policy (for iwlst-exps) "we sample (X' , Y') uniformly from all examples with similar target Y' lengths (|Y |−|Y'| ≤ 5)."? (tips: this line is from supplement of this paper, C.1-paragraph)
maybe this is a factor accounting for why your result is a little bit lower than that reported.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants