Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions regarding calculation of the loss #406

Open
NLKaiser opened this issue Jan 30, 2025 · 0 comments
Open

Questions regarding calculation of the loss #406

NLKaiser opened this issue Jan 30, 2025 · 0 comments

Comments

@NLKaiser
Copy link

Dear Bonito developers,

I have some questions about the functions used to calculate the loss in Bonito, which unfortunately I cannot answer myself from the Bonito or Koi code or other issues like #101.

Can you explain in more detail how the forward and backward calculations work?
Do you use a transition matrix somewhere, similar to how it is done in HMMs?
Are the stay_scores and move_scores interleaved at any point, as it is done with the blanks and labels in CTC loss?

I do not yet understand how the normalisation is calculated via 'logZ_cu_sparse'. Can you provide further information here?

It is very interesting how quickly Bonito reaches high accuracies after only a few epochs. Do you set any special constraints for the forward and backward calculations to achieve this?

For decoding the output during inference, the Dorado implementation appears to use a special variant of Beam Search. Does your version of the Viterbi algorithm work similarly during training with Bonito?

Are there any papers or other resources that explain the specific idea behind the functions used to calculate the loss?

I hope that these questions can also help other developers. Thank you very much for your time.

With kind regards

Nick

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant