You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm experimenting with chain of thoughts and finetuning, is there a way to use this library to assign different reward for each steps towards the problem resolution of the LLM? Like the new technique proposed in the step by step paper from OpenAI.
Would be amazing to experiment with it
The text was updated successfully, but these errors were encountered:
Currently only a reward can be given to the last token and it discounted towards the beginning. The KL-div however is added per token and you could adapt the compute_rewards method to accept lists of scores for each sample.
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
I'm experimenting with chain of thoughts and finetuning, is there a way to use this library to assign different reward for each steps towards the problem resolution of the LLM? Like the new technique proposed in the step by step paper from OpenAI.
Would be amazing to experiment with it
The text was updated successfully, but these errors were encountered: