Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Possible bug in fine-tuning baseline implementation #41

Open
drd13 opened this issue Feb 2, 2024 · 1 comment
Open

Possible bug in fine-tuning baseline implementation #41

drd13 opened this issue Feb 2, 2024 · 1 comment

Comments

@drd13
Copy link

drd13 commented Feb 2, 2024

I've found what I believe to be a bug in the implementation of the fine-tuning baseline which would yield incorrect results when the target is longer than one token.

Looking at the code, the fine-tuning baseline seems to get the logits on which to backpropagate by calling model(**inputs) where inputs are the prompt with the subject but excluding the target. It then maximises the probability of the target by taking the logits associated to the last token in the input, and maximising the probability of all the target tokens as simultaneous direct continuations. This is not the regular fine-tuning behaviour which would be to maximise the probability of the first token in the target being a continuation to the input and then maximising the probability of the second token in the target being a continuation to the first token in the target.

Thank you for your assistance and look forwards to hearing back and understanding whether I may have misunderstood an aspect in the implementation.

@drd13
Copy link
Author

drd13 commented Mar 18, 2024

Just wanted to point out that there is some relevant discussion around this issue in zjunlp/EasyEdit#173

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant