-
Notifications
You must be signed in to change notification settings - Fork 2.2k
Dft #3960
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Dft #3960
Conversation
Tgt ent control
thanks @1485840691 can we just have the sft script example only in this PR? |
@kashif . Thanks, aligned with main branch. The code is from a fork from the trl:main branch. But I created another PR and submit to the main branch of the fork. Seems my account could only create a single fork from trl. So this results in the un-related commits in the change history. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks a lot for this contribution!
I'd prefer having this directly in SFTTrainer.compute_loss
. Let's say with a new arg loss_type="dft"
(default to "cross_entropy"
). What do you think?
Closing via #4042 |
What does this PR do?
Support dynamic fine tuning
Fixes #3877 (issue)
Before submitting
Pull Request section?
to it if that's the case.
Who can review?
Anyone in the community is free to review the PR once the tests have passed. Feel free to tag
members/contributors who may be interested in your PR.