Skip to content

Conversation

@mnoukhov
Copy link
Contributor

@mnoukhov mnoukhov commented Jun 1, 2023

fixes #368 #386 #297 and partially #327
correctly merge weights with peft's merge_and_unload
if merging a reward model, load sequence classification

Note that trl-lib/llama-7b-se-rm-peft does not contain the score weight so I will release my own pretrained reward model adapter weights

correctly merge weights with peft's merge_and_unload
load sequence classification model for reward models
@mnoukhov mnoukhov mentioned this pull request Jun 1, 2023
@HuggingFaceDocBuilderDev
Copy link

HuggingFaceDocBuilderDev commented Jun 2, 2023

The documentation is not available anymore as the PR was closed or merged.

Copy link
Contributor

@younesbelkada younesbelkada left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Amazing work @mnoukhov !
Could you run the styling checks?

make style && make quality

After that we should be good for merging
Thanks!

@mnoukhov
Copy link
Contributor Author

mnoukhov commented Jun 4, 2023

Done!

@mnoukhov mnoukhov requested a review from younesbelkada June 4, 2023 03:22
Copy link
Contributor

@younesbelkada younesbelkada left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks a lot!

@younesbelkada younesbelkada requested a review from lvwerra June 5, 2023 08:16
@younesbelkada younesbelkada merged commit 0ddf9f6 into huggingface:main Jun 5, 2023
yxliu-TAMU pushed a commit to mincheolseong/ECEN743-GRPO-Project-Proposal that referenced this pull request Apr 20, 2025
* correctly merge stackllama models

correctly merge weights with peft's merge_and_unload
load sequence classification model for reward models

* style, black line length 119

* flake8
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Llama Reward Model is incorrectly merged

4 participants