Skip to content

Conversation

@mnoukhov
Copy link
Contributor

@mnoukhov mnoukhov commented Jun 1, 2023

added steps argument and break to respect max training epochs added more PPOConfig args to script args
removed llama tokenizer hacks
black + isort

switched to LlamaTokenizer from AutoTokenizer added return_token_type_ids=False to pipeline kwargs because

added steps argument and break to respect max training epochs
added more PPOConfig args to script args
removed llama tokenizer hacks
removed extra args in dataset
changed to llamatokenizer from autotokenizer
black + isort
@mnoukhov mnoukhov mentioned this pull request Jun 1, 2023
@ArthurZucker
Copy link
Contributor

You can use return_token_type_ids=False as a quick fix to prevent them from being passed

@HuggingFaceDocBuilderDev
Copy link

HuggingFaceDocBuilderDev commented Jun 2, 2023

The documentation is not available anymore as the PR was closed or merged.

Copy link
Contributor

@younesbelkada younesbelkada left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

THank you so much for your contribution! Could you just run the styling checks? After that we should be good for merging

make style && make quality

@mnoukhov
Copy link
Contributor Author

mnoukhov commented Jun 4, 2023

Fixed style and quality and Switched back to Autotokenizer, thanks for the tip @ArthurZucker

Copy link
Contributor

@younesbelkada younesbelkada left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks so much!

@younesbelkada younesbelkada requested a review from lvwerra June 5, 2023 08:27
@younesbelkada younesbelkada merged commit a4793c2 into huggingface:main Jun 5, 2023
yxliu-TAMU pushed a commit to mincheolseong/ECEN743-GRPO-Project-Proposal that referenced this pull request Apr 20, 2025
* fixed rl training args

added steps argument and break to respect max training epochs
added more PPOConfig args to script args
removed llama tokenizer hacks
removed extra args in dataset
changed to llamatokenizer from autotokenizer
black + isort

* black and flake8

* style, quality, and switch back to AutoTokenizer
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants