-
Notifications
You must be signed in to change notification settings - Fork 2.4k
StackLlama: fixed RL training and added args #400
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
added steps argument and break to respect max training epochs added more PPOConfig args to script args removed llama tokenizer hacks removed extra args in dataset changed to llamatokenizer from autotokenizer black + isort
|
You can use |
|
The documentation is not available anymore as the PR was closed or merged. |
younesbelkada
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
THank you so much for your contribution! Could you just run the styling checks? After that we should be good for merging
make style && make quality|
Fixed style and quality and Switched back to Autotokenizer, thanks for the tip @ArthurZucker |
younesbelkada
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks so much!
* fixed rl training args added steps argument and break to respect max training epochs added more PPOConfig args to script args removed llama tokenizer hacks removed extra args in dataset changed to llamatokenizer from autotokenizer black + isort * black and flake8 * style, quality, and switch back to AutoTokenizer
added steps argument and break to respect max training epochs added more PPOConfig args to script args
removed llama tokenizer hacks
black + isort
switched toaddedLlamaTokenizerfromAutoTokenizerreturn_token_type_ids=Falseto pipeline kwargs becauseLlamaTokenizerFastLlamaTokenizerFastwill outputtoken_type_idssee LLaMATokenizerFast works abnormally transformers#23818 and 🚨🚨 🚨🚨 [Tokenizer] attemp to fix add_token issues🚨🚨 🚨🚨 transformers#23909token_type_idscause an error in our reward modelpipeline, namelyTypeError: LlamaForSequenceClassification.forward() got an unexpected keyword argument 'token_type_ids'