-
Notifications
You must be signed in to change notification settings - Fork 315
Description
Thanks a lot for the great works,
I want to ask about the performance of GRACE with LLAMA-3-8B model.
My reproduced results with GRACE, as well as the results reported by WISE, suggest that GRACE has very good performance (in terms of rewrite accuracy and localization accuracy) with both Llama-2-7b and Mistral-7b.
However, currently on my side it has very low performance with LLAMA-3-8B, even with only 1 edit sample (my accuracy for T=1 is less than 40% rewrite accuracy). Meanwhile, the training loss drops to nearly 0 so I could not find the reason for the collapsing performance. I have extensively tuned the hyper-parameters but nothing has worked yet.
Do you have any intuitions / suggestions to fix the performance issue with GRACE on LLAMA-3?
Thanks in advance.