Skip to content

GRACE's performance collapes on LLAMA-3-8B #543

@wang-kee

Description

@wang-kee

Thanks a lot for the great works,

I want to ask about the performance of GRACE with LLAMA-3-8B model.

My reproduced results with GRACE, as well as the results reported by WISE, suggest that GRACE has very good performance (in terms of rewrite accuracy and localization accuracy) with both Llama-2-7b and Mistral-7b.

However, currently on my side it has very low performance with LLAMA-3-8B, even with only 1 edit sample (my accuracy for T=1 is less than 40% rewrite accuracy). Meanwhile, the training loss drops to nearly 0 so I could not find the reason for the collapsing performance. I have extensively tuned the hyper-parameters but nothing has worked yet.

Do you have any intuitions / suggestions to fix the performance issue with GRACE on LLAMA-3?

Thanks in advance.

Metadata

Metadata

Assignees

No one assigned

    Labels

    questionFurther information is requested

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions