-
Notifications
You must be signed in to change notification settings - Fork 3
Open
Labels
audience/technicalIssue primarily for technical review and service.Issue primarily for technical review and service.kind/increase-accuracykind/performancekind/reproducibilitykind/tuning-configurationkind/validationtraige/good first issueGood for newcomersGood for newcomerstriage/help wantedExtra attention is neededExtra attention is neededtriage/high-priority
Description
Kind of issue: [bug | feature-request-or-enhancement | Process Change | Support request]: Final fine tuning to the merge candidate: 233e882
Additional context
Last commit looks like our cold start performance is at parity with GPT2's pre-trained performance. It did run for 2 1/2 hours. Goals here:
- Fine tune the model search to a constrained range of at or near optimal values.
- Reduce the number of sub-trials and epochs.
- Maybe, see if we can get away with increasing the sequence length further and if it is worth it in terms of embedding performance.
Suggested Labels (If you don't know, that's ok): kind/performance kind/hpc hind/scientific
Metadata
Metadata
Assignees
Labels
audience/technicalIssue primarily for technical review and service.Issue primarily for technical review and service.kind/increase-accuracykind/performancekind/reproducibilitykind/tuning-configurationkind/validationtraige/good first issueGood for newcomersGood for newcomerstriage/help wantedExtra attention is neededExtra attention is neededtriage/high-priority