You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi! Thank you for sharing a good initial implementation of lookahead decoding!
I'm just curious on how to set initial W tokens when starting lookahead decoding.
(I tried to find a code related to it but couldn't.)
It seems like this initial setup can impact decoding speed somehow since if those words are chosen randomly, no real match might be retrieved from the n-gram pools.
The text was updated successfully, but these errors were encountered:
Actually, our method includes a multi-level lookahead window, and we only set one level at the beginning. It will run several warmup steps to fill the whole 2-D window.
Currently, we are copying the init tokens from the prompt. Actually, althrough tokens are randomly selected from the vocabulary, lookahead decoding can still converge and achieve speedup, thanks to the magic Jacobi iteration.
Hi! Thank you for sharing a good initial implementation of lookahead decoding!
I'm just curious on how to set initial W tokens when starting lookahead decoding.
(I tried to find a code related to it but couldn't.)
It seems like this initial setup can impact decoding speed somehow since if those words are chosen randomly, no real match might be retrieved from the n-gram pools.
The text was updated successfully, but these errors were encountered: