|
57 | 57 | from executorch.examples.models.llama.source_transformation.quantize import ( |
58 | 58 | get_quant_embedding_transform, |
59 | 59 | ) |
60 | | -from executorch.examples.models.llama.tokenizer.tiktoken import Tokenizer as Tiktoken |
61 | 60 | from executorch.examples.qualcomm.oss_scripts.llama.model.static_llama import ( |
62 | 61 | LlamaModel, |
63 | 62 | ModelArgs, |
|
77 | 76 | from executorch.extension.llm.export.builder import DType |
78 | 77 | from pytorch_tokenizers import get_tokenizer |
79 | 78 | from pytorch_tokenizers.llama2c import Llama2cTokenizer as SentencePieceTokenizer |
| 79 | +from pytorch_tokenizers import TiktokenTokenizer |
80 | 80 |
|
81 | 81 | from torch.ao.quantization.observer import MinMaxObserver |
82 | 82 | from torch.ao.quantization.quantize_pt2e import convert_pt2e, prepare_pt2e |
@@ -141,7 +141,7 @@ def _kv_calibrate( |
141 | 141 | # Llama2 tokenizer has no special tokens |
142 | 142 | if isinstance(tokenizer, SentencePieceTokenizer): |
143 | 143 | token_list = tokenizer.encode(user_prompts, bos=True, eos=False) |
144 | | - elif isinstance(tokenizer, Tiktoken): |
| 144 | + elif isinstance(tokenizer, TiktokenTokenizer): |
145 | 145 | token_list = tokenizer.encode( |
146 | 146 | user_prompts, bos=True, eos=False, allowed_special="all" |
147 | 147 | ) |
@@ -213,7 +213,7 @@ def _prefill_calibrate( |
213 | 213 | # Llama2 tokenizer has no special tokens |
214 | 214 | if isinstance(tokenizer, SentencePieceTokenizer): |
215 | 215 | token_list = tokenizer.encode(user_prompts, bos=True, eos=False) |
216 | | - elif isinstance(tokenizer, Tiktoken): |
| 216 | + elif isinstance(tokenizer, TiktokenTokenizer): |
217 | 217 | token_list = tokenizer.encode( |
218 | 218 | user_prompts, bos=True, eos=False, allowed_special="all" |
219 | 219 | ) |
@@ -1111,7 +1111,7 @@ def export_llama(args) -> None: |
1111 | 1111 | runtime_tokenizer_path = args.tokenizer_bin |
1112 | 1112 | elif args.llama_model == "llama3_2": |
1113 | 1113 | assert isinstance( |
1114 | | - tokenizer, Tiktoken |
| 1114 | + tokenizer, TiktokenTokenizer |
1115 | 1115 | ), f"Wrong tokenizer provided for llama3_2." |
1116 | 1116 | runtime_tokenizer_path = args.tokenizer_model |
1117 | 1117 | else: |
|
0 commit comments