Skip to content

Commit 4ba1ffb

Browse files
mergennachinmalfet
authored andcommitted
Bug fix in tokenizer config (pytorch#305)
Summary: It handles the case hhen params.json is explicitly given and the json doesn't contain mention about tokenizer config, but the command line does. Test Plan: python torchchat.py generate --device cpu --checkpoint-path /Users/mnachin/models/Meta-Llama-3-8B/original/consolidated.00.pth --params-path=/Users/mnachin/models/Meta-Llama-3-8B/original/params.json --temperature 0 --tiktoken
1 parent 80ffe93 commit 4ba1ffb

File tree

2 files changed

+5
-2
lines changed

2 files changed

+5
-2
lines changed

build/builder.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -367,7 +367,10 @@ def tokenizer_setting_to_name(tiktoken: bool = False) -> str:
367367
def validate_args(model: Transformer, tokenizer_args: TokenizerArgs):
368368
use_tiktoken = model.config.use_tiktoken
369369
is_tiktoken = tokenizer_args.is_tiktoken
370-
if use_tiktoken != is_tiktoken:
370+
371+
if use_tiktoken is None:
372+
model.config.use_tiktoken = is_tiktoken
373+
elif use_tiktoken != is_tiktoken:
371374
raise RuntimeError(f"model-specified tokenizer ({tokenizer_setting_to_name(use_tiktoken)} does not match provided tokenizer ({tokenizer_setting_to_name(is_tiktoken)}")
372375

373376
def resolve_model_name(model: str) -> str:

build/model.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ class ModelArgs:
3737
norm_eps: float = 1e-5
3838
multiple_of: int = 256
3939
ffn_dim_multiplier: Optional[int] = None
40-
use_tiktoken: bool = False
40+
use_tiktoken: Optional[bool] = None
4141

4242
def __post_init__(self):
4343
if self.n_local_heads == -1:

0 commit comments

Comments
 (0)