File tree Expand file tree Collapse file tree 1 file changed +2
-1
lines changed
Filter options
Expand file tree Collapse file tree 1 file changed +2
-1
lines changed
Original file line number Diff line number Diff line change 6
6
7
7
import llama_cpp
8
8
import llama_cpp .llama_speculative as llama_speculative
9
+ import llama_cpp .llama_tokenizer as llama_tokenizer
9
10
10
11
from llama_cpp .server .settings import ModelSettings
11
12
@@ -95,7 +96,7 @@ def load_llama_from_model_settings(settings: ModelSettings) -> llama_cpp.Llama:
95
96
96
97
tokenizer : Optional [llama_cpp .BaseLlamaTokenizer ] = None
97
98
if settings .hf_pretrained_model_name_or_path is not None :
98
- tokenizer = llama_cpp .LlamaHFTokenizer .from_pretrained (settings .hf_pretrained_model_name_or_path )
99
+ tokenizer = llama_tokenizer .LlamaHFTokenizer .from_pretrained (settings .hf_pretrained_model_name_or_path )
99
100
100
101
draft_model = None
101
102
if settings .draft_model is not None :
You can’t perform that action at this time.
0 commit comments