Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Commit 389e09c

Browse filesBrowse files
authored
misc: Remove unnecessary metadata lookups (abetlen#1448)
Special tokens are already mapped from metadata by llama.cpp
1 parent 4b54f79 commit 389e09c
Copy full SHA for 389e09c

File tree

Expand file treeCollapse file tree

1 file changed

+5
-5
lines changed
Filter options
Expand file treeCollapse file tree

1 file changed

+5
-5
lines changed

‎llama_cpp/llama.py

Copy file name to clipboardExpand all lines: llama_cpp/llama.py
+5-5Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -410,8 +410,8 @@ def __init__(
410410
if self.verbose:
411411
print(f"Model metadata: {self.metadata}", file=sys.stderr)
412412

413-
eos_token_id = int(self.metadata.get("tokenizer.ggml.eos_token_id", self.token_eos()))
414-
bos_token_id = int(self.metadata.get("tokenizer.ggml.bos_token_id", self.token_bos()))
413+
eos_token_id = self.token_eos()
414+
bos_token_id = self.token_bos()
415415

416416
eos_token = self._model.token_get_text(eos_token_id)
417417
bos_token = self._model.token_get_text(bos_token_id)
@@ -961,9 +961,9 @@ def _create_completion(
961961

962962
completion_id: str = f"cmpl-{str(uuid.uuid4())}"
963963
created: int = int(time.time())
964-
prefix_token_id: int = int(self.metadata.get("tokenizer.ggml.prefix_token_id", self._model.token_prefix()))
965-
middle_token_id: int = int(self.metadata.get("tokenizer.ggml.middle_token_id", self._model.token_middle()))
966-
suffix_token_id: int = int(self.metadata.get("tokenizer.ggml.suffix_token_id", self._model.token_suffix()))
964+
prefix_token_id: int = self._model.token_prefix()
965+
middle_token_id: int = self._model.token_middle()
966+
suffix_token_id: int = self._model.token_suffix()
967967
# If prompt is empty, initialize completion with BOS token to avoid
968968
# detokenization including a space at the beginning of the completion
969969
completion_tokens: List[int] = [] if len(prompt) > 0 else [self.token_bos()]

0 commit comments

Comments
0 (0)
Morty Proxy This is a proxified and sanitized view of the page, visit original site.