Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Commit bfb42b7

Browse filesBrowse files
committed
Merge branch 'main' of github.com:abetlen/llama-cpp-python into main
2 parents c5de5d3 + 5e39a85 commit bfb42b7
Copy full SHA for bfb42b7

File tree

Expand file treeCollapse file tree

3 files changed

+7
-6
lines changed
Filter options
Expand file treeCollapse file tree

3 files changed

+7
-6
lines changed

‎.github/workflows/build-and-release.yaml

Copy file name to clipboardExpand all lines: .github/workflows/build-and-release.yaml
+2-2Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ jobs:
2929
python -m pip install -e .[all]
3030
3131
- name: Build wheels
32-
uses: pypa/cibuildwheel@v2.19.2
32+
uses: pypa/cibuildwheel@v2.20.0
3333
env:
3434
# disable repair
3535
CIBW_REPAIR_WHEEL_COMMAND: ""
@@ -56,7 +56,7 @@ jobs:
5656
platforms: linux/arm64
5757

5858
- name: Build wheels
59-
uses: pypa/cibuildwheel@v2.19.2
59+
uses: pypa/cibuildwheel@v2.20.0
6060
env:
6161
CIBW_SKIP: "*musllinux* pp*"
6262
CIBW_REPAIR_WHEEL_COMMAND: ""

‎.github/workflows/build-wheels-metal.yaml

Copy file name to clipboardExpand all lines: .github/workflows/build-wheels-metal.yaml
+1-1Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ jobs:
3030
python -m pip install -e .[all]
3131
3232
- name: Build wheels
33-
uses: pypa/cibuildwheel@v2.19.2
33+
uses: pypa/cibuildwheel@v2.20.0
3434
env:
3535
# disable repair
3636
CIBW_REPAIR_WHEEL_COMMAND: ""

‎llama_cpp/llama.py

Copy file name to clipboardExpand all lines: llama_cpp/llama.py
+4-3Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -777,11 +777,12 @@ def generate(
777777
else:
778778
break
779779
if longest_prefix > 0:
780-
if self.verbose:
781-
print("Llama.generate: prefix-match hit", file=sys.stderr)
782780
reset = False
783781
tokens = tokens[longest_prefix:]
784782
self.n_tokens = longest_prefix
783+
if self.verbose:
784+
print(f"Llama.generate: {longest_prefix} prefix-match hit, "
785+
f"remaining {len(tokens)} prompt tokens to eval", file=sys.stderr)
785786

786787
# Reset the model state
787788
if reset:
@@ -2159,7 +2160,7 @@ def from_pretrained(
21592160

21602161
files = [
21612162
file["name"] if isinstance(file, dict) else file
2162-
for file in hffs.ls(repo_id)
2163+
for file in hffs.ls(repo_id, recursive=True))
21632164
]
21642165

21652166
# split each file into repo_id, subfolder, filename

0 commit comments

Comments
0 (0)
Morty Proxy This is a proxified and sanitized view of the page, visit original site.