Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Commit 8740ddc

Browse filesBrowse files
committed
Only support generating one prompt at a time.
1 parent 8895b90 commit 8740ddc
Copy full SHA for 8740ddc

File tree

Expand file treeCollapse file tree

1 file changed

+4
-3
lines changed
Filter options
Expand file treeCollapse file tree

1 file changed

+4
-3
lines changed

‎llama_cpp/server/app.py

Copy file name to clipboardExpand all lines: llama_cpp/server/app.py
+4-3Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -166,10 +166,10 @@ def get_llama():
166166
description="Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.",
167167
)
168168

169+
169170
class CreateCompletionRequest(BaseModel):
170171
prompt: Union[str, List[str]] = Field(
171-
default="",
172-
description="The prompt to generate completions for."
172+
default="", description="The prompt to generate completions for."
173173
)
174174
suffix: Optional[str] = Field(
175175
default=None,
@@ -224,7 +224,8 @@ def create_completion(
224224
request: CreateCompletionRequest, llama: llama_cpp.Llama = Depends(get_llama)
225225
):
226226
if isinstance(request.prompt, list):
227-
request.prompt = "".join(request.prompt)
227+
assert len(request.prompt) <= 1
228+
request.prompt = request.prompt[0] if len(request.prompt) > 0 else ""
228229

229230
completion_or_chunks = llama(
230231
**request.dict(

0 commit comments

Comments
0 (0)
Morty Proxy This is a proxified and sanitized view of the page, visit original site.