@@ -260,18 +260,18 @@ class CreateCompletionRequest(BaseModel):
260
260
presence_penalty : Optional [float ] = presence_penalty_field
261
261
frequency_penalty : Optional [float ] = frequency_penalty_field
262
262
logit_bias : Optional [Dict [str , float ]] = Field (None )
263
- logit_bias_type : Optional [Literal [ "input_ids" , "tokens" ] ] = Field (None )
263
+ logprobs : Optional [int ] = Field (None )
264
264
265
265
# ignored or currently unsupported
266
266
model : Optional [str ] = model_field
267
267
n : Optional [int ] = 1
268
- logprobs : Optional [int ] = Field (None )
269
268
best_of : Optional [int ] = 1
270
269
user : Optional [str ] = Field (None )
271
270
272
271
# llama.cpp specific parameters
273
272
top_k : int = top_k_field
274
273
repeat_penalty : float = repeat_penalty_field
274
+ logit_bias_type : Optional [Literal ["input_ids" , "tokens" ]] = Field (None )
275
275
276
276
class Config :
277
277
schema_extra = {
@@ -424,7 +424,6 @@ class CreateChatCompletionRequest(BaseModel):
424
424
presence_penalty : Optional [float ] = presence_penalty_field
425
425
frequency_penalty : Optional [float ] = frequency_penalty_field
426
426
logit_bias : Optional [Dict [str , float ]] = Field (None )
427
- logit_bias_type : Optional [Literal ["input_ids" , "tokens" ]] = Field (None )
428
427
429
428
# ignored or currently unsupported
430
429
model : Optional [str ] = model_field
@@ -434,6 +433,7 @@ class CreateChatCompletionRequest(BaseModel):
434
433
# llama.cpp specific parameters
435
434
top_k : int = top_k_field
436
435
repeat_penalty : float = repeat_penalty_field
436
+ logit_bias_type : Optional [Literal ["input_ids" , "tokens" ]] = Field (None )
437
437
438
438
class Config :
439
439
schema_extra = {
0 commit comments