File tree Expand file tree Collapse file tree 2 files changed +5
-1
lines changed
Filter options
Expand file tree Collapse file tree 2 files changed +5
-1
lines changed
Original file line number Diff line number Diff line change @@ -526,6 +526,7 @@ class llama_context_params(Structure):
526
526
# bool quantize_output_tensor; // quantize output.weight
527
527
# bool only_copy; // only copy tensors - ftype, allow_requantize and quantize_output_tensor are ignored
528
528
# bool pure; // disable k-quant mixtures and quantize all tensors to the same type
529
+ # void * imatrix; // pointer to importance matrix data
529
530
# } llama_model_quantize_params;
530
531
class llama_model_quantize_params (Structure ):
531
532
"""Parameters for llama_model_quantize
@@ -537,6 +538,7 @@ class llama_model_quantize_params(Structure):
537
538
quantize_output_tensor (bool): quantize output.weight
538
539
only_copy (bool): only copy tensors - ftype, allow_requantize and quantize_output_tensor are ignored
539
540
pure (bool): disable k-quant mixtures and quantize all tensors to the same type
541
+ imatrix (ctypes.c_void_p): pointer to importance matrix data
540
542
"""
541
543
542
544
_fields_ = [
@@ -545,6 +547,8 @@ class llama_model_quantize_params(Structure):
545
547
("allow_requantize" , c_bool ),
546
548
("quantize_output_tensor" , c_bool ),
547
549
("only_copy" , c_bool ),
550
+ ("pure" , c_bool ),
551
+ ("imatrix" , c_void_p ),
548
552
]
549
553
550
554
You can’t perform that action at this time.
0 commit comments