File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed
Filter options
Expand file tree Collapse file tree 1 file changed +3
-3
lines changed
Original file line number Diff line number Diff line change @@ -102,10 +102,10 @@ CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install llama-cpp-
102
102
<details >
103
103
<summary >cuBLAS (CUDA)</summary >
104
104
105
- To install with cuBLAS, set the ` LLAMA_CUBLAS =on` environment variable before installing:
105
+ To install with cuBLAS, set the ` LLAMA_CUDA =on` environment variable before installing:
106
106
107
107
``` bash
108
- CMAKE_ARGS=" -DLLAMA_CUBLAS =on" pip install llama-cpp-python
108
+ CMAKE_ARGS=" -DLLAMA_CUDA =on" pip install llama-cpp-python
109
109
```
110
110
111
111
</details >
@@ -569,7 +569,7 @@ python3 -m llama_cpp.server --model models/7B/llama-model.gguf
569
569
Similar to Hardware Acceleration section above, you can also install with GPU (cuBLAS) support like this:
570
570
571
571
``` bash
572
- CMAKE_ARGS=" -DLLAMA_CUBLAS =on" FORCE_CMAKE=1 pip install ' llama-cpp-python[server]'
572
+ CMAKE_ARGS=" -DLLAMA_CUDA =on" FORCE_CMAKE=1 pip install ' llama-cpp-python[server]'
573
573
python3 -m llama_cpp.server --model models/7B/llama-model.gguf --n_gpu_layers 35
574
574
```
575
575
You can’t perform that action at this time.
0 commit comments