File tree Expand file tree Collapse file tree 1 file changed +9
-7
lines changed
Filter options
Expand file tree Collapse file tree 1 file changed +9
-7
lines changed
Original file line number Diff line number Diff line change @@ -89,16 +89,18 @@ deploy.pyinstaller.mac:
89
89
@if [ ` uname -m` != " arm64" ]; then echo " Must be on aarch64" ; exit 1; fi
90
90
@if [ ` uname` != " Darwin" ]; then echo " Must be on MacOS" ; exit 1; fi
91
91
@echo " Building and installing with proper env vars for aarch64-specific ops"
92
- CMAKE_ARGS=" -DGGML_METAL=off -DGGML_LLAMAFILE=OFF -DGGML_BLAS=OFF -DCMAKE_BUILD_TYPE=Release" python3 -m pip install -v -e .[server,dev]
92
+
93
+ # This still builds with metal support (I think b/c GGML_NATIVE=ON). Not an
94
+ # issue since can still run Q4_0 models w/ repacking support on CPU if `-ngl 0`.
95
+ CMAKE_ARGS="-DGGML_METAL=OFF -DGGML_LLAMAFILE=OFF -DGGML_BLAS=OFF \
96
+ -DGGML_NATIVE=ON -DGGML_CPU_AARCH64=ON \
97
+ -DCMAKE_BUILD_TYPE=Release" python3 -m pip install -v -e .[server,dev]
93
98
@server_path=$$(python -c 'import llama_cpp.server; print(llama_cpp.server.__file__)' | sed s/init/main/) ; \
94
99
echo "Server path: $$server_path" ; \
95
- libllama_path=$$(python -c 'import llama_cpp.llama_cpp; print(llama_cpp.llama_cpp._load_shared_library("llama" ) ._name)' ) ; \
96
- libggml_path=$$(python -c 'import llama_cpp.llama_cpp; print(llama_cpp.llama_cpp._load_shared_library("ggml" ) ._name)' ) ; \
97
- echo " libllama path: $$ libllama_path" ; \
98
- echo " libggml path: $$ libggml_path" ; \
100
+ base_path=$$(python -c 'from llama_cpp._ggml import libggml_base_path; print(str(libggml_base_path))') ; \
101
+ echo "Base path: $$base_path" ; \
99
102
pyinstaller -DF $$server_path \
100
- --add-data $$ libllama_path:llama_cpp/lib \
101
- --add-data $$ libggml_path:llama_cpp/lib \
103
+ --add-data $$base_path:llama_cpp/lib \
102
104
-n llama-cpp-py-server
103
105
104
106
test :
You can’t perform that action at this time.
0 commit comments