mirror of
https://github.com/ParisNeo/lollms-webui.git
synced 2024-12-19 04:17:52 +00:00
7 lines
261 B
Batchfile
7 lines
261 B
Batchfile
echo this will recompile llapacpp to use your hardware with gpu enabled.
|
|
pip uninstall llama-cpp-python -y
|
|
rem First we need to purge any old installation
|
|
pip cache purge
|
|
set CMAKE_ARGS=-DLLAMA_CUBLAS=on
|
|
set FORCE_CMAKE=1
|
|
pip install llama-cpp-python --upgrade |