mirror of
https://github.com/ParisNeo/lollms-webui.git
synced 2024-12-21 05:13:10 +00:00
8 lines
269 B
Bash
8 lines
269 B
Bash
|
echo "this will recompile llapacpp to use your hardware with gpu enabled."
|
||
|
pip uninstall llama-cpp-python -y
|
||
|
# First we need to purge any old installation
|
||
|
pip cache purge
|
||
|
export CMAKE_ARGS="-DLLAMA_CUBLAS=on"
|
||
|
export FORCE_CMAKE=1
|
||
|
pip install llama-cpp-python --upgrade
|