mirror of
https://github.com/ParisNeo/lollms-webui.git
synced 2024-12-19 04:17:52 +00:00
8 lines
269 B
Bash
8 lines
269 B
Bash
echo "this will recompile llapacpp to use your hardware with gpu enabled."
|
|
pip uninstall llama-cpp-python -y
|
|
# First we need to purge any old installation
|
|
pip cache purge
|
|
export CMAKE_ARGS="-DLLAMA_CUBLAS=on"
|
|
export FORCE_CMAKE=1
|
|
pip install llama-cpp-python --upgrade
|