mirror of
https://github.com/ParisNeo/lollms-webui.git
synced 2024-12-23 22:32:24 +00:00
36 lines
1.5 KiB
Python
36 lines
1.5 KiB
Python
import subprocess
|
|
from pathlib import Path
|
|
import requests
|
|
from tqdm import tqdm
|
|
|
|
class Install:
|
|
def __init__(self, api):
|
|
# Get the current directory
|
|
current_dir = Path(__file__).resolve().parent
|
|
install_file = current_dir / ".installed"
|
|
|
|
if not install_file.exists():
|
|
print("-------------- llama_cpp_official backend -------------------------------")
|
|
print("This is the first time you are using this backend.")
|
|
print("Installing ...")
|
|
try:
|
|
print("Checking pytorch")
|
|
import torch
|
|
import torchvision
|
|
if torch.cuda.is_available():
|
|
print("CUDA is supported.")
|
|
else:
|
|
print("CUDA is not supported. Reinstalling PyTorch with CUDA support.")
|
|
self.reinstall_pytorch_with_cuda()
|
|
except Exception as ex:
|
|
self.reinstall_pytorch_with_cuda()
|
|
# Step 2: Install dependencies using pip from requirements.txt
|
|
requirements_file = current_dir / "requirements.txt"
|
|
subprocess.run(["pip", "install", "--no-cache-dir", "-r", str(requirements_file)])
|
|
with open(install_file,"w") as f:
|
|
f.write("ok")
|
|
print("Installed successfully")
|
|
|
|
def reinstall_pytorch_with_cuda(self):
|
|
subprocess.run(["pip", "install", "torch", "torchvision", "torchaudio", "--no-cache-dir", "--index-url", "https://download.pytorch.org/whl/cu117"])
|
|
|