lollms-webui/bindings/gpt_4all/install.py

46 lines
1.8 KiB
Python
Raw Normal View History

2023-05-21 23:43:11 +00:00
import subprocess
from pathlib import Path
import requests
from tqdm import tqdm
class Install:
def __init__(self, api):
# Get the current directory
current_dir = Path(__file__).resolve().parent
install_file = current_dir / ".installed"
if not install_file.exists():
2023-05-25 21:24:14 +00:00
print("-------------- GPT4All binding by nomic-ai -------------------------------")
print("This is the first time you are using this binding.")
2023-05-21 23:43:11 +00:00
print("Installing ...")
2023-05-22 10:27:45 +00:00
"""
2023-05-21 23:43:11 +00:00
try:
print("Checking pytorch")
import torch
import torchvision
if torch.cuda.is_available():
print("CUDA is supported.")
else:
print("CUDA is not supported. Reinstalling PyTorch with CUDA support.")
self.reinstall_pytorch_with_cuda()
except Exception as ex:
2023-05-22 10:27:45 +00:00
self.reinstall_pytorch_with_cuda()
"""
2023-05-21 23:43:11 +00:00
# Step 2: Install dependencies using pip from requirements.txt
requirements_file = current_dir / "requirements.txt"
subprocess.run(["pip", "install", "--no-cache-dir", "-r", str(requirements_file)])
# Create ther models folder
models_folder = Path("./models/c_transformers")
models_folder.mkdir(exist_ok=True, parents=True)
#Create the install file
with open(install_file,"w") as f:
f.write("ok")
print("Installed successfully")
def reinstall_pytorch_with_cuda(self):
subprocess.run(["pip", "install", "torch", "torchvision", "torchaudio", "--no-cache-dir", "--index-url", "https://download.pytorch.org/whl/cu117"])