2023-04-09 15:53:35 +00:00
#!/usr/bin/bash
echo HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHH .HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHH. ,HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHH.## HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHHH#.HHHHH/*,*,*,*,*,*,*,*,***,*,**#HHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHHHHH.*,,***,***,***,***,***,***,*******HHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHHH*,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,*,,,,,HHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHH.,,,***,***,***,***,***,***,***,***,***,***/HHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHH*,,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*HHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHH#,***,***,***,***,***,***,***,***,***,***,***,**HHHHHHHHHHHHHHHHH
echo HHHHHHHHHH..HHH,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,*#HHHHHHHHHHHHHHHH
echo HHHHHHH,,,**,/H*,***,***,***,,,*,***,***,***,**,,,**,***,***,***H,,*,***HHHHHHHH
echo HHHHHH.*,,,*,,,,,*,*,*,***#HHHHH.,,*,*,*,*,**/HHHHH.,*,*,*,*,*,*,*,*****HHHHHHHH
echo HHHHHH.*,***,*,*,***,***,.HHHHHHH/**,***,****HHHHHHH.***,***,***,*******HHHHHHHH
echo HHHHHH.,,,,,,,,,,,,,,,,,,,.HHHHH.,,,,,,,,,,,,.HHHHHH,,,,,,,,,,,,,,,,,***HHHHHHHH
echo HHHHHH.,,,,,,/H,,,**,***,***,,,*,***,***,***,**,,,,*,***,***,***H***,***HHHHHHHH
echo HHHHHHH.,,,,*.H,,,,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,***H*,,,,/HHHHHHHHH
echo HHHHHHHHHHHHHHH*,***,***,**,,***,***,***,***,***,***,***,***,**.HHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHH,,,,,,,,*,,#H#,,,,,*,,,*,,,,,,,,*#H*,,,,,,,,,**HHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHH,,*,***,***,**/.HHHHHHHHHHHHH#*,,,*,***,***,*HHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHH,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*HHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHH**,***,***,***,***,***,***,***,***,***,***,*.HHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHH*,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,*HHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHH**,***,***,*******/..HHHHHHHHH.#/*,*,,,***,***HHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHH*,*,*,******#HHHHHHHHHHHHHHHHHHHHHHHHHHHH./**,,,.HHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHH.,,*,***.HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH.*#HHHHHHHHHHHH
echo HHHHHHHHHHHHHHH/,,,*.HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHH,,#HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHH.HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
echo HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH
# Install Python 3.11 and pip
echo -n "Checking for python3.11..."
if command -v python3.11 > /dev/null 2>& 1; then
echo "OK"
else
read -p "Python3.11 is not installed. Would you like to install Python3.11? [Y/N] " choice
if [ " $choice " = "Y" ] || [ " $choice " = "y" ] ; then
echo "Installing Python3.11..."
brew install python@3.11
else
echo "Please install Python3.11 and try again."
exit 1
fi
fi
#check if cmake is installed, if not install it
if [ " $( command -v cmake) " = "" ] ; then
echo "cmake not found, installing cmake ..."
brew install cmake
fi
# check if nproc is installed, if not install it
if [ " $( command -v nproc) " = "" ] ; then
echo "nproc not found, installing nproc ..."
brew install coreutils
fi
# Install venv module
echo -n "Checking for venv module..."
if python3.11 -m venv env > /dev/null 2>& 1; then
echo "OK"
else
read -p "venv module is not available. Would you like to install it? [Y/N] " choice
if [ " $choice " = "Y" ] || [ " $choice " = "y" ] ; then
echo "Installing venv module..."
python3.11 -m ensurepip --upgrade
python3.11 -m pip install virtualenv
else
echo "Please install venv module and try again."
exit 1
fi
fi
# Create a new virtual environment
echo -n "Creating virtual environment..."
python3.11 -m venv env
if [ $? -ne 0 ] ; then
echo "Failed to create virtual environment. Please check your Python installation and try again."
exit 1
else
echo "OK"
fi
# Activate the virtual environment
echo -n "Activating virtual environment..."
source env/bin/activate
echo "OK"
# Install the required packages
echo "Installing requirements..."
export DS_BUILD_OPS = 0
export DS_BUILD_AIO = 0
python3.11 -m pip install pip --upgrade
python3.11 -m pip install -r requirements.txt
if [ $? -ne 0 ] ; then
echo "Failed to install required packages. Please check your internet connection and try again."
exit 1
fi
echo Downloading latest model
curl -o models/gpt4all-lora-quantized-ggml.bin https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-quantized-ggml.bin
if [ $? -ne 0 ] ; then
echo "Failed to download model. Please check your internet connection and try again."
exit 1
fi
2023-04-13 06:08:49 +00:00
echo "Downloading latest model"
if [ ! -d "models" ] ; then
mkdir models
fi
if [ ! -f "models/gpt4all-lora-quantized-ggml.bin" ] ; then
echo ""
read -p "The default model file (gpt4all-lora-quantized-ggml.bin) does not exist. Do you want to download it? Press Y to download it with a browser (faster)." yn
case $yn in
[ Yy] * ) open "https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-quantized-ggml.bin"
echo "Link has been opened with the default web browser, make sure to save it into the models folder before continuing. Press any key to continue..."
read -n 1 -s; ;
* ) echo "Skipping download of model file..." ; ;
esac
else
echo ""
read -p "The default model file (gpt4all-lora-quantized-ggml.bin) already exists. Do you want to replace it? Press Y to download it with a browser (faster)." yn
case $yn in
[ Yy] * ) open "https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-quantized-ggml.bin"
echo "Link has been opened with the default web browser, make sure to save it into the models folder before continuing. Press any key to continue..."
read -n 1 -s; ;
* ) echo "Skipping download of model file..." ; ;
esac
fi
2023-04-13 22:22:40 +00:00
#!/bin/bash
2023-04-13 06:08:49 +00:00
echo ""
echo "Downloading latest model..."
curl -o "models/gpt4all-lora-quantized-ggml.bin" "https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-quantized-ggml.bin"
if [ $? -ne 0 ] ; then
echo "Failed to download model. Please check your internet connection."
read -p "Do you want to try downloading again? Press Y to download." yn
case $yn in
[ Yy] * ) echo "Downloading latest model..."
curl -o "models/gpt4all-lora-quantized-ggml.bin" "https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-quantized-ggml.bin" ; ;
* ) echo "Skipping download of model file..." ; ;
esac
else
echo "Model successfully downloaded."
fi
echo ""
echo "In order to make a model work, it needs to go through the LLaMA tokenizer, this will fix errors with the model in run.bat. Do you want to convert the model?"
read -p "Press Y to convert or N to skip: " yn
case $yn in
[ Yy] * )
echo ""
echo "Select a model to convert:"
count = 0
for f in models/*; do
count = $(( count+1))
file[ $count ] = $f
echo " [ $count ] $f "
done
2023-04-13 22:22:40 +00:00
# Prompt user to choose a model to convert
2023-04-13 06:08:49 +00:00
read -p "Enter the number of the model you want to convert: " modelNumber
if [ -z " ${ file [modelNumber] } " ] ; then
2023-04-13 22:22:40 +00:00
echo ""
echo "Invalid option. Restarting..."
exit 1
2023-04-13 06:08:49 +00:00
fi
modelPath = " ${ file [modelNumber] } "
echo ""
echo " You selected $modelPath "
2023-04-13 22:22:40 +00:00
# Ask user if they want to convert the model
2023-04-13 06:08:49 +00:00
echo ""
read -p "Do you want to convert the selected model to the new format? (Y/N)" choice
if [ " $choice " = = "N" ] ; then
2023-04-13 22:22:40 +00:00
echo ""
echo "Model conversion cancelled. Skipping..."
else
# Convert the model
echo "Converting the model to the new format..."
if [ ! -d "tmp/llama.cpp" ] ; then
git clone https://github.com/ggerganov/llama.cpp.git tmp/llama.cpp
fi
mv -f " ${ modelPath } " " ${ modelPath } .original "
python tmp/llama.cpp/migrate-ggml-2023-03-30-pr613.py " ${ modelPath } .original " " ${ modelPath } "
if [ $? -ne 0 ] ; then
echo ""
echo "Error during model conversion. Restarting..."
mv -f " ${ modelPath } .original " " ${ modelPath } "
exit 1
else
echo ""
echo " The model file ( ${ modelPath } ) has been converted to the new format. "
fi
2023-04-13 06:08:49 +00:00
fi
2023-04-13 22:22:40 +00:00
; ;
* ) echo ""
echo "Conversion cancelled. Skipping..."
; ;
2023-04-13 06:08:49 +00:00
esac
echo ""
echo "Cleaning tmp folder"
rm -rf "./tmp"
2023-04-09 15:53:35 +00:00
echo "Virtual environment created and packages installed successfully."
echo "Everything is set up. Just run run.sh"
exit 0