lollms-webui/scripts/wsl/petals_server.sh
2023-09-27 13:48:48 +02:00

21 lines
658 B
Bash

#!/bin/bash
echo changing dir to lollms-webui
cd ~/lollms-webui
# activate conda environment
echo sourcing miniconda
source ~/miniconda/etc/profile.d/conda.sh
echo activating environment
conda activate ./env
echo running server
# Run petals server
# Prompt the user for a model name
read -p "Enter the model name (press Enter for default petals-team/StableBeluga2): " modelName
# Use the default model name if no input is provided
if [ -z "$modelName" ]; then
modelName="petals-team/StableBeluga2"
fi
# Run the Python command with the chosen model name
python3 -m petals.cli.run_server "$modelName" --public_name https://github.com/ParisNeo/lollms-webui