#!/bin/sh printf "whisper.cpp: this script hasn't been maintained and is not functional atm\n" exit 1 # This script downloads Whisper model files that have already been converted to Core ML format. # This way you don't have to convert them yourself. src="https://huggingface.co/datasets/ggerganov/whisper.cpp-coreml" pfx="resolve/main/ggml" # get the path of this script get_script_path() { if [ -x "$(command -v realpath)" ]; then dirname "$(realpath "$0")" else _ret="$(cd -- "$(dirname "$0")" >/dev/null 2>&1 || exit ; pwd -P)" echo "$_ret" fi } script_path="$(get_script_path)" # Check if the script is inside a /bin/ directory case "$script_path" in */bin) default_download_path="$PWD" ;; # Use current directory as default download path if in /bin/ *) default_download_path="$script_path" ;; # Otherwise, use script directory esac models_path="${2:-$default_download_path}" # Whisper models models="tiny.en tiny base.en base small.en small medium.en medium large-v1 large-v2 large-v3 large-v3-turbo" # list available models list_models() { printf "\n" printf " Available models:" for model in $models; do printf " %s" "$models" done printf "\n\n" } if [ "$#" -lt 1 ] || [ "$#" -gt 2 ]; then printf "Usage: %s [models_path]\n" "$0" list_models exit 1 fi model=$1 if ! echo "$models" | grep -q -w "$model"; then printf "Invalid model: %s\n" "$model" list_models exit 1 fi # download Core ML model printf "Downloading Core ML model %s from '%s' ...\n" "$model" "$src" cd "$models_path" || exit if [ -f "ggml-$model.mlmodel" ]; then printf "Model %s already exists. Skipping download.\n" "$model" exit 0 fi if [ -x "$(command -v wget)" ]; then wget --quiet --show-progress -O ggml-"$model".mlmodel $src/$pfx-"$model".mlmodel elif [ -x "$(command -v curl)" ]; then curl -L --output ggml-"$model".mlmodel $src/$pfx-"$model".mlmodel else printf "Either wget or curl is required to download models.\n" exit 1 fi if [ $? -ne 0 ]; then printf "Failed to download Core ML model %s \n" "$model" printf "Please try again later or download the original Whisper model files and convert them yourself.\n" exit 1 fi # Check if 'whisper-cli' is available in the system PATH if command -v whisper-cli >/dev/null 2>&1; then # If found, use 'whisper-cli' (relying on PATH resolution) whisper_cmd="whisper-cli" else # If not found, use the local build version whisper_cmd="./build/bin/whisper-cli" fi printf "Done! Model '%s' saved in '%s/ggml-%s.bin'\n" "$model" "$models_path" "$model" printf "Run the following command to compile it:\n\n" printf " $ xcrun coremlc compile %s/ggml-%s.mlmodel %s\n\n" "$models_path" "$model" "$models_path" printf "You can now use it like this:\n\n" printf " $ %s -m %s/ggml-%s.bin -f samples/jfk.wav\n" "$whisper_cmd" "$models_path" "$model" printf "\n"