package main import ( "context" "encoding/json" "errors" "fmt" "os" "os/signal" "path/filepath" "strings" "syscall" "time" "github.com/go-skynet/LocalAI/core/backend" "github.com/go-skynet/LocalAI/core/config" "github.com/go-skynet/LocalAI/core/http" "github.com/go-skynet/LocalAI/core/startup" "github.com/go-skynet/LocalAI/internal" "github.com/go-skynet/LocalAI/pkg/gallery" model "github.com/go-skynet/LocalAI/pkg/model" "github.com/rs/zerolog" "github.com/rs/zerolog/log" progressbar "github.com/schollz/progressbar/v3" "github.com/urfave/cli/v2" ) const ( remoteLibraryURL = "https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/model_library.yaml" ) func main() { log.Logger = log.Output(zerolog.ConsoleWriter{Out: os.Stderr}) // clean up process go func() { c := make(chan os.Signal, 1) // we need to reserve to buffer size 1, so the notifier are not blocked signal.Notify(c, os.Interrupt, syscall.SIGTERM) <-c os.Exit(1) }() path, err := os.Getwd() if err != nil { log.Error().Msgf("error: %s", err.Error()) os.Exit(1) } app := &cli.App{ Name: "LocalAI", Version: internal.PrintableVersion(), Usage: "OpenAI, OSS alternative. Drop-in compatible API for running LLM, GPT and genAI models locally on CPU, GPUs with consumer grade hardware. Supported server endpoints: OpenAI, Elevenlabs", Flags: []cli.Flag{ &cli.BoolFlag{ Name: "f16", EnvVars: []string{"F16"}, }, &cli.BoolFlag{ Name: "autoload-galleries", EnvVars: []string{"AUTOLOAD_GALLERIES"}, }, &cli.BoolFlag{ Name: "debug", EnvVars: []string{"DEBUG"}, }, &cli.BoolFlag{ Name: "single-active-backend", EnvVars: []string{"SINGLE_ACTIVE_BACKEND"}, Usage: "Allow only one backend to be running.", }, &cli.BoolFlag{ Name: "parallel-requests", EnvVars: []string{"PARALLEL_REQUESTS"}, Usage: "Enable backends to handle multiple requests in parallel. This is for backends that supports multiple requests in parallel, like llama.cpp or vllm", }, &cli.BoolFlag{ Name: "cors", EnvVars: []string{"CORS"}, }, &cli.StringFlag{ Name: "cors-allow-origins", EnvVars: []string{"CORS_ALLOW_ORIGINS"}, }, &cli.IntFlag{ Name: "threads", Usage: "Number of threads used for parallel computation. Usage of the number of physical cores in the system is suggested.", EnvVars: []string{"THREADS"}, Value: 4, }, &cli.StringFlag{ Name: "models-path", Usage: "Path containing models used for inferencing", EnvVars: []string{"MODELS_PATH"}, Value: filepath.Join(path, "models"), }, &cli.StringFlag{ Name: "galleries", Usage: "JSON list of galleries", EnvVars: []string{"GALLERIES"}, }, &cli.StringFlag{ Name: "remote-library", Usage: "A LocalAI remote library URL", EnvVars: []string{"REMOTE_LIBRARY"}, Value: remoteLibraryURL, }, &cli.StringFlag{ Name: "preload-models", Usage: "A List of models to apply in JSON at start", EnvVars: []string{"PRELOAD_MODELS"}, }, &cli.StringSliceFlag{ Name: "models", Usage: "A List of models URLs configurations.", EnvVars: []string{"MODELS"}, }, &cli.StringFlag{ Name: "preload-models-config", Usage: "A List of models to apply at startup. Path to a YAML config file", EnvVars: []string{"PRELOAD_MODELS_CONFIG"}, }, &cli.StringFlag{ Name: "config-file", Usage: "Config file", EnvVars: []string{"CONFIG_FILE"}, }, &cli.StringFlag{ Name: "address", Usage: "Bind address for the API server.", EnvVars: []string{"ADDRESS"}, Value: ":8080", }, &cli.StringFlag{ Name: "image-path", Usage: "Image directory", EnvVars: []string{"IMAGE_PATH"}, Value: "/tmp/generated/images", }, &cli.StringFlag{ Name: "audio-path", Usage: "audio directory", EnvVars: []string{"AUDIO_PATH"}, Value: "/tmp/generated/audio", }, &cli.StringFlag{ Name: "upload-path", Usage: "Path to store uploads from files api", EnvVars: []string{"UPLOAD_PATH"}, Value: "/tmp/localai/upload", }, &cli.StringFlag{ Name: "config-path", Usage: "Path to store uploads from files api", EnvVars: []string{"CONFIG_PATH"}, Value: "/tmp/localai/config", }, &cli.StringFlag{ Name: "backend-assets-path", Usage: "Path used to extract libraries that are required by some of the backends in runtime.", EnvVars: []string{"BACKEND_ASSETS_PATH"}, Value: "/tmp/localai/backend_data", }, &cli.StringSliceFlag{ Name: "external-grpc-backends", Usage: "A list of external grpc backends", EnvVars: []string{"EXTERNAL_GRPC_BACKENDS"}, }, &cli.IntFlag{ Name: "context-size", Usage: "Default context size of the model", EnvVars: []string{"CONTEXT_SIZE"}, Value: 512, }, &cli.IntFlag{ Name: "upload-limit", Usage: "Default upload-limit. MB", EnvVars: []string{"UPLOAD_LIMIT"}, Value: 15, }, &cli.StringSliceFlag{ Name: "api-keys", Usage: "List of API Keys to enable API authentication. When this is set, all the requests must be authenticated with one of these API keys.", EnvVars: []string{"API_KEY"}, }, &cli.BoolFlag{ Name: "enable-watchdog-idle", Usage: "Enable watchdog for stopping idle backends. This will stop the backends if are in idle state for too long.", EnvVars: []string{"WATCHDOG_IDLE"}, Value: false, }, &cli.BoolFlag{ Name: "enable-watchdog-busy", Usage: "Enable watchdog for stopping busy backends that exceed a defined threshold.", EnvVars: []string{"WATCHDOG_BUSY"}, Value: false, }, &cli.StringFlag{ Name: "watchdog-busy-timeout", Usage: "Watchdog timeout. This will restart the backend if it crashes.", EnvVars: []string{"WATCHDOG_BUSY_TIMEOUT"}, Value: "5m", }, &cli.StringFlag{ Name: "watchdog-idle-timeout", Usage: "Watchdog idle timeout. This will restart the backend if it crashes.", EnvVars: []string{"WATCHDOG_IDLE_TIMEOUT"}, Value: "15m", }, &cli.BoolFlag{ Name: "preload-backend-only", Usage: "If set, the api is NOT launched, and only the preloaded models / backends are started. This is intended for multi-node setups.", EnvVars: []string{"PRELOAD_BACKEND_ONLY"}, Value: false, }, &cli.StringFlag{ Name: "localai-config-dir", Usage: "Directory to use for the configuration files of LocalAI itself. This is NOT where model files should be placed.", EnvVars: []string{"LOCALAI_CONFIG_DIR"}, Value: "./configuration", }, }, Description: ` LocalAI is a drop-in replacement OpenAI API which runs inference locally. Some of the models compatible are: - Vicuna - Koala - GPT4ALL - GPT4ALL-J - Cerebras - Alpaca - StableLM (ggml quantized) For a list of compatible model, check out: https://localai.io/model-compatibility/index.html `, UsageText: `local-ai [options]`, Copyright: "Ettore Di Giacinto", Action: func(ctx *cli.Context) error { opts := []config.AppOption{ config.WithConfigFile(ctx.String("config-file")), config.WithJSONStringPreload(ctx.String("preload-models")), config.WithYAMLConfigPreload(ctx.String("preload-models-config")), config.WithModelPath(ctx.String("models-path")), config.WithContextSize(ctx.Int("context-size")), config.WithDebug(ctx.Bool("debug")), config.WithImageDir(ctx.String("image-path")), config.WithAudioDir(ctx.String("audio-path")), config.WithUploadDir(ctx.String("upload-path")), config.WithConfigsDir(ctx.String("config-path")), config.WithF16(ctx.Bool("f16")), config.WithStringGalleries(ctx.String("galleries")), config.WithModelLibraryURL(ctx.String("remote-library")), config.WithDisableMessage(false), config.WithCors(ctx.Bool("cors")), config.WithCorsAllowOrigins(ctx.String("cors-allow-origins")), config.WithThreads(ctx.Int("threads")), config.WithBackendAssets(backendAssets), config.WithBackendAssetsOutput(ctx.String("backend-assets-path")), config.WithUploadLimitMB(ctx.Int("upload-limit")), config.WithApiKeys(ctx.StringSlice("api-keys")), config.WithModelsURL(append(ctx.StringSlice("models"), ctx.Args().Slice()...)...), } idleWatchDog := ctx.Bool("enable-watchdog-idle") busyWatchDog := ctx.Bool("enable-watchdog-busy") if idleWatchDog || busyWatchDog { opts = append(opts, config.EnableWatchDog) if idleWatchDog { opts = append(opts, config.EnableWatchDogIdleCheck) dur, err := time.ParseDuration(ctx.String("watchdog-idle-timeout")) if err != nil { return err } opts = append(opts, config.SetWatchDogIdleTimeout(dur)) } if busyWatchDog { opts = append(opts, config.EnableWatchDogBusyCheck) dur, err := time.ParseDuration(ctx.String("watchdog-busy-timeout")) if err != nil { return err } opts = append(opts, config.SetWatchDogBusyTimeout(dur)) } } if ctx.Bool("parallel-requests") { opts = append(opts, config.EnableParallelBackendRequests) } if ctx.Bool("single-active-backend") { opts = append(opts, config.EnableSingleBackend) } externalgRPC := ctx.StringSlice("external-grpc-backends") // split ":" to get backend name and the uri for _, v := range externalgRPC { backend := v[:strings.IndexByte(v, ':')] uri := v[strings.IndexByte(v, ':')+1:] opts = append(opts, config.WithExternalBackend(backend, uri)) } if ctx.Bool("autoload-galleries") { opts = append(opts, config.EnableGalleriesAutoload) } if ctx.Bool("preload-backend-only") { _, _, _, err := startup.Startup(opts...) return err } cl, ml, options, err := startup.Startup(opts...) if err != nil { return fmt.Errorf("failed basic startup tasks with error %s", err.Error()) } configdir := ctx.String("localai-config-dir") // Watch the configuration directory // If the directory does not exist, we don't watch it if _, err := os.Stat(configdir); err == nil { closeConfigWatcherFn, err := startup.WatchConfigDirectory(ctx.String("localai-config-dir"), options) defer closeConfigWatcherFn() if err != nil { return fmt.Errorf("failed while watching configuration directory %s", ctx.String("localai-config-dir")) } } appHTTP, err := http.App(cl, ml, options) if err != nil { log.Error().Msg("Error during HTTP App constructor") return err } return appHTTP.Listen(ctx.String("address")) }, Commands: []*cli.Command{ { Name: "models", Usage: "List or install models", Subcommands: []*cli.Command{ { Name: "list", Usage: "List the models avaiable in your galleries", Action: func(ctx *cli.Context) error { var galleries []gallery.Gallery if err := json.Unmarshal([]byte(ctx.String("galleries")), &galleries); err != nil { log.Error().Msgf("unable to load galleries: %s", err.Error()) } models, err := gallery.AvailableGalleryModels(galleries, ctx.String("models-path")) if err != nil { return err } for _, model := range models { if model.Installed { fmt.Printf(" * %s@%s (installed)\n", model.Gallery.Name, model.Name) } else { fmt.Printf(" - %s@%s\n", model.Gallery.Name, model.Name) } } return nil }, }, { Name: "install", Usage: "Install a model from the gallery", Action: func(ctx *cli.Context) error { modelName := ctx.Args().First() var galleries []gallery.Gallery if err := json.Unmarshal([]byte(ctx.String("galleries")), &galleries); err != nil { log.Error().Msgf("unable to load galleries: %s", err.Error()) } progressBar := progressbar.NewOptions( 1000, progressbar.OptionSetDescription(fmt.Sprintf("downloading model %s", modelName)), progressbar.OptionShowBytes(false), progressbar.OptionClearOnFinish(), ) progressCallback := func(fileName string, current string, total string, percentage float64) { progressBar.Set(int(percentage * 10)) } err = gallery.InstallModelFromGallery(galleries, modelName, ctx.String("models-path"), gallery.GalleryModel{}, progressCallback) if err != nil { return err } return nil }, }, }, }, { Name: "tts", Usage: "Convert text to speech", Flags: []cli.Flag{ &cli.StringFlag{ Name: "backend", Value: "piper", Aliases: []string{"b"}, Usage: "Backend to run the TTS model", }, &cli.StringFlag{ Name: "model", Aliases: []string{"m"}, Usage: "Model name to run the TTS", Required: true, }, &cli.StringFlag{ Name: "voice", Aliases: []string{"v"}, Usage: "Voice name to run the TTS (optional)", Required: true, }, &cli.StringFlag{ Name: "output-file", Aliases: []string{"o"}, Usage: "The path to write the output wav file", }, }, Action: func(ctx *cli.Context) error { modelOption := ctx.String("model") if modelOption == "" { return errors.New("--model parameter is required") } backendOption := ctx.String("backend") if backendOption == "" { backendOption = "piper" } outputFile := ctx.String("output-file") outputDir := ctx.String("backend-assets-path") if outputFile != "" { outputDir = filepath.Dir(outputFile) } text := strings.Join(ctx.Args().Slice(), " ") opts := &config.ApplicationConfig{ ModelPath: ctx.String("models-path"), Context: context.Background(), AudioDir: outputDir, AssetsDestination: ctx.String("backend-assets-path"), } ml := model.NewModelLoader(opts.ModelPath) defer ml.StopAllGRPC() filePath, _, err := backend.ModelTTS(backendOption, text, modelOption, ctx.String("voice"), ml, opts, config.BackendConfig{}) if err != nil { return err } if outputFile != "" { if err := os.Rename(filePath, outputFile); err != nil { return err } fmt.Printf("Generate file %s\n", outputFile) } else { fmt.Printf("Generate file %s\n", filePath) } return nil }, }, { Name: "transcript", Usage: "Convert audio to text", Flags: []cli.Flag{ &cli.StringFlag{ Name: "backend", Value: "whisper", Aliases: []string{"b"}, Usage: "Backend to run the transcription model", }, &cli.StringFlag{ Name: "model", Aliases: []string{"m"}, Usage: "Model name to run the transcription", }, &cli.StringFlag{ Name: "language", Aliases: []string{"l"}, Usage: "Language of the audio file", }, &cli.IntFlag{ Name: "threads", Aliases: []string{"t"}, Usage: "Threads to use", Value: 1, }, &cli.StringFlag{ Name: "output-file", Aliases: []string{"o"}, Usage: "The path to write the output wav file", }, }, Action: func(ctx *cli.Context) error { modelOption := ctx.String("model") filename := ctx.Args().First() language := ctx.String("language") threads := ctx.Int("threads") opts := &config.ApplicationConfig{ ModelPath: ctx.String("models-path"), Context: context.Background(), AssetsDestination: ctx.String("backend-assets-path"), } cl := config.NewBackendConfigLoader() ml := model.NewModelLoader(opts.ModelPath) if err := cl.LoadBackendConfigsFromPath(ctx.String("models-path")); err != nil { return err } c, exists := cl.GetBackendConfig(modelOption) if !exists { return errors.New("model not found") } c.Threads = &threads defer ml.StopAllGRPC() tr, err := backend.ModelTranscription(filename, language, ml, c, opts) if err != nil { return err } for _, segment := range tr.Segments { fmt.Println(segment.Start.String(), "-", segment.Text) } return nil }, }, }, } err = app.Run(os.Args) if err != nil { log.Error().Msgf("error: %s", err.Error()) os.Exit(1) } }