mirror of
https://github.com/mudler/LocalAI.git
synced 2024-12-21 21:47:51 +00:00
1d0ed95a54
This finally makes everything more consistent Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
167 lines
4.0 KiB
Go
167 lines
4.0 KiB
Go
package model
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"os"
|
|
"path/filepath"
|
|
"strings"
|
|
"sync"
|
|
"text/template"
|
|
|
|
"github.com/go-skynet/LocalAI/pkg/grpc"
|
|
process "github.com/mudler/go-processmanager"
|
|
"github.com/rs/zerolog/log"
|
|
)
|
|
|
|
type ModelLoader struct {
|
|
ModelPath string
|
|
mu sync.Mutex
|
|
// TODO: this needs generics
|
|
models map[string]*grpc.Client
|
|
grpcProcesses map[string]*process.Process
|
|
promptsTemplates map[string]*template.Template
|
|
}
|
|
|
|
func NewModelLoader(modelPath string) *ModelLoader {
|
|
return &ModelLoader{
|
|
ModelPath: modelPath,
|
|
models: make(map[string]*grpc.Client),
|
|
promptsTemplates: make(map[string]*template.Template),
|
|
grpcProcesses: make(map[string]*process.Process),
|
|
}
|
|
}
|
|
|
|
func (ml *ModelLoader) ExistsInModelPath(s string) bool {
|
|
_, err := os.Stat(filepath.Join(ml.ModelPath, s))
|
|
return err == nil
|
|
}
|
|
|
|
func (ml *ModelLoader) ListModels() ([]string, error) {
|
|
files, err := ioutil.ReadDir(ml.ModelPath)
|
|
if err != nil {
|
|
return []string{}, err
|
|
}
|
|
|
|
models := []string{}
|
|
for _, file := range files {
|
|
// Skip templates, YAML and .keep files
|
|
if strings.HasSuffix(file.Name(), ".tmpl") || strings.HasSuffix(file.Name(), ".keep") || strings.HasSuffix(file.Name(), ".yaml") || strings.HasSuffix(file.Name(), ".yml") {
|
|
continue
|
|
}
|
|
|
|
models = append(models, file.Name())
|
|
}
|
|
|
|
return models, nil
|
|
}
|
|
|
|
func (ml *ModelLoader) TemplatePrefix(modelName string, in interface{}) (string, error) {
|
|
ml.mu.Lock()
|
|
defer ml.mu.Unlock()
|
|
|
|
m, ok := ml.promptsTemplates[modelName]
|
|
if !ok {
|
|
modelFile := filepath.Join(ml.ModelPath, modelName)
|
|
if err := ml.loadTemplateIfExists(modelName, modelFile); err != nil {
|
|
return "", err
|
|
}
|
|
|
|
t, exists := ml.promptsTemplates[modelName]
|
|
if exists {
|
|
m = t
|
|
}
|
|
}
|
|
if m == nil {
|
|
return "", fmt.Errorf("failed loading any template")
|
|
}
|
|
|
|
var buf bytes.Buffer
|
|
|
|
if err := m.Execute(&buf, in); err != nil {
|
|
return "", err
|
|
}
|
|
return buf.String(), nil
|
|
}
|
|
|
|
func (ml *ModelLoader) loadTemplateIfExists(modelName, modelFile string) error {
|
|
// Check if the template was already loaded
|
|
if _, ok := ml.promptsTemplates[modelName]; ok {
|
|
return nil
|
|
}
|
|
|
|
// Check if the model path exists
|
|
// skip any error here - we run anyway if a template does not exist
|
|
modelTemplateFile := fmt.Sprintf("%s.tmpl", modelName)
|
|
|
|
if !ml.ExistsInModelPath(modelTemplateFile) {
|
|
return nil
|
|
}
|
|
|
|
dat, err := os.ReadFile(filepath.Join(ml.ModelPath, modelTemplateFile))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Parse the template
|
|
tmpl, err := template.New("prompt").Parse(string(dat))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
ml.promptsTemplates[modelName] = tmpl
|
|
|
|
return nil
|
|
}
|
|
|
|
func (ml *ModelLoader) LoadModel(modelName string, loader func(string) (*grpc.Client, error)) (*grpc.Client, error) {
|
|
ml.mu.Lock()
|
|
defer ml.mu.Unlock()
|
|
|
|
// Check if we already have a loaded model
|
|
if model := ml.checkIsLoaded(modelName); model != nil {
|
|
log.Debug().Msgf("Model already loaded in memory: %s", modelName)
|
|
return model, nil
|
|
}
|
|
|
|
// Load the model and keep it in memory for later use
|
|
modelFile := filepath.Join(ml.ModelPath, modelName)
|
|
log.Debug().Msgf("Loading model in memory from file: %s", modelFile)
|
|
|
|
model, err := loader(modelFile)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// If there is a prompt template, load it
|
|
if err := ml.loadTemplateIfExists(modelName, modelFile); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
ml.models[modelName] = model
|
|
return model, nil
|
|
}
|
|
|
|
func (ml *ModelLoader) checkIsLoaded(s string) *grpc.Client {
|
|
if m, ok := ml.models[s]; ok {
|
|
log.Debug().Msgf("Model already loaded in memory: %s", s)
|
|
|
|
if !m.HealthCheck(context.Background()) {
|
|
log.Debug().Msgf("GRPC Model not responding", s)
|
|
if !ml.grpcProcesses[s].IsAlive() {
|
|
log.Debug().Msgf("GRPC Process is not responding", s)
|
|
// stop and delete the process, this forces to re-load the model and re-create again the service
|
|
ml.grpcProcesses[s].Stop()
|
|
delete(ml.grpcProcesses, s)
|
|
delete(ml.models, s)
|
|
return nil
|
|
}
|
|
}
|
|
|
|
return m
|
|
}
|
|
|
|
return nil
|
|
}
|