package grpc import ( "github.com/mudler/LocalAI/core/schema" pb "github.com/mudler/LocalAI/pkg/grpc/proto" ) type LLM interface { Busy() bool Lock() Unlock() Locking() bool Predict(*pb.PredictOptions) (string, error) PredictStream(*pb.PredictOptions, chan string) error Load(*pb.ModelOptions) error Embeddings(*pb.PredictOptions) ([]float32, error) GenerateImage(*pb.GenerateImageRequest) error AudioTranscription(*pb.TranscriptRequest) (schema.TranscriptionResult, error) TTS(*pb.TTSRequest) error TokenizeString(*pb.PredictOptions) (pb.TokenizationResponse, error) Status() (pb.StatusResponse, error) StoresSet(*pb.StoresSetOptions) error StoresDelete(*pb.StoresDeleteOptions) error StoresGet(*pb.StoresGetOptions) (pb.StoresGetResult, error) StoresFind(*pb.StoresFindOptions) (pb.StoresFindResult, error) } func newReply(s string) *pb.Reply { return &pb.Reply{Message: []byte(s)} }