2024-04-17 21:33:49 +00:00
|
|
|
package openai
|
|
|
|
|
|
|
|
import (
|
2024-06-23 08:24:36 +00:00
|
|
|
"github.com/mudler/LocalAI/core/backend"
|
|
|
|
"github.com/mudler/LocalAI/core/config"
|
2024-04-17 21:33:49 +00:00
|
|
|
|
2024-06-23 08:24:36 +00:00
|
|
|
"github.com/mudler/LocalAI/core/schema"
|
|
|
|
model "github.com/mudler/LocalAI/pkg/model"
|
2024-04-17 21:33:49 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
func ComputeChoices(
|
|
|
|
req *schema.OpenAIRequest,
|
|
|
|
predInput string,
|
|
|
|
config *config.BackendConfig,
|
|
|
|
o *config.ApplicationConfig,
|
|
|
|
loader *model.ModelLoader,
|
|
|
|
cb func(string, *[]schema.Choice),
|
|
|
|
tokenCallback func(string, backend.TokenUsage) bool) ([]schema.Choice, backend.TokenUsage, error) {
|
|
|
|
n := req.N // number of completions to return
|
|
|
|
result := []schema.Choice{}
|
|
|
|
|
|
|
|
if n == 0 {
|
|
|
|
n = 1
|
|
|
|
}
|
|
|
|
|
|
|
|
images := []string{}
|
|
|
|
for _, m := range req.Messages {
|
|
|
|
images = append(images, m.StringImages...)
|
|
|
|
}
|
2024-09-19 09:21:59 +00:00
|
|
|
videos := []string{}
|
|
|
|
for _, m := range req.Messages {
|
|
|
|
videos = append(videos, m.StringVideos...)
|
|
|
|
}
|
2024-09-19 10:26:53 +00:00
|
|
|
audios := []string{}
|
|
|
|
for _, m := range req.Messages {
|
|
|
|
audios = append(audios, m.StringAudios...)
|
|
|
|
}
|
2024-04-17 21:33:49 +00:00
|
|
|
|
|
|
|
// get the model function to call for the result
|
2024-09-19 10:26:53 +00:00
|
|
|
predFunc, err := backend.ModelInference(req.Context, predInput, req.Messages, images, videos, audios, loader, *config, o, tokenCallback)
|
2024-04-17 21:33:49 +00:00
|
|
|
if err != nil {
|
|
|
|
return result, backend.TokenUsage{}, err
|
|
|
|
}
|
|
|
|
|
|
|
|
tokenUsage := backend.TokenUsage{}
|
|
|
|
|
|
|
|
for i := 0; i < n; i++ {
|
|
|
|
prediction, err := predFunc()
|
|
|
|
if err != nil {
|
|
|
|
return result, backend.TokenUsage{}, err
|
|
|
|
}
|
|
|
|
|
|
|
|
tokenUsage.Prompt += prediction.Usage.Prompt
|
|
|
|
tokenUsage.Completion += prediction.Usage.Completion
|
|
|
|
|
|
|
|
finetunedResponse := backend.Finetune(*config, predInput, prediction.Response)
|
|
|
|
cb(finetunedResponse, &result)
|
|
|
|
|
|
|
|
//result = append(result, Choice{Text: prediction})
|
|
|
|
|
|
|
|
}
|
|
|
|
return result, tokenUsage, err
|
|
|
|
}
|