2024-10-01 12:41:20 +00:00
|
|
|
package backend
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"fmt"
|
|
|
|
|
|
|
|
"github.com/mudler/LocalAI/core/config"
|
|
|
|
"github.com/mudler/LocalAI/pkg/grpc/proto"
|
|
|
|
model "github.com/mudler/LocalAI/pkg/model"
|
|
|
|
)
|
|
|
|
|
|
|
|
func TokenMetrics(
|
|
|
|
modelFile string,
|
|
|
|
loader *model.ModelLoader,
|
|
|
|
appConfig *config.ApplicationConfig,
|
|
|
|
backendConfig config.BackendConfig) (*proto.MetricsResponse, error) {
|
|
|
|
|
2024-10-02 06:55:58 +00:00
|
|
|
opts := ModelOptions(backendConfig, appConfig, []model.Option{
|
2024-10-01 12:41:20 +00:00
|
|
|
model.WithModel(modelFile),
|
|
|
|
})
|
|
|
|
model, err := loader.BackendLoader(opts...)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
if model == nil {
|
|
|
|
return nil, fmt.Errorf("could not loadmodel model")
|
|
|
|
}
|
|
|
|
|
|
|
|
res, err := model.GetTokenMetrics(context.Background(), &proto.MetricsRequest{})
|
|
|
|
|
|
|
|
return res, err
|
|
|
|
}
|