package localai import ( "github.com/gofiber/fiber/v2" "github.com/mudler/LocalAI/core/backend" "github.com/mudler/LocalAI/core/config" fiberContext "github.com/mudler/LocalAI/core/http/ctx" "github.com/mudler/LocalAI/core/schema" "github.com/mudler/LocalAI/pkg/model" "github.com/rs/zerolog/log" ) // TokenizeEndpoint exposes a REST API to tokenize the content // @Summary Tokenize the input. // @Success 200 {object} schema.TokenizeResponse "Response" // @Router /v1/tokenize [post] func TokenizeEndpoint(cl *config.BackendConfigLoader, ml *model.ModelLoader, appConfig *config.ApplicationConfig) func(c *fiber.Ctx) error { return func(c *fiber.Ctx) error { input := new(schema.TokenizeRequest) // Get input data from the request body if err := c.BodyParser(input); err != nil { return err } modelFile, err := fiberContext.ModelFromContext(c, cl, ml, input.Model, false) if err != nil { modelFile = input.Model log.Warn().Msgf("Model not found in context: %s", input.Model) } cfg, err := cl.LoadBackendConfigFileByName(modelFile, appConfig.ModelPath, config.LoadOptionDebug(appConfig.Debug), config.LoadOptionThreads(appConfig.Threads), config.LoadOptionContextSize(appConfig.ContextSize), config.LoadOptionF16(appConfig.F16), ) if err != nil { log.Err(err) modelFile = input.Model log.Warn().Msgf("Model not found in context: %s", input.Model) } else { modelFile = cfg.Model } log.Debug().Msgf("Request for model: %s", modelFile) tokenResponse, err := backend.ModelTokenize(input.Content, ml, *cfg, appConfig) if err != nil { return err } c.JSON(tokenResponse) return nil } }