mirror of
https://github.com/mudler/LocalAI.git
synced 2024-12-19 04:37:53 +00:00
1c312685aa
* core 1 * api/openai/files fix * core 2 - core/config * move over core api.go and tests to the start of core/http * move over localai specific endpoints to core/http, begin the service/endpoint split there * refactor big chunk on the plane * refactor chunk 2 on plane, next step: port and modify changes to request.go * easy fixes for request.go, major changes not done yet * lintfix * json tag lintfix? * gitignore and .keep files * strange fix attempt: rename the config dir?
62 lines
2.0 KiB
Go
62 lines
2.0 KiB
Go
package backend
|
|
|
|
import (
|
|
"github.com/go-skynet/LocalAI/core/config"
|
|
|
|
"github.com/go-skynet/LocalAI/pkg/grpc/proto"
|
|
model "github.com/go-skynet/LocalAI/pkg/model"
|
|
)
|
|
|
|
func ImageGeneration(height, width, mode, step, seed int, positive_prompt, negative_prompt, src, dst string, loader *model.ModelLoader, backendConfig config.BackendConfig, appConfig *config.ApplicationConfig) (func() error, error) {
|
|
|
|
opts := modelOpts(backendConfig, appConfig, []model.Option{
|
|
model.WithBackendString(backendConfig.Backend),
|
|
model.WithAssetDir(appConfig.AssetsDestination),
|
|
model.WithThreads(uint32(backendConfig.Threads)),
|
|
model.WithContext(appConfig.Context),
|
|
model.WithModel(backendConfig.Model),
|
|
model.WithLoadGRPCLoadModelOpts(&proto.ModelOptions{
|
|
CUDA: backendConfig.CUDA || backendConfig.Diffusers.CUDA,
|
|
SchedulerType: backendConfig.Diffusers.SchedulerType,
|
|
PipelineType: backendConfig.Diffusers.PipelineType,
|
|
CFGScale: backendConfig.Diffusers.CFGScale,
|
|
LoraAdapter: backendConfig.LoraAdapter,
|
|
LoraScale: backendConfig.LoraScale,
|
|
LoraBase: backendConfig.LoraBase,
|
|
IMG2IMG: backendConfig.Diffusers.IMG2IMG,
|
|
CLIPModel: backendConfig.Diffusers.ClipModel,
|
|
CLIPSubfolder: backendConfig.Diffusers.ClipSubFolder,
|
|
CLIPSkip: int32(backendConfig.Diffusers.ClipSkip),
|
|
ControlNet: backendConfig.Diffusers.ControlNet,
|
|
}),
|
|
})
|
|
|
|
inferenceModel, err := loader.BackendLoader(
|
|
opts...,
|
|
)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
fn := func() error {
|
|
_, err := inferenceModel.GenerateImage(
|
|
appConfig.Context,
|
|
&proto.GenerateImageRequest{
|
|
Height: int32(height),
|
|
Width: int32(width),
|
|
Mode: int32(mode),
|
|
Step: int32(step),
|
|
Seed: int32(seed),
|
|
CLIPSkip: int32(backendConfig.Diffusers.ClipSkip),
|
|
PositivePrompt: positive_prompt,
|
|
NegativePrompt: negative_prompt,
|
|
Dst: dst,
|
|
Src: src,
|
|
EnableParameters: backendConfig.Diffusers.EnableParameters,
|
|
})
|
|
return err
|
|
}
|
|
|
|
return fn, nil
|
|
}
|