mirror of
https://github.com/mudler/LocalAI.git
synced 2025-02-17 16:10:16 +00:00
Add LowVRAM option parameter (#642)
This commit is contained in:
parent
4e3c319e83
commit
2f5feb4841
@ -35,6 +35,7 @@ type Config struct {
|
|||||||
NGPULayers int `yaml:"gpu_layers"`
|
NGPULayers int `yaml:"gpu_layers"`
|
||||||
MMap bool `yaml:"mmap"`
|
MMap bool `yaml:"mmap"`
|
||||||
MMlock bool `yaml:"mmlock"`
|
MMlock bool `yaml:"mmlock"`
|
||||||
|
LowVRAM bool `yaml:"low_vram"`
|
||||||
|
|
||||||
TensorSplit string `yaml:"tensor_split"`
|
TensorSplit string `yaml:"tensor_split"`
|
||||||
MainGPU string `yaml:"main_gpu"`
|
MainGPU string `yaml:"main_gpu"`
|
||||||
|
@ -48,6 +48,10 @@ func defaultLLamaOpts(c Config) []llama.ModelOption {
|
|||||||
llamaOpts = append(llamaOpts, llama.SetNBatch(512))
|
llamaOpts = append(llamaOpts, llama.SetNBatch(512))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if c.LowVRAM {
|
||||||
|
llamaOpts = append(llamaOpts, llama.EnabelLowVRAM)
|
||||||
|
}
|
||||||
|
|
||||||
return llamaOpts
|
return llamaOpts
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user