mirror of
https://github.com/mudler/LocalAI.git
synced 2025-01-30 08:04:13 +00:00
models(gallery): add Codestral (#2442)
models(gallery): add Coderstral Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
This commit is contained in:
parent
0c40f545d4
commit
5b75bf16c7
@ -112,7 +112,8 @@ func CompletionEndpoint(cl *config.BackendConfigLoader, ml *model.ModelLoader, a
|
|||||||
|
|
||||||
if templateFile != "" {
|
if templateFile != "" {
|
||||||
templatedInput, err := ml.EvaluateTemplateForPrompt(model.CompletionPromptTemplate, templateFile, model.PromptTemplateData{
|
templatedInput, err := ml.EvaluateTemplateForPrompt(model.CompletionPromptTemplate, templateFile, model.PromptTemplateData{
|
||||||
Input: predInput,
|
Input: predInput,
|
||||||
|
SystemPrompt: config.SystemPrompt,
|
||||||
})
|
})
|
||||||
if err == nil {
|
if err == nil {
|
||||||
predInput = templatedInput
|
predInput = templatedInput
|
||||||
|
@ -1799,6 +1799,30 @@
|
|||||||
- filename: "codellama-7b.Q4_0.gguf"
|
- filename: "codellama-7b.Q4_0.gguf"
|
||||||
sha256: "33052f6dd41436db2f83bd48017b6fff8ce0184e15a8a227368b4230f1da97b5"
|
sha256: "33052f6dd41436db2f83bd48017b6fff8ce0184e15a8a227368b4230f1da97b5"
|
||||||
uri: "huggingface://TheBloke/CodeLlama-7B-GGUF/codellama-7b.Q4_0.gguf"
|
uri: "huggingface://TheBloke/CodeLlama-7B-GGUF/codellama-7b.Q4_0.gguf"
|
||||||
|
- !!merge <<: *codellama
|
||||||
|
name: "codestral-22b-v0.1"
|
||||||
|
license: mnpl
|
||||||
|
description: |
|
||||||
|
Codestral-22B-v0.1 is trained on a diverse dataset of 80+ programming languages, including the most popular ones, such as Python, Java, C, C++, JavaScript, and Bash (more details in the Blogpost). The model can be queried:
|
||||||
|
|
||||||
|
As instruct, for instance to answer any questions about a code snippet (write documentation, explain, factorize) or to generate code following specific indications
|
||||||
|
As Fill in the Middle (FIM), to predict the middle tokens between a prefix and a suffix (very useful for software development add-ons like in VS Code)
|
||||||
|
urls:
|
||||||
|
- https://huggingface.co/mistralai/Codestral-22B-v0.1
|
||||||
|
- https://huggingface.co/bartowski/Codestral-22B-v0.1-GGUF
|
||||||
|
tags:
|
||||||
|
- llm
|
||||||
|
- gguf
|
||||||
|
- gpu
|
||||||
|
- code
|
||||||
|
- cpu
|
||||||
|
overrides:
|
||||||
|
parameters:
|
||||||
|
model: Codestral-22B-v0.1-Q4_K_M.gguf
|
||||||
|
files:
|
||||||
|
- filename: "Codestral-22B-v0.1-Q4_K_M.gguf"
|
||||||
|
sha256: "defc9e0a1bb42857558d43df4e7f0f3d0a29d06a953e498e967d763f45d10431"
|
||||||
|
uri: "huggingface://bartowski/Codestral-22B-v0.1-GGUF/Codestral-22B-v0.1-Q4_K_M.gguf"
|
||||||
- &openvino
|
- &openvino
|
||||||
### START OpenVINO
|
### START OpenVINO
|
||||||
url: "github:mudler/LocalAI/gallery/openvino.yaml@master"
|
url: "github:mudler/LocalAI/gallery/openvino.yaml@master"
|
||||||
|
Loading…
x
Reference in New Issue
Block a user