mirror of
https://github.com/mudler/LocalAI.git
synced 2024-12-19 20:57:54 +00:00
models(gallery): add tess-r1-limerick-llama-3.1-70b (#4095)
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
This commit is contained in:
parent
d454118887
commit
a0cdd19038
@ -2120,6 +2120,37 @@
|
|||||||
- filename: Llama-3.1-WhiteRabbitNeo-2-8B-Q4_K_M.gguf
|
- filename: Llama-3.1-WhiteRabbitNeo-2-8B-Q4_K_M.gguf
|
||||||
sha256: dbaf619312e706c5440214d324d8f304717866675fc9728e3901c75ef5bbfeca
|
sha256: dbaf619312e706c5440214d324d8f304717866675fc9728e3901c75ef5bbfeca
|
||||||
uri: huggingface://bartowski/Llama-3.1-WhiteRabbitNeo-2-8B-GGUF/Llama-3.1-WhiteRabbitNeo-2-8B-Q4_K_M.gguf
|
uri: huggingface://bartowski/Llama-3.1-WhiteRabbitNeo-2-8B-GGUF/Llama-3.1-WhiteRabbitNeo-2-8B-Q4_K_M.gguf
|
||||||
|
- !!merge <<: *llama31
|
||||||
|
name: "tess-r1-limerick-llama-3.1-70b"
|
||||||
|
icon: https://huggingface.co/migtissera/Tess-R1-Llama-3.1-70B/resolve/main/Tess-R1-2.jpg
|
||||||
|
urls:
|
||||||
|
- https://huggingface.co/migtissera/Tess-R1-Limerick-Llama-3.1-70B
|
||||||
|
- https://huggingface.co/bartowski/Tess-R1-Limerick-Llama-3.1-70B-GGUF
|
||||||
|
description: |
|
||||||
|
Welcome to the Tess-Reasoning-1 (Tess-R1) series of models. Tess-R1 is designed with test-time compute in mind, and has the capabilities to produce a Chain-of-Thought (CoT) reasoning before producing the final output.
|
||||||
|
|
||||||
|
The model is trained to first think step-by-step, and contemplate on its answers. It can also write alternatives after contemplating. Once all the steps have been thought through, it writes the final output.
|
||||||
|
|
||||||
|
Step-by-step, Chain-of-Thought thinking process. Uses <thinking> </thinking> tags to indicate when the model is performing CoT.
|
||||||
|
<contemplation> </contemplation> tags are used when the model contemplate on its answers.
|
||||||
|
<alternatively> </alternatively> tags are used for alternate suggestions.
|
||||||
|
Finally, <output> </output> tags are used for the final output
|
||||||
|
|
||||||
|
Important Note:
|
||||||
|
|
||||||
|
In a multi-turn conversation, only the contents between the <output> </output> tags (discarding the tags) should be carried forward. Otherwise the model will see out of distribution input data and will fail.
|
||||||
|
|
||||||
|
The model was trained mostly with Chain-of-Thought reasoning data, including the XML tags. However, to generalize model generations, some single-turn and multi-turn data without XML tags were also included. Due to this, in some instances the model does not produce XML tags and does not fully utilize test-time compute capabilities. There is two ways to get around this:
|
||||||
|
|
||||||
|
Include a try/catch statement in your inference script, and only pass on the contents between the <output> </output> tags if it's available.
|
||||||
|
Use the <thinking> tag as the seed in the generation, and force the model to produce outputs with XML tags. i.e: f"{conversation}{user_input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n<thinking>"
|
||||||
|
overrides:
|
||||||
|
parameters:
|
||||||
|
model: Tess-R1-Limerick-Llama-3.1-70B-Q4_K_M.gguf
|
||||||
|
files:
|
||||||
|
- filename: Tess-R1-Limerick-Llama-3.1-70B-Q4_K_M.gguf
|
||||||
|
sha256: 92da5dad8a36ed5060becf78a83537d776079b7eaa4de73733d3ca57156286ab
|
||||||
|
uri: huggingface://bartowski/Tess-R1-Limerick-Llama-3.1-70B-GGUF/Tess-R1-Limerick-Llama-3.1-70B-Q4_K_M.gguf
|
||||||
- &deepseek
|
- &deepseek
|
||||||
## Deepseek
|
## Deepseek
|
||||||
url: "github:mudler/LocalAI/gallery/deepseek.yaml@master"
|
url: "github:mudler/LocalAI/gallery/deepseek.yaml@master"
|
||||||
|
Loading…
Reference in New Issue
Block a user