diff --git a/gallery/index.yaml b/gallery/index.yaml
index 833c4ccc..dabde476 100644
--- a/gallery/index.yaml
+++ b/gallery/index.yaml
@@ -2120,6 +2120,37 @@
- filename: Llama-3.1-WhiteRabbitNeo-2-8B-Q4_K_M.gguf
sha256: dbaf619312e706c5440214d324d8f304717866675fc9728e3901c75ef5bbfeca
uri: huggingface://bartowski/Llama-3.1-WhiteRabbitNeo-2-8B-GGUF/Llama-3.1-WhiteRabbitNeo-2-8B-Q4_K_M.gguf
+- !!merge <<: *llama31
+ name: "tess-r1-limerick-llama-3.1-70b"
+ icon: https://huggingface.co/migtissera/Tess-R1-Llama-3.1-70B/resolve/main/Tess-R1-2.jpg
+ urls:
+ - https://huggingface.co/migtissera/Tess-R1-Limerick-Llama-3.1-70B
+ - https://huggingface.co/bartowski/Tess-R1-Limerick-Llama-3.1-70B-GGUF
+ description: |
+ Welcome to the Tess-Reasoning-1 (Tess-R1) series of models. Tess-R1 is designed with test-time compute in mind, and has the capabilities to produce a Chain-of-Thought (CoT) reasoning before producing the final output.
+
+ The model is trained to first think step-by-step, and contemplate on its answers. It can also write alternatives after contemplating. Once all the steps have been thought through, it writes the final output.
+
+ Step-by-step, Chain-of-Thought thinking process. Uses tags to indicate when the model is performing CoT.
+ tags are used when the model contemplate on its answers.
+ tags are used for alternate suggestions.
+ Finally, tags are used for the final output
+
+ Important Note:
+
+ In a multi-turn conversation, only the contents between the tags (discarding the tags) should be carried forward. Otherwise the model will see out of distribution input data and will fail.
+
+ The model was trained mostly with Chain-of-Thought reasoning data, including the XML tags. However, to generalize model generations, some single-turn and multi-turn data without XML tags were also included. Due to this, in some instances the model does not produce XML tags and does not fully utilize test-time compute capabilities. There is two ways to get around this:
+
+ Include a try/catch statement in your inference script, and only pass on the contents between the tags if it's available.
+ Use the tag as the seed in the generation, and force the model to produce outputs with XML tags. i.e: f"{conversation}{user_input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ overrides:
+ parameters:
+ model: Tess-R1-Limerick-Llama-3.1-70B-Q4_K_M.gguf
+ files:
+ - filename: Tess-R1-Limerick-Llama-3.1-70B-Q4_K_M.gguf
+ sha256: 92da5dad8a36ed5060becf78a83537d776079b7eaa4de73733d3ca57156286ab
+ uri: huggingface://bartowski/Tess-R1-Limerick-Llama-3.1-70B-GGUF/Tess-R1-Limerick-Llama-3.1-70B-Q4_K_M.gguf
- &deepseek
## Deepseek
url: "github:mudler/LocalAI/gallery/deepseek.yaml@master"