fix: yamlint warnings and errors (#2131)

fix yamlint warnings and errors

Signed-off-by: Dave Lee <dave@gray101.com>
This commit is contained in:
Dave 2024-04-25 13:25:56 -04:00 committed by GitHub
parent 4ae4e44506
commit 45761f8be2
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
22 changed files with 340 additions and 326 deletions

4
.yamllint Normal file
View File

@ -0,0 +1,4 @@
extends: default
rules:
line-length: disable

View File

@ -1,3 +1,4 @@
---
name: "bert-embeddings" name: "bert-embeddings"
config_file: | config_file: |
@ -6,6 +7,6 @@ config_file: |
backend: bert-embeddings backend: bert-embeddings
embeddings: true embeddings: true
files: files:
- filename: "bert-MiniLM-L6-v2q4_0.bin" - filename: "bert-MiniLM-L6-v2q4_0.bin"
sha256: "a5a174d8772c8a569faf9f3136c441f2c3855b5bf35ed32274294219533feaad" sha256: "a5a174d8772c8a569faf9f3136c441f2c3855b5bf35ed32274294219533feaad"
uri: "https://huggingface.co/mudler/all-MiniLM-L6-v2/resolve/main/ggml-model-q4_0.bin" uri: "https://huggingface.co/mudler/all-MiniLM-L6-v2/resolve/main/ggml-model-q4_0.bin"

View File

@ -1,3 +1,4 @@
---
config_file: | config_file: |
backend: llama-cpp backend: llama-cpp
context_size: 8192 context_size: 8192

View File

@ -1,3 +1,4 @@
---
name: "codellama" name: "codellama"
config_file: | config_file: |

View File

@ -1,6 +1,6 @@
---
name: "dreamshaper" name: "dreamshaper"
config_file: | config_file: |
backend: diffusers backend: diffusers
step: 25 step: 25

View File

@ -1,6 +1,6 @@
---
name: "hermes-2-pro-mistral" name: "hermes-2-pro-mistral"
config_file: | config_file: |
mmap: true mmap: true
template: template:
@ -52,4 +52,3 @@ config_file: |
- <dummy32000> - <dummy32000>
- "\n</tool_call>" - "\n</tool_call>"
- "\n\n\n" - "\n\n\n"

View File

@ -1,4 +1,4 @@
---
### START parler-tts ### START parler-tts
- &parler-tts - &parler-tts
url: "github:mudler/LocalAI/gallery/parler-tts.yaml@master" url: "github:mudler/LocalAI/gallery/parler-tts.yaml@master"
@ -9,13 +9,13 @@
description: | description: |
Parler-TTS is a lightweight text-to-speech (TTS) model that can generate high-quality, natural sounding speech in the style of a given speaker (gender, pitch, speaking style, etc). It is a reproduction of work from the paper Natural language guidance of high-fidelity text-to-speech with synthetic annotations by Dan Lyth and Simon King, from Stability AI and Edinburgh University respectively. Parler-TTS is a lightweight text-to-speech (TTS) model that can generate high-quality, natural sounding speech in the style of a given speaker (gender, pitch, speaking style, etc). It is a reproduction of work from the paper Natural language guidance of high-fidelity text-to-speech with synthetic annotations by Dan Lyth and Simon King, from Stability AI and Edinburgh University respectively.
urls: urls:
- https://github.com/huggingface/parler-tts - https://github.com/huggingface/parler-tts
tags: tags:
- tts - tts
- gpu - gpu
- cpu - cpu
- text-to-speech - text-to-speech
- python - python
### START rerankers ### START rerankers
- &rerankers - &rerankers
url: "github:mudler/LocalAI/gallery/rerankers.yaml@master" url: "github:mudler/LocalAI/gallery/rerankers.yaml@master"
@ -26,9 +26,9 @@
description: | description: |
A cross-encoder model that can be used for reranking A cross-encoder model that can be used for reranking
tags: tags:
- reranker - reranker
- gpu - gpu
- python - python
## LLMs ## LLMs
### START LLAMA3 ### START LLAMA3
- &llama3 - &llama3
@ -49,43 +49,43 @@
Model Architecture Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety. Model Architecture Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
urls: urls:
- https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct - https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct
- https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF - https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF
tags: tags:
- llm - llm
- gguf - gguf
- gpu - gpu
- cpu - cpu
- llama3 - llama3
overrides: overrides:
parameters: parameters:
model: Meta-Llama-3-8B-Instruct-Q5_K_M.gguf model: Meta-Llama-3-8B-Instruct-Q5_K_M.gguf
files: files:
- filename: Meta-Llama-3-8B-Instruct.Q4_0.gguf - filename: Meta-Llama-3-8B-Instruct.Q4_0.gguf
sha256: 19ded996fe6c60254dc7544d782276eff41046ed42aa5f2d0005dc457e5c0895 sha256: 19ded996fe6c60254dc7544d782276eff41046ed42aa5f2d0005dc457e5c0895
uri: huggingface://QuantFactory/Meta-Llama-3-8B-Instruct-GGUF/Meta-Llama-3-8B-Instruct.Q4_0.gguf uri: huggingface://QuantFactory/Meta-Llama-3-8B-Instruct-GGUF/Meta-Llama-3-8B-Instruct.Q4_0.gguf
- <<: *llama3 - <<: *llama3
name: "llama3-8b-instruct:Q6_K" name: "llama3-8b-instruct:Q6_K"
overrides: overrides:
parameters: parameters:
model: Meta-Llama-3-8B-Instruct.Q6_K.gguf model: Meta-Llama-3-8B-Instruct.Q6_K.gguf
files: files:
- filename: Meta-Llama-3-8B-Instruct.Q6_K.gguf - filename: Meta-Llama-3-8B-Instruct.Q6_K.gguf
sha256: b7bad45618e2a76cc1e89a0fbb93a2cac9bf410e27a619c8024ed6db53aa9b4a sha256: b7bad45618e2a76cc1e89a0fbb93a2cac9bf410e27a619c8024ed6db53aa9b4a
uri: huggingface://QuantFactory/Meta-Llama-3-8B-Instruct-GGUF/Meta-Llama-3-8B-Instruct.Q6_K.gguf uri: huggingface://QuantFactory/Meta-Llama-3-8B-Instruct-GGUF/Meta-Llama-3-8B-Instruct.Q6_K.gguf
- <<: *llama3 - <<: *llama3
name: "llama3-70b-instruct" name: "llama3-70b-instruct"
overrides: overrides:
parameters: parameters:
model: Meta-Llama-3-70B-Instruct.Q4_K_M.gguf model: Meta-Llama-3-70B-Instruct.Q4_K_M.gguf
files: files:
- filename: Meta-Llama-3-70B-Instruct.Q4_K_M.gguf - filename: Meta-Llama-3-70B-Instruct.Q4_K_M.gguf
sha256: d559de8dd806a76dbd29f8d8bd04666f2b29e7c7872d8e8481abd07805884d72 sha256: d559de8dd806a76dbd29f8d8bd04666f2b29e7c7872d8e8481abd07805884d72
uri: huggingface://MaziyarPanahi/Meta-Llama-3-70B-Instruct-GGUF/Meta-Llama-3-70B-Instruct.Q4_K_M.gguf uri: huggingface://MaziyarPanahi/Meta-Llama-3-70B-Instruct-GGUF/Meta-Llama-3-70B-Instruct.Q4_K_M.gguf
- <<: *llama3 - <<: *llama3
name: "llama-3-sauerkrautlm-8b-instruct" name: "llama-3-sauerkrautlm-8b-instruct"
urls: urls:
- https://huggingface.co/bartowski/Llama-3-SauerkrautLM-8b-Instruct-GGUF - https://huggingface.co/bartowski/Llama-3-SauerkrautLM-8b-Instruct-GGUF
icon: https://vago-solutions.ai/wp-content/uploads/2024/04/Llama3-Pic.png icon: https://vago-solutions.ai/wp-content/uploads/2024/04/Llama3-Pic.png
description: | description: |
SauerkrautLM-llama-3-8B-Instruct SauerkrautLM-llama-3-8B-Instruct
@ -96,13 +96,13 @@
parameters: parameters:
model: Llama-3-SauerkrautLM-8b-Instruct-Q4_K_M.gguf model: Llama-3-SauerkrautLM-8b-Instruct-Q4_K_M.gguf
files: files:
- filename: Llama-3-SauerkrautLM-8b-Instruct-Q4_K_M.gguf - filename: Llama-3-SauerkrautLM-8b-Instruct-Q4_K_M.gguf
sha256: 5833d99d5596cade0d02e61cddaa6dac49170864ee56d0b602933c6f9fbae314 sha256: 5833d99d5596cade0d02e61cddaa6dac49170864ee56d0b602933c6f9fbae314
uri: huggingface://bartowski/Llama-3-SauerkrautLM-8b-Instruct-GGUF/Llama-3-SauerkrautLM-8b-Instruct-Q4_K_M.gguf uri: huggingface://bartowski/Llama-3-SauerkrautLM-8b-Instruct-GGUF/Llama-3-SauerkrautLM-8b-Instruct-Q4_K_M.gguf
- <<: *llama3 - <<: *llama3
name: "llama-3-13b-instruct-v0.1" name: "llama-3-13b-instruct-v0.1"
urls: urls:
- https://huggingface.co/MaziyarPanahi/Llama-3-13B-Instruct-v0.1-GGUF - https://huggingface.co/MaziyarPanahi/Llama-3-13B-Instruct-v0.1-GGUF
icon: https://huggingface.co/MaziyarPanahi/Llama-3-13B-Instruct-v0.1/resolve/main/llama-3-merges.webp icon: https://huggingface.co/MaziyarPanahi/Llama-3-13B-Instruct-v0.1/resolve/main/llama-3-merges.webp
description: | description: |
This model is a self-merge of meta-llama/Meta-Llama-3-8B-Instruct model. This model is a self-merge of meta-llama/Meta-Llama-3-8B-Instruct model.
@ -110,13 +110,13 @@
parameters: parameters:
model: Llama-3-13B-Instruct-v0.1.Q4_K_M.gguf model: Llama-3-13B-Instruct-v0.1.Q4_K_M.gguf
files: files:
- filename: Llama-3-13B-Instruct-v0.1.Q4_K_M.gguf - filename: Llama-3-13B-Instruct-v0.1.Q4_K_M.gguf
sha256: 071a28043c271d259b5ffa883d19a9e0b33269b55148c4abaf5f95da4d084266 sha256: 071a28043c271d259b5ffa883d19a9e0b33269b55148c4abaf5f95da4d084266
uri: huggingface://MaziyarPanahi/Llama-3-13B-Instruct-v0.1-GGUF/Llama-3-13B-Instruct-v0.1.Q4_K_M.gguf uri: huggingface://MaziyarPanahi/Llama-3-13B-Instruct-v0.1-GGUF/Llama-3-13B-Instruct-v0.1.Q4_K_M.gguf
- <<: *llama3 - <<: *llama3
name: "llama-3-smaug-8b" name: "llama-3-smaug-8b"
urls: urls:
- https://huggingface.co/MaziyarPanahi/Llama-3-Smaug-8B-GGUF - https://huggingface.co/MaziyarPanahi/Llama-3-Smaug-8B-GGUF
icon: https://cdn-uploads.huggingface.co/production/uploads/64c14f95cac5f9ba52bbcd7f/OrcJyTaUtD2HxJOPPwNva.png icon: https://cdn-uploads.huggingface.co/production/uploads/64c14f95cac5f9ba52bbcd7f/OrcJyTaUtD2HxJOPPwNva.png
description: | description: |
This model was built using the Smaug recipe for improving performance on real world multi-turn conversations applied to meta-llama/Meta-Llama-3-8B. This model was built using the Smaug recipe for improving performance on real world multi-turn conversations applied to meta-llama/Meta-Llama-3-8B.
@ -124,13 +124,13 @@
parameters: parameters:
model: Llama-3-Smaug-8B.Q4_K_M.gguf model: Llama-3-Smaug-8B.Q4_K_M.gguf
files: files:
- filename: Llama-3-Smaug-8B.Q4_K_M.gguf - filename: Llama-3-Smaug-8B.Q4_K_M.gguf
sha256: b17c4c1144768ead9e8a96439165baf49e98c53d458b4da8827f137fbabf38c1 sha256: b17c4c1144768ead9e8a96439165baf49e98c53d458b4da8827f137fbabf38c1
uri: huggingface://MaziyarPanahi/Llama-3-Smaug-8B-GGUF/Llama-3-Smaug-8B.Q4_K_M.gguf uri: huggingface://MaziyarPanahi/Llama-3-Smaug-8B-GGUF/Llama-3-Smaug-8B.Q4_K_M.gguf
- <<: *llama3 - <<: *llama3
name: "llama-3-8b-openhermes-dpo" name: "llama-3-8b-openhermes-dpo"
urls: urls:
- https://huggingface.co/mradermacher/Llama3-8B-OpenHermes-DPO-GGUF - https://huggingface.co/mradermacher/Llama3-8B-OpenHermes-DPO-GGUF
icon: https://cdn-uploads.huggingface.co/production/uploads/64fc6d81d75293f417fee1d1/QF2OsDu9DJKP4QYPBu4aK.png icon: https://cdn-uploads.huggingface.co/production/uploads/64fc6d81d75293f417fee1d1/QF2OsDu9DJKP4QYPBu4aK.png
description: | description: |
Llama3-8B-OpenHermes-DPO is DPO-Finetuned model of Llama3-8B, on the OpenHermes-2.5 preference dataset using QLoRA. Llama3-8B-OpenHermes-DPO is DPO-Finetuned model of Llama3-8B, on the OpenHermes-2.5 preference dataset using QLoRA.
@ -138,13 +138,13 @@
parameters: parameters:
model: Llama3-8B-OpenHermes-DPO.Q4_K_M.gguf model: Llama3-8B-OpenHermes-DPO.Q4_K_M.gguf
files: files:
- filename: Llama3-8B-OpenHermes-DPO.Q4_K_M.gguf - filename: Llama3-8B-OpenHermes-DPO.Q4_K_M.gguf
sha256: 1147e5881cb1d67796916e6cab7dab0ae0f532a4c1e626c9e92861e5f67752ca sha256: 1147e5881cb1d67796916e6cab7dab0ae0f532a4c1e626c9e92861e5f67752ca
uri: huggingface://mradermacher/Llama3-8B-OpenHermes-DPO-GGUF/Llama3-8B-OpenHermes-DPO.Q4_K_M.gguf uri: huggingface://mradermacher/Llama3-8B-OpenHermes-DPO-GGUF/Llama3-8B-OpenHermes-DPO.Q4_K_M.gguf
- <<: *llama3 - <<: *llama3
name: "llama-3-unholy-8b" name: "llama-3-unholy-8b"
urls: urls:
- https://huggingface.co/Undi95/Llama-3-Unholy-8B-GGUF - https://huggingface.co/Undi95/Llama-3-Unholy-8B-GGUF
icon: https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/JmdBlOHlBHVmX1IbZzWSv.png icon: https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/JmdBlOHlBHVmX1IbZzWSv.png
description: | description: |
Use at your own risk, I'm not responsible for any usage of this model, don't try to do anything this model tell you to do. Use at your own risk, I'm not responsible for any usage of this model, don't try to do anything this model tell you to do.
@ -156,13 +156,13 @@
parameters: parameters:
model: Llama-3-Unholy-8B.q4_k_m.gguf model: Llama-3-Unholy-8B.q4_k_m.gguf
files: files:
- filename: Llama-3-Unholy-8B.q4_k_m.gguf - filename: Llama-3-Unholy-8B.q4_k_m.gguf
sha256: 17b7f716bce1b34d4aa99ee730a19a834f8c77ddb36090dde5a1eda963f93602 sha256: 17b7f716bce1b34d4aa99ee730a19a834f8c77ddb36090dde5a1eda963f93602
uri: huggingface://Undi95/Llama-3-Unholy-8B-GGUF/Llama-3-Unholy-8B.q4_k_m.gguf uri: huggingface://Undi95/Llama-3-Unholy-8B-GGUF/Llama-3-Unholy-8B.q4_k_m.gguf
- <<: *llama3 - <<: *llama3
name: "lexi-llama-3-8b-uncensored" name: "lexi-llama-3-8b-uncensored"
urls: urls:
- https://huggingface.co/NikolayKozloff/Lexi-Llama-3-8B-Uncensored-Q6_K-GGUF - https://huggingface.co/NikolayKozloff/Lexi-Llama-3-8B-Uncensored-Q6_K-GGUF
icon: https://cdn-uploads.huggingface.co/production/uploads/644ad182f434a6a63b18eee6/H6axm5mlmiOWnbIFvx_em.png icon: https://cdn-uploads.huggingface.co/production/uploads/644ad182f434a6a63b18eee6/H6axm5mlmiOWnbIFvx_em.png
description: | description: |
Lexi is uncensored, which makes the model compliant. You are advised to implement your own alignment layer before exposing the model as a service. It will be highly compliant with any requests, even unethical ones. Lexi is uncensored, which makes the model compliant. You are advised to implement your own alignment layer before exposing the model as a service. It will be highly compliant with any requests, even unethical ones.
@ -174,13 +174,13 @@
parameters: parameters:
model: lexi-llama-3-8b-uncensored.Q6_K.gguf model: lexi-llama-3-8b-uncensored.Q6_K.gguf
files: files:
- filename: lexi-llama-3-8b-uncensored.Q6_K.gguf - filename: lexi-llama-3-8b-uncensored.Q6_K.gguf
sha256: 5805f3856cc18a769fae0b7c5659fe6778574691c370c910dad6eeec62c62436 sha256: 5805f3856cc18a769fae0b7c5659fe6778574691c370c910dad6eeec62c62436
uri: huggingface://NikolayKozloff/Lexi-Llama-3-8B-Uncensored-Q6_K-GGUF/lexi-llama-3-8b-uncensored.Q6_K.gguf uri: huggingface://NikolayKozloff/Lexi-Llama-3-8B-Uncensored-Q6_K-GGUF/lexi-llama-3-8b-uncensored.Q6_K.gguf
- <<: *llama3 - <<: *llama3
name: "chaos-rp_l3_b-iq-imatrix" name: "chaos-rp_l3_b-iq-imatrix"
urls: urls:
- https://huggingface.co/Lewdiculous/Chaos_RP_l3_8B-GGUF-IQ-Imatrix - https://huggingface.co/Lewdiculous/Chaos_RP_l3_8B-GGUF-IQ-Imatrix
icon: https://cdn-uploads.huggingface.co/production/uploads/626dfb8786671a29c715f8a9/u5p9kdbXT2QQA3iMU0vF1.png icon: https://cdn-uploads.huggingface.co/production/uploads/626dfb8786671a29c715f8a9/u5p9kdbXT2QQA3iMU0vF1.png
description: | description: |
A chaotic force beckons for you, will you heed her call? A chaotic force beckons for you, will you heed her call?
@ -192,13 +192,13 @@
parameters: parameters:
model: Chaos_RP_l3_8B-Q4_K_M-imat.gguf model: Chaos_RP_l3_8B-Q4_K_M-imat.gguf
files: files:
- filename: Chaos_RP_l3_8B-Q4_K_M-imat.gguf - filename: Chaos_RP_l3_8B-Q4_K_M-imat.gguf
sha256: 4273c5a8f23d49bf6294e620a5aa1fcd78d491ea0b90d0ec63ad708eedb83893 sha256: 4273c5a8f23d49bf6294e620a5aa1fcd78d491ea0b90d0ec63ad708eedb83893
uri: huggingface://Lewdiculous/Chaos_RP_l3_8B-GGUF-IQ-Imatrix/Chaos_RP_l3_8B-Q4_K_M-imat.gguf uri: huggingface://Lewdiculous/Chaos_RP_l3_8B-GGUF-IQ-Imatrix/Chaos_RP_l3_8B-Q4_K_M-imat.gguf
- <<: *llama3 - <<: *llama3
name: "llama-3-unholy-8b:Q8_0" name: "llama-3-unholy-8b:Q8_0"
urls: urls:
- https://huggingface.co/Undi95/Llama-3-Unholy-8B-GGUF - https://huggingface.co/Undi95/Llama-3-Unholy-8B-GGUF
icon: https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/JmdBlOHlBHVmX1IbZzWSv.png icon: https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/JmdBlOHlBHVmX1IbZzWSv.png
description: | description: |
Use at your own risk, I'm not responsible for any usage of this model, don't try to do anything this model tell you to do. Use at your own risk, I'm not responsible for any usage of this model, don't try to do anything this model tell you to do.
@ -210,13 +210,13 @@
parameters: parameters:
model: Llama-3-Unholy-8B.q8_0.gguf model: Llama-3-Unholy-8B.q8_0.gguf
files: files:
- filename: Llama-3-Unholy-8B.q8_0.gguf - filename: Llama-3-Unholy-8B.q8_0.gguf
sha256: 8d4137018acdcd57df4beccc84d9ad3f7f08cac50588f76370afc16c85752702 sha256: 8d4137018acdcd57df4beccc84d9ad3f7f08cac50588f76370afc16c85752702
uri: huggingface://Undi95/Llama-3-Unholy-8B-GGUF/Llama-3-Unholy-8B.q8_0.gguf uri: huggingface://Undi95/Llama-3-Unholy-8B-GGUF/Llama-3-Unholy-8B.q8_0.gguf
- <<: *llama3 - <<: *llama3
name: "therapyllama-8b-v1" name: "therapyllama-8b-v1"
urls: urls:
- https://huggingface.co/victunes/TherapyLlama-8B-v1-GGUF - https://huggingface.co/victunes/TherapyLlama-8B-v1-GGUF
icon: https://cdn-uploads.huggingface.co/production/uploads/65f07d05279d2d8f725bf0c3/A-ckcZ9H0Ee1n_ls2FM41.png icon: https://cdn-uploads.huggingface.co/production/uploads/65f07d05279d2d8f725bf0c3/A-ckcZ9H0Ee1n_ls2FM41.png
description: | description: |
Trained on Llama 3 8B using a modified version of jerryjalapeno/nart-100k-synthetic. Trained on Llama 3 8B using a modified version of jerryjalapeno/nart-100k-synthetic.
@ -243,13 +243,13 @@
parameters: parameters:
model: TherapyLlama-8B-v1-Q4_K_M.gguf model: TherapyLlama-8B-v1-Q4_K_M.gguf
files: files:
- filename: TherapyLlama-8B-v1-Q4_K_M.gguf - filename: TherapyLlama-8B-v1-Q4_K_M.gguf
sha256: 3d5a16d458e074a7bc7e706a493d8e95e8a7b2cb16934c851aece0af9d1da14a sha256: 3d5a16d458e074a7bc7e706a493d8e95e8a7b2cb16934c851aece0af9d1da14a
uri: huggingface://victunes/TherapyLlama-8B-v1-GGUF/TherapyLlama-8B-v1-Q4_K_M.gguf uri: huggingface://victunes/TherapyLlama-8B-v1-GGUF/TherapyLlama-8B-v1-Q4_K_M.gguf
- <<: *llama3 - <<: *llama3
name: "aura-uncensored-l3-8b-iq-imatrix" name: "aura-uncensored-l3-8b-iq-imatrix"
urls: urls:
- https://huggingface.co/Lewdiculous/Aura_Uncensored_l3_8B-GGUF-IQ-Imatrix - https://huggingface.co/Lewdiculous/Aura_Uncensored_l3_8B-GGUF-IQ-Imatrix
icon: https://cdn-uploads.huggingface.co/production/uploads/626dfb8786671a29c715f8a9/oiYHWIEHqmgUkY0GsVdDx.png icon: https://cdn-uploads.huggingface.co/production/uploads/626dfb8786671a29c715f8a9/oiYHWIEHqmgUkY0GsVdDx.png
description: | description: |
This is another better atempt at a less censored Llama-3 with hopefully more stable formatting. This is another better atempt at a less censored Llama-3 with hopefully more stable formatting.
@ -257,20 +257,20 @@
parameters: parameters:
model: Aura_Uncensored_l3_8B-Q4_K_M-imat.gguf model: Aura_Uncensored_l3_8B-Q4_K_M-imat.gguf
files: files:
- filename: Aura_Uncensored_l3_8B-Q4_K_M-imat.gguf - filename: Aura_Uncensored_l3_8B-Q4_K_M-imat.gguf
sha256: 265ded6a4f439bec160f394e3083a4a20e32ebb9d1d2d85196aaab23dab87fb2 sha256: 265ded6a4f439bec160f394e3083a4a20e32ebb9d1d2d85196aaab23dab87fb2
uri: huggingface://Lewdiculous/Aura_Uncensored_l3_8B-GGUF-IQ-Imatrix/Aura_Uncensored_l3_8B-Q4_K_M-imat.gguf uri: huggingface://Lewdiculous/Aura_Uncensored_l3_8B-GGUF-IQ-Imatrix/Aura_Uncensored_l3_8B-Q4_K_M-imat.gguf
- &dolphin - &dolphin
name: "dolphin-2.9-llama3-8b" name: "dolphin-2.9-llama3-8b"
url: "github:mudler/LocalAI/gallery/hermes-2-pro-mistral.yaml@master" url: "github:mudler/LocalAI/gallery/hermes-2-pro-mistral.yaml@master"
urls: urls:
- https://huggingface.co/cognitivecomputations/dolphin-2.9-llama3-8b-gguf - https://huggingface.co/cognitivecomputations/dolphin-2.9-llama3-8b-gguf
tags: tags:
- llm - llm
- gguf - gguf
- gpu - gpu
- cpu - cpu
- llama3 - llama3
license: llama3 license: llama3
description: | description: |
Dolphin-2.9 has a variety of instruction, conversational, and coding skills. It also has initial agentic abilities and supports function calling. Dolphin-2.9 has a variety of instruction, conversational, and coding skills. It also has initial agentic abilities and supports function calling.
@ -281,18 +281,18 @@
parameters: parameters:
model: dolphin-2.9-llama3-8b-q4_K_M.gguf model: dolphin-2.9-llama3-8b-q4_K_M.gguf
files: files:
- filename: dolphin-2.9-llama3-8b-q4_K_M.gguf - filename: dolphin-2.9-llama3-8b-q4_K_M.gguf
sha256: be988199ce28458e97205b11ae9d9cf4e3d8e18ff4c784e75bfc12f54407f1a1 sha256: be988199ce28458e97205b11ae9d9cf4e3d8e18ff4c784e75bfc12f54407f1a1
uri: huggingface://cognitivecomputations/dolphin-2.9-llama3-8b-gguf/dolphin-2.9-llama3-8b-q4_K_M.gguf uri: huggingface://cognitivecomputations/dolphin-2.9-llama3-8b-gguf/dolphin-2.9-llama3-8b-q4_K_M.gguf
- <<: *dolphin - <<: *dolphin
name: "dolphin-2.9-llama3-8b:Q6_K" name: "dolphin-2.9-llama3-8b:Q6_K"
overrides: overrides:
parameters: parameters:
model: dolphin-2.9-llama3-8b-q6_K.gguf model: dolphin-2.9-llama3-8b-q6_K.gguf
files: files:
- filename: dolphin-2.9-llama3-8b-q6_K.gguf - filename: dolphin-2.9-llama3-8b-q6_K.gguf
sha256: 8aac72a0bd72c075ba7be1aa29945e47b07d39cd16be9a80933935f51b57fb32 sha256: 8aac72a0bd72c075ba7be1aa29945e47b07d39cd16be9a80933935f51b57fb32
uri: huggingface://cognitivecomputations/dolphin-2.9-llama3-8b-gguf/dolphin-2.9-llama3-8b-q6_K.gguf uri: huggingface://cognitivecomputations/dolphin-2.9-llama3-8b-gguf/dolphin-2.9-llama3-8b-q6_K.gguf
## LLama2 and derivatives ## LLama2 and derivatives
### Start Fimbulvetr ### Start Fimbulvetr
- &vicuna-chat - &vicuna-chat
@ -304,20 +304,20 @@
description: | description: |
Cute girl to catch your attention. Cute girl to catch your attention.
urls: urls:
- https://huggingface.co/Sao10K/Fimbulvetr-11B-v2-GGUF - https://huggingface.co/Sao10K/Fimbulvetr-11B-v2-GGUF
tags: tags:
- llm - llm
- gguf - gguf
- gpu - gpu
- cpu - cpu
- llama3 - llama3
overrides: overrides:
parameters: parameters:
model: Fimbulvetr-11B-v2-Test-14.q4_K_M.gguf model: Fimbulvetr-11B-v2-Test-14.q4_K_M.gguf
files: files:
- filename: Fimbulvetr-11B-v2-Test-14.q4_K_M.gguf - filename: Fimbulvetr-11B-v2-Test-14.q4_K_M.gguf
sha256: 3597dacfb0ab717d565d8a4d6067f10dcb0e26cc7f21c832af1a10a87882a8fd sha256: 3597dacfb0ab717d565d8a4d6067f10dcb0e26cc7f21c832af1a10a87882a8fd
uri: huggingface://Sao10K/Fimbulvetr-11B-v2-GGUF/Fimbulvetr-11B-v2-Test-14.q4_K_M.gguf uri: huggingface://Sao10K/Fimbulvetr-11B-v2-GGUF/Fimbulvetr-11B-v2-Test-14.q4_K_M.gguf
### Start noromaid ### Start noromaid
- &noromaid - &noromaid
url: "github:mudler/LocalAI/gallery/noromaid.yaml@master" url: "github:mudler/LocalAI/gallery/noromaid.yaml@master"
@ -325,20 +325,20 @@
icon: https://cdn-uploads.huggingface.co/production/uploads/630dfb008df86f1e5becadc3/VKX2Z2yjZX5J8kXzgeCYO.png icon: https://cdn-uploads.huggingface.co/production/uploads/630dfb008df86f1e5becadc3/VKX2Z2yjZX5J8kXzgeCYO.png
license: cc-by-nc-4.0 license: cc-by-nc-4.0
urls: urls:
- https://huggingface.co/NeverSleep/Noromaid-13B-0.4-DPO-GGUF - https://huggingface.co/NeverSleep/Noromaid-13B-0.4-DPO-GGUF
tags: tags:
- llm - llm
- llama2 - llama2
- gguf - gguf
- gpu - gpu
- cpu - cpu
overrides: overrides:
parameters: parameters:
model: Noromaid-13B-0.4-DPO.q4_k_m.gguf model: Noromaid-13B-0.4-DPO.q4_k_m.gguf
files: files:
- filename: Noromaid-13B-0.4-DPO.q4_k_m.gguf - filename: Noromaid-13B-0.4-DPO.q4_k_m.gguf
sha256: cb28e878d034fae3d0b43326c5fc1cfb4ab583b17c56e41d6ce023caec03c1c1 sha256: cb28e878d034fae3d0b43326c5fc1cfb4ab583b17c56e41d6ce023caec03c1c1
uri: huggingface://NeverSleep/Noromaid-13B-0.4-DPO-GGUF/Noromaid-13B-0.4-DPO.q4_k_m.gguf uri: huggingface://NeverSleep/Noromaid-13B-0.4-DPO-GGUF/Noromaid-13B-0.4-DPO.q4_k_m.gguf
### START LLaVa ### START LLaVa
- &llava - &llava
url: "github:mudler/LocalAI/gallery/llava.yaml@master" url: "github:mudler/LocalAI/gallery/llava.yaml@master"
@ -348,25 +348,25 @@
LLaVA represents a novel end-to-end trained large multimodal model that combines a vision encoder and Vicuna for general-purpose visual and language understanding, achieving impressive chat capabilities mimicking spirits of the multimodal GPT-4 and setting a new state-of-the-art accuracy on Science QA. LLaVA represents a novel end-to-end trained large multimodal model that combines a vision encoder and Vicuna for general-purpose visual and language understanding, achieving impressive chat capabilities mimicking spirits of the multimodal GPT-4 and setting a new state-of-the-art accuracy on Science QA.
urls: urls:
- https://llava-vl.github.io/ - https://llava-vl.github.io/
tags: tags:
- llm - llm
- multimodal - multimodal
- gguf - gguf
- gpu - gpu
- llama2 - llama2
- cpu - cpu
name: "llava-1.6-vicuna" name: "llava-1.6-vicuna"
overrides: overrides:
mmproj: mmproj-vicuna7b-f16.gguf mmproj: mmproj-vicuna7b-f16.gguf
parameters: parameters:
model: vicuna-7b-q5_k.gguf model: vicuna-7b-q5_k.gguf
files: files:
- filename: vicuna-7b-q5_k.gguf - filename: vicuna-7b-q5_k.gguf
uri: https://huggingface.co/cmp-nct/llava-1.6-gguf/resolve/main/vicuna-7b-q5_k.gguf uri: https://huggingface.co/cmp-nct/llava-1.6-gguf/resolve/main/vicuna-7b-q5_k.gguf
- filename: mmproj-vicuna7b-f16.gguf - filename: mmproj-vicuna7b-f16.gguf
uri: https://huggingface.co/cmp-nct/llava-1.6-gguf/resolve/main/mmproj-vicuna7b-f16.gguf uri: https://huggingface.co/cmp-nct/llava-1.6-gguf/resolve/main/mmproj-vicuna7b-f16.gguf
- <<: *llava - <<: *llava
name: "llava-1.6-mistral" name: "llava-1.6-mistral"
overrides: overrides:
@ -374,12 +374,12 @@
parameters: parameters:
model: llava-v1.6-mistral-7b.gguf model: llava-v1.6-mistral-7b.gguf
files: files:
- filename: llava-v1.6-mistral-7b.gguf - filename: llava-v1.6-mistral-7b.gguf
sha256: 31826170ffa2e8080bbcd74cac718f906484fd5a59895550ef94c1baa4997595 sha256: 31826170ffa2e8080bbcd74cac718f906484fd5a59895550ef94c1baa4997595
uri: huggingface://cjpais/llava-1.6-mistral-7b-gguf/llava-v1.6-mistral-7b.Q6_K.gguf uri: huggingface://cjpais/llava-1.6-mistral-7b-gguf/llava-v1.6-mistral-7b.Q6_K.gguf
- filename: llava-v1.6-7b-mmproj-f16.gguf - filename: llava-v1.6-7b-mmproj-f16.gguf
sha256: 00205ee8a0d7a381900cd031e43105f86aa0d8c07bf329851e85c71a26632d16 sha256: 00205ee8a0d7a381900cd031e43105f86aa0d8c07bf329851e85c71a26632d16
uri: huggingface://cjpais/llava-1.6-mistral-7b-gguf/mmproj-model-f16.gguf uri: huggingface://cjpais/llava-1.6-mistral-7b-gguf/mmproj-model-f16.gguf
- <<: *llava - <<: *llava
name: "llava-1.5" name: "llava-1.5"
overrides: overrides:
@ -387,12 +387,12 @@
parameters: parameters:
model: llava-v1.5-7b-Q4_K.gguf model: llava-v1.5-7b-Q4_K.gguf
files: files:
- filename: llava-v1.5-7b-Q4_K.gguf - filename: llava-v1.5-7b-Q4_K.gguf
sha256: c91ebf0a628ceb25e374df23ad966cc1bf1514b33fecf4f0073f9619dec5b3f9 sha256: c91ebf0a628ceb25e374df23ad966cc1bf1514b33fecf4f0073f9619dec5b3f9
uri: huggingface://jartine/llava-v1.5-7B-GGUF/llava-v1.5-7b-Q4_K.gguf uri: huggingface://jartine/llava-v1.5-7B-GGUF/llava-v1.5-7b-Q4_K.gguf
- filename: llava-v1.5-7b-mmproj-Q8_0.gguf - filename: llava-v1.5-7b-mmproj-Q8_0.gguf
sha256: 09c230de47f6f843e4841656f7895cac52c6e7ec7392acb5e8527de8b775c45a sha256: 09c230de47f6f843e4841656f7895cac52c6e7ec7392acb5e8527de8b775c45a
uri: huggingface://jartine/llava-v1.5-7B-GGUF/llava-v1.5-7b-mmproj-Q8_0.gguf uri: huggingface://jartine/llava-v1.5-7B-GGUF/llava-v1.5-7b-mmproj-Q8_0.gguf
### START Phi-2 ### START Phi-2
- &phi-2-chat - &phi-2-chat
url: "github:mudler/LocalAI/gallery/phi-2-chat.yaml@master" url: "github:mudler/LocalAI/gallery/phi-2-chat.yaml@master"
@ -416,32 +416,32 @@
Finetuned from model: Phi-2 Finetuned from model: Phi-2
urls: urls:
- https://huggingface.co/l3utterfly/phi-2-layla-v1-chatml - https://huggingface.co/l3utterfly/phi-2-layla-v1-chatml
- https://huggingface.co/l3utterfly/phi-2-layla-v1-chatml-gguf - https://huggingface.co/l3utterfly/phi-2-layla-v1-chatml-gguf
tags: tags:
- llm - llm
- gguf - gguf
- gpu - gpu
- llama2 - llama2
- cpu - cpu
name: "phi-2-chat:Q8_0" name: "phi-2-chat:Q8_0"
overrides: overrides:
parameters: parameters:
model: phi-2-layla-v1-chatml-Q8_0.gguf model: phi-2-layla-v1-chatml-Q8_0.gguf
files: files:
- filename: "phi-2-layla-v1-chatml-Q8_0.gguf" - filename: "phi-2-layla-v1-chatml-Q8_0.gguf"
sha256: "0cf542a127c2c835066a78028009b7eddbaf773cc2a26e1cb157ce5e09c1a2e0" sha256: "0cf542a127c2c835066a78028009b7eddbaf773cc2a26e1cb157ce5e09c1a2e0"
uri: "huggingface://l3utterfly/phi-2-layla-v1-chatml-gguf/phi-2-layla-v1-chatml-Q8_0.gguf" uri: "huggingface://l3utterfly/phi-2-layla-v1-chatml-gguf/phi-2-layla-v1-chatml-Q8_0.gguf"
- <<: *phi-2-chat - <<: *phi-2-chat
name: "phi-2-chat" name: "phi-2-chat"
overrides: overrides:
parameters: parameters:
model: phi-2-layla-v1-chatml-Q4_K.gguf model: phi-2-layla-v1-chatml-Q4_K.gguf
files: files:
- filename: "phi-2-layla-v1-chatml-Q4_K.gguf" - filename: "phi-2-layla-v1-chatml-Q4_K.gguf"
sha256: "b071e5624b60b8911f77261398802c4b4079c6c689e38e2ce75173ed62bc8a48" sha256: "b071e5624b60b8911f77261398802c4b4079c6c689e38e2ce75173ed62bc8a48"
uri: "huggingface://l3utterfly/phi-2-layla-v1-chatml-gguf/phi-2-layla-v1-chatml-Q4_K.gguf" uri: "huggingface://l3utterfly/phi-2-layla-v1-chatml-gguf/phi-2-layla-v1-chatml-Q4_K.gguf"
- <<: *phi-2-chat - <<: *phi-2-chat
license: mit license: mit
icon: "https://huggingface.co/rhysjones/phi-2-orange/resolve/main/phi-2-orange.jpg" icon: "https://huggingface.co/rhysjones/phi-2-orange/resolve/main/phi-2-orange.jpg"
@ -450,23 +450,23 @@
There is an updated model at rhysjones/phi-2-orange-v2 which has higher evals, if you wish to test. There is an updated model at rhysjones/phi-2-orange-v2 which has higher evals, if you wish to test.
urls: urls:
- https://huggingface.co/rhysjones/phi-2-orange - https://huggingface.co/rhysjones/phi-2-orange
- https://huggingface.co/TheBloke/phi-2-orange-GGUF - https://huggingface.co/TheBloke/phi-2-orange-GGUF
tags: tags:
- llm - llm
- gguf - gguf
- llama2 - llama2
- gpu - gpu
- cpu - cpu
name: "phi-2-orange" name: "phi-2-orange"
overrides: overrides:
parameters: parameters:
model: phi-2-orange.Q4_0.gguf model: phi-2-orange.Q4_0.gguf
files: files:
- filename: "phi-2-orange.Q4_0.gguf" - filename: "phi-2-orange.Q4_0.gguf"
sha256: "49cb710ae688e1b19b1b299087fa40765a0cd677e3afcc45e5f7ef6750975dcf" sha256: "49cb710ae688e1b19b1b299087fa40765a0cd677e3afcc45e5f7ef6750975dcf"
uri: "huggingface://TheBloke/phi-2-orange-GGUF/phi-2-orange.Q4_0.gguf" uri: "huggingface://TheBloke/phi-2-orange-GGUF/phi-2-orange.Q4_0.gguf"
### START Phi-3 ### START Phi-3
- &phi-3 - &phi-3
url: "github:mudler/LocalAI/gallery/phi-3-chat.yaml@master" url: "github:mudler/LocalAI/gallery/phi-3-chat.yaml@master"
@ -477,30 +477,30 @@
The Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets that includes both synthetic data and the filtered publicly available websites data with a focus on high-quality and reasoning dense properties. The model belongs to the Phi-3 family with the Mini version in two variants 4K and 128K which is the context length (in tokens) it can support. The model has underwent a post-training process that incorporates both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures. When assessed against benchmarks testing common sense, language understanding, math, code, long context and logical reasoning, Phi-3 Mini-4K-Instruct showcased a robust and state-of-the-art performance among models with less than 13 billion parameters. The Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets that includes both synthetic data and the filtered publicly available websites data with a focus on high-quality and reasoning dense properties. The model belongs to the Phi-3 family with the Mini version in two variants 4K and 128K which is the context length (in tokens) it can support. The model has underwent a post-training process that incorporates both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures. When assessed against benchmarks testing common sense, language understanding, math, code, long context and logical reasoning, Phi-3 Mini-4K-Instruct showcased a robust and state-of-the-art performance among models with less than 13 billion parameters.
urls: urls:
- https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf - https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf
tags: tags:
- llm - llm
- gguf - gguf
- gpu - gpu
- llama2 - llama2
- cpu - cpu
overrides: overrides:
parameters: parameters:
model: Phi-3-mini-4k-instruct-q4.gguf model: Phi-3-mini-4k-instruct-q4.gguf
files: files:
- filename: "Phi-3-mini-4k-instruct-q4.gguf" - filename: "Phi-3-mini-4k-instruct-q4.gguf"
sha256: "4fed7364ee3e0c7cb4fe0880148bfdfcd1b630981efa0802a6b62ee52e7da97e" sha256: "4fed7364ee3e0c7cb4fe0880148bfdfcd1b630981efa0802a6b62ee52e7da97e"
uri: "huggingface://microsoft/Phi-3-mini-4k-instruct-gguf/Phi-3-mini-4k-instruct-q4.gguf" uri: "huggingface://microsoft/Phi-3-mini-4k-instruct-gguf/Phi-3-mini-4k-instruct-q4.gguf"
- <<: *phi-3 - <<: *phi-3
name: "phi-3-mini-4k-instruct:fp16" name: "phi-3-mini-4k-instruct:fp16"
overrides: overrides:
parameters: parameters:
model: Phi-3-mini-4k-instruct-fp16.gguf model: Phi-3-mini-4k-instruct-fp16.gguf
files: files:
- filename: "Phi-3-mini-4k-instruct-fp16.gguf" - filename: "Phi-3-mini-4k-instruct-fp16.gguf"
sha256: "ad9f8ff11cd096115adc8ff50befa22fc3da2718672ddd2ab30faccd70488605" sha256: "ad9f8ff11cd096115adc8ff50befa22fc3da2718672ddd2ab30faccd70488605"
uri: "huggingface://microsoft/Phi-3-mini-4k-instruct-gguf/Phi-3-mini-4k-instruct-fp16.gguf" uri: "huggingface://microsoft/Phi-3-mini-4k-instruct-gguf/Phi-3-mini-4k-instruct-fp16.gguf"
### START Hermes-2-Pro-Mistral ### START Hermes-2-Pro-Mistral
- &hermes-2-pro-mistral - &hermes-2-pro-mistral
url: "github:mudler/LocalAI/gallery/hermes-2-pro-mistral.yaml@master" url: "github:mudler/LocalAI/gallery/hermes-2-pro-mistral.yaml@master"
@ -520,39 +520,39 @@
Learn more about the function calling on our github repo here: https://github.com/NousResearch/Hermes-Function-Calling/tree/main Learn more about the function calling on our github repo here: https://github.com/NousResearch/Hermes-Function-Calling/tree/main
urls: urls:
- https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B-GGUF - https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B-GGUF
tags: tags:
- llm - llm
- gguf - gguf
- gpu - gpu
- llama2 - llama2
- cpu - cpu
overrides: overrides:
parameters: parameters:
model: Hermes-2-Pro-Mistral-7B.Q4_0.gguf model: Hermes-2-Pro-Mistral-7B.Q4_0.gguf
files: files:
- filename: "Hermes-2-Pro-Mistral-7B.Q4_0.gguf" - filename: "Hermes-2-Pro-Mistral-7B.Q4_0.gguf"
sha256: "f446c3125026f7af6757dd097dda02280adc85e908c058bd6f1c41a118354745" sha256: "f446c3125026f7af6757dd097dda02280adc85e908c058bd6f1c41a118354745"
uri: "huggingface://NousResearch/Hermes-2-Pro-Mistral-7B-GGUF/Hermes-2-Pro-Mistral-7B.Q4_0.gguf" uri: "huggingface://NousResearch/Hermes-2-Pro-Mistral-7B-GGUF/Hermes-2-Pro-Mistral-7B.Q4_0.gguf"
- <<: *hermes-2-pro-mistral - <<: *hermes-2-pro-mistral
name: "hermes-2-pro-mistral:Q6_K" name: "hermes-2-pro-mistral:Q6_K"
overrides: overrides:
parameters: parameters:
model: Hermes-2-Pro-Mistral-7B.Q6_K.gguf model: Hermes-2-Pro-Mistral-7B.Q6_K.gguf
files: files:
- filename: "Hermes-2-Pro-Mistral-7B.Q6_K.gguf" - filename: "Hermes-2-Pro-Mistral-7B.Q6_K.gguf"
sha256: "40adc3b227bc36764de148fdda4df5df385adc06650d58d4dbe726ee0214eeff" sha256: "40adc3b227bc36764de148fdda4df5df385adc06650d58d4dbe726ee0214eeff"
uri: "huggingface://NousResearch/Hermes-2-Pro-Mistral-7B-GGUF/Hermes-2-Pro-Mistral-7B.Q6_K.gguf" uri: "huggingface://NousResearch/Hermes-2-Pro-Mistral-7B-GGUF/Hermes-2-Pro-Mistral-7B.Q6_K.gguf"
- <<: *hermes-2-pro-mistral - <<: *hermes-2-pro-mistral
name: "hermes-2-pro-mistral:Q8_0" name: "hermes-2-pro-mistral:Q8_0"
overrides: overrides:
parameters: parameters:
model: Hermes-2-Pro-Mistral-7B.Q8_0.gguf model: Hermes-2-Pro-Mistral-7B.Q8_0.gguf
files: files:
- filename: "Hermes-2-Pro-Mistral-7B.Q8_0.gguf" - filename: "Hermes-2-Pro-Mistral-7B.Q8_0.gguf"
sha256: "b6d95d7ec9a395b7568cc94b0447fd4f90b6f69d6e44794b1fbb84e3f732baca" sha256: "b6d95d7ec9a395b7568cc94b0447fd4f90b6f69d6e44794b1fbb84e3f732baca"
uri: "huggingface://NousResearch/Hermes-2-Pro-Mistral-7B-GGUF/Hermes-2-Pro-Mistral-7B.Q8_0.gguf" uri: "huggingface://NousResearch/Hermes-2-Pro-Mistral-7B-GGUF/Hermes-2-Pro-Mistral-7B.Q8_0.gguf"
### END Hermes-2-Pro-Mistral ### END Hermes-2-Pro-Mistral
### START Cerbero ### START Cerbero
- url: "github:mudler/LocalAI/gallery/cerbero.yaml@master" - url: "github:mudler/LocalAI/gallery/cerbero.yaml@master"
@ -560,20 +560,20 @@
description: | description: |
cerbero-7b is specifically crafted to fill the void in Italy's AI landscape. cerbero-7b is specifically crafted to fill the void in Italy's AI landscape.
urls: urls:
- https://huggingface.co/galatolo/cerbero-7b - https://huggingface.co/galatolo/cerbero-7b
tags: tags:
- llm - llm
- gguf - gguf
- gpu - gpu
- cpu - cpu
- mistral - mistral
overrides: overrides:
parameters: parameters:
model: galatolo-Q4_K.gguf model: galatolo-Q4_K.gguf
files: files:
- filename: "galatolo-Q4_K.gguf" - filename: "galatolo-Q4_K.gguf"
sha256: "ca0cfd5a9ad40dc16416aa3a277015d0299b62c0803b67f5709580042202c172" sha256: "ca0cfd5a9ad40dc16416aa3a277015d0299b62c0803b67f5709580042202c172"
uri: "huggingface://galatolo/cerbero-7b-gguf/ggml-model-Q4_K.gguf" uri: "huggingface://galatolo/cerbero-7b-gguf/ggml-model-Q4_K.gguf"
### START Codellama ### START Codellama
- &codellama - &codellama
url: "github:mudler/LocalAI/gallery/codellama.yaml@master" url: "github:mudler/LocalAI/gallery/codellama.yaml@master"
@ -584,34 +584,34 @@
Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 34 billion parameters. This model is designed for general code synthesis and understanding. Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 34 billion parameters. This model is designed for general code synthesis and understanding.
urls: urls:
- https://huggingface.co/TheBloke/CodeLlama-7B-GGUF - https://huggingface.co/TheBloke/CodeLlama-7B-GGUF
- https://huggingface.co/meta-llama/CodeLlama-7b-hf - https://huggingface.co/meta-llama/CodeLlama-7b-hf
tags: tags:
- llm - llm
- gguf - gguf
- gpu - gpu
- llama2 - llama2
- cpu - cpu
overrides: overrides:
parameters: parameters:
model: codellama-7b.Q4_0.gguf model: codellama-7b.Q4_0.gguf
files: files:
- filename: "codellama-7b.Q4_0.gguf" - filename: "codellama-7b.Q4_0.gguf"
sha256: "33052f6dd41436db2f83bd48017b6fff8ce0184e15a8a227368b4230f1da97b5" sha256: "33052f6dd41436db2f83bd48017b6fff8ce0184e15a8a227368b4230f1da97b5"
uri: "huggingface://TheBloke/CodeLlama-7B-GGUF/codellama-7b.Q4_0.gguf" uri: "huggingface://TheBloke/CodeLlama-7B-GGUF/codellama-7b.Q4_0.gguf"
### START Embeddings ### START Embeddings
- &sentencentransformers - &sentencentransformers
description: | description: |
This framework provides an easy method to compute dense vector representations for sentences, paragraphs, and images. The models are based on transformer networks like BERT / RoBERTa / XLM-RoBERTa etc. and achieve state-of-the-art performance in various tasks. Text is embedded in vector space such that similar text are closer and can efficiently be found using cosine similarity. This framework provides an easy method to compute dense vector representations for sentences, paragraphs, and images. The models are based on transformer networks like BERT / RoBERTa / XLM-RoBERTa etc. and achieve state-of-the-art performance in various tasks. Text is embedded in vector space such that similar text are closer and can efficiently be found using cosine similarity.
urls: urls:
- https://github.com/UKPLab/sentence-transformers - https://github.com/UKPLab/sentence-transformers
tags: tags:
- gpu - gpu
- cpu - cpu
- embeddings - embeddings
- python - python
name: "all-MiniLM-L6-v2" name: "all-MiniLM-L6-v2"
url: "github:mudler/LocalAI/gallery/sentencetransformers.yaml@master" url: "github:mudler/LocalAI/gallery/sentencetransformers.yaml@master"
overrides: overrides:
@ -628,30 +628,30 @@
A text-to-image model that uses Stable Diffusion 1.5 to generate images from text prompts. This model is DreamShaper model by Lykon. A text-to-image model that uses Stable Diffusion 1.5 to generate images from text prompts. This model is DreamShaper model by Lykon.
urls: urls:
- https://civitai.com/models/4384/dreamshaper - https://civitai.com/models/4384/dreamshaper
tags: tags:
- text-to-image - text-to-image
- stablediffusion - stablediffusion
- python - python
- sd-1.5 - sd-1.5
- gpu - gpu
url: "github:mudler/LocalAI/gallery/dreamshaper.yaml@master" url: "github:mudler/LocalAI/gallery/dreamshaper.yaml@master"
overrides: overrides:
parameters: parameters:
model: DreamShaper_8_pruned.safetensors model: DreamShaper_8_pruned.safetensors
files: files:
- filename: DreamShaper_8_pruned.safetensors - filename: DreamShaper_8_pruned.safetensors
uri: huggingface://Lykon/DreamShaper/DreamShaper_8_pruned.safetensors uri: huggingface://Lykon/DreamShaper/DreamShaper_8_pruned.safetensors
sha256: 879db523c30d3b9017143d56705015e15a2cb5628762c11d086fed9538abd7fd sha256: 879db523c30d3b9017143d56705015e15a2cb5628762c11d086fed9538abd7fd
## Whisper ## Whisper
- url: "github:mudler/LocalAI/gallery/whisper-base.yaml@master" - url: "github:mudler/LocalAI/gallery/whisper-base.yaml@master"
name: "whisper-1" name: "whisper-1"
license: "MIT" license: "MIT"
urls: urls:
- https://github.com/ggerganov/whisper.cpp - https://github.com/ggerganov/whisper.cpp
- https://huggingface.co/ggerganov/whisper.cpp - https://huggingface.co/ggerganov/whisper.cpp
description: | description: |
Port of OpenAI's Whisper model in C/C++ Port of OpenAI's Whisper model in C/C++
@ -661,9 +661,9 @@
name: "bert-embeddings" name: "bert-embeddings"
license: "Apache 2.0" license: "Apache 2.0"
urls: urls:
- https://huggingface.co/skeskinen/ggml - https://huggingface.co/skeskinen/ggml
tags: tags:
- embeddings - embeddings
description: | description: |
Bert model that can be used for embeddings Bert model that can be used for embeddings
@ -671,8 +671,8 @@
- url: github:mudler/LocalAI/gallery/stablediffusion.yaml@master - url: github:mudler/LocalAI/gallery/stablediffusion.yaml@master
license: "BSD-3" license: "BSD-3"
urls: urls:
- https://github.com/EdVince/Stable-Diffusion-NCNN - https://github.com/EdVince/Stable-Diffusion-NCNN
- https://github.com/EdVince/Stable-Diffusion-NCNN/blob/main/LICENSE - https://github.com/EdVince/Stable-Diffusion-NCNN/blob/main/LICENSE
description: | description: |
Stable Diffusion in NCNN with c++, supported txt2img and img2img Stable Diffusion in NCNN with c++, supported txt2img and img2img
@ -702,9 +702,9 @@
A fast, local neural text to speech system that sounds great and is optimized for the Raspberry Pi 4. Piper is used in a variety of [projects](https://github.com/rhasspy/piper#people-using-piper). A fast, local neural text to speech system that sounds great and is optimized for the Raspberry Pi 4. Piper is used in a variety of [projects](https://github.com/rhasspy/piper#people-using-piper).
tags: tags:
- tts - tts
- text-to-speech - text-to-speech
- cpu - cpu
override: override:
parameters: parameters:

View File

@ -1,6 +1,6 @@
---
name: "llama3-instruct" name: "llama3-instruct"
config_file: | config_file: |
mmap: true mmap: true
template: template:

View File

@ -1,6 +1,6 @@
---
name: "llava" name: "llava"
config_file: | config_file: |
backend: llama-cpp backend: llama-cpp
context_size: 4096 context_size: 4096

View File

@ -1,3 +1,4 @@
---
config_file: | config_file: |
mmap: true mmap: true
backend: llama-cpp backend: llama-cpp
@ -50,4 +51,3 @@ config_file: |
- <dummy32000> - <dummy32000>
- "\n</tool_call>" - "\n</tool_call>"
- "\n\n\n" - "\n\n\n"

View File

@ -1,2 +1,3 @@
---
config_file: | config_file: |
backend: parler-tts backend: parler-tts

View File

@ -1,6 +1,6 @@
---
name: "phi-2-chatml" name: "phi-2-chatml"
config_file: | config_file: |
mmap: true mmap: true
template: template:
@ -16,4 +16,3 @@ config_file: |
f16: true f16: true
stopwords: stopwords:
- <|im_end|> - <|im_end|>

View File

@ -1,3 +1,4 @@
---
name: "phi-2-orange" name: "phi-2-orange"
config_file: | config_file: |

View File

@ -1,3 +1,4 @@
---
name: "phi-3-chat" name: "phi-3-chat"
config_file: | config_file: |
@ -15,4 +16,3 @@ config_file: |
f16: true f16: true
stopwords: stopwords:
- <|end|> - <|end|>

View File

@ -1,2 +1,3 @@
---
config_file: | config_file: |
backend: piper backend: piper

View File

@ -1,2 +1,3 @@
---
config_file: | config_file: |
backend: rerankers backend: rerankers

View File

@ -1,3 +1,4 @@
---
name: "sentencetransformers" name: "sentencetransformers"
config_file: | config_file: |

View File

@ -1,3 +1,4 @@
---
name: "stablediffusion-cpp" name: "stablediffusion-cpp"
config_file: | config_file: |
@ -7,42 +8,42 @@ config_file: |
model: stablediffusion_assets model: stablediffusion_assets
files: files:
- filename: "stablediffusion_assets/AutoencoderKL-256-256-fp16-opt.param" - filename: "stablediffusion_assets/AutoencoderKL-256-256-fp16-opt.param"
sha256: "18ca4b66685e21406bcf64c484b3b680b4949900415536d599cc876579c85c82" sha256: "18ca4b66685e21406bcf64c484b3b680b4949900415536d599cc876579c85c82"
uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/AutoencoderKL-256-256-fp16-opt.param" uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/AutoencoderKL-256-256-fp16-opt.param"
- filename: "stablediffusion_assets/AutoencoderKL-512-512-fp16-opt.param" - filename: "stablediffusion_assets/AutoencoderKL-512-512-fp16-opt.param"
sha256: "cf45f63aacf3dbbab0f59ed92a6f2c14d9a1801314631cd3abe91e3c85639a20" sha256: "cf45f63aacf3dbbab0f59ed92a6f2c14d9a1801314631cd3abe91e3c85639a20"
uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/AutoencoderKL-512-512-fp16-opt.param" uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/AutoencoderKL-512-512-fp16-opt.param"
- filename: "stablediffusion_assets/AutoencoderKL-base-fp16.param" - filename: "stablediffusion_assets/AutoencoderKL-base-fp16.param"
sha256: "0254a056dce61b0c27dc9ec1b78b53bcf55315c540f55f051eb841aa992701ba" sha256: "0254a056dce61b0c27dc9ec1b78b53bcf55315c540f55f051eb841aa992701ba"
uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/AutoencoderKL-base-fp16.param" uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/AutoencoderKL-base-fp16.param"
- filename: "stablediffusion_assets/AutoencoderKL-encoder-512-512-fp16.bin" - filename: "stablediffusion_assets/AutoencoderKL-encoder-512-512-fp16.bin"
sha256: "ddcb79a9951b9f91e05e087739ed69da2c1c4ae30ba4168cce350b49d617c9fa" sha256: "ddcb79a9951b9f91e05e087739ed69da2c1c4ae30ba4168cce350b49d617c9fa"
uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/releases/download/naifu/AutoencoderKL-encoder-512-512-fp16.bin" uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/releases/download/naifu/AutoencoderKL-encoder-512-512-fp16.bin"
- filename: "stablediffusion_assets/AutoencoderKL-fp16.bin" - filename: "stablediffusion_assets/AutoencoderKL-fp16.bin"
sha256: "f02e71f80e70252734724bbfaed5c4ddd3a8ed7e61bb2175ff5f53099f0e35dd" sha256: "f02e71f80e70252734724bbfaed5c4ddd3a8ed7e61bb2175ff5f53099f0e35dd"
uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/releases/download/naifu/AutoencoderKL-fp16.bin" uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/releases/download/naifu/AutoencoderKL-fp16.bin"
- filename: "stablediffusion_assets/FrozenCLIPEmbedder-fp16.bin" - filename: "stablediffusion_assets/FrozenCLIPEmbedder-fp16.bin"
sha256: "1c9a12f4e1dd1b295a388045f7f28a2352a4d70c3dc96a542189a3dd7051fdd6" sha256: "1c9a12f4e1dd1b295a388045f7f28a2352a4d70c3dc96a542189a3dd7051fdd6"
uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/releases/download/naifu/FrozenCLIPEmbedder-fp16.bin" uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/releases/download/naifu/FrozenCLIPEmbedder-fp16.bin"
- filename: "stablediffusion_assets/FrozenCLIPEmbedder-fp16.param" - filename: "stablediffusion_assets/FrozenCLIPEmbedder-fp16.param"
sha256: "471afbe678dd1fd3fe764ef9c6eccaccb0a7d7e601f27b462aa926b20eb368c9" sha256: "471afbe678dd1fd3fe764ef9c6eccaccb0a7d7e601f27b462aa926b20eb368c9"
uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/FrozenCLIPEmbedder-fp16.param" uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/FrozenCLIPEmbedder-fp16.param"
- filename: "stablediffusion_assets/log_sigmas.bin" - filename: "stablediffusion_assets/log_sigmas.bin"
sha256: "a2089f8aa4c61f9c200feaec541ab3f5c94233b28deb6d5e8bcd974fa79b68ac" sha256: "a2089f8aa4c61f9c200feaec541ab3f5c94233b28deb6d5e8bcd974fa79b68ac"
uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/raw/main/x86/linux/assets/log_sigmas.bin" uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/raw/main/x86/linux/assets/log_sigmas.bin"
- filename: "stablediffusion_assets/UNetModel-256-256-MHA-fp16-opt.param" - filename: "stablediffusion_assets/UNetModel-256-256-MHA-fp16-opt.param"
sha256: "a58c380229f09491776df837b7aa7adffc0a87821dc4708b34535da2e36e3da1" sha256: "a58c380229f09491776df837b7aa7adffc0a87821dc4708b34535da2e36e3da1"
uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/UNetModel-256-256-MHA-fp16-opt.param" uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/UNetModel-256-256-MHA-fp16-opt.param"
- filename: "stablediffusion_assets/UNetModel-512-512-MHA-fp16-opt.param" - filename: "stablediffusion_assets/UNetModel-512-512-MHA-fp16-opt.param"
sha256: "f12034067062827bd7f43d1d21888d1f03905401acf6c6eea22be23c259636fa" sha256: "f12034067062827bd7f43d1d21888d1f03905401acf6c6eea22be23c259636fa"
uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/UNetModel-512-512-MHA-fp16-opt.param" uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/UNetModel-512-512-MHA-fp16-opt.param"
- filename: "stablediffusion_assets/UNetModel-base-MHA-fp16.param" - filename: "stablediffusion_assets/UNetModel-base-MHA-fp16.param"
sha256: "696f6975de49f4325b53ce32aff81861a6d6c07cd9ce3f0aae2cc405350af38d" sha256: "696f6975de49f4325b53ce32aff81861a6d6c07cd9ce3f0aae2cc405350af38d"
uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/UNetModel-base-MHA-fp16.param" uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/UNetModel-base-MHA-fp16.param"
- filename: "stablediffusion_assets/UNetModel-MHA-fp16.bin" - filename: "stablediffusion_assets/UNetModel-MHA-fp16.bin"
sha256: "d618918d011bfc1f644c0f2a33bf84931bd53b28a98492b0a8ed6f3a818852c3" sha256: "d618918d011bfc1f644c0f2a33bf84931bd53b28a98492b0a8ed6f3a818852c3"
uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/releases/download/naifu/UNetModel-MHA-fp16.bin" uri: "https://github.com/EdVince/Stable-Diffusion-NCNN/releases/download/naifu/UNetModel-MHA-fp16.bin"
- filename: "stablediffusion_assets/vocab.txt" - filename: "stablediffusion_assets/vocab.txt"
sha256: "e30e57b6f1e47616982ef898d8922be24e535b4fa3d0110477b3a6f02ebbae7d" sha256: "e30e57b6f1e47616982ef898d8922be24e535b4fa3d0110477b3a6f02ebbae7d"
uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/vocab.txt" uri: "https://raw.githubusercontent.com/EdVince/Stable-Diffusion-NCNN/main/x86/linux/assets/vocab.txt"

View File

@ -1,3 +1,4 @@
---
name: "tinydream" name: "tinydream"
config_file: | config_file: |

View File

@ -1,3 +1,4 @@
---
name: "vicuna-chat" name: "vicuna-chat"
description: | description: |

View File

@ -1,3 +1,4 @@
---
name: "virtual" name: "virtual"
description: | description: |

View File

@ -1,12 +1,12 @@
---
name: "whisper-base" name: "whisper-base"
config_file: | config_file: |
backend: whisper backend: whisper
parameters: parameters:
model: ggml-whisper-base.bin model: ggml-whisper-base.bin
files: files:
- filename: "ggml-whisper-base.bin" - filename: "ggml-whisper-base.bin"
sha256: "60ed5bc3dd14eea856493d334349b405782ddcaf0028d4b5df4088345fba2efe" sha256: "60ed5bc3dd14eea856493d334349b405782ddcaf0028d4b5df4088345fba2efe"
uri: "https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-base.bin" uri: "https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-base.bin"