From af41436f1bf40fca937990ae6bede9dd3f6f0cd0 Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Fri, 31 Jan 2025 09:57:58 +0100 Subject: [PATCH] fix(tests): pin to branch for config used in tests (#4721) Signed-off-by: Ettore Di Giacinto --- core/config/backend_config_test.go | 4 ++-- core/http/app_test.go | 4 ++-- docs/content/docs/features/model-gallery.md | 4 ++-- 3 files changed, 6 insertions(+), 6 deletions(-) diff --git a/core/config/backend_config_test.go b/core/config/backend_config_test.go index 04eacb7e..e6a54b89 100644 --- a/core/config/backend_config_test.go +++ b/core/config/backend_config_test.go @@ -48,9 +48,9 @@ parameters: Expect(config.Name).To(Equal("bar-baz")) Expect(config.Validate()).To(BeTrue()) - // download https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/hermes-2-pro-mistral.yaml + // download https://raw.githubusercontent.com/mudler/LocalAI/v2.25.0/embedded/models/hermes-2-pro-mistral.yaml httpClient := http.Client{} - resp, err := httpClient.Get("https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/hermes-2-pro-mistral.yaml") + resp, err := httpClient.Get("https://raw.githubusercontent.com/mudler/LocalAI/v2.25.0/embedded/models/hermes-2-pro-mistral.yaml") Expect(err).To(BeNil()) defer resp.Body.Close() tmp, err = os.CreateTemp("", "config.yaml") diff --git a/core/http/app_test.go b/core/http/app_test.go index f57a3ea7..bc4ecfae 100644 --- a/core/http/app_test.go +++ b/core/http/app_test.go @@ -476,7 +476,7 @@ var _ = Describe("API test", func() { }) It("apply models from config", func() { response := postModelApplyRequest("http://127.0.0.1:9090/models/apply", modelApplyRequest{ - ConfigURL: "https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/hermes-2-pro-mistral.yaml", + ConfigURL: "https://raw.githubusercontent.com/mudler/LocalAI/v2.25.0/embedded/models/hermes-2-pro-mistral.yaml", }) Expect(response["uuid"]).ToNot(BeEmpty(), fmt.Sprint(response)) @@ -600,7 +600,7 @@ var _ = Describe("API test", func() { modelName := "hermes-2-pro-mistral" response := postModelApplyRequest("http://127.0.0.1:9090/models/apply", modelApplyRequest{ - ConfigURL: "https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/hermes-2-pro-mistral.yaml", + ConfigURL: "https://raw.githubusercontent.com/mudler/LocalAI/v2.25.0/embedded/models/hermes-2-pro-mistral.yaml", }) Expect(response["uuid"]).ToNot(BeEmpty(), fmt.Sprint(response)) diff --git a/docs/content/docs/features/model-gallery.md b/docs/content/docs/features/model-gallery.md index c17a5946..6943866a 100644 --- a/docs/content/docs/features/model-gallery.md +++ b/docs/content/docs/features/model-gallery.md @@ -134,12 +134,12 @@ curl $LOCALAI/models/apply -H "Content-Type: application/json" -d '{ }' ``` -An example that installs openllama can be: +An example that installs hermes-2-pro-mistral can be: ```bash LOCALAI=http://localhost:8080 curl $LOCALAI/models/apply -H "Content-Type: application/json" -d '{ - "config_url": "https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/hermes-2-pro-mistral.yaml" + "config_url": "https://raw.githubusercontent.com/mudler/LocalAI/v2.25.0/embedded/models/hermes-2-pro-mistral.yaml" }' ```