LocalAI/.github/workflows
Sertaç Özercan a670318a9f
feat: auto select llama-cpp cuda runtime (#2306)
* auto select cpu variant

Signed-off-by: Sertac Ozercan <sozercan@gmail.com>

* remove cuda target for now

Signed-off-by: Sertac Ozercan <sozercan@gmail.com>

* fix metal

Signed-off-by: Sertac Ozercan <sozercan@gmail.com>

* fix path

Signed-off-by: Sertac Ozercan <sozercan@gmail.com>

* cuda

Signed-off-by: Sertac Ozercan <sozercan@gmail.com>

* auto select cuda

Signed-off-by: Sertac Ozercan <sozercan@gmail.com>

* update test

Signed-off-by: Sertac Ozercan <sozercan@gmail.com>

* select CUDA backend only if present

Signed-off-by: mudler <mudler@localai.io>

* ci: keep cuda bin in path

Signed-off-by: mudler <mudler@localai.io>

* Makefile: make dist now builds also cuda

Signed-off-by: mudler <mudler@localai.io>

* Keep pushing fallback in case auto-flagset/nvidia fails

There could be other reasons for which the default binary may fail. For example we might have detected an Nvidia GPU,
however the user might not have the drivers/cuda libraries installed in the system, and so it would fail to start.

We keep the fallback of llama.cpp at the end of the llama.cpp backends to try to fallback loading in case things go wrong

Signed-off-by: mudler <mudler@localai.io>

* Do not build cuda on MacOS

Signed-off-by: mudler <mudler@localai.io>

* cleanup

Signed-off-by: Sertac Ozercan <sozercan@gmail.com>

* Apply suggestions from code review

Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>

---------

Signed-off-by: Sertac Ozercan <sozercan@gmail.com>
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
Signed-off-by: mudler <mudler@localai.io>
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
Co-authored-by: mudler <mudler@localai.io>
2024-05-14 19:40:18 +02:00
..
disabled feat(conda): conda environments (#1144) 2023-11-04 15:30:32 +01:00
bump_deps.yaml build(deps): bump peter-evans/create-pull-request from 5 to 6 (#2005) 2024-04-11 13:58:04 +00:00
bump_docs.yaml build(deps): bump peter-evans/create-pull-request from 5 to 6 (#2005) 2024-04-11 13:58:04 +00:00
checksum_checker.yaml ci: get latest git version 2024-05-09 11:33:16 +02:00
dependabot_auto.yml build(deps): bump dependabot/fetch-metadata from 2.0.0 to 2.1.0 (#2186) 2024-04-29 21:12:37 +00:00
generate_grpc_cache.yaml fix: bring everything onto the same GRPC version to fix tests (#2199) 2024-04-30 19:12:15 +00:00
image_build.yml fix: bring everything onto the same GRPC version to fix tests (#2199) 2024-04-30 19:12:15 +00:00
image-pr.yml feat: update ROCM and use smaller image (#2196) 2024-05-03 18:46:49 +02:00
image.yml feat: update ROCM and use smaller image (#2196) 2024-05-03 18:46:49 +02:00
labeler.yml fix(seed): generate random seed per-request if -1 is set (#1952) 2024-04-03 22:25:47 +02:00
localaibot_automerge.yml fix - correct checkout versions (#2029) 2024-04-13 19:01:17 +02:00
release.yaml feat: auto select llama-cpp cuda runtime (#2306) 2024-05-14 19:40:18 +02:00
secscan.yaml build(deps): bump github/codeql-action from 2 to 3 (#2041) 2024-04-15 22:02:44 +00:00
test-extra.yml feat: migrate python backends from conda to uv (#2215) 2024-05-10 15:08:08 +02:00
test.yml feat: auto select llama-cpp cuda runtime (#2306) 2024-05-14 19:40:18 +02:00
update_swagger.yaml ci: add swagger pipeline 2024-04-25 15:11:01 +02:00
yaml-check.yml Update yaml-check.yml 2024-04-25 16:05:02 +02:00