Dave
57f7900210
feat: Initial Version of vscode DevContainer ( #3217 )
...
initial version of devcontainer
---------
Signed-off-by: Dave Lee <dave@gray101.com>
2024-08-14 09:06:41 +02:00
dependabot[bot]
5bb2321fe0
chore(deps): Bump openai from 1.39.0 to 1.40.4 in /examples/functions ( #3235 )
...
Bumps [openai](https://github.com/openai/openai-python ) from 1.39.0 to 1.40.4.
- [Release notes](https://github.com/openai/openai-python/releases )
- [Changelog](https://github.com/openai/openai-python/blob/main/CHANGELOG.md )
- [Commits](https://github.com/openai/openai-python/compare/v1.39.0...v1.40.4 )
---
updated-dependencies:
- dependency-name: openai
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-13 23:47:52 +00:00
LocalAI [bot]
10324d9ad2
chore: ⬆️ Update ggerganov/llama.cpp to 06943a69f678fb32829ff06d9c18367b17d4b361
( #3245 )
...
⬆️ Update ggerganov/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-13 21:45:01 +00:00
Ettore Di Giacinto
02de274e00
feat(federated): allow to pickup a specific worker, improve loadbalancing ( #3243 )
...
* feat(explorer): allow to specify a worker target
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(explorer): correctly load balance requests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(explorer): mark load balanced by default
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* fix: make sure to delete tunnels that might not exist anymore
If a worker goes off and on might change tunnel address, and we want to
load balance only on the active tunnels.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-13 16:17:18 +02:00
dependabot[bot]
7d92936e1a
chore(deps): Bump openai from 1.39.0 to 1.40.6 in /examples/langchain/langchainpy-localai-example ( #3244 )
...
chore(deps): Bump openai
Bumps [openai](https://github.com/openai/openai-python ) from 1.39.0 to 1.40.6.
- [Release notes](https://github.com/openai/openai-python/releases )
- [Changelog](https://github.com/openai/openai-python/blob/main/CHANGELOG.md )
- [Commits](https://github.com/openai/openai-python/compare/v1.39.0...v1.40.6 )
---
updated-dependencies:
- dependency-name: openai
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-13 03:59:16 +00:00
dependabot[bot]
447d9f844b
chore(deps): Bump aiohttp from 3.10.2 to 3.10.3 in /examples/langchain/langchainpy-localai-example ( #3234 )
...
chore(deps): Bump aiohttp
Bumps [aiohttp](https://github.com/aio-libs/aiohttp ) from 3.10.2 to 3.10.3.
- [Release notes](https://github.com/aio-libs/aiohttp/releases )
- [Changelog](https://github.com/aio-libs/aiohttp/blob/master/CHANGES.rst )
- [Commits](https://github.com/aio-libs/aiohttp/compare/v3.10.2...v3.10.3 )
---
updated-dependencies:
- dependency-name: aiohttp
dependency-type: direct:production
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-13 02:18:44 +00:00
LocalAI [bot]
89979da33f
chore: ⬆️ Update ggerganov/whisper.cpp to 22fcd5fd110ba1ff592b4e23013d870831756259
( #3239 )
...
⬆️ Update ggerganov/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-13 02:01:26 +00:00
dependabot[bot]
71f3fa653a
chore(deps): Bump openai from 1.39.0 to 1.40.5 in /examples/langchain-chroma ( #3241 )
...
chore(deps): Bump openai in /examples/langchain-chroma
Bumps [openai](https://github.com/openai/openai-python ) from 1.39.0 to 1.40.5.
- [Release notes](https://github.com/openai/openai-python/releases )
- [Changelog](https://github.com/openai/openai-python/blob/main/CHANGELOG.md )
- [Commits](https://github.com/openai/openai-python/compare/v1.39.0...v1.40.5 )
---
updated-dependencies:
- dependency-name: openai
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-13 01:12:11 +00:00
LocalAI [bot]
cd385c2720
chore: ⬆️ Update ggerganov/llama.cpp to fc4ca27b25464a11b3b86c9dbb5b6ed6065965c2
( #3240 )
...
⬆️ Update ggerganov/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-13 00:59:04 +00:00
dependabot[bot]
83ffd626dc
chore(deps): Bump llama-index from 0.10.59 to 0.10.65 in /examples/chainlit ( #3238 )
...
chore(deps): Bump llama-index in /examples/chainlit
Bumps [llama-index](https://github.com/run-llama/llama_index ) from 0.10.59 to 0.10.65.
- [Release notes](https://github.com/run-llama/llama_index/releases )
- [Changelog](https://github.com/run-llama/llama_index/blob/main/CHANGELOG.md )
- [Commits](https://github.com/run-llama/llama_index/compare/v0.10.59...v0.10.65 )
---
updated-dependencies:
- dependency-name: llama-index
dependency-type: direct:production
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-13 00:23:31 +00:00
dependabot[bot]
121ffe61c5
chore(deps): Bump pyyaml from 6.0.1 to 6.0.2 in /examples/langchain/langchainpy-localai-example ( #3231 )
...
chore(deps): Bump pyyaml
Bumps [pyyaml](https://github.com/yaml/pyyaml ) from 6.0.1 to 6.0.2.
- [Release notes](https://github.com/yaml/pyyaml/releases )
- [Changelog](https://github.com/yaml/pyyaml/blob/main/CHANGES )
- [Commits](https://github.com/yaml/pyyaml/compare/6.0.1...6.0.2 )
---
updated-dependencies:
- dependency-name: pyyaml
dependency-type: direct:production
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-12 23:31:45 +00:00
dependabot[bot]
710f566553
chore(deps): Bump attrs from 23.2.0 to 24.2.0 in /examples/langchain/langchainpy-localai-example ( #3232 )
...
chore(deps): Bump attrs
Bumps [attrs](https://github.com/sponsors/hynek ) from 23.2.0 to 24.2.0.
- [Commits](https://github.com/sponsors/hynek/commits )
---
updated-dependencies:
- dependency-name: attrs
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-12 22:36:11 +00:00
dependabot[bot]
bd57ebf042
chore(deps): Bump langchain-community from 0.2.9 to 0.2.11 in /examples/langchain/langchainpy-localai-example ( #3230 )
...
chore(deps): Bump langchain-community
Bumps [langchain-community](https://github.com/langchain-ai/langchain ) from 0.2.9 to 0.2.11.
- [Release notes](https://github.com/langchain-ai/langchain/releases )
- [Commits](https://github.com/langchain-ai/langchain/compare/langchain-community==0.2.9...langchain-community==0.2.11 )
---
updated-dependencies:
- dependency-name: langchain-community
dependency-type: direct:production
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-12 21:13:01 +00:00
dependabot[bot]
ae4b67fb56
chore(deps): Bump llama-index from 0.10.61 to 0.10.65 in /examples/langchain-chroma ( #3225 )
...
chore(deps): Bump llama-index in /examples/langchain-chroma
Bumps [llama-index](https://github.com/run-llama/llama_index ) from 0.10.61 to 0.10.65.
- [Release notes](https://github.com/run-llama/llama_index/releases )
- [Changelog](https://github.com/run-llama/llama_index/blob/main/CHANGELOG.md )
- [Commits](https://github.com/run-llama/llama_index/compare/v0.10.61...v0.10.65 )
---
updated-dependencies:
- dependency-name: llama-index
dependency-type: direct:production
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-12 21:00:30 +00:00
Ettore Di Giacinto
9729d2ae37
feat(explorer): make possible to run sync in a separate process ( #3224 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-12 19:25:44 +02:00
Ettore Di Giacinto
4dfa085339
models(gallery): add cathallama-70b ( #3221 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-12 09:59:17 +02:00
Ettore Di Giacinto
7137c32f8f
models(gallery): add infinity-instruct-7m-gen-llama3_1-70b ( #3220 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-12 09:56:31 +02:00
LocalAI [bot]
e30114a4a4
chore: ⬆️ Update ggerganov/llama.cpp to 4134999e01f31256b15342b41c4de9e2477c4a6c
( #3218 )
...
⬆️ Update ggerganov/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-11 21:46:30 +00:00
Ettore Di Giacinto
a92b3b13e9
chore: fix gosum missing entry
2024-08-11 11:22:00 +02:00
Ettore Di Giacinto
c4534cd908
chore(deps): update edgevpn ( #3214 )
...
* chore(deps): update edgevpn
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* fix: initialize failure map
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-11 10:46:17 +02:00
Ettore Di Giacinto
9f61ac8acc
models(gallery): add flux.1-dev and flux.1-schnell ( #3215 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-11 10:19:02 +02:00
Ettore Di Giacinto
74eaf02484
feat(diffusers): support flux models ( #3129 )
...
* feat(diffusers): support flux models
This adds support for FLUX models. For instance:
https://huggingface.co/black-forest-labs/FLUX.1-dev
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(diffusers): support FluxTransformer2DModel
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Small fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-11 01:31:53 +02:00
Ettore Di Giacinto
7ba4a78fcc
fix(explorer): reset counter when network is active ( #3213 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
2024-08-11 00:59:58 +02:00
LocalAI [bot]
f3357a17b8
chore: ⬆️ Update ggerganov/llama.cpp to 6e02327e8b7837358e0406bf90a4632e18e27846
( #3212 )
...
⬆️ Update ggerganov/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-10 22:16:51 +00:00
Ettore Di Giacinto
8627bc2dd4
feat(explorer): relax token deletion with error threshold ( #3211 )
...
feat(explorer): relax token deletion with error threashold
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
2024-08-10 20:50:57 +02:00
Ettore Di Giacinto
0c0bc18c94
fix(diffusers): pin torch and torchvision ( #1592 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-10 10:10:47 +02:00
LocalAI [bot]
63ee689f21
chore(model-gallery): ⬆️ update checksum ( #3210 )
...
⬆️ Checksum updates in gallery/index.yaml
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-09 23:02:22 +00:00
LocalAI [bot]
a0e0804f25
chore: ⬆️ Update ggerganov/whisper.cpp to 81c999fe0a25c4ebbfef10ed8a1a96df9cfc10fd
( #3209 )
...
⬆️ Update ggerganov/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-09 22:35:22 +00:00
LocalAI [bot]
71b8232076
chore: ⬆️ Update ggerganov/llama.cpp to b72942fac998672a79a1ae3c03b340f7e629980b
( #3208 )
...
⬆️ Update ggerganov/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-09 22:20:51 +00:00
Ettore Di Giacinto
2e2a0dffbc
fix(diffusers-hipblas): pin to rocm6.1
...
As per https://rocm.docs.amd.com/projects/install-on-linux/en/latest/install/3rd-party/pytorch-install.html
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
2024-08-09 22:36:10 +02:00
dependabot[bot]
6d20f38510
chore(deps): Bump aiohttp from 3.9.5 to 3.10.2 in /examples/langchain/langchainpy-localai-example in the pip group ( #3207 )
...
chore(deps): Bump aiohttp
Bumps the pip group in /examples/langchain/langchainpy-localai-example with 1 update: [aiohttp](https://github.com/aio-libs/aiohttp ).
Updates `aiohttp` from 3.9.5 to 3.10.2
- [Release notes](https://github.com/aio-libs/aiohttp/releases )
- [Changelog](https://github.com/aio-libs/aiohttp/blob/master/CHANGES.rst )
- [Commits](https://github.com/aio-libs/aiohttp/compare/v3.9.5...v3.10.2 )
---
updated-dependencies:
- dependency-name: aiohttp
dependency-type: direct:production
dependency-group: pip
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-08-09 20:08:24 +00:00
Ettore Di Giacinto
9e3e892ac7
feat(p2p): add network explorer and community pools ( #3125 )
...
* WIP
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Wire up a simple explorer DB
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* wip
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* WIP
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* refactor: group services id so can be identified easily in the ledger table
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(discovery): discovery service now gather worker informations correctly
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(explorer): display network token
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(explorer): display form to add new networks
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(explorer): stop from overwriting networks
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(explorer): display only networks with active workers
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(explorer): list only clusters in a network if it has online workers
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* remove invalid and inactive networks
if networks have no workers delete them from the database, similarly,
if invalid.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* ci: add workflow to deploy new explorer versions automatically
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* build-api: build with p2p tag
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Allow to specify a connection timeout
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* logging
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Better p2p defaults
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Set loglevel
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Fix dht enable
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Default to info for loglevel
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Add navbar
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Slightly improve rendering
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* Allow to copy the token easily
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* ci fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-09 20:12:01 +02:00
Ettore Di Giacinto
5fcafc3d1e
fix(diffusers): allow pre-releases for requirements
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
2024-08-09 11:07:38 +02:00
LocalAI [bot]
74f8785047
chore: ⬆️ Update ggerganov/llama.cpp to 3a14e00366399040a139c67dd5951177a8cb5695
( #3204 )
...
⬆️ Update ggerganov/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-08 22:36:08 +00:00
LocalAI [bot]
b1773e33d5
chore: ⬆️ Update ggerganov/whisper.cpp to 6eac06759b87b50132a01be019e9250a3ffc8969
( #3203 )
...
⬆️ Update ggerganov/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-08 22:18:00 +00:00
Ettore Di Giacinto
a507c13f8e
fix(diffusers): do not specify --pre
as with pip
...
drop --pre as it is not supported by `uv`
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
2024-08-08 22:21:05 +02:00
Ettore Di Giacinto
8317839ca5
fix(diffusers): use nightly rocm for hipblas builds ( #3202 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
2024-08-08 17:28:07 +02:00
Ettore Di Giacinto
4a1a3a56ba
models(gallery): add calme-2.3-legalkit-8b ( #3200 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-08 11:59:42 +02:00
Ettore Di Giacinto
f7ffa9cd58
fix(vall-e-x): pin hipblas deps ( #3201 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-08 11:59:31 +02:00
Ettore Di Giacinto
60117ec057
fix(apple): disable BUILD_TYPE metal on fallback ( #3199 )
...
When compiling the single-binary on Apple, we enforce BUILD_TYPE=metal,
however, we want still to have the fallback vanilla such as if llama.cpp fails
to load metal (e.g. if Acceleration framework is missing, or MacOS version is too old)
we can still run by offloading to the CPU. The default backend is still using metal as usual.
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
2024-08-08 00:59:10 -04:00
LocalAI [bot]
1c708d21de
chore: ⬆️ Update ggerganov/llama.cpp to 15fa07a5c564d3ed7e7eb64b73272cedb27e73ec
( #3197 )
...
⬆️ Update ggerganov/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-07 22:19:20 +00:00
LocalAI [bot]
1d94aaa10f
feat(swagger): update swagger ( #3196 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-07 21:54:27 +00:00
Ettore Di Giacinto
8814b31805
chore: drop gpt4all.cpp ( #3106 )
...
chore: drop gpt4all
gpt4all is already supported in llama.cpp - the backend was kept for
keeping compatibility with old gpt4all models (prior to gguf format).
It is good time now to clean up and remove it to slim the compilation
process.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-07 23:35:55 +02:00
Ettore Di Giacinto
36e185ba63
feat(p2p): allow to run multiple clusters in the same p2p network ( #3128 )
...
feat(p2p): allow to run multiple clusters in the same network
Allow to specify a network ID via CLI which allows to run multiple
clusters, logically separated within the same network (by using the same
shared token).
Note: This segregation is not "secure" by any means, anyone having the
network token can see the services available in all the network,
however, this provides a way to separate the inference endpoints.
This allows for instance to have a node which is both federated and
having attached a set of llama.cpp workers.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-07 23:35:44 +02:00
Ettore Di Giacinto
2c8623dbb4
fix(python): move vllm to after deps, drop diffusers main deps
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-07 23:34:37 +02:00
Ettore Di Giacinto
e198347886
feat(openai): add json_schema
format type and strict mode ( #3193 )
...
* feat(openai): add json_schema and strict mode
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* handle err vs _
security scanners prefer if we put these branches in, and I tend to agree.
Signed-off-by: Dave <dave@gray101.com>
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Signed-off-by: Dave <dave@gray101.com>
Co-authored-by: Dave <dave@gray101.com>
2024-08-07 15:27:02 -04:00
Ettore Di Giacinto
66cf38b0b3
feat(venv): shared env ( #3195 )
...
* feat(venv): allow to share veenvs
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* fix(vllm): add back flash-attn
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-07 19:45:14 +02:00
Ettore Di Giacinto
11b2adae0c
fix(vllm): drop flash-attn installation afterwards
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
2024-08-07 18:08:26 +02:00
Ettore Di Giacinto
61b5602111
fix(python): move accelerate and GPU-specific libs to build-type ( #3194 )
...
Some of the dependencies in `requirements.txt`, even if generic, pulls
down the line CUDA libraries.
This changes moves mostly all GPU-specific libs to the build-type, and
tries a safer approach. In `requirements.txt` now are listed only
"first-level" dependencies, for instance, grpc, but libs-dependencies
are moved down to the respective build-type `requirements.txt` to avoid
any mixin.
This should fix #2737 and #1592 .
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-07 17:02:32 +02:00
Ettore Di Giacinto
abcf0ff000
chore: ⬆️ Update ggerganov/llama.cpp to 1e6f6554aa11fa10160a5fda689e736c3c34169f
( #3189 )
...
* arrow_up: Update ggerganov/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
* fix(llama.cpp): adapt to upstream naming changes
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
---------
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com>
2024-08-07 01:10:21 +02:00