Georgi Gerganov
2948c740a2
sync : ggml ( #2001 )
...
* sync : update scripts
* sync : ggml
* talk-llama : sync llama.cpp
* make : WHISPER_CUBLAS -> WHISPER_CUDA
* ci : try to fix sycl build
* talk-llama : fix make build
2024-03-27 18:55:10 +02:00
Neo Zhang Jianyu
bae7c23fbf
Revert "[SYCL] fix error when set main gpu to non-zero (llama/5901)" (llama/5918)
...
This reverts commit ceca1aef0738b57951cd12c603c3477e75312dec.
2024-03-08 11:38:33 +02:00
Neo Zhang Jianyu
18ea187d42
fix error when set main gpu to non-zero (llama/5901)
...
* fix error when set main gpu to non-zero
* fix delete condition
2024-03-08 11:38:33 +02:00
Neo Zhang Jianyu
c3bfc9bfda
Support multiple GPUs (split mode) on SYCL backend (llama/5806)
...
* suport multiple cards: split-mode - layer|row
* rm warning
* rebase with master, support tow new OPs, close feature for -sm=row, fix for unit test
* update news
* fix merge error
* update according to review comments
2024-03-08 11:38:32 +02:00
Georgi Gerganov
8b17a2f776
src : relocate new backend sources
2024-02-10 09:55:47 +02:00