whisper.cpp/ggml/include
Georgi Gerganov d0b8335789 metal : optimize FA kernels (llama/10171)
* ggml : add ggml_flash_attn_ext_get_prec

* metal : use F16 precision in FA kernels

ggml-ci

* metal : minor clean-up

* metal : compile-guard bf16 FA kernels

ggml-ci

* build : remove obsolete compile flag [no ci]

* metal : prevent int overflows [no ci]

* cuda : disable BF16 FA

ggml-ci

* metal : fix BF16 requirement for FA kernels

ggml-ci

* make : clean-up [no ci]
2024-11-15 15:21:04 +02:00
..
ggml-alloc.h ggml : fix typo in example usage ggml_gallocr_new (ggml/984) 2024-10-05 15:23:51 +03:00
ggml-amx.h ggml : add AMX backend (llama/8998) 2024-11-01 10:19:05 +02:00
ggml-backend.h ggml : move CPU backend to a separate file (llama/10144) 2024-11-15 15:21:04 +02:00
ggml-blas.h ggml : add backend registry / device interfaces to BLAS backend (llama/9752) 2024-11-01 10:19:05 +02:00
ggml-cann.h Adapt to dynamically loadable backends mechanism (llama/9970) 2024-11-01 10:19:05 +02:00
ggml-cpp.h llama : use smart pointers for ggml resources (llama/10117) 2024-11-15 15:21:04 +02:00
ggml-cpu.h ggml : move CPU backend to a separate file (llama/10144) 2024-11-15 15:21:04 +02:00
ggml-cuda.h llama : refactor model loader with backend registry (llama/10026) 2024-11-15 15:21:04 +02:00
ggml-kompute.h kompute: add backend registry / device interfaces (llama/10045) 2024-11-15 15:21:04 +02:00
ggml-metal.h ggml : add metal backend registry / device (llama/9713) 2024-11-01 10:19:05 +02:00
ggml-rpc.h rpc : add backend registry / device interfaces (llama/9812) 2024-11-01 10:19:05 +02:00
ggml-sycl.h Add SYCL Backend registry, device and Event Interfaces (llama/9705) 2024-11-01 10:19:05 +02:00
ggml-vulkan.h vulkan : add backend registry / device interfaces (llama/9721) 2024-11-01 10:19:05 +02:00
ggml.h metal : optimize FA kernels (llama/10171) 2024-11-15 15:21:04 +02:00