Georgi Gerganov
|
3a5302108d
|
sync : ggml (ggml_scale, ggml_row_size, etc.) (#1677)
* sync : ggml
* sync : llama.cpp
* talk-llama : fix obsolete param
* ggml-alloc : fix ggml_tallocr_is_own
* talk.wasm : update to new ggml
* ggml : fix type punning in ggml_scale
* ggml : cuda jetson + arm quants warnings
|
2023-12-22 17:53:39 +02:00 |
|
Georgi Gerganov
|
59a3d0cb57
|
ggml : sync (ggml-alloc, GPU, eps, etc.) (#1220)
* ggml : sync (ggml-alloc, GPU, eps, etc.)
* ggml : fix build
* wasm : fix build
|
2023-09-05 13:54:40 +03:00 |
|
Georgi Gerganov
|
794b162a46
|
whisper : add integer quantization support (#540)
* whisper : add integer quantization support
* examples : add common-ggml + prepare to add "quantize" tool
* whisper : quantization tool ready
* whisper : fix F32 support
* whisper : try to fix shared lib linkage
* wasm : update quantized models to Q5
* bench.wasm : remove "medium" button
* bench.wasm : fix custom model button
* ggml : add Q5_0 and Q5_1 WASM SIMD
* wasm : add quantized models to all WASM examples
* wasm : bump DB version number to 2
* talk-llama : update example to latest llama.cpp
* node : increase test timeout to 10s
* readme : add information for model quantization
* wasm : add links to other examples
|
2023-04-30 18:51:57 +03:00 |
|
Bader-eddine Ouaich
|
2c856fb9e5
|
whisper : fix potential memory leaks (#740)
* fix potential memory leak if whisper_init_state failed
* fix potential memory leak if gpt2_init failed
|
2023-04-14 20:05:56 +03:00 |
|
Georgi Gerganov
|
aa6adda26e
|
talk : make compatible with c++11 (part 2)
|
2022-12-11 20:34:04 +02:00 |
|
Georgi Gerganov
|
444349f4ec
|
talk : make compatible with c++11
|
2022-12-11 20:19:17 +02:00 |
|
Georgi Gerganov
|
ff36415a86
|
talk.wasm : update video link + some minor fixes
|
2022-11-24 20:15:24 +02:00 |
|
Georgi Gerganov
|
37422ed733
|
talk.wasm : add audio pre-processing + bump memory
|
2022-11-24 00:34:00 +02:00 |
|
Georgi Gerganov
|
be3b720f96
|
talk.wasm : refactoring + update README.md
|
2022-11-24 00:08:57 +02:00 |
|