Georgi Gerganov
1f50a7d29f
sync : llama.cpp
2024-01-17 21:23:33 +02:00
Benjamin Heiniger
f6614155e4
talk-llama : optional wake-up command and audio confirmation ( #1765 )
...
* talk-llama: add optional wake-word detection from command
* talk-llama: add optional audio confirmation before generating answer
* talk-llama: fix small formatting issue in output
* talk-llama.cpp: fix Windows build
2024-01-16 15:52:01 +02:00
Przemysław Pawełczyk
f5f159c320
server : fix building and simplify lib deps on Windows ( #1772 )
...
* make : fix server example building on MSYS2 environments (Windows)
It was not working since commit eff3570f78742dfd56024328ed93d4f442434280
when server was introduced.
* cmake : simplify server example lib deps on Windows
server uses httplib::Server, not httplib::SSLServer, so there is no need
to mention cryptographic libraries in target_link_libraries.
Winsock (ws2_32) suffices here.
Also use plain library names like we use in other places.
2024-01-15 15:48:13 +02:00
Georgi Gerganov
6ebba525f1
talk-llama : sync llama.cpp
2024-01-14 18:08:20 +02:00
Georgi Gerganov
2a5874441d
talk-llama : llama.cpp
2024-01-14 11:06:28 +02:00
Georgi Gerganov
d08445c9ad
sync : ggml
2024-01-14 10:55:18 +02:00
Georgi Gerganov
f001a3b7b6
talk-llama : sync llama.cpp
2024-01-14 00:13:17 +02:00
RhinoDevel
db078a9ba8
talk-llama : add optional CLI arg to set the bot name ( #1764 )
2024-01-13 20:51:35 +02:00
james wolf
a13a7da5ad
examples : add python example for transcription ( #1744 )
...
* rebase and add simple python interface
* moved python files to examples/python
2024-01-13 19:37:18 +02:00
Georgi Gerganov
40ae0962f4
talk-llama : sync llama.cpp
2024-01-12 22:04:51 +02:00
George Hindle
fbcb52d3cd
server : add more parameters to server api ( #1754 )
...
* feat(server): add more parameters to server api
* fix(server): reset params to original parsed values for each request
2024-01-12 13:42:52 +02:00
George Hindle
f7908f9bb8
params : don't compute timestamps when not printing them ( #1755 )
2024-01-12 13:24:38 +02:00
Georgi Gerganov
00b7a4be02
talk-llama : sync llama.cpp
2024-01-11 22:10:10 +02:00
Georgi Gerganov
32e71a1861
sync : ggml
2024-01-11 21:54:17 +02:00
Georgi Gerganov
9c857cf280
sync : llama.cpp
2024-01-11 21:50:01 +02:00
RhinoDevel
bcc1658cd0
talk-llama : add optional Piper TTS support ( #1749 )
...
Add commented-out command to optionally use Piper (https://github.com/rhasspy/piper ) as text-to-speech solution for the talk-llama example. Piper voices sound almost like real people which is a big improvement (e.g.) from something like espeak.
2024-01-10 16:15:28 +02:00
Emmanuel Schmidbauer
c46886f599
server : add request path option( #1741 )
2024-01-08 22:39:51 +00:00
Georgi Gerganov
29f78392c1
main : add cli option to disable system prints ( #1740 )
2024-01-08 16:41:28 +02:00
Georgi Gerganov
022756a872
server : fix server temperature + add temperature_inc ( #1729 )
...
* server : fix server temperature + add temperature_inc
* server : change dashes to underscores in parameter names
2024-01-07 13:35:14 +02:00
Georgi Gerganov
3b8c2dff57
talk-llama : sync latest llama.cpp
2024-01-06 17:22:57 +02:00
Georgi Gerganov
ab0a8593c5
whisper.swiftui : add .gitignore
2024-01-04 15:00:27 +02:00
Tamotsu Takahashi
d87de61ae6
ci : build with CLBlast + ggml-opencl use GGML_API ( #1576 )
...
* Build with CLBlast
* Declare GGML_API
After rebasing, examples/talk-llama failed:
"D:\a\whisper.cpp\whisper.cpp\build\ALL_BUILD.vcxproj" (build target) (1) ->
"D:\a\whisper.cpp\whisper.cpp\build\examples\talk-llama\talk-llama.vcxproj" (default target) (14) ->
(Link target) ->
llama.obj : error LNK2019: unresolved external symbol ggml_cl_free_data referenced in function "public: __cdecl llama_model::~llama_model(void)" (??1llama_model@@QEAA@XZ) [D:\a\whisper.cpp\whisper.cpp\build\examples\talk-llama\talk-llama.vcxproj]
llama.obj : error LNK2019: unresolved external symbol ggml_cl_transform_tensor referenced in function "public: void __cdecl llama_model_loader::load_all_data(struct ggml_context *,void (__cdecl*)(float,void *),void *,struct llama_mlock *)" (?load_all_data@llama_model_loader@@QEAAXPEAUggml_context@@P6AXMPEAX@Z1PEAUllama_mlock@@@Z) [D:\a\whisper.cpp\whisper.cpp\build\examples\talk-llama\talk-llama.vcxproj]
D:\a\whisper.cpp\whisper.cpp\build\bin\Release\talk-llama.exe : fatal error LNK1120: 2 unresolved externals [D:\a\whisper.cpp\whisper.cpp\build\examples\talk-llama\talk-llama.vcxproj]
2023-12-29 12:23:27 +02:00
Georgi Gerganov
3a5302108d
sync : ggml (ggml_scale, ggml_row_size, etc.) ( #1677 )
...
* sync : ggml
* sync : llama.cpp
* talk-llama : fix obsolete param
* ggml-alloc : fix ggml_tallocr_is_own
* talk.wasm : update to new ggml
* ggml : fix type punning in ggml_scale
* ggml : cuda jetson + arm quants warnings
2023-12-22 17:53:39 +02:00
bobqianic
d2419030b0
examples : Revert CMakeLists.txt for talk-llama ( #1669 )
2023-12-21 22:48:52 +02:00
Georgi Gerganov
940de9dbe9
wchess : update README.md
2023-12-14 22:00:47 +02:00
Georgi Gerganov
375585c07c
wchess : update readme
2023-12-14 17:51:14 +02:00
fraxy-v
fd99ece8e3
wchess : whisper assisted chess ( #1595 )
...
* wchess: whisper assisted chess
* wchess: fix allowed moves in check
* wchess: touchstart, touchend events
* wchess: css, disabled button
* wchess : html touches
* wchess : minor fixes and code style
* wchess : bump encoder context to 1280
* wchess : index.html
* wchess : fix CI warnings
* wchess : add array header
* wchess : build static library
* wchess : display grammar
* wchess : update UX
* wchess : add comment
* wchess : add README
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
2023-12-14 15:58:26 +02:00
Kreijstal
ec03661b20
cmake : target windows 8 or above for prefetchVirtualMemory in llama-talk ( #1617 )
...
Since we use prefetchVirtualMemory we specify we target win 8 or above, otherwise other compilers will refuse to use the prefetchVirtualMemory api, (I understand you are loading it dynamically but the header definition has this limitation)
2023-12-12 11:35:00 +00:00
Kreijstal
6335933a5b
cmake : Fix bug in httplib.h for mingw ( #1615 )
...
Fix bug in httlib.h for mingw, please see https://github.com/yhirose/cpp-httplib/issues/1669
2023-12-10 17:47:52 +00:00
Georgi Gerganov
9521ba6801
whisper.objc : disable timestamps for real-time transcription
2023-12-08 13:43:37 +02:00
Oleg Sidorov
3163090d89
server : pass max-len argument to the server ( #1574 )
...
This commit fixes the missing parameter binding for max-len between the input
arguments and wparams.
2023-12-05 23:01:45 +02:00
Aleksander Andrzejewski
a0ec3fac54
Server : Add support for .vtt format to Whisper server ( #1578 )
...
- The code comes from examples/main
- The output mimetype is set to text/vtt
Example usage:
```shell
curl 127.0.0.1:8080/inference \
-H "Content-Type: multipart/form-data" \
-F file="@samples/jfk.wav" \
-F temperature="0.2" \
-F response-format="vtt"
```
2023-11-30 23:44:26 +00:00
Oleg Sidorov
6559b538e5
server : backport .srt output format ( #1565 )
...
This commit adds a support of .srt format to Whisper server. The code is
effectively backported from examples/main. The output mimetype is set to
application/x-subrip as per https://en.wikipedia.org/wiki/SubRip .
Example usage:
curl 127.0.0.1:8080/inference \
-H "Content-Type: multipart/form-data" \
-F file="@<file-path>" \
-F temperature="0.2" \
-F response-format="srt"
2023-11-28 15:42:58 +02:00
Kasumi
6b094b6dfe
server : set default CORS headers to allow all ( #1567 )
2023-11-28 11:55:20 +02:00
Hang
641f2f4282
readme : update help ( #1560 )
2023-11-27 12:04:08 +02:00
Ismatulla Mansurov
23c21e92eb
server : automatically convert audio on the server ( #1539 )
...
* server : automatically convert audio on the server
* server : remove rebundant comments
* server : automatic conversion refactor
* server : update server readme
* server : remove unnecessary comments and tabs
* server : put back remove calling
* server : apply suggestions from code review
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
* server : check ffmpeg before the server lunch
* server : fix indentation
* Apply suggestions from code review
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
* server : fix function typo calling
* server : fix function typo calling
* server : add warning in readme
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
2023-11-27 11:28:34 +02:00
ecneladis
a5881d619c
server : add --print-realtime param ( #1541 )
...
* server : add --print-realtime param
* Fix duplicate realtime output
2023-11-24 09:35:02 +02:00
Okabintaro
8328d1900f
fix(server): typo in temperature parameter ( #1545 )
...
Also fixed another typo in comments.
2023-11-23 20:59:36 +02:00
Felix
5c7be85fdc
Change temp file name for server application ( #1535 )
...
Avoid issue of removing file if it exists in the current working
directory
2023-11-22 09:23:36 +01:00
Felix
9ac88f2b57
Close file after writing in server application ( #1533 )
...
Fix of mistake leaving file open while reading it again as wav
2023-11-21 20:36:10 +01:00
Georgi Gerganov
46f5b6cb08
server : add video to readme
2023-11-21 17:30:43 +02:00
Felix
eff3570f78
server : add a REST Whisper server example with OAI-like API ( #1380 )
...
* Add first draft of server
* Added json support and base funcs for server.cpp
* Add more user input via api-request
also some clean up
* Add reqest params and load post function
Also some general clean up
* Remove unused function
* Add readme
* Add exception handlers
* Update examples/server/server.cpp
* make : add server target
* Add magic curl syntax
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
2023-11-20 21:40:24 +02:00
Georgi Gerganov
a01b2e0971
sdl : fix audio callback ( #1523 )
2023-11-20 13:16:38 +02:00
Georgi Gerganov
bebf0da983
quantize : add support for K-quant types
2023-11-16 16:18:24 +02:00
Sam Pullara
7883d1cae4
talk-llama : improve quote and backtick handling ( #1364 )
...
* ISSUE-1329: replace " with ' so it doesn't try to execute code in backticks.
* Typo
* Update to keep possessives in the output
Closes the ' then puts a ' in quotes then reopens the ' to escape the ' characters.
2023-11-16 10:34:05 +02:00
Georgi Gerganov
ccc85b4ff8
talk-llama : enable GPU by default
2023-11-15 21:33:00 +02:00
Georgi Gerganov
bfbaa4dce5
whisper : make large version explicit + fix data size units ( #1493 )
2023-11-15 19:42:25 +02:00
Georgi Gerganov
b6c5f49b78
whisper : add batched decoding ( #1486 )
...
* whisper : add whisper_batch
* whisper : move kv_self to whisper_state
* whisper : full batched decoding support
* whisper : fix memory leak in whisper_batch
* whisper : fix mem leak again + remove oboslete function
* whisper : clear kv cache when using whisper_decode API
* whisper : speed-up sampling
* whisper : fix decoders initializer
* bench : add batch size 5 bench
* whisper : add comment about the KV cache size
* whisper : add check for max number of decoders
* whisper : avoid starting sampling threads with bs=1
* whisper : enable beam-search by default
* cuda : sync llama.cpp fixes
2023-11-15 16:12:52 +02:00
Evan Jones
3e5c7feeff
whisper : add grammar-based sampling ( #1229 )
...
* whisper : add grammar-based sampling
* build : fix after master merge
* command : fix exception when recognizing the command
* whisper : fine-tuning grammar functionality
* command : grammar-related improvements
- option to read grammar from file
- add sample grammars for colors and chess moves
- fine-tune the performance further
* grammars : add assistant + update comments
* command : enable beam-search, add "no_timestamps", add "context", add p
* whisper : remove comment
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
2023-11-13 10:51:34 +02:00
rlapray
c23598e4ca
talk-llama : add n_gpu_layers parameter ( #1475 )
2023-11-13 10:04:16 +02:00