Commit Graph

775 Commits

Author SHA1 Message Date
ea19ed33f1 Update README.md (#46)
Add references to the new Android app
2022-12-16 19:28:51 +02:00
675e787171 Add Android sample (#277)
* Add Android sample

* Use main project C files

* Stop existing playback before starting new playback

* Make text scrollable

* Stop playback when starting to record

* Remove extra var
2022-12-16 19:20:13 +02:00
c6c3ad5a98 ci : add Windows build without OpenBLAS + change to Release (#85) (#282) 2022-12-16 18:51:46 +02:00
6a7c82501e whisper : improve decoding strategy (#244)
- Clear past prompt when there is very short audio left for processing.
  My observation is that in these cases the decoding tends to repeat and
  hallucinate stuff and I think this is induced by the existing prompt
- When we fail to sample timestamp token, retry by clearing the past
  prompt. If it fails again, then we advance the window by 1 second
2022-12-16 18:34:35 +02:00
a82d331034 stream : update README.md + comments 2022-12-16 18:04:19 +02:00
c37c2443c1 Update README.md (#56) 2022-12-16 18:01:05 +02:00
0f11759406 ggml : make more compatible with c99 (#262) 2022-12-16 18:00:12 +02:00
5a5c5ddcca Update README.md 2022-12-15 20:38:08 +02:00
34e0b4b9ef stream : fix build 2022-12-15 20:15:36 +02:00
b0f8013eb9 stream : add sliding window mode 2022-12-15 19:59:17 +02:00
124c718c73 whisper : fix UB when reading buffer of length 0 bytes (#265) 2022-12-13 23:14:47 +02:00
f66ac6dc4f ggml : fix indentation 2022-12-13 23:09:21 +02:00
9955fa4ed7 ggml : make compatible with c99 (#262) 2022-12-13 23:07:49 +02:00
a613f16aec talk : improve prompting 2022-12-12 23:44:36 +02:00
930c693989 release : v1.0.3
Fixed whisper.spm tests
1.0.3
2022-12-12 20:36:52 +02:00
d8a0dde31a Update README.md 2022-12-12 20:33:09 +02:00
9e3e6f253a release : v1.0.2 2022-12-12 20:29:30 +02:00
57ccd7cc4f Update README.md 2022-12-12 20:23:10 +02:00
812ae3ffbd Update README.md 2022-12-12 20:20:51 +02:00
f309f97df6 Node.js package (#260)
* npm : preparing infra for node package

* npm : package infra ready

* npm : initial version ready

* npm : change name to whisper.cpp

whisper.js is taken
2022-12-12 20:17:27 +02:00
aa6adda26e talk : make compatible with c++11 (part 2) 2022-12-11 20:34:04 +02:00
444349f4ec talk : make compatible with c++11 2022-12-11 20:19:17 +02:00
37a93d2459 cmake : require c++11 instead of c++20 2022-12-11 20:04:05 +02:00
e70d47baab Remove C++20 requirement (#257)
* Remove C++20 requirement

* Roll back C features not supported in VS2017
2022-12-11 20:03:07 +02:00
6ed786957e Add newline per segment for text output (#254) 2022-12-11 20:00:29 +02:00
ea38ad6e70 bench : more concise representation of the results (#89) 2022-12-11 11:56:13 +02:00
054940e1f6 minor : fix .gitignore to not ignore examples 2022-12-11 11:39:46 +02:00
fcf515de60 bench.wasm : same as "bench" but runs in the browser (#89) 2022-12-11 11:09:10 +02:00
85c9ac18b5 Update README.md 2022-12-10 16:54:57 +02:00
b7c85d1ea6 talk : fix build for MSVC 2022-12-10 16:51:58 +02:00
3b1aacbe6d talk : talk with AI in the terminal 2022-12-10 16:51:58 +02:00
d1da35de06 fix potential bug reading model data into a small size optimized string which could lead to memory corruption. In an SSO string, you can't write data to &str[0] and expect it to work well.
Also added a small wrapper function to more safely read model data without having to get the sizeof right. I tested this on tiny, base and large models, there was no change in behaviour.
2022-12-10 16:20:48 +02:00
603f97ba11 whisper : minor improvemnt in decoding strategy (#244)
Do not allow for text segments to go beyond end of audio.
This partially mitigates some issues when the last audio window is 1-2
seconds just before the end of the audio file and the decoding spirals
into a repetition of the last transcribed phrase.
2022-12-10 13:38:26 +02:00
50a061b313 ggml : add alternative cblas_sgemm call 2022-12-08 23:48:04 +02:00
832b4f34c9 make : indentation + .gitignore 2022-12-08 19:42:06 +02:00
0f98755fc5 Flag for Position Independent Code 2022-12-08 19:41:01 +02:00
56822621a8 twitch.sh : various fixes and polishing
- check if streamlink is installed
- fix audio chunking
- change default threads to 4
2022-12-08 19:20:04 +02:00
9e5f3ddc16 Allow for Twitch.tv live transcription
We rely on streamlink library to give us a stream, then we proceed similarly to
the radio livestream example.
2022-12-08 19:20:04 +02:00
d91c001120 Fix paths echoed after the download
Was using models path instead of root path
2022-12-08 09:23:52 +02:00
04a16bbf11 fix compilation on haiku 2022-12-08 09:20:57 +02:00
47afb93c3c yt-wsp.sh : improve usage instructions 2022-12-07 22:12:08 +02:00
575c53dc41 yt-wsp.sh : fix usage instruction + comment 2022-12-07 21:12:55 +02:00
3996ecc156 Update README.md 2022-12-07 05:15:46 +02:00
faa85f9840 livestream.sh : remove obsolete comment 2022-12-07 04:41:43 +02:00
b6597539f9 ggml : fix typo in previous commit 2022-12-06 22:12:57 +02:00
9a4b7a916e ggml : use macros to inline FP16 <-> FP32 conversions 2022-12-06 22:09:26 +02:00
f8ec718b76 ggml : add F16C CPU flag check 2022-12-06 21:56:56 +02:00
35b40a93b9 add fp16/fp32 convert intrinsics 2022-12-06 21:44:24 +02:00
9fe7306f4b models : add the new "large" model release by OpenAI
The old "large" model is now renamed "large-v1".
If you have been using it, make sure to rename it and download the new
"large" model for best results.
2022-12-06 18:48:57 +02:00
13e8eb2346 bench : add commit hash to bench-all.sh results 2022-12-06 18:47:48 +02:00