-
Notifications
You must be signed in to change notification settings - Fork 3.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add fields to verbose_json
response and show examples on the home page
#1802
Add fields to verbose_json
response and show examples on the home page
#1802
Conversation
Indeed, we opt for |
Wow! That sounds great! I believe it can definitely improve the use of whisper.cpp. I think we can merge this pull request first and then either add a new issue for that or include some TODO comments in the code. What do you think? |
The problem may not related to the changes in this PR, which only changed the frontend page and the output format. The mechanism of file processing is not involved. I think the problem is about the changes in #1781, which allows passing audio file content as file name. (I think that's why you got |
sorry it was compilation mistake from my side thanks |
* ggerganov/master: (60 commits) sync : ggml (#0) ggml : fix IQ3_XXS on Metal (llama/5219) sync : ggml (llama/0) Faster AVX2 dot product for IQ2_XS (llama/5187) SOTA 3-bit quants (llama/5196) ggml alloc: Fix for null dereference on alloc failure (llama/5200) Nomic Vulkan backend (llama/4456) ggml : add max buffer sizes to opencl and metal backends (llama/5181) metal : free metal objects (llama/5161) gguf : fix comparison (ggml/715) `ggml_cuda_cpy` support for 4d tensors and float16->float32 upcasting (ggml/686) gguf : add input validation, prevent integer overflows (ggml/709) ci : fix yolo URLs + fix metal capture (ggml/712) metal : add debug capture backend function (ggml/694) common : fix wav buffer detection (ggerganov#1819) server : add fields to `verbose_json` response (ggerganov#1802) make : update MSYS_NT (ggerganov#1813) talk-llama : sync llama.cpp sync : ggml ggml : add Vulkan backend (llama/2059) ...
* server: include additional fields in the verbose_json response as OpenAI does * server: show request examples on home page * server: todo note for compression_ratio and no_speech_prob * server: add simple demo form to the homepage
* server: include additional fields in the verbose_json response as OpenAI does * server: show request examples on home page * server: todo note for compression_ratio and no_speech_prob * server: add simple demo form to the homepage
* server: include additional fields in the verbose_json response as OpenAI does * server: show request examples on home page * server: todo note for compression_ratio and no_speech_prob * server: add simple demo form to the homepage
* server: include additional fields in the verbose_json response as OpenAI does * server: show request examples on home page * server: todo note for compression_ratio and no_speech_prob * server: add simple demo form to the homepage
The
verbose_json
format seems to aim to match the same format as OpenAI's, but it is currently missing several fields.task
,language
,duration
,temperature
, andavg_logprob
.Some observations:
segment.words
do not match OpenAI's (OpenAI does not provide this one), but I think that's okay; it can be an extension to showcase the capabilities of whisper.cpp at the word level.compression_ratio
andno_speech_prob
, if I have not missed anything, currently, they cannot be obtained from whisper.cpp?OpenAI's
verbose_json
whisper.cpp's
verbose_json
I also replaced the
hello
on the homepage with request examples. I think the first action after running the server with./server
is to open the URL in the terminal and check out the homepage. It may be easier for users to try the server this way. (A web interface for direct interaction would be better, but it requires some time to design...)