mirror of
https://github.com/ggerganov/whisper.cpp.git
synced 2025-06-17 06:18:08 +00:00
talk-llama : sync llama.cpp
This commit is contained in:
@ -177,7 +177,7 @@ static bool ggml_graph_compute_helper(
|
||||
int n_threads,
|
||||
ggml_abort_callback abort_callback,
|
||||
void * abort_callback_data) {
|
||||
struct ggml_cplan plan = ggml_graph_plan(graph, n_threads);
|
||||
struct ggml_cplan plan = ggml_graph_plan(graph, n_threads, nullptr);
|
||||
|
||||
plan.abort_callback = abort_callback;
|
||||
plan.abort_callback_data = abort_callback_data;
|
||||
@ -2894,7 +2894,7 @@ static bool whisper_decode_internal(
|
||||
ggml_backend_tensor_set(KQ_mask, wstate.inp_mask.data(), 0, ggml_nelements(KQ_mask)*sizeof(float));
|
||||
}
|
||||
|
||||
logits = gf->nodes[gf->n_nodes - 1];
|
||||
logits = ggml_graph_node(gf, -1);
|
||||
|
||||
if (!ggml_graph_compute_helper(sched, gf, n_threads)) {
|
||||
return false;
|
||||
|
Reference in New Issue
Block a user