Skip to content

Commit

Permalink
whisper : remove extra backend instance (huh?)
Browse files Browse the repository at this point in the history
  • Loading branch information
ggerganov committed May 14, 2024
1 parent 2877b02 commit 4caa64b
Showing 1 changed file with 4 additions and 15 deletions.
19 changes: 4 additions & 15 deletions whisper.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -819,8 +819,6 @@ struct whisper_state {

whisper_decoder decoders[WHISPER_MAX_DECODERS];

ggml_backend_t backend = nullptr;

// ggml-alloc:
// - stores meta info about the intermediate tensors into the `meta` buffers
// - stores the actual tensor data into the `data` buffers
Expand Down Expand Up @@ -2240,7 +2238,7 @@ static bool whisper_encode_internal(
}

if (!whisper_encode_external(wstate)) {
if (!ggml_graph_compute_helper(wstate.backend, gf, n_threads)) {
if (!ggml_graph_compute_helper(wctx.backend, gf, n_threads)) {
return false;
}
} else {
Expand All @@ -2263,7 +2261,7 @@ static bool whisper_encode_internal(
return false;
}

if (!ggml_graph_compute_helper(wstate.backend, gf, n_threads)) {
if (!ggml_graph_compute_helper(wctx.backend, gf, n_threads)) {
return false;
}
}
Expand All @@ -2279,7 +2277,7 @@ static bool whisper_encode_internal(
return false;
}

if (!ggml_graph_compute_helper(wstate.backend, gf, n_threads)) {
if (!ggml_graph_compute_helper(wctx.backend, gf, n_threads)) {
return false;
}
}
Expand Down Expand Up @@ -2744,7 +2742,7 @@ static bool whisper_decode_internal(

logits = gf->nodes[gf->n_nodes - 1];

if (!ggml_graph_compute_helper(wstate.backend, gf, n_threads)) {
if (!ggml_graph_compute_helper(wctx.backend, gf, n_threads)) {
return false;
}
}
Expand Down Expand Up @@ -3191,13 +3189,6 @@ struct whisper_state * whisper_init_state(whisper_context * ctx) {

whisper_state * state = new whisper_state;

state->backend = whisper_backend_init(ctx->params);
if (!state->backend) {
WHISPER_LOG_ERROR("%s: whisper_backend_init() failed\n", __func__);
whisper_free_state(state);
return nullptr;
}

// at this point, we don't know yet how many decoders will be used, so we overallocate 3x ctx
// in theory, there can be a case where this is not enough, but in practice it should always be enough
const int factor = 3;
Expand Down Expand Up @@ -3623,8 +3614,6 @@ void whisper_free_state(struct whisper_state * state) {
ggml_gallocr_free(state->alloc_cross.alloc);
ggml_gallocr_free(state->alloc_decode.alloc);

ggml_backend_free(state->backend);

// [EXPERIMENTAL] Token-level timestamps with DTW
aheads_masks_free(state->aheads_masks);

Expand Down

0 comments on commit 4caa64b

Please sign in to comment.