Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Revert "whisper : remove extra backend instance (huh?)" #2182

Merged
merged 1 commit into from
May 27, 2024
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
19 changes: 15 additions & 4 deletions whisper.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -818,6 +818,8 @@ struct whisper_state {

whisper_decoder decoders[WHISPER_MAX_DECODERS];

ggml_backend_t backend = nullptr;

// ggml-alloc:
// - stores meta info about the intermediate tensors into the `meta` buffers
// - stores the actual tensor data into the `data` buffers
Expand Down Expand Up @@ -2261,7 +2263,7 @@ static bool whisper_encode_internal(
}

if (!whisper_encode_external(wstate)) {
if (!ggml_graph_compute_helper(wctx.backend, gf, n_threads)) {
if (!ggml_graph_compute_helper(wstate.backend, gf, n_threads)) {
return false;
}
} else {
Expand All @@ -2284,7 +2286,7 @@ static bool whisper_encode_internal(
return false;
}

if (!ggml_graph_compute_helper(wctx.backend, gf, n_threads)) {
if (!ggml_graph_compute_helper(wstate.backend, gf, n_threads)) {
return false;
}
}
Expand All @@ -2300,7 +2302,7 @@ static bool whisper_encode_internal(
return false;
}

if (!ggml_graph_compute_helper(wctx.backend, gf, n_threads)) {
if (!ggml_graph_compute_helper(wstate.backend, gf, n_threads)) {
return false;
}
}
Expand Down Expand Up @@ -2801,7 +2803,7 @@ static bool whisper_decode_internal(

logits = gf->nodes[gf->n_nodes - 1];

if (!ggml_graph_compute_helper(wctx.backend, gf, n_threads)) {
if (!ggml_graph_compute_helper(wstate.backend, gf, n_threads)) {
return false;
}
}
Expand Down Expand Up @@ -3248,6 +3250,13 @@ struct whisper_state * whisper_init_state(whisper_context * ctx) {

whisper_state * state = new whisper_state;

state->backend = whisper_backend_init(ctx->params);
if (!state->backend) {
WHISPER_LOG_ERROR("%s: whisper_backend_init() failed\n", __func__);
whisper_free_state(state);
return nullptr;
}

// at this point, we don't know yet how many decoders will be used, so we overallocate 3x ctx
// in theory, there can be a case where this is not enough, but in practice it should always be enough
const int factor = 3;
Expand Down Expand Up @@ -3684,6 +3693,8 @@ void whisper_free_state(struct whisper_state * state) {
ggml_gallocr_free(state->alloc_cross.alloc);
ggml_gallocr_free(state->alloc_decode.alloc);

ggml_backend_free(state->backend);

// [EXPERIMENTAL] Token-level timestamps with DTW
aheads_masks_free(state->aheads_masks);

Expand Down
Loading