Skip to content

Commit

Permalink
mamba : more correctly update the "used" field of the KV cache
Browse files Browse the repository at this point in the history
  • Loading branch information
compilade committed Mar 2, 2024
1 parent ed7c4c2 commit 8646535
Showing 1 changed file with 17 additions and 14 deletions.
31 changes: 17 additions & 14 deletions llama.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -2209,22 +2209,19 @@ static bool llama_kv_cache_find_slot(
// For recurrent state architectures (like Mamba),
// each KV cache cell can store the state for a whole sequence.

// starting point to find the minimum seq_id used in the batch
cache.head = cache.size - 1;
// likewise, to find the max seq_id in the batch
cache.used = 0;
llama_seq_id min = cache.size - 1;
llama_seq_id max = 0;

for (uint32_t i = 0; i < n_tokens; ++i) {
for (int32_t j = 0; j < batch.n_seq_id[i]; ++j) {
llama_seq_id seq_id = batch.seq_id[i][j];
// make sure it's a valid seq_id
if ((uint32_t)seq_id < cache.size) {
// the number of "used" cells is simply the biggest seq_id
if (cache.used < (uint32_t)seq_id) {
cache.used = seq_id;
if ((uint32_t) seq_id < cache.size) {
if (seq_id > max) {
max = seq_id;
}
// the "head" is the smallest seq_id
if (cache.head > (uint32_t)seq_id) {
cache.head = seq_id;
if (seq_id < min) {
min = seq_id;
}
// Assuming the tokens are in-order
if (batch.pos[i] != cache.cells[seq_id].pos + 1) {
Expand All @@ -2233,6 +2230,9 @@ static bool llama_kv_cache_find_slot(
LLAMA_LOG_WARN("%s: non-consecutive token position %d after %d for sequence %d\n",
__func__, batch.pos[i], cache.cells[seq_id].pos, seq_id);
}
if (cache.cells[seq_id].pos < 0 && 0 <= batch.pos[i]) {
cache.used += 1;
}
cache.cells[seq_id].pos = batch.pos[i];
// NOTE: seq_ids are not inserted here; they are handled when the input tensors are set
} else {
Expand All @@ -2244,9 +2244,12 @@ static bool llama_kv_cache_find_slot(
}
}

cache.n = cache.used - cache.head + 1;
// sanity check (max >= min)
return cache.used >= cache.head;
// allow getting the range of used cells, from head to head + n
cache.head = min;
cache.n = max - min + 1;

// sanity check
return max >= min;
}
// otherwise, one cell per token.

Expand Down

0 comments on commit 8646535

Please sign in to comment.