Skip to content

Commit 5b0373e

Browse files
committed
mamba : recurrent inference WORKS!!!
1 parent 711c77b commit 5b0373e

File tree

2 files changed

+1
-4
lines changed

2 files changed

+1
-4
lines changed

ggml.c

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5148,8 +5148,6 @@ static struct ggml_tensor * ggml_soft_plus_impl(
51485148
struct ggml_tensor * a,
51495149
bool inplace) {
51505150

5151-
// TODO: does `a` need to be contiguous?
5152-
51535151
bool is_node = false;
51545152

51555153
if (a->grad) {
@@ -11736,7 +11734,7 @@ static void ggml_compute_forward_soft_plus_f32(
1173611734
float * x = (float *) ((char *) dst->data + i*( dst->nb[1]));
1173711735
float * y = (float *) ((char *) src0->data + i*(src0->nb[1]));
1173811736
for (int j = 0; j < nc; ++j) {
11739-
x[j] = logf(1.0f + expf(y[i]));
11737+
x[j] = logf(1.0f + expf(y[j]));
1174011738
}
1174111739
}
1174211740
}

llama.cpp

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6946,7 +6946,6 @@ struct llm_build_context {
69466946
cur = llm_build_norm(ctx0, inpL, hparams,
69476947
model.layers[il].attn_norm, NULL,
69486948
LLM_NORM_RMS, cb, il);
6949-
// TODO: that's probably the wrong name.
69506949
cb(cur, "attn_norm", il);
69516950

69526951
// {n_embd, batch} * {n_embd, 2*d_inner} = {batch, 2*d_inner}

0 commit comments

Comments
 (0)