Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Commit dc39012

Browse files
authored
llama : fix op mul check with command-r-plus (ggml-org#10476)
1 parent 9336db4 commit dc39012

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

src/llama.cpp

+2-2
Original file line numberDiff line numberDiff line change
@@ -7181,12 +7181,12 @@ static bool weight_buft_supported(const llama_hparams & hparams, ggml_tensor * w
71817181
} break;
71827182
case GGML_OP_ADD:
71837183
{
7184-
ggml_tensor * a = ggml_new_tensor_2d(ctx, GGML_TYPE_F32, w->ne[0], 512);
7184+
ggml_tensor * a = ggml_new_tensor_4d(ctx, GGML_TYPE_F32, w->ne[0], w->ne[1], w->ne[2], w->ne[3]);
71857185
op_tensor = ggml_add(ctx, a, w);
71867186
} break;
71877187
case GGML_OP_MUL:
71887188
{
7189-
ggml_tensor * a = ggml_new_tensor_2d(ctx, GGML_TYPE_F32, w->ne[0], 512);
7189+
ggml_tensor * a = ggml_new_tensor_4d(ctx, GGML_TYPE_F32, w->ne[0], w->ne[1], w->ne[2], w->ne[3]);
71907190
op_tensor = ggml_mul(ctx, a, w);
71917191
} break;
71927192
case GGML_OP_DIV:

0 commit comments

Comments
 (0)