@@ -1334,7 +1334,7 @@ void ggml_cl_free_data(const struct ggml_tensor* tensor) {
1334
1334
return ;
1335
1335
}
1336
1336
1337
- cl_mem mem = (cl_mem)tensor->data ;
1337
+ cl_mem mem = (cl_mem)tensor->extra ;
1338
1338
clReleaseMemObject (mem);
1339
1339
}
1340
1340
@@ -1393,7 +1393,7 @@ static void ggml_cl_mul_f32(const ggml_tensor * src0, const ggml_tensor * src1,
1393
1393
size_t d_size;
1394
1394
1395
1395
cl_mem d_X = ggml_cl_pool_malloc (ne0 * sizeof (float ), &x_size); // src0
1396
- cl_mem d_Y = (cl_mem) src1->data ; // src1 is already on device, broadcasted.
1396
+ cl_mem d_Y = (cl_mem) src1->extra ; // src1 is already on device, broadcasted.
1397
1397
cl_mem d_D = ggml_cl_pool_malloc (ne0 * sizeof (float ), &d_size); // dst
1398
1398
1399
1399
@@ -1491,7 +1491,7 @@ static void ggml_cl_mul_mat_f32(const ggml_tensor * src0, const ggml_tensor * sr
1491
1491
size_t d_size;
1492
1492
cl_mem d_X;
1493
1493
if (src0->backend == GGML_BACKEND_GPU) { // NOLINT
1494
- d_X = (cl_mem) src0->data ;
1494
+ d_X = (cl_mem) src0->extra ;
1495
1495
} else {
1496
1496
d_X = ggml_cl_pool_malloc (sizeof (float ) * x_ne, &x_size);
1497
1497
}
@@ -1567,7 +1567,7 @@ static void ggml_cl_mul_mat_f16(const ggml_tensor * src0, const ggml_tensor * sr
1567
1567
size_t d_size;
1568
1568
cl_mem d_X;
1569
1569
if (src0->backend == GGML_BACKEND_GPU) { // NOLINT
1570
- d_X = (cl_mem) src0->data ;
1570
+ d_X = (cl_mem) src0->extra ;
1571
1571
} else {
1572
1572
d_X = ggml_cl_pool_malloc (sizeof (ggml_fp16_t ) * x_ne, &x_size);
1573
1573
}
@@ -1697,7 +1697,7 @@ static void ggml_cl_mul_mat_q_f32(const ggml_tensor * src0, const ggml_tensor *
1697
1697
events.emplace_back ();
1698
1698
CL_CHECK (ggml_cl_h2d_tensor_2d (queue, d_Q, 0 , src0, i03, i02, events.data () + ev_idx++));
1699
1699
} else if (src0->backend == GGML_BACKEND_GPU) {
1700
- d_Q = (cl_mem) src0->data ;
1700
+ d_Q = (cl_mem) src0->extra ;
1701
1701
} else {
1702
1702
GGML_ASSERT (false );
1703
1703
}
@@ -1860,6 +1860,6 @@ void ggml_cl_transform_tensor(void * data, ggml_tensor * tensor) {
1860
1860
1861
1861
CL_CHECK (clFinish (queue));
1862
1862
1863
- tensor->data = dst;
1863
+ tensor->extra = dst;
1864
1864
GGML_ASSERT (tensor->backend == GGML_BACKEND_GPU);
1865
1865
}
0 commit comments