extern const char *arithm_bitwise_xor_scalar_mask;
extern const char *arithm_compare_eq;
extern const char *arithm_compare_ne;
- extern const char *arithm_sub;
- extern const char *arithm_sub_scalar;
- extern const char *arithm_sub_scalar_mask;
extern const char *arithm_mul;
extern const char *arithm_div;
extern const char *arithm_absdiff;
/////////////////////// add subtract multiply divide /////////////////////////
//////////////////////////////////////////////////////////////////////////////
template<typename T>
-void arithmetic_run(const oclMat &src1, const oclMat &src2, oclMat &dst, string kernelName, const char **kernelString, void *_scalar)
+void arithmetic_run(const oclMat &src1, const oclMat &src2, oclMat &dst,
+ string kernelName, const char **kernelString, void *_scalar, int op_type = 0)
{
if(!src1.clCxt->supportsFeature(Context::CL_DOUBLE) && src1.type() == CV_64F)
{
scalar = (T)scalar1;
args.push_back( make_pair( sizeof(T), (void *)&scalar ));
}
-
- openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, -1, depth);
+ switch(op_type)
+ {
+ case MAT_ADD:
+ openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, -1, depth, "-D ARITHM_ADD");
+ break;
+ case MAT_SUB:
+ openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, -1, depth, "-D ARITHM_SUB");
+ break;
+ default:
+ openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, -1, depth);
+ }
}
-static void arithmetic_run(const oclMat &src1, const oclMat &src2, oclMat &dst, string kernelName, const char **kernelString)
+static void arithmetic_run(const oclMat &src1, const oclMat &src2, oclMat &dst,
+ string kernelName, const char **kernelString, int op_type = 0)
{
- arithmetic_run<char>(src1, src2, dst, kernelName, kernelString, (void *)NULL);
+ arithmetic_run<char>(src1, src2, dst, kernelName, kernelString, (void *)NULL, op_type);
}
-static void arithmetic_run(const oclMat &src1, const oclMat &src2, oclMat &dst, const oclMat &mask, string kernelName, const char **kernelString)
+static void arithmetic_run(const oclMat &src1, const oclMat &src2, oclMat &dst, const oclMat &mask,
+ string kernelName, const char **kernelString, int op_type = 0)
{
if(!src1.clCxt->supportsFeature(Context::CL_DOUBLE) && src1.type() == CV_64F)
{
args.push_back( make_pair( sizeof(cl_int), (void *)&cols ));
args.push_back( make_pair( sizeof(cl_int), (void *)&dst_step1 ));
- openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, channels, depth);
+ switch (op_type)
+ {
+ case MAT_ADD:
+ openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, channels, depth, "-D ARITHM_ADD");
+ break;
+ case MAT_SUB:
+ openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, channels, depth, "-D ARITHM_SUB");
+ break;
+ default:
+ openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, channels, depth);
+ }
}
void cv::ocl::add(const oclMat &src1, const oclMat &src2, oclMat &dst)
{
- arithmetic_run(src1, src2, dst, "arithm_add", &arithm_add);
+ arithmetic_run(src1, src2, dst, "arithm_add", &arithm_add, MAT_ADD);
}
void cv::ocl::add(const oclMat &src1, const oclMat &src2, oclMat &dst, const oclMat &mask)
{
- arithmetic_run(src1, src2, dst, mask, "arithm_add_with_mask", &arithm_add);
+ arithmetic_run(src1, src2, dst, mask, "arithm_add_with_mask", &arithm_add, MAT_ADD);
}
void cv::ocl::subtract(const oclMat &src1, const oclMat &src2, oclMat &dst)
{
- arithmetic_run(src1, src2, dst, "arithm_sub", &arithm_sub);
+ arithmetic_run(src1, src2, dst, "arithm_add", &arithm_add, MAT_SUB);
}
void cv::ocl::subtract(const oclMat &src1, const oclMat &src2, oclMat &dst, const oclMat &mask)
{
- arithmetic_run(src1, src2, dst, mask, "arithm_sub_with_mask", &arithm_sub);
+ arithmetic_run(src1, src2, dst, mask, "arithm_add_with_mask", &arithm_add, MAT_SUB);
}
typedef void (*MulDivFunc)(const oclMat &src1, const oclMat &src2, oclMat &dst, string kernelName,
const char **kernelString, void *scalar);
args.push_back( make_pair( sizeof(cl_int) , (void *)&cols ));
args.push_back( make_pair( sizeof(cl_int) , (void *)&dst_step1 ));
if(isMatSubScalar != 0)
- {
- isMatSubScalar = isMatSubScalar > 0 ? 1 : 0;
- args.push_back( make_pair( sizeof(cl_int) , (void *)&isMatSubScalar));
- }
-
- openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, channels, depth);
+ openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, channels, depth, "-D ARITHM_SUB");
+ else
+ openCLExecuteKernel(clCxt, kernelString, kernelName, globalThreads, localThreads, args, channels, depth, "-D ARITHM_ADD");
}
static void arithmetic_scalar_run(const oclMat &src, oclMat &dst, string kernelName, const char **kernelString, double scalar)
void cv::ocl::subtract(const oclMat &src1, const Scalar &src2, oclMat &dst, const oclMat &mask)
{
- string kernelName = mask.data ? "arithm_s_sub_with_mask" : "arithm_s_sub";
- const char **kernelString = mask.data ? &arithm_sub_scalar_mask : &arithm_sub_scalar;
+ string kernelName = mask.data ? "arithm_s_add_with_mask" : "arithm_s_add";
+ const char **kernelString = mask.data ? &arithm_add_scalar_mask : &arithm_add_scalar;
arithmetic_scalar( src1, src2, dst, mask, kernelName, kernelString, 1);
}
void cv::ocl::subtract(const Scalar &src2, const oclMat &src1, oclMat &dst, const oclMat &mask)
{
- string kernelName = mask.data ? "arithm_s_sub_with_mask" : "arithm_s_sub";
- const char **kernelString = mask.data ? &arithm_sub_scalar_mask : &arithm_sub_scalar;
+ string kernelName = mask.data ? "arithm_s_add_with_mask" : "arithm_s_add";
+ const char **kernelString = mask.data ? &arithm_add_scalar_mask : &arithm_add_scalar;
arithmetic_scalar( src1, src2, dst, mask, kernelName, kernelString, -1);
}
void cv::ocl::divide(double scalar, const oclMat &src, oclMat &dst)
#endif
#endif
+#ifdef ARITHM_ADD
+ #define ARITHM_OP(A,B) ((A)+(B))
+#elif defined ARITHM_SUB
+ #define ARITHM_OP(A,B) ((A)-(B))
+#endif
//////////////////////////////////////////////////////////////////////////////////////////////////////
/////////////////////////////////////////////ADD////////////////////////////////////////////////////
///////////////////////////////////////////////////////////////////////////////////////////////////////
src2_data.xyzw = (src2_index == -1) ? src2_data.wxyz:tmp.xyzw;
}
uchar4 dst_data = *((__global uchar4 *)(dst + dst_index));
- short4 tmp = convert_short4_sat(src1_data) + convert_short4_sat(src2_data);
+ short4 tmp = ARITHM_OP(convert_short4_sat(src1_data), convert_short4_sat(src2_data));
uchar4 tmp_data = convert_uchar4_sat(tmp);
dst_data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : dst_data.x;
ushort4 src2_data = vload4(0, (__global ushort *)((__global char *)src2 + src2_index));
ushort4 dst_data = *((__global ushort4 *)((__global char *)dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) + convert_int4_sat(src2_data);
+ int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), convert_int4_sat(src2_data));
ushort4 tmp_data = convert_ushort4_sat(tmp);
dst_data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : dst_data.x;
short4 src2_data = vload4(0, (__global short *)((__global char *)src2 + src2_index));
short4 dst_data = *((__global short4 *)((__global char *)dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) + convert_int4_sat(src2_data);
+ int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), convert_int4_sat(src2_data));
short4 tmp_data = convert_short4_sat(tmp);
dst_data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : dst_data.x;
int data1 = *((__global int *)((__global char *)src1 + src1_index));
int data2 = *((__global int *)((__global char *)src2 + src2_index));
- long tmp = (long)(data1) + (long)(data2);
+ long tmp = ARITHM_OP((long)(data1), (long)(data2));
*((__global int *)((__global char *)dst + dst_index)) = convert_int_sat(tmp);
}
float data1 = *((__global float *)((__global char *)src1 + src1_index));
float data2 = *((__global float *)((__global char *)src2 + src2_index));
- float tmp = data1 + data2;
+ float tmp = ARITHM_OP(data1, data2);
*((__global float *)((__global char *)dst + dst_index)) = tmp;
}
double data1 = *((__global double *)((__global char *)src1 + src1_index));
double data2 = *((__global double *)((__global char *)src2 + src2_index));
- *((__global double *)((__global char *)dst + dst_index)) = data1 + data2;
+ *((__global double *)((__global char *)dst + dst_index)) = ARITHM_OP(data1, data2);
}
}
#endif
}
uchar4 data = *((__global uchar4 *)(dst + dst_index));
- short4 tmp = convert_short4_sat(src1_data) + convert_short4_sat(src2_data);
+ short4 tmp = ARITHM_OP(convert_short4_sat(src1_data), convert_short4_sat(src2_data));
uchar4 tmp_data = convert_uchar4_sat(tmp);
data.x = ((mask_data.x) && (dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : data.x;
uchar2 mask_data = vload2(0, mask + mask_index);
ushort2 data = *((__global ushort2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) + convert_int2_sat(src2_data);
+ int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), convert_int2_sat(src2_data));
ushort2 tmp_data = convert_ushort2_sat(tmp);
data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
uchar2 mask_data = vload2(0, mask + mask_index);
short2 data = *((__global short2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) + convert_int2_sat(src2_data);
+ int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), convert_int2_sat(src2_data));
short2 tmp_data = convert_short2_sat(tmp);
data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
int src_data2 = *((__global int *)((__global char *)src2 + src2_index));
int dst_data = *((__global int *)((__global char *)dst + dst_index));
- int data = convert_int_sat((long)src_data1 + (long)src_data2);
+ int data = convert_int_sat(ARITHM_OP((long)src_data1, (long)src_data2));
data = mask_data ? data : dst_data;
*((__global int *)((__global char *)dst + dst_index)) = data;
float src_data2 = *((__global float *)((__global char *)src2 + src2_index));
float dst_data = *((__global float *)((__global char *)dst + dst_index));
- float data = src_data1 + src_data2;
+ float data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global float *)((__global char *)dst + dst_index)) = data;
double src_data2 = *((__global double *)((__global char *)src2 + src2_index));
double dst_data = *((__global double *)((__global char *)dst + dst_index));
- double data = src_data1 + src_data2;
+ double data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global double *)((__global char *)dst + dst_index)) = data;
uchar2 mask_data = vload2(0, mask + mask_index);
uchar4 data = *((__global uchar4 *)(dst + dst_index));
- short4 tmp = convert_short4_sat(src1_data) + convert_short4_sat(src2_data);
+ short4 tmp = ARITHM_OP(convert_short4_sat(src1_data), convert_short4_sat(src2_data));
uchar4 tmp_data = convert_uchar4_sat(tmp);
data.xy = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.xy : data.xy;
ushort2 src_data2 = *((__global ushort2 *)((__global char *)src2 + src2_index));
ushort2 dst_data = *((__global ushort2 *)((__global char *)dst + dst_index));
- int2 tmp = convert_int2_sat(src_data1) + convert_int2_sat(src_data2);
+ int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), convert_int2_sat(src_data2));
ushort2 data = convert_ushort2_sat(tmp);
data = mask_data ? data : dst_data;
short2 src_data2 = *((__global short2 *)((__global char *)src2 + src2_index));
short2 dst_data = *((__global short2 *)((__global char *)dst + dst_index));
- int2 tmp = convert_int2_sat(src_data1) + convert_int2_sat(src_data2);
+ int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), convert_int2_sat(src_data2));
short2 data = convert_short2_sat(tmp);
data = mask_data ? data : dst_data;
int2 src_data2 = *((__global int2 *)((__global char *)src2 + src2_index));
int2 dst_data = *((__global int2 *)((__global char *)dst + dst_index));
- int2 data = convert_int2_sat(convert_long2_sat(src_data1) + convert_long2_sat(src_data2));
+ int2 data = convert_int2_sat(ARITHM_OP(convert_long2_sat(src_data1), convert_long2_sat(src_data2)));
data = mask_data ? data : dst_data;
*((__global int2 *)((__global char *)dst + dst_index)) = data;
float2 src_data2 = *((__global float2 *)((__global char *)src2 + src2_index));
float2 dst_data = *((__global float2 *)((__global char *)dst + dst_index));
- float2 data = src_data1 + src_data2;
+ float2 data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global float2 *)((__global char *)dst + dst_index)) = data;
double2 src_data2 = *((__global double2 *)((__global char *)src2 + src2_index));
double2 dst_data = *((__global double2 *)((__global char *)dst + dst_index));
- double2 data = src_data1 + src_data2;
+ double2 data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global double2 *)((__global char *)dst + dst_index)) = data;
uchar4 src_data2 = *((__global uchar4 *)(src2 + src2_index));
uchar4 dst_data = *((__global uchar4 *)(dst + dst_index));
- uchar4 data = convert_uchar4_sat(convert_ushort4_sat(src_data1) + convert_ushort4_sat(src_data2));
+ uchar4 data = convert_uchar4_sat(ARITHM_OP(convert_short4_sat(src_data1), convert_short4_sat(src_data2)));
data = mask_data ? data : dst_data;
*((__global uchar4 *)(dst + dst_index)) = data;
ushort4 src_data2 = *((__global ushort4 *)((__global char *)src2 + src2_index));
ushort4 dst_data = *((__global ushort4 *)((__global char *)dst + dst_index));
- ushort4 data = convert_ushort4_sat(convert_int4_sat(src_data1) + convert_int4_sat(src_data2));
+ ushort4 data = convert_ushort4_sat(ARITHM_OP(convert_int4_sat(src_data1), convert_int4_sat(src_data2)));
data = mask_data ? data : dst_data;
*((__global ushort4 *)((__global char *)dst + dst_index)) = data;
short4 src_data2 = *((__global short4 *)((__global char *)src2 + src2_index));
short4 dst_data = *((__global short4 *)((__global char *)dst + dst_index));
- short4 data = convert_short4_sat(convert_int4_sat(src_data1) + convert_int4_sat(src_data2));
+ short4 data = convert_short4_sat(ARITHM_OP(convert_int4_sat(src_data1), convert_int4_sat(src_data2)));
data = mask_data ? data : dst_data;
*((__global short4 *)((__global char *)dst + dst_index)) = data;
int4 src_data2 = *((__global int4 *)((__global char *)src2 + src2_index));
int4 dst_data = *((__global int4 *)((__global char *)dst + dst_index));
- int4 data = convert_int4_sat(convert_long4_sat(src_data1) + convert_long4_sat(src_data2));
+ int4 data = convert_int4_sat(ARITHM_OP(convert_long4_sat(src_data1), convert_long4_sat(src_data2)));
data = mask_data ? data : dst_data;
*((__global int4 *)((__global char *)dst + dst_index)) = data;
float4 src_data2 = *((__global float4 *)((__global char *)src2 + src2_index));
float4 dst_data = *((__global float4 *)((__global char *)dst + dst_index));
- float4 data = src_data1 + src_data2;
+ float4 data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global float4 *)((__global char *)dst + dst_index)) = data;
double4 src_data2 = *((__global double4 *)((__global char *)src2 + src2_index));
double4 dst_data = *((__global double4 *)((__global char *)dst + dst_index));
- double4 data = src_data1 + src_data2;
+ double4 data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global double4 *)((__global char *)dst + dst_index)) = data;
#elif defined (cl_amd_fp64)
#pragma OPENCL EXTENSION cl_amd_fp64:enable
#endif
+#endif
+#ifdef ARITHM_ADD
+ #define ARITHM_OP(A,B) ((A)+(B))
+#elif defined ARITHM_SUB
+ #define ARITHM_OP(A,B) ((A)-(B))
#endif
/**************************************add with scalar without mask**************************************/
__kernel void arithm_s_add_C1_D0 (__global uchar *src1, int src1_step, int src1_offset,
}
uchar4 data = *((__global uchar4 *)(dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) + src2_data;
+ int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), src2_data);
uchar4 tmp_data = convert_uchar4_sat(tmp);
data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : data.x;
int2 src2_data = (int2)(src2.x, src2.x);
ushort2 data = *((__global ushort2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) + src2_data;
+ int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), src2_data);
ushort2 tmp_data = convert_ushort2_sat(tmp);
data.x = (dst_index + 0 >= dst_start) ? tmp_data.x : data.x;
int2 src2_data = (int2)(src2.x, src2.x);
short2 data = *((__global short2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) + src2_data;
+ int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), src2_data);
short2 tmp_data = convert_short2_sat(tmp);
data.x = (dst_index + 0 >= dst_start) ? tmp_data.x : data.x;
int src_data2 = src2.x;
int dst_data = *((__global int *)((__global char *)dst + dst_index));
- int data = convert_int_sat((long)src_data1 + (long)src_data2);
+ int data = convert_int_sat(ARITHM_OP((long)src_data1, (long)src_data2));
*((__global int *)((__global char *)dst + dst_index)) = data;
}
float src_data2 = src2.x;
float dst_data = *((__global float *)((__global char *)dst + dst_index));
- float data = src_data1 + src_data2;
+ float data = ARITHM_OP(src_data1, src_data2);
*((__global float *)((__global char *)dst + dst_index)) = data;
}
double src2_data = src2.x;
double dst_data = *((__global double *)((__global char *)dst + dst_index));
- double data = src_data1 + src2_data;
+ double data = ARITHM_OP(src_data1, src2_data);
*((__global double *)((__global char *)dst + dst_index)) = data;
}
int4 src2_data = (int4)(src2.x, src2.y, src2.x, src2.y);
uchar4 data = *((__global uchar4 *)(dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) + src2_data;
+ int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), src2_data);
uchar4 tmp_data = convert_uchar4_sat(tmp);
data.xy = (dst_index + 0 >= dst_start) ? tmp_data.xy : data.xy;
int2 src_data2 = (int2)(src2.x, src2.y);
ushort2 dst_data = *((__global ushort2 *)((__global char *)dst + dst_index));
- int2 tmp = convert_int2_sat(src_data1) + src_data2;
+ int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), src_data2);
ushort2 data = convert_ushort2_sat(tmp);
*((__global ushort2 *)((__global char *)dst + dst_index)) = data;
int2 src_data2 = (int2)(src2.x, src2.y);
short2 dst_data = *((__global short2 *)((__global char *)dst + dst_index));
- int2 tmp = convert_int2_sat(src_data1) + src_data2;
+ int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), src_data2);
short2 data = convert_short2_sat(tmp);
*((__global short2 *)((__global char *)dst + dst_index)) = data;
int2 src_data2 = (int2)(src2.x, src2.y);
int2 dst_data = *((__global int2 *)((__global char *)dst + dst_index));
- int2 data = convert_int2_sat(convert_long2_sat(src_data1) + convert_long2_sat(src_data2));
+ int2 data = convert_int2_sat(ARITHM_OP(convert_long2_sat(src_data1), convert_long2_sat(src_data2)));
*((__global int2 *)((__global char *)dst + dst_index)) = data;
}
}
float2 src_data2 = (float2)(src2.x, src2.y);
float2 dst_data = *((__global float2 *)((__global char *)dst + dst_index));
- float2 data = src_data1 + src_data2;
+ float2 data = ARITHM_OP(src_data1, src_data2);
*((__global float2 *)((__global char *)dst + dst_index)) = data;
}
}
double2 src_data2 = (double2)(src2.x, src2.y);
double2 dst_data = *((__global double2 *)((__global char *)dst + dst_index));
- double2 data = src_data1 + src_data2;
+ double2 data = ARITHM_OP(src_data1, src_data2);
*((__global double2 *)((__global char *)dst + dst_index)) = data;
}
uchar4 src_data1 = *((__global uchar4 *)(src1 + src1_index));
- uchar4 data = convert_uchar4_sat(convert_int4_sat(src_data1) + src2);
+ uchar4 data = convert_uchar4_sat(ARITHM_OP(convert_int4_sat(src_data1), src2));
*((__global uchar4 *)(dst + dst_index)) = data;
}
ushort4 src_data1 = *((__global ushort4 *)((__global char *)src1 + src1_index));
- ushort4 data = convert_ushort4_sat(convert_int4_sat(src_data1) + src2);
+ ushort4 data = convert_ushort4_sat(ARITHM_OP(convert_int4_sat(src_data1), src2));
*((__global ushort4 *)((__global char *)dst + dst_index)) = data;
}
short4 src_data1 = *((__global short4 *)((__global char *)src1 + src1_index));
- short4 data = convert_short4_sat(convert_int4_sat(src_data1) + src2);
+ short4 data = convert_short4_sat(ARITHM_OP(convert_int4_sat(src_data1), src2));
*((__global short4 *)((__global char *)dst + dst_index)) = data;
}
int4 src_data1 = *((__global int4 *)((__global char *)src1 + src1_index));
- int4 data = convert_int4_sat(convert_long4_sat(src_data1) + convert_long4_sat(src2));
+ int4 data = convert_int4_sat(ARITHM_OP(convert_long4_sat(src_data1), convert_long4_sat(src2)));
*((__global int4 *)((__global char *)dst + dst_index)) = data;
}
float4 src_data1 = *((__global float4 *)((__global char *)src1 + src1_index));
- float4 data = src_data1 + src2;
+ float4 data = ARITHM_OP(src_data1, src2);
*((__global float4 *)((__global char *)dst + dst_index)) = data;
}
double4 src_data1 = *((__global double4 *)((__global char *)src1 + src1_index));
- double4 data = src_data1 + src2;
+ double4 data = ARITHM_OP(src_data1, src2);
*((__global double4 *)((__global char *)dst + dst_index)) = data;
}
#endif
#endif
+#ifdef ARITHM_ADD
+ #define ARITHM_OP(A,B) ((A)+(B))
+#elif defined ARITHM_SUB
+ #define ARITHM_OP(A,B) ((A)-(B))
+#endif
/**************************************add with scalar with mask**************************************/
__kernel void arithm_s_add_with_mask_C1_D0 (__global uchar *src1, int src1_step, int src1_offset,
__global uchar *dst, int dst_step, int dst_offset,
}
uchar4 data = *((__global uchar4 *)(dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) + src2_data;
+ int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), src2_data);
uchar4 tmp_data = convert_uchar4_sat(tmp);
data.x = ((mask_data.x) && (dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : data.x;
uchar2 mask_data = vload2(0, mask + mask_index);
ushort2 data = *((__global ushort2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) + src2_data;
+ int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), src2_data);
ushort2 tmp_data = convert_ushort2_sat(tmp);
data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
uchar2 mask_data = vload2(0, mask + mask_index);
short2 data = *((__global short2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) + src2_data;
+ int2 tmp = ARITHM_OP(convert_int2_sat(src1_data), src2_data);
short2 tmp_data = convert_short2_sat(tmp);
data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
int src_data2 = src2.x;
int dst_data = *((__global int *)((__global char *)dst + dst_index));
- int data = convert_int_sat((long)src_data1 + (long)src_data2);
+ int data = convert_int_sat(ARITHM_OP((long)src_data1, (long)src_data2));
data = mask_data ? data : dst_data;
*((__global int *)((__global char *)dst + dst_index)) = data;
float src_data2 = src2.x;
float dst_data = *((__global float *)((__global char *)dst + dst_index));
- float data = src_data1 + src_data2;
+ float data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global float *)((__global char *)dst + dst_index)) = data;
double src_data2 = src2.x;
double dst_data = *((__global double *)((__global char *)dst + dst_index));
- double data = src_data1 + src_data2;
+ double data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global double *)((__global char *)dst + dst_index)) = data;
uchar2 mask_data = vload2(0, mask + mask_index);
uchar4 data = *((__global uchar4 *)(dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) + src2_data;
+ int4 tmp = ARITHM_OP(convert_int4_sat(src1_data), src2_data);
uchar4 tmp_data = convert_uchar4_sat(tmp);
data.xy = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.xy : data.xy;
int2 src_data2 = (int2)(src2.x, src2.y);
ushort2 dst_data = *((__global ushort2 *)((__global char *)dst + dst_index));
- int2 tmp = convert_int2_sat(src_data1) + src_data2;
+ int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), src_data2);
ushort2 data = convert_ushort2_sat(tmp);
data = mask_data ? data : dst_data;
int2 src_data2 = (int2)(src2.x, src2.y);
short2 dst_data = *((__global short2 *)((__global char *)dst + dst_index));
- int2 tmp = convert_int2_sat(src_data1) + src_data2;
+ int2 tmp = ARITHM_OP(convert_int2_sat(src_data1), src_data2);
short2 data = convert_short2_sat(tmp);
data = mask_data ? data : dst_data;
int2 src_data2 = (int2)(src2.x, src2.y);
int2 dst_data = *((__global int2 *)((__global char *)dst + dst_index));
- int2 data = convert_int2_sat(convert_long2_sat(src_data1) + convert_long2_sat(src_data2));
+ int2 data = convert_int2_sat(ARITHM_OP(convert_long2_sat(src_data1), convert_long2_sat(src_data2)));
data = mask_data ? data : dst_data;
*((__global int2 *)((__global char *)dst + dst_index)) = data;
float2 src_data2 = (float2)(src2.x, src2.y);
float2 dst_data = *((__global float2 *)((__global char *)dst + dst_index));
- float2 data = src_data1 + src_data2;
+ float2 data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global float2 *)((__global char *)dst + dst_index)) = data;
double2 src_data2 = (double2)(src2.x, src2.y);
double2 dst_data = *((__global double2 *)((__global char *)dst + dst_index));
- double2 data = src_data1 + src_data2;
+ double2 data = ARITHM_OP(src_data1, src_data2);
data = mask_data ? data : dst_data;
*((__global double2 *)((__global char *)dst + dst_index)) = data;
__global uchar *mask, int mask_step, int mask_offset,
int4 src2, int rows, int cols, int dst_step1)
{
-
int x = get_global_id(0);
int y = get_global_id(1);
uchar4 src_data1 = *((__global uchar4 *)(src1 + src1_index));
uchar4 dst_data = *((__global uchar4 *)(dst + dst_index));
- uchar4 data = convert_uchar4_sat(convert_int4_sat(src_data1) + src2);
+ uchar4 data = convert_uchar4_sat(ARITHM_OP(convert_int4_sat(src_data1), src2));
data = mask_data ? data : dst_data;
*((__global uchar4 *)(dst + dst_index)) = data;
ushort4 src_data1 = *((__global ushort4 *)((__global char *)src1 + src1_index));
ushort4 dst_data = *((__global ushort4 *)((__global char *)dst + dst_index));
- ushort4 data = convert_ushort4_sat(convert_int4_sat(src_data1) + src2);
+ ushort4 data = convert_ushort4_sat(ARITHM_OP(convert_int4_sat(src_data1), src2));
data = mask_data ? data : dst_data;
*((__global ushort4 *)((__global char *)dst + dst_index)) = data;
short4 src_data1 = *((__global short4 *)((__global char *)src1 + src1_index));
short4 dst_data = *((__global short4 *)((__global char *)dst + dst_index));
- short4 data = convert_short4_sat(convert_int4_sat(src_data1) + src2);
+ short4 data = convert_short4_sat(ARITHM_OP(convert_int4_sat(src_data1), src2));
data = mask_data ? data : dst_data;
*((__global short4 *)((__global char *)dst + dst_index)) = data;
int4 src_data1 = *((__global int4 *)((__global char *)src1 + src1_index));
int4 dst_data = *((__global int4 *)((__global char *)dst + dst_index));
- int4 data = convert_int4_sat(convert_long4_sat(src_data1) + convert_long4_sat(src2));
+ int4 data = convert_int4_sat(ARITHM_OP(convert_long4_sat(src_data1), convert_long4_sat(src2)));
data = mask_data ? data : dst_data;
*((__global int4 *)((__global char *)dst + dst_index)) = data;
float4 src_data1 = *((__global float4 *)((__global char *)src1 + src1_index));
float4 dst_data = *((__global float4 *)((__global char *)dst + dst_index));
- float4 data = src_data1 + src2;
+ float4 data = ARITHM_OP(src_data1, src2);
data = mask_data ? data : dst_data;
*((__global float4 *)((__global char *)dst + dst_index)) = data;
double4 src_data1 = *((__global double4 *)((__global char *)src1 + src1_index));
double4 dst_data = *((__global double4 *)((__global char *)dst + dst_index));
- double4 data = src_data1 + src2;
+ double4 data = ARITHM_OP(src_data1, src2);
data = mask_data ? data : dst_data;
*((__global double4 *)((__global char *)dst + dst_index)) = data;
+++ /dev/null
-/*M///////////////////////////////////////////////////////////////////////////////////////
-//
-// IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
-//
-// By downloading, copying, installing or using the software you agree to this license.
-// If you do not agree to this license, do not download, install,
-// copy or use the software.
-//
-//
-// License Agreement
-// For Open Source Computer Vision Library
-//
-// Copyright (C) 2010-2012, Institute Of Software Chinese Academy Of Science, all rights reserved.
-// Copyright (C) 2010-2012, Advanced Micro Devices, Inc., all rights reserved.
-// Third party copyrights are property of their respective owners.
-//
-// @Authors
-// Jia Haipeng, jiahaipeng95@gmail.com
-//
-// Redistribution and use in source and binary forms, with or without modification,
-// are permitted provided that the following conditions are met:
-//
-// * Redistribution's of source code must retain the above copyright notice,
-// this list of conditions and the following disclaimer.
-//
-// * Redistribution's in binary form must reproduce the above copyright notice,
-// this list of conditions and the following disclaimer in the documentation
-// and/or other GpuMaterials provided with the distribution.
-//
-// * The name of the copyright holders may not be used to endorse or promote products
-// derived from this software without specific prior written permission.
-//
-// This software is provided by the copyright holders and contributors as is and
-// any express or implied warranties, including, but not limited to, the implied
-// warranties of merchantability and fitness for a particular purpose are disclaimed.
-// In no event shall the Intel Corporation or contributors be liable for any direct,
-// indirect, incidental, special, exemplary, or consequential damages
-// (including, but not limited to, procurement of substitute goods or services;
-// loss of use, data, or profits; or business interruption) however caused
-// and on any theory of liability, whether in contract, strict liability,
-// or tort (including negligence or otherwise) arising in any way out of
-// the use of this software, even if advised of the possibility of such damage.
-//
-//M*/
-
-#if defined (DOUBLE_SUPPORT)
-#ifdef cl_khr_fp64
-#pragma OPENCL EXTENSION cl_khr_fp64:enable
-#elif defined (cl_amd_fp64)
-#pragma OPENCL EXTENSION cl_amd_fp64:enable
-#endif
-#endif
-
-//////////////////////////////////////////////////////////////////////////////////////////////////////
-/////////////////////////////////////////////SUB////////////////////////////////////////////////////
-///////////////////////////////////////////////////////////////////////////////////////////////////////
-/**************************************sub without mask**************************************/
-__kernel void arithm_sub_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *src2, int src2_step, int src2_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 2;
-
- #define dst_align (dst_offset & 3)
- int src1_index = mad24(y, src1_step, x + src1_offset - dst_align);
- int src2_index = mad24(y, src2_step, x + src2_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + x & (int)0xfffffffc);
-
- uchar4 src1_data = vload4(0, src1 + src1_index);
- uchar4 src2_data = vload4(0, src2 + src2_index);
-
- uchar4 dst_data = *((__global uchar4 *)(dst + dst_index));
- short4 tmp = convert_short4_sat(src1_data) - convert_short4_sat(src2_data);
- uchar4 tmp_data = convert_uchar4_sat(tmp);
-
- dst_data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : dst_data.x;
- dst_data.y = ((dst_index + 1 >= dst_start) && (dst_index + 1 < dst_end)) ? tmp_data.y : dst_data.y;
- dst_data.z = ((dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.z : dst_data.z;
- dst_data.w = ((dst_index + 3 >= dst_start) && (dst_index + 3 < dst_end)) ? tmp_data.w : dst_data.w;
-
- *((__global uchar4 *)(dst + dst_index)) = dst_data;
- }
-}
-__kernel void arithm_sub_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *src2, int src2_step, int src2_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 2;
-
- #define dst_align ((dst_offset >> 1) & 3)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
- int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1));
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffff8);
-
- ushort4 src1_data = vload4(0, (__global ushort *)((__global char *)src1 + src1_index));
- ushort4 src2_data = vload4(0, (__global ushort *)((__global char *)src2 + src2_index));
-
- ushort4 dst_data = *((__global ushort4 *)((__global char *)dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) - convert_int4_sat(src2_data);
- ushort4 tmp_data = convert_ushort4_sat(tmp);
-
- dst_data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : dst_data.x;
- dst_data.y = ((dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.y : dst_data.y;
- dst_data.z = ((dst_index + 4 >= dst_start) && (dst_index + 4 < dst_end)) ? tmp_data.z : dst_data.z;
- dst_data.w = ((dst_index + 6 >= dst_start) && (dst_index + 6 < dst_end)) ? tmp_data.w : dst_data.w;
-
- *((__global ushort4 *)((__global char *)dst + dst_index)) = dst_data;
- }
-}
-__kernel void arithm_sub_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *src2, int src2_step, int src2_offset,
- __global short *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 2;
-
- #define dst_align ((dst_offset >> 1) & 3)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
- int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1));
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffff8);
-
- short4 src1_data = vload4(0, (__global short *)((__global char *)src1 + src1_index));
- short4 src2_data = vload4(0, (__global short *)((__global char *)src2 + src2_index));
-
- short4 dst_data = *((__global short4 *)((__global char *)dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) - convert_int4_sat(src2_data);
- short4 tmp_data = convert_short4_sat(tmp);
-
- dst_data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : dst_data.x;
- dst_data.y = ((dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.y : dst_data.y;
- dst_data.z = ((dst_index + 4 >= dst_start) && (dst_index + 4 < dst_end)) ? tmp_data.z : dst_data.z;
- dst_data.w = ((dst_index + 6 >= dst_start) && (dst_index + 6 < dst_end)) ? tmp_data.w : dst_data.w;
-
- *((__global short4 *)((__global char *)dst + dst_index)) = dst_data;
- }
-}
-
-__kernel void arithm_sub_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *src2, int src2_step, int src2_offset,
- __global int *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 2) + src2_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- int data1 = *((__global int *)((__global char *)src1 + src1_index));
- int data2 = *((__global int *)((__global char *)src2 + src2_index));
- long tmp = (long)(data1) - (long)(data2);
-
- *((__global int *)((__global char *)dst + dst_index)) = convert_int_sat(tmp);
- }
-}
-__kernel void arithm_sub_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *src2, int src2_step, int src2_offset,
- __global float *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 2) + src2_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- float data1 = *((__global float *)((__global char *)src1 + src1_index));
- float data2 = *((__global float *)((__global char *)src2 + src2_index));
- float tmp = data1 - data2;
-
- *((__global float *)((__global char *)dst + dst_index)) = tmp;
- }
-}
-
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_sub_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *src2, int src2_step, int src2_offset,
- __global double *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 3) + src2_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- double data1 = *((__global double *)((__global char *)src1 + src1_index));
- double data2 = *((__global double *)((__global char *)src2 + src2_index));
-
- *((__global double *)((__global char *)dst + dst_index)) = data1 - data2;
- }
-}
-#endif
-
-/**************************************sub with mask**************************************/
-__kernel void arithm_sub_with_mask_C1_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 2;
-
- #define dst_align (dst_offset & 3)
- int src1_index = mad24(y, src1_step, x + src1_offset - dst_align);
- int src2_index = mad24(y, src2_step, x + src2_offset - dst_align);
- int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + x & (int)0xfffffffc);
-
- uchar4 src1_data = vload4(0, src1 + src1_index);
- uchar4 src2_data = vload4(0, src2 + src2_index);
- uchar4 mask_data = vload4(0, mask + mask_index);
-
- uchar4 data = *((__global uchar4 *)(dst + dst_index));
- short4 tmp = convert_short4_sat(src1_data) - convert_short4_sat(src2_data);
- uchar4 tmp_data = convert_uchar4_sat(tmp);
-
- data.x = ((mask_data.x) && (dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : data.x;
- data.y = ((mask_data.y) && (dst_index + 1 >= dst_start) && (dst_index + 1 < dst_end)) ? tmp_data.y : data.y;
- data.z = ((mask_data.z) && (dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.z : data.z;
- data.w = ((mask_data.w) && (dst_index + 3 >= dst_start) && (dst_index + 3 < dst_end)) ? tmp_data.w : data.w;
-
- *((__global uchar4 *)(dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C1_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 1;
-
- #define dst_align ((dst_offset >> 1) & 1)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
- int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1));
- int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
-
- ushort2 src1_data = vload2(0, (__global ushort *)((__global char *)src1 + src1_index));
- ushort2 src2_data = vload2(0, (__global ushort *)((__global char *)src2 + src2_index));
- uchar2 mask_data = vload2(0, mask + mask_index);
-
- ushort2 data = *((__global ushort2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) - convert_int2_sat(src2_data);
- ushort2 tmp_data = convert_ushort2_sat(tmp);
-
- data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
- data.y = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.y : data.y;
-
- *((__global ushort2 *)((__global uchar *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C1_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global short *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 1;
-
- #define dst_align ((dst_offset >> 1) & 1)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
- int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1));
- int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
-
- short2 src1_data = vload2(0, (__global short *)((__global char *)src1 + src1_index));
- short2 src2_data = vload2(0, (__global short *)((__global char *)src2 + src2_index));
- uchar2 mask_data = vload2(0, mask + mask_index);
-
- short2 data = *((__global short2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) - convert_int2_sat(src2_data);
- short2 tmp_data = convert_short2_sat(tmp);
-
- data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
- data.y = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.y : data.y;
-
- *((__global short2 *)((__global uchar *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C1_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global int *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 2) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- int src_data1 = *((__global int *)((__global char *)src1 + src1_index));
- int src_data2 = *((__global int *)((__global char *)src2 + src2_index));
- int dst_data = *((__global int *)((__global char *)dst + dst_index));
-
- int data = convert_int_sat((long)src_data1 - (long)src_data2);
- data = mask_data ? data : dst_data;
-
- *((__global int *)((__global char *)dst + dst_index)) = data;
- }
-}
-
-__kernel void arithm_sub_with_mask_C1_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global float *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 2) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- float src_data1 = *((__global float *)((__global char *)src1 + src1_index));
- float src_data2 = *((__global float *)((__global char *)src2 + src2_index));
- float dst_data = *((__global float *)((__global char *)dst + dst_index));
-
- float data = src_data1 - src_data2;
- data = mask_data ? data : dst_data;
-
- *((__global float *)((__global char *)dst + dst_index)) = data;
- }
-}
-
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_sub_with_mask_C1_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global double *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 3) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- double src_data1 = *((__global double *)((__global char *)src1 + src1_index));
- double src_data2 = *((__global double *)((__global char *)src2 + src2_index));
- double dst_data = *((__global double *)((__global char *)dst + dst_index));
-
- double data = src_data1 - src_data2;
- data = mask_data ? data : dst_data;
-
- *((__global double *)((__global char *)dst + dst_index)) = data;
- }
-}
-#endif
-
-__kernel void arithm_sub_with_mask_C2_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 1;
-
- #define dst_align ((dst_offset >> 1) & 1)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
- int src2_index = mad24(y, src2_step, (x << 1) + src2_offset - (dst_align << 1));
- int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
-
- uchar4 src1_data = vload4(0, src1 + src1_index);
- uchar4 src2_data = vload4(0, src2 + src2_index);
- uchar2 mask_data = vload2(0, mask + mask_index);
-
- uchar4 data = *((__global uchar4 *)(dst + dst_index));
- short4 tmp = convert_short4_sat(src1_data) - convert_short4_sat(src2_data);
- uchar4 tmp_data = convert_uchar4_sat(tmp);
-
- data.xy = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.xy : data.xy;
- data.zw = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.zw : data.zw;
-
- *((__global uchar4 *)(dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C2_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 2) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- ushort2 src_data1 = *((__global ushort2 *)((__global char *)src1 + src1_index));
- ushort2 src_data2 = *((__global ushort2 *)((__global char *)src2 + src2_index));
- ushort2 dst_data = *((__global ushort2 *)((__global char *)dst + dst_index));
-
- int2 tmp = convert_int2_sat(src_data1) - convert_int2_sat(src_data2);
- ushort2 data = convert_ushort2_sat(tmp);
- data = mask_data ? data : dst_data;
-
- *((__global ushort2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C2_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global short *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 2) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- short2 src_data1 = *((__global short2 *)((__global char *)src1 + src1_index));
- short2 src_data2 = *((__global short2 *)((__global char *)src2 + src2_index));
- short2 dst_data = *((__global short2 *)((__global char *)dst + dst_index));
-
- int2 tmp = convert_int2_sat(src_data1) - convert_int2_sat(src_data2);
- short2 data = convert_short2_sat(tmp);
- data = mask_data ? data : dst_data;
-
- *((__global short2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C2_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global int *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 3) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- int2 src_data1 = *((__global int2 *)((__global char *)src1 + src1_index));
- int2 src_data2 = *((__global int2 *)((__global char *)src2 + src2_index));
- int2 dst_data = *((__global int2 *)((__global char *)dst + dst_index));
-
- int2 data = convert_int2_sat(convert_long2_sat(src_data1) - convert_long2_sat(src_data2));
- data = mask_data ? data : dst_data;
-
- *((__global int2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C2_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global float *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 3) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- float2 src_data1 = *((__global float2 *)((__global char *)src1 + src1_index));
- float2 src_data2 = *((__global float2 *)((__global char *)src2 + src2_index));
- float2 dst_data = *((__global float2 *)((__global char *)dst + dst_index));
-
- float2 data = src_data1 - src_data2;
- data = mask_data ? data : dst_data;
-
- *((__global float2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_sub_with_mask_C2_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global double *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 4) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- double2 src_data1 = *((__global double2 *)((__global char *)src1 + src1_index));
- double2 src_data2 = *((__global double2 *)((__global char *)src2 + src2_index));
- double2 dst_data = *((__global double2 *)((__global char *)dst + dst_index));
-
- double2 data = src_data1 - src_data2;
- data = mask_data ? data : dst_data;
-
- *((__global double2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-#endif
-
-__kernel void arithm_sub_with_mask_C4_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 2) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- uchar4 src_data1 = *((__global uchar4 *)(src1 + src1_index));
- uchar4 src_data2 = *((__global uchar4 *)(src2 + src2_index));
- uchar4 dst_data = *((__global uchar4 *)(dst + dst_index));
-
- uchar4 data = convert_uchar4_sat(convert_short4_sat(src_data1) - convert_short4_sat(src_data2));
- data = mask_data ? data : dst_data;
-
- *((__global uchar4 *)(dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C4_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 3) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- ushort4 src_data1 = *((__global ushort4 *)((__global char *)src1 + src1_index));
- ushort4 src_data2 = *((__global ushort4 *)((__global char *)src2 + src2_index));
- ushort4 dst_data = *((__global ushort4 *)((__global char *)dst + dst_index));
-
- ushort4 data = convert_ushort4_sat(convert_int4_sat(src_data1) - convert_int4_sat(src_data2));
- data = mask_data ? data : dst_data;
-
- *((__global ushort4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C4_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global short *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 3) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- short4 src_data1 = *((__global short4 *)((__global char *)src1 + src1_index));
- short4 src_data2 = *((__global short4 *)((__global char *)src2 + src2_index));
- short4 dst_data = *((__global short4 *)((__global char *)dst + dst_index));
-
- short4 data = convert_short4_sat(convert_int4_sat(src_data1) - convert_int4_sat(src_data2));
- data = mask_data ? data : dst_data;
-
- *((__global short4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C4_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global int *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 4) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- int4 src_data1 = *((__global int4 *)((__global char *)src1 + src1_index));
- int4 src_data2 = *((__global int4 *)((__global char *)src2 + src2_index));
- int4 dst_data = *((__global int4 *)((__global char *)dst + dst_index));
-
- int4 data = convert_int4_sat(convert_long4_sat(src_data1) - convert_long4_sat(src_data2));
- data = mask_data ? data : dst_data;
-
- *((__global int4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_sub_with_mask_C4_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global float *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 4) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- float4 src_data1 = *((__global float4 *)((__global char *)src1 + src1_index));
- float4 src_data2 = *((__global float4 *)((__global char *)src2 + src2_index));
- float4 dst_data = *((__global float4 *)((__global char *)dst + dst_index));
-
- float4 data = src_data1 - src_data2;
- data = mask_data ? data : dst_data;
-
- *((__global float4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_sub_with_mask_C4_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *src2, int src2_step, int src2_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- __global double *dst, int dst_step, int dst_offset,
- int rows, int cols, int dst_step1)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 5) + src1_offset);
- int src2_index = mad24(y, src2_step, (x << 5) + src2_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 5) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- double4 src_data1 = *((__global double4 *)((__global char *)src1 + src1_index));
- double4 src_data2 = *((__global double4 *)((__global char *)src2 + src2_index));
- double4 dst_data = *((__global double4 *)((__global char *)dst + dst_index));
-
- double4 data = src_data1 - src_data2;
- data = mask_data ? data : dst_data;
-
- *((__global double4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-#endif
+++ /dev/null
-/*M///////////////////////////////////////////////////////////////////////////////////////
-//
-// IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
-//
-// By downloading, copying, installing or using the software you agree to this license.
-// If you do not agree to this license, do not download, install,
-// copy or use the software.
-//
-//
-// License Agreement
-// For Open Source Computer Vision Library
-//
-// Copyright (C) 2010-2012, Institute Of Software Chinese Academy Of Science, all rights reserved.
-// Copyright (C) 2010-2012, Advanced Micro Devices, Inc., all rights reserved.
-// Third party copyrights are property of their respective owners.
-//
-// @Authors
-// Jia Haipeng, jiahaipeng95@gmail.com
-//
-// Redistribution and use in source and binary forms, with or without modification,
-// are permitted provided that the following conditions are met:
-//
-// * Redistribution's of source code must retain the above copyright notice,
-// this list of conditions and the following disclaimer.
-//
-// * Redistribution's in binary form must reproduce the above copyright notice,
-// this list of conditions and the following disclaimer in the documentation
-// and/or other oclMaterials provided with the distribution.
-//
-// * The name of the copyright holders may not be used to endorse or promote products
-// derived from this software without specific prior written permission.
-//
-// This software is provided by the copyright holders and contributors as is and
-// any express or implied warranties, including, but not limited to, the implied
-// warranties of merchantability and fitness for a particular purpose are disclaimed.
-// In no event shall the Intel Corporation or contributors be liable for any direct,
-// indirect, incidental, special, exemplary, or consequential damages
-// (including, but not limited to, procurement of substitute goods or services;
-// loss of use, data, or profits; or business interruption) however caused
-// and on any theory of liability, whether in contract, strict liability,
-// or tort (including negligence or otherwise) arising in any way out of
-// the use of this software, even if advised of the possibility of such damage.
-//
-//M*/
-#if defined (DOUBLE_SUPPORT)
-#ifdef cl_khr_fp64
-#pragma OPENCL EXTENSION cl_khr_fp64:enable
-#elif defined (cl_amd_fp64)
-#pragma OPENCL EXTENSION cl_amd_fp64:enable
-#endif
-#endif
-/**************************************sub with scalar without mask**************************************/
-__kernel void arithm_s_sub_C1_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 2;
-
- #define dst_align (dst_offset & 3)
- int src1_index = mad24(y, src1_step, x + src1_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + x & (int)0xfffffffc);
-
- uchar4 src1_data = vload4(0, src1 + src1_index);
- int4 src2_data = (int4)(src2.x, src2.x, src2.x, src2.x);
-
- uchar4 data = *((__global uchar4 *)(dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) - src2_data;
- tmp = isMatSubScalar ? tmp : -tmp;
- uchar4 tmp_data = convert_uchar4_sat(tmp);
-
- data.x = ((dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : data.x;
- data.y = ((dst_index + 1 >= dst_start) && (dst_index + 1 < dst_end)) ? tmp_data.y : data.y;
- data.z = ((dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.z : data.z;
- data.w = ((dst_index + 3 >= dst_start) && (dst_index + 3 < dst_end)) ? tmp_data.w : data.w;
-
- *((__global uchar4 *)(dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C1_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 1;
-
- #define dst_align ((dst_offset >> 1) & 1)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
-
- ushort2 src1_data = vload2(0, (__global ushort *)((__global char *)src1 + src1_index));
- int2 src2_data = (int2)(src2.x, src2.x);
-
- ushort2 data = *((__global ushort2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) - src2_data;
- tmp = isMatSubScalar ? tmp : -tmp;
- ushort2 tmp_data = convert_ushort2_sat(tmp);
-
- data.x = (dst_index + 0 >= dst_start) ? tmp_data.x : data.x;
- data.y = (dst_index + 2 < dst_end ) ? tmp_data.y : data.y;
-
- *((__global ushort2 *)((__global uchar *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C1_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 1;
-
- #define dst_align ((dst_offset >> 1) & 1)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
-
- short2 src1_data = vload2(0, (__global short *)((__global char *)src1 + src1_index));
- int2 src2_data = (int2)(src2.x, src2.x);
- short2 data = *((__global short2 *)((__global uchar *)dst + dst_index));
-
- int2 tmp = convert_int2_sat(src1_data) - src2_data;
- tmp = isMatSubScalar ? tmp : -tmp;
- short2 tmp_data = convert_short2_sat(tmp);
-
- data.x = (dst_index + 0 >= dst_start) ? tmp_data.x : data.x;
- data.y = (dst_index + 2 < dst_end ) ? tmp_data.y : data.y;
-
- *((__global short2 *)((__global uchar *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C1_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- int src_data1 = *((__global int *)((__global char *)src1 + src1_index));
- int src_data2 = src2.x;
-
- long tmp = (long)src_data1 - (long)src_data2;
- tmp = isMatSubScalar ? tmp : -tmp;
- int data = convert_int_sat(tmp);
-
- *((__global int *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C1_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *dst, int dst_step, int dst_offset,
- float4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- float src_data1 = *((__global float *)((__global char *)src1 + src1_index));
- float src_data2 = src2.x;
-
- float tmp = src_data1 - src_data2;
- tmp = isMatSubScalar ? tmp : -tmp;
-
- *((__global float *)((__global char *)dst + dst_index)) = tmp;
- }
-}
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_s_sub_C1_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *dst, int dst_step, int dst_offset,
- double4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- double src_data1 = *((__global double *)((__global char *)src1 + src1_index));
- double src2_data = src2.x;
-
- double data = src_data1 - src2_data;
- data = isMatSubScalar ? data : -data;
-
- *((__global double *)((__global char *)dst + dst_index)) = data;
- }
-}
-#endif
-
-__kernel void arithm_s_sub_C2_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 1;
-
- #define dst_align ((dst_offset >> 1) & 1)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
-
- uchar4 src1_data = vload4(0, src1 + src1_index);
- int4 src2_data = (int4)(src2.x, src2.y, src2.x, src2.y);
-
- uchar4 data = *((__global uchar4 *)(dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) - src2_data;
- tmp = isMatSubScalar ? tmp : -tmp;
- uchar4 tmp_data = convert_uchar4_sat(tmp);
-
- data.xy = (dst_index + 0 >= dst_start) ? tmp_data.xy : data.xy;
- data.zw = (dst_index + 2 < dst_end ) ? tmp_data.zw : data.zw;
-
- *((__global uchar4 *)(dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C2_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- ushort2 src_data1 = *((__global ushort2 *)((__global char *)src1 + src1_index));
- int2 src_data2 = (int2)(src2.x, src2.y);
- ushort2 dst_data = *((__global ushort2 *)((__global char *)dst + dst_index));
-
- int2 tmp = convert_int2_sat(src_data1) - src_data2;
- tmp = isMatSubScalar ? tmp : -tmp;
- ushort2 data = convert_ushort2_sat(tmp);
-
- *((__global ushort2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C2_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- short2 src_data1 = *((__global short2 *)((__global char *)src1 + src1_index));
- int2 src_data2 = (int2)(src2.x, src2.y);
- short2 dst_data = *((__global short2 *)((__global char *)dst + dst_index));
-
- int2 tmp = convert_int2_sat(src_data1) - src_data2;
- tmp = isMatSubScalar ? tmp : -tmp;
- short2 data = convert_short2_sat(tmp);
-
- *((__global short2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C2_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- int2 src_data1 = *((__global int2 *)((__global char *)src1 + src1_index));
- int2 src_data2 = (int2)(src2.x, src2.y);
- int2 dst_data = *((__global int2 *)((__global char *)dst + dst_index));
-
- long2 tmp = convert_long2_sat(src_data1) - convert_long2_sat(src_data2);
- tmp = isMatSubScalar ? tmp : -tmp;
- int2 data = convert_int2_sat(tmp);
-
- *((__global int2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C2_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *dst, int dst_step, int dst_offset,
- float4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- float2 src_data1 = *((__global float2 *)((__global char *)src1 + src1_index));
- float2 src_data2 = (float2)(src2.x, src2.y);
- float2 dst_data = *((__global float2 *)((__global char *)dst + dst_index));
-
- float2 tmp = src_data1 - src_data2;
- tmp = isMatSubScalar ? tmp : -tmp;
-
- *((__global float2 *)((__global char *)dst + dst_index)) = tmp;
- }
-}
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_s_sub_C2_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *dst, int dst_step, int dst_offset,
- double4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
-
- double2 src_data1 = *((__global double2 *)((__global char *)src1 + src1_index));
- double2 src_data2 = (double2)(src2.x, src2.y);
- double2 dst_data = *((__global double2 *)((__global char *)dst + dst_index));
-
- double2 data = src_data1 - src_data2;
- data = isMatSubScalar ? data : -data;
-
- *((__global double2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-#endif
-
-__kernel void arithm_s_sub_C4_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar4 src_data1 = *((__global uchar4 *)(src1 + src1_index));
-
- int4 tmp = convert_int4_sat(src_data1) - src2;
- tmp = isMatSubScalar ? tmp : -tmp;
- uchar4 data = convert_uchar4_sat(tmp);
-
- *((__global uchar4 *)(dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C4_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- ushort4 src_data1 = *((__global ushort4 *)((__global char *)src1 + src1_index));
-
- int4 tmp = convert_int4_sat(src_data1) - src2;
- tmp = isMatSubScalar ? tmp : -tmp;
- ushort4 data = convert_ushort4_sat(tmp);
-
- *((__global ushort4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C4_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- short4 src_data1 = *((__global short4 *)((__global char *)src1 + src1_index));
-
- int4 tmp = convert_int4_sat(src_data1) - src2;
- tmp = isMatSubScalar ? tmp : -tmp;
- short4 data = convert_short4_sat(tmp);
-
- *((__global short4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C4_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *dst, int dst_step, int dst_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
-
- int4 src_data1 = *((__global int4 *)((__global char *)src1 + src1_index));
-
- long4 tmp = convert_long4_sat(src_data1) - convert_long4_sat(src2);
- tmp = isMatSubScalar ? tmp : -tmp;
- int4 data = convert_int4_sat(tmp);
-
- *((__global int4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_C4_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *dst, int dst_step, int dst_offset,
- float4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
-
- float4 src_data1 = *((__global float4 *)((__global char *)src1 + src1_index));
-
- float4 tmp = src_data1 - src2;
- tmp = isMatSubScalar ? tmp : -tmp;
-
- *((__global float4 *)((__global char *)dst + dst_index)) = tmp;
- }
-}
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_s_sub_C4_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *dst, int dst_step, int dst_offset,
- double4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 5) + src1_offset);
- int dst_index = mad24(y, dst_step, (x << 5) + dst_offset);
-
- double4 src_data1 = *((__global double4 *)((__global char *)src1 + src1_index));
-
- double4 data = src_data1 - src2;
- data = isMatSubScalar ? data : -data;
-
- *((__global double4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-#endif
+++ /dev/null
-/*M///////////////////////////////////////////////////////////////////////////////////////
-//
-// IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
-//
-// By downloading, copying, installing or using the software you agree to this license.
-// If you do not agree to this license, do not download, install,
-// copy or use the software.
-//
-//
-// License Agreement
-// For Open Source Computer Vision Library
-//
-// Copyright (C) 2010-2012, Institute Of Software Chinese Academy Of Science, all rights reserved.
-// Copyright (C) 2010-2012, Advanced Micro Devices, Inc., all rights reserved.
-// Third party copyrights are property of their respective owners.
-//
-// @Authors
-// Jia Haipeng, jiahaipeng95@gmail.com
-//
-// Redistribution and use in source and binary forms, with or without modification,
-// are permitted provided that the following conditions are met:
-//
-// * Redistribution's of source code must retain the above copyright notice,
-// this list of conditions and the following disclaimer.
-//
-// * Redistribution's in binary form must reproduce the above copyright notice,
-// this list of conditions and the following disclaimer in the documentation
-// and/or other GpuMaterials provided with the distribution.
-//
-// * The name of the copyright holders may not be used to endorse or promote products
-// derived from this software without specific prior written permission.
-//
-// This software is provided by the copyright holders and contributors as is and
-// any express or implied warranties, including, but not limited to, the implied
-// warranties of merchantability and fitness for a particular purpose are disclaimed.
-// In no event shall the Intel Corporation or contributors be liable for any direct,
-// indirect, incidental, special, exemplary, or consequential damages
-// (including, but not limited to, procurement of substitute goods or services;
-// loss of use, data, or profits; or business interruption) however caused
-// and on any theory of liability, whether in contract, strict liability,
-// or tort (including negligence or otherwise) arising in any way out of
-// the use of this software, even if advised of the possibility of such damage.
-//
-//M*/
-
-#if defined (DOUBLE_SUPPORT)
-#ifdef cl_khr_fp64
-#pragma OPENCL EXTENSION cl_khr_fp64:enable
-#elif defined (cl_amd_fp64)
-#pragma OPENCL EXTENSION cl_amd_fp64:enable
-#endif
-#endif
-
-/**************************************sub with scalar with mask**************************************/
-__kernel void arithm_s_sub_with_mask_C1_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 2;
-
- #define dst_align (dst_offset & 3)
- int src1_index = mad24(y, src1_step, x + src1_offset - dst_align);
- int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + x & (int)0xfffffffc);
-
- uchar4 src1_data = vload4(0, src1 + src1_index);
- int4 src2_data = (int4)(src2.x, src2.x, src2.x, src2.x);
- uchar4 mask_data = vload4(0, mask + mask_index);
-
- uchar4 data = *((__global uchar4 *)(dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) - src2_data;
- tmp = isMatSubScalar ? tmp : -tmp;
- uchar4 tmp_data = convert_uchar4_sat(tmp);
-
- data.x = ((mask_data.x) && (dst_index + 0 >= dst_start) && (dst_index + 0 < dst_end)) ? tmp_data.x : data.x;
- data.y = ((mask_data.y) && (dst_index + 1 >= dst_start) && (dst_index + 1 < dst_end)) ? tmp_data.y : data.y;
- data.z = ((mask_data.z) && (dst_index + 2 >= dst_start) && (dst_index + 2 < dst_end)) ? tmp_data.z : data.z;
- data.w = ((mask_data.w) && (dst_index + 3 >= dst_start) && (dst_index + 3 < dst_end)) ? tmp_data.w : data.w;
-
- *((__global uchar4 *)(dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C1_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 1;
-
- #define dst_align ((dst_offset >> 1) & 1)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
- int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
-
- ushort2 src1_data = vload2(0, (__global ushort *)((__global char *)src1 + src1_index));
- int2 src2_data = (int2)(src2.x, src2.x);
- uchar2 mask_data = vload2(0, mask + mask_index);
-
- ushort2 data = *((__global ushort2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) - src2_data;
- tmp = isMatSubScalar ? tmp : -tmp;
- ushort2 tmp_data = convert_ushort2_sat(tmp);
-
- data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
- data.y = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.y : data.y;
-
- *((__global ushort2 *)((__global uchar *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C1_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 1;
-
- #define dst_align ((dst_offset >> 1) & 1)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
- int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
-
- short2 src1_data = vload2(0, (__global short *)((__global char *)src1 + src1_index));
- int2 src2_data = (int2)(src2.x, src2.x);
- uchar2 mask_data = vload2(0, mask + mask_index);
-
- short2 data = *((__global short2 *)((__global uchar *)dst + dst_index));
- int2 tmp = convert_int2_sat(src1_data) - src2_data;
- tmp = isMatSubScalar ? tmp : -tmp;
- short2 tmp_data = convert_short2_sat(tmp);
-
- data.x = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.x : data.x;
- data.y = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.y : data.y;
-
- *((__global short2 *)((__global uchar *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C1_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- int src_data1 = *((__global int *)((__global char *)src1 + src1_index));
- int src_data2 = src2.x;
- int dst_data = *((__global int *)((__global char *)dst + dst_index));
-
- long tmp = (long)src_data1 - (long)src_data2;
- tmp = isMatSubScalar ? tmp : - tmp;
- int data = convert_int_sat(tmp);
- data = mask_data ? data : dst_data;
-
- *((__global int *)((__global char *)dst + dst_index)) = data;
- }
-}
-
-__kernel void arithm_s_sub_with_mask_C1_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- float4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- float src_data1 = *((__global float *)((__global char *)src1 + src1_index));
- float src_data2 = src2.x;
- float dst_data = *((__global float *)((__global char *)dst + dst_index));
-
- float data = src_data1 - src_data2;
- data = isMatSubScalar ? data : -data;
- data = mask_data ? data : dst_data;
-
- *((__global float *)((__global char *)dst + dst_index)) = data;
- }
-}
-
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_s_sub_with_mask_C1_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- double4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- double src_data1 = *((__global double *)((__global char *)src1 + src1_index));
- double src_data2 = src2.x;
- double dst_data = *((__global double *)((__global char *)dst + dst_index));
-
- double data = src_data1 - src_data2;
- data = isMatSubScalar ? data : -data;
- data = mask_data ? data : dst_data;
-
- *((__global double *)((__global char *)dst + dst_index)) = data;
- }
-}
-#endif
-__kernel void arithm_s_sub_with_mask_C2_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- x = x << 1;
-
- #define dst_align ((dst_offset >> 1) & 1)
- int src1_index = mad24(y, src1_step, (x << 1) + src1_offset - (dst_align << 1));
- int mask_index = mad24(y, mask_step, x + mask_offset - dst_align);
-
- int dst_start = mad24(y, dst_step, dst_offset);
- int dst_end = mad24(y, dst_step, dst_offset + dst_step1);
- int dst_index = mad24(y, dst_step, dst_offset + (x << 1) & (int)0xfffffffc);
-
- uchar4 src1_data = vload4(0, src1 + src1_index);
- int4 src2_data = (int4)(src2.x, src2.y, src2.x, src2.y);
- uchar2 mask_data = vload2(0, mask + mask_index);
-
- uchar4 data = *((__global uchar4 *)(dst + dst_index));
- int4 tmp = convert_int4_sat(src1_data) - src2_data;
- tmp = isMatSubScalar ? tmp : -tmp;
- uchar4 tmp_data = convert_uchar4_sat(tmp);
-
- data.xy = ((mask_data.x) && (dst_index + 0 >= dst_start)) ? tmp_data.xy : data.xy;
- data.zw = ((mask_data.y) && (dst_index + 2 < dst_end )) ? tmp_data.zw : data.zw;
-
- *((__global uchar4 *)(dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C2_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- ushort2 src_data1 = *((__global ushort2 *)((__global char *)src1 + src1_index));
- int2 src_data2 = (int2)(src2.x, src2.y);
- ushort2 dst_data = *((__global ushort2 *)((__global char *)dst + dst_index));
-
- int2 tmp = convert_int2_sat(src_data1) - src_data2;
- tmp = isMatSubScalar ? tmp : -tmp;
- ushort2 data = convert_ushort2_sat(tmp);
- data = mask_data ? data : dst_data;
-
- *((__global ushort2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C2_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- short2 src_data1 = *((__global short2 *)((__global char *)src1 + src1_index));
- int2 src_data2 = (int2)(src2.x, src2.y);
- short2 dst_data = *((__global short2 *)((__global char *)dst + dst_index));
-
- int2 tmp = convert_int2_sat(src_data1) - src_data2;
- tmp = isMatSubScalar ? tmp : -tmp;
- short2 data = convert_short2_sat(tmp);
- data = mask_data ? data : dst_data;
-
- *((__global short2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C2_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- int2 src_data1 = *((__global int2 *)((__global char *)src1 + src1_index));
- int2 src_data2 = (int2)(src2.x, src2.y);
- int2 dst_data = *((__global int2 *)((__global char *)dst + dst_index));
-
- long2 tmp = convert_long2_sat(src_data1) - convert_long2_sat(src_data2);
- tmp = isMatSubScalar ? tmp : -tmp;
- int2 data = convert_int2_sat(tmp);
- data = mask_data ? data : dst_data;
-
- *((__global int2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C2_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- float4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- float2 src_data1 = *((__global float2 *)((__global char *)src1 + src1_index));
- float2 src_data2 = (float2)(src2.x, src2.y);
- float2 dst_data = *((__global float2 *)((__global char *)dst + dst_index));
-
- float2 data = src_data1 - src_data2;
- data = isMatSubScalar ? data : -data;
- data = mask_data ? data : dst_data;
-
- *((__global float2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_s_sub_with_mask_C2_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- double4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- double2 src_data1 = *((__global double2 *)((__global char *)src1 + src1_index));
- double2 src_data2 = (double2)(src2.x, src2.y);
- double2 dst_data = *((__global double2 *)((__global char *)dst + dst_index));
-
- double2 data = src_data1 - src_data2;
- data = isMatSubScalar ? data : -data;
- data = mask_data ? data : dst_data;
-
- *((__global double2 *)((__global char *)dst + dst_index)) = data;
- }
-}
-#endif
-
-__kernel void arithm_s_sub_with_mask_C4_D0 (__global uchar *src1, int src1_step, int src1_offset,
- __global uchar *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 2) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 2) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- uchar4 src_data1 = *((__global uchar4 *)(src1 + src1_index));
- uchar4 dst_data = *((__global uchar4 *)(dst + dst_index));
-
- int4 tmp = convert_int4_sat(src_data1) - src2;
- tmp = isMatSubScalar ? tmp : -tmp;
- uchar4 data = convert_uchar4_sat(tmp);
-
- data = mask_data ? data : dst_data;
-
- *((__global uchar4 *)(dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C4_D2 (__global ushort *src1, int src1_step, int src1_offset,
- __global ushort *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- ushort4 src_data1 = *((__global ushort4 *)((__global char *)src1 + src1_index));
- ushort4 dst_data = *((__global ushort4 *)((__global char *)dst + dst_index));
-
- int4 tmp = convert_int4_sat(src_data1) - src2;
- tmp = isMatSubScalar ? tmp : -tmp;
- ushort4 data = convert_ushort4_sat(tmp);
-
- data = mask_data ? data : dst_data;
-
- *((__global ushort4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C4_D3 (__global short *src1, int src1_step, int src1_offset,
- __global short *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 3) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 3) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- short4 src_data1 = *((__global short4 *)((__global char *)src1 + src1_index));
- short4 dst_data = *((__global short4 *)((__global char *)dst + dst_index));
-
- int4 tmp = convert_int4_sat(src_data1) - src2;
- tmp = isMatSubScalar ? tmp : -tmp;
- short4 data = convert_short4_sat(tmp);
-
- data = mask_data ? data : dst_data;
-
- *((__global short4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C4_D4 (__global int *src1, int src1_step, int src1_offset,
- __global int *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- int4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- int4 src_data1 = *((__global int4 *)((__global char *)src1 + src1_index));
- int4 dst_data = *((__global int4 *)((__global char *)dst + dst_index));
-
- long4 tmp = convert_long4_sat(src_data1) - convert_long4_sat(src2);
- tmp = isMatSubScalar ? tmp : -tmp;
- int4 data = convert_int4_sat(tmp);
-
- data = mask_data ? data : dst_data;
-
- *((__global int4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-__kernel void arithm_s_sub_with_mask_C4_D5 (__global float *src1, int src1_step, int src1_offset,
- __global float *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- float4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 4) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 4) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- float4 src_data1 = *((__global float4 *)((__global char *)src1 + src1_index));
- float4 dst_data = *((__global float4 *)((__global char *)dst + dst_index));
-
- float4 data = src_data1 - src2;
- data = isMatSubScalar ? data : -data;
-
- data = mask_data ? data : dst_data;
-
- *((__global float4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-
-#if defined (DOUBLE_SUPPORT)
-__kernel void arithm_s_sub_with_mask_C4_D6 (__global double *src1, int src1_step, int src1_offset,
- __global double *dst, int dst_step, int dst_offset,
- __global uchar *mask, int mask_step, int mask_offset,
- double4 src2, int rows, int cols, int dst_step1, int isMatSubScalar)
-{
-
- int x = get_global_id(0);
- int y = get_global_id(1);
-
- if (x < cols && y < rows)
- {
- int src1_index = mad24(y, src1_step, (x << 5) + src1_offset);
- int mask_index = mad24(y, mask_step, x + mask_offset);
- int dst_index = mad24(y, dst_step, (x << 5) + dst_offset);
-
- uchar mask_data = *(mask + mask_index);
-
- double4 src_data1 = *((__global double4 *)((__global char *)src1 + src1_index));
- double4 dst_data = *((__global double4 *)((__global char *)dst + dst_index));
-
- double4 data = src_data1 - src2;
- data = isMatSubScalar ? data : -data;
- data = mask_data ? data : dst_data;
-
- *((__global double4 *)((__global char *)dst + dst_index)) = data;
- }
-}
-#endif