1 // This file is part of OpenCV project.
2 // It is subject to the license terms in the LICENSE file found in the top-level directory
3 // of this distribution and at http://opencv.org/license.html
7 #include "opencl_kernels_core.hpp"
9 namespace cv { namespace hal {
12 template<typename T> struct VMerge2;
13 template<typename T> struct VMerge3;
14 template<typename T> struct VMerge4;
16 #define MERGE2_KERNEL_TEMPLATE(name, data_type, reg_type, load_func, store_func) \
18 struct name<data_type>{ \
19 void operator()(const data_type* src0, const data_type* src1, \
22 r.val[0] = load_func(src0); \
23 r.val[1] = load_func(src1); \
28 #define MERGE3_KERNEL_TEMPLATE(name, data_type, reg_type, load_func, store_func) \
30 struct name<data_type>{ \
31 void operator()(const data_type* src0, const data_type* src1, \
32 const data_type* src2, data_type* dst){ \
34 r.val[0] = load_func(src0); \
35 r.val[1] = load_func(src1); \
36 r.val[2] = load_func(src2); \
41 #define MERGE4_KERNEL_TEMPLATE(name, data_type, reg_type, load_func, store_func) \
43 struct name<data_type>{ \
44 void operator()(const data_type* src0, const data_type* src1, \
45 const data_type* src2, const data_type* src3, \
48 r.val[0] = load_func(src0); \
49 r.val[1] = load_func(src1); \
50 r.val[2] = load_func(src2); \
51 r.val[3] = load_func(src3); \
56 MERGE2_KERNEL_TEMPLATE(VMerge2, uchar , uint8x16x2_t, vld1q_u8 , vst2q_u8 );
57 MERGE2_KERNEL_TEMPLATE(VMerge2, ushort, uint16x8x2_t, vld1q_u16, vst2q_u16);
58 MERGE2_KERNEL_TEMPLATE(VMerge2, int , int32x4x2_t, vld1q_s32, vst2q_s32);
59 MERGE2_KERNEL_TEMPLATE(VMerge2, int64 , int64x1x2_t, vld1_s64 , vst2_s64 );
61 MERGE3_KERNEL_TEMPLATE(VMerge3, uchar , uint8x16x3_t, vld1q_u8 , vst3q_u8 );
62 MERGE3_KERNEL_TEMPLATE(VMerge3, ushort, uint16x8x3_t, vld1q_u16, vst3q_u16);
63 MERGE3_KERNEL_TEMPLATE(VMerge3, int , int32x4x3_t, vld1q_s32, vst3q_s32);
64 MERGE3_KERNEL_TEMPLATE(VMerge3, int64 , int64x1x3_t, vld1_s64 , vst3_s64 );
66 MERGE4_KERNEL_TEMPLATE(VMerge4, uchar , uint8x16x4_t, vld1q_u8 , vst4q_u8 );
67 MERGE4_KERNEL_TEMPLATE(VMerge4, ushort, uint16x8x4_t, vld1q_u16, vst4q_u16);
68 MERGE4_KERNEL_TEMPLATE(VMerge4, int , int32x4x4_t, vld1q_s32, vst4q_s32);
69 MERGE4_KERNEL_TEMPLATE(VMerge4, int64 , int64x1x4_t, vld1_s64 , vst4_s64 );
76 VMerge2() : support(false) { }
77 void operator()(const T *, const T *, T *) const { }
85 VMerge3() : support(false) { }
86 void operator()(const T *, const T *, const T *, T *) const { }
94 VMerge4() : support(false) { }
95 void operator()(const T *, const T *, const T *, const T *, T *) const { }
100 #define MERGE2_KERNEL_TEMPLATE(data_type, reg_type, cast_type, _mm_interleave, flavor, se) \
102 struct VMerge2<data_type> \
106 ELEMS_IN_VEC = 16 / sizeof(data_type) \
111 support = checkHardwareSupport(se); \
114 void operator()(const data_type * src0, const data_type * src1, \
115 data_type * dst) const \
117 reg_type v_src0 = _mm_loadu_##flavor((const cast_type *)(src0)); \
118 reg_type v_src1 = _mm_loadu_##flavor((const cast_type *)(src0 + ELEMS_IN_VEC)); \
119 reg_type v_src2 = _mm_loadu_##flavor((const cast_type *)(src1)); \
120 reg_type v_src3 = _mm_loadu_##flavor((const cast_type *)(src1 + ELEMS_IN_VEC)); \
122 _mm_interleave(v_src0, v_src1, v_src2, v_src3); \
124 _mm_storeu_##flavor((cast_type *)(dst), v_src0); \
125 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC), v_src1); \
126 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 2), v_src2); \
127 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 3), v_src3); \
133 #define MERGE3_KERNEL_TEMPLATE(data_type, reg_type, cast_type, _mm_interleave, flavor, se) \
135 struct VMerge3<data_type> \
139 ELEMS_IN_VEC = 16 / sizeof(data_type) \
144 support = checkHardwareSupport(se); \
147 void operator()(const data_type * src0, const data_type * src1, const data_type * src2,\
148 data_type * dst) const \
150 reg_type v_src0 = _mm_loadu_##flavor((const cast_type *)(src0)); \
151 reg_type v_src1 = _mm_loadu_##flavor((const cast_type *)(src0 + ELEMS_IN_VEC)); \
152 reg_type v_src2 = _mm_loadu_##flavor((const cast_type *)(src1)); \
153 reg_type v_src3 = _mm_loadu_##flavor((const cast_type *)(src1 + ELEMS_IN_VEC)); \
154 reg_type v_src4 = _mm_loadu_##flavor((const cast_type *)(src2)); \
155 reg_type v_src5 = _mm_loadu_##flavor((const cast_type *)(src2 + ELEMS_IN_VEC)); \
157 _mm_interleave(v_src0, v_src1, v_src2, \
158 v_src3, v_src4, v_src5); \
160 _mm_storeu_##flavor((cast_type *)(dst), v_src0); \
161 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC), v_src1); \
162 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 2), v_src2); \
163 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 3), v_src3); \
164 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 4), v_src4); \
165 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 5), v_src5); \
171 #define MERGE4_KERNEL_TEMPLATE(data_type, reg_type, cast_type, _mm_interleave, flavor, se) \
173 struct VMerge4<data_type> \
177 ELEMS_IN_VEC = 16 / sizeof(data_type) \
182 support = checkHardwareSupport(se); \
185 void operator()(const data_type * src0, const data_type * src1, \
186 const data_type * src2, const data_type * src3, \
187 data_type * dst) const \
189 reg_type v_src0 = _mm_loadu_##flavor((const cast_type *)(src0)); \
190 reg_type v_src1 = _mm_loadu_##flavor((const cast_type *)(src0 + ELEMS_IN_VEC)); \
191 reg_type v_src2 = _mm_loadu_##flavor((const cast_type *)(src1)); \
192 reg_type v_src3 = _mm_loadu_##flavor((const cast_type *)(src1 + ELEMS_IN_VEC)); \
193 reg_type v_src4 = _mm_loadu_##flavor((const cast_type *)(src2)); \
194 reg_type v_src5 = _mm_loadu_##flavor((const cast_type *)(src2 + ELEMS_IN_VEC)); \
195 reg_type v_src6 = _mm_loadu_##flavor((const cast_type *)(src3)); \
196 reg_type v_src7 = _mm_loadu_##flavor((const cast_type *)(src3 + ELEMS_IN_VEC)); \
198 _mm_interleave(v_src0, v_src1, v_src2, v_src3, \
199 v_src4, v_src5, v_src6, v_src7); \
201 _mm_storeu_##flavor((cast_type *)(dst), v_src0); \
202 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC), v_src1); \
203 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 2), v_src2); \
204 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 3), v_src3); \
205 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 4), v_src4); \
206 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 5), v_src5); \
207 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 6), v_src6); \
208 _mm_storeu_##flavor((cast_type *)(dst + ELEMS_IN_VEC * 7), v_src7); \
214 MERGE2_KERNEL_TEMPLATE( uchar, __m128i, __m128i, _mm_interleave_epi8, si128, CV_CPU_SSE2);
215 MERGE3_KERNEL_TEMPLATE( uchar, __m128i, __m128i, _mm_interleave_epi8, si128, CV_CPU_SSE2);
216 MERGE4_KERNEL_TEMPLATE( uchar, __m128i, __m128i, _mm_interleave_epi8, si128, CV_CPU_SSE2);
219 MERGE2_KERNEL_TEMPLATE(ushort, __m128i, __m128i, _mm_interleave_epi16, si128, CV_CPU_SSE4_1);
220 MERGE3_KERNEL_TEMPLATE(ushort, __m128i, __m128i, _mm_interleave_epi16, si128, CV_CPU_SSE4_1);
221 MERGE4_KERNEL_TEMPLATE(ushort, __m128i, __m128i, _mm_interleave_epi16, si128, CV_CPU_SSE4_1);
224 MERGE2_KERNEL_TEMPLATE( int, __m128, float, _mm_interleave_ps, ps, CV_CPU_SSE2);
225 MERGE3_KERNEL_TEMPLATE( int, __m128, float, _mm_interleave_ps, ps, CV_CPU_SSE2);
226 MERGE4_KERNEL_TEMPLATE( int, __m128, float, _mm_interleave_ps, ps, CV_CPU_SSE2);
230 template<typename T> static void
231 merge_( const T** src, T* dst, int len, int cn )
233 int k = cn % 4 ? cn % 4 : 4;
237 const T* src0 = src[0];
238 for( i = j = 0; i < len; i++, j += cn )
243 const T *src0 = src[0], *src1 = src[1];
248 int inc_i = (sizeof(T) == 8)? 1: 16/sizeof(T);
249 int inc_j = 2 * inc_i;
252 for( ; i < len - inc_i; i += inc_i, j += inc_j)
253 vmerge(src0 + i, src1 + i, dst + j);
258 int inc_i = 32/sizeof(T);
259 int inc_j = 2 * inc_i;
263 for( ; i < len - inc_i; i += inc_i, j += inc_j)
264 vmerge(src0 + i, src1 + i, dst + j);
267 for( ; i < len; i++, j += cn )
275 const T *src0 = src[0], *src1 = src[1], *src2 = src[2];
280 int inc_i = (sizeof(T) == 8)? 1: 16/sizeof(T);
281 int inc_j = 3 * inc_i;
284 for( ; i < len - inc_i; i += inc_i, j += inc_j)
285 vmerge(src0 + i, src1 + i, src2 + i, dst + j);
290 int inc_i = 32/sizeof(T);
291 int inc_j = 3 * inc_i;
295 for( ; i < len - inc_i; i += inc_i, j += inc_j)
296 vmerge(src0 + i, src1 + i, src2 + i, dst + j);
299 for( ; i < len; i++, j += cn )
308 const T *src0 = src[0], *src1 = src[1], *src2 = src[2], *src3 = src[3];
313 int inc_i = (sizeof(T) == 8)? 1: 16/sizeof(T);
314 int inc_j = 4 * inc_i;
317 for( ; i < len - inc_i; i += inc_i, j += inc_j)
318 vmerge(src0 + i, src1 + i, src2 + i, src3 + i, dst + j);
323 int inc_i = 32/sizeof(T);
324 int inc_j = 4 * inc_i;
328 for( ; i < len - inc_i; i += inc_i, j += inc_j)
329 vmerge(src0 + i, src1 + i, src2 + i, src3 + i, dst + j);
332 for( ; i < len; i++, j += cn )
334 dst[j] = src0[i]; dst[j+1] = src1[i];
335 dst[j+2] = src2[i]; dst[j+3] = src3[i];
339 for( ; k < cn; k += 4 )
341 const T *src0 = src[k], *src1 = src[k+1], *src2 = src[k+2], *src3 = src[k+3];
342 for( i = 0, j = k; i < len; i++, j += cn )
344 dst[j] = src0[i]; dst[j+1] = src1[i];
345 dst[j+2] = src2[i]; dst[j+3] = src3[i];
351 void merge8u(const uchar** src, uchar* dst, int len, int cn )
353 CALL_HAL(merge8u, cv_hal_merge8u, src, dst, len, cn)
354 merge_(src, dst, len, cn);
357 void merge16u(const ushort** src, ushort* dst, int len, int cn )
359 CALL_HAL(merge16u, cv_hal_merge16u, src, dst, len, cn)
360 merge_(src, dst, len, cn);
363 void merge32s(const int** src, int* dst, int len, int cn )
365 CALL_HAL(merge32s, cv_hal_merge32s, src, dst, len, cn)
366 merge_(src, dst, len, cn);
369 void merge64s(const int64** src, int64* dst, int len, int cn )
371 CALL_HAL(merge64s, cv_hal_merge64s, src, dst, len, cn)
372 merge_(src, dst, len, cn);
378 typedef void (*MergeFunc)(const uchar** src, uchar* dst, int len, int cn);
380 static MergeFunc getMergeFunc(int depth)
382 static MergeFunc mergeTab[] =
384 (MergeFunc)GET_OPTIMIZED(cv::hal::merge8u), (MergeFunc)GET_OPTIMIZED(cv::hal::merge8u), (MergeFunc)GET_OPTIMIZED(cv::hal::merge16u), (MergeFunc)GET_OPTIMIZED(cv::hal::merge16u),
385 (MergeFunc)GET_OPTIMIZED(cv::hal::merge32s), (MergeFunc)GET_OPTIMIZED(cv::hal::merge32s), (MergeFunc)GET_OPTIMIZED(cv::hal::merge64s), 0
388 return mergeTab[depth];
394 static bool ipp_merge(const Mat* mv, Mat& dst, int channels)
397 CV_INSTRUMENT_REGION_IPP()
399 if(channels != 3 && channels != 4)
404 IppiSize size = ippiSize(mv[0].size());
405 const void *srcPtrs[4] = {NULL};
406 size_t srcStep = mv[0].step;
407 for(int i = 0; i < channels; i++)
409 srcPtrs[i] = mv[i].ptr();
410 if(srcStep != mv[i].step)
414 return CV_INSTRUMENT_FUN_IPP(llwiCopyMerge, srcPtrs, (int)srcStep, dst.ptr(), (int)dst.step, size, (int)mv[0].elemSize1(), channels, 0) >= 0;
418 const Mat *arrays[5] = {NULL};
419 uchar *ptrs[5] = {NULL};
422 for(int i = 1; i < channels; i++)
424 arrays[i] = &mv[i-1];
427 NAryMatIterator it(arrays, ptrs);
428 IppiSize size = { (int)it.size, 1 };
430 for( size_t i = 0; i < it.nplanes; i++, ++it )
432 if(CV_INSTRUMENT_FUN_IPP(llwiCopyMerge, (const void**)&ptrs[1], 0, ptrs[0], 0, size, (int)mv[0].elemSize1(), channels, 0) < 0)
438 CV_UNUSED(dst); CV_UNUSED(mv); CV_UNUSED(channels);
445 void cv::merge(const Mat* mv, size_t n, OutputArray _dst)
447 CV_INSTRUMENT_REGION()
449 CV_Assert( mv && n > 0 );
451 int depth = mv[0].depth();
456 for( i = 0; i < n; i++ )
458 CV_Assert(mv[i].size == mv[0].size && mv[i].depth() == depth);
459 allch1 = allch1 && mv[i].channels() == 1;
460 cn += mv[i].channels();
463 CV_Assert( 0 < cn && cn <= CV_CN_MAX );
464 _dst.create(mv[0].dims, mv[0].size, CV_MAKETYPE(depth, cn));
465 Mat dst = _dst.getMat();
473 CV_IPP_RUN_FAST(ipp_merge(mv, dst, (int)n));
477 AutoBuffer<int> pairs(cn*2);
480 for( i = 0, j = 0; i < n; i++, j += ni )
482 ni = mv[i].channels();
483 for( k = 0; k < ni; k++ )
485 pairs[(j+k)*2] = j + k;
486 pairs[(j+k)*2+1] = j + k;
489 mixChannels( mv, n, &dst, 1, &pairs[0], cn );
493 MergeFunc func = getMergeFunc(depth);
494 CV_Assert( func != 0 );
496 size_t esz = dst.elemSize(), esz1 = dst.elemSize1();
497 size_t blocksize0 = (int)((BLOCK_SIZE + esz-1)/esz);
498 AutoBuffer<uchar> _buf((cn+1)*(sizeof(Mat*) + sizeof(uchar*)) + 16);
499 const Mat** arrays = (const Mat**)_buf.data();
500 uchar** ptrs = (uchar**)alignPtr(arrays + cn + 1, 16);
503 for( k = 0; k < cn; k++ )
504 arrays[k+1] = &mv[k];
506 NAryMatIterator it(arrays, ptrs, cn+1);
507 size_t total = (int)it.size;
508 size_t blocksize = std::min((size_t)CV_SPLIT_MERGE_MAX_BLOCK_SIZE(cn), cn <= 4 ? total : std::min(total, blocksize0));
510 for( i = 0; i < it.nplanes; i++, ++it )
512 for( size_t j = 0; j < total; j += blocksize )
514 size_t bsz = std::min(total - j, blocksize);
515 func( (const uchar**)&ptrs[1], ptrs[0], (int)bsz, cn );
517 if( j + blocksize < total )
520 for( int t = 0; t < cn; t++ )
521 ptrs[t+1] += bsz*esz1;
531 static bool ocl_merge( InputArrayOfArrays _mv, OutputArray _dst )
533 std::vector<UMat> src, ksrc;
534 _mv.getUMatVector(src);
535 CV_Assert(!src.empty());
537 int type = src[0].type(), depth = CV_MAT_DEPTH(type),
538 rowsPerWI = ocl::Device::getDefault().isIntel() ? 4 : 1;
539 Size size = src[0].size();
541 for (size_t i = 0, srcsize = src.size(); i < srcsize; ++i)
543 int itype = src[i].type(), icn = CV_MAT_CN(itype), idepth = CV_MAT_DEPTH(itype),
544 esz1 = CV_ELEM_SIZE1(idepth);
548 CV_Assert(size == src[i].size() && depth == idepth);
550 for (int cn = 0; cn < icn; ++cn)
553 tsrc.offset += cn * esz1;
554 ksrc.push_back(tsrc);
557 int dcn = (int)ksrc.size();
559 String srcargs, processelem, cndecl, indexdecl;
560 for (int i = 0; i < dcn; ++i)
562 srcargs += format("DECLARE_SRC_PARAM(%d)", i);
563 processelem += format("PROCESS_ELEM(%d)", i);
564 indexdecl += format("DECLARE_INDEX(%d)", i);
565 cndecl += format(" -D scn%d=%d", i, ksrc[i].channels());
568 ocl::Kernel k("merge", ocl::core::split_merge_oclsrc,
569 format("-D OP_MERGE -D cn=%d -D T=%s -D DECLARE_SRC_PARAMS_N=%s"
570 " -D DECLARE_INDEX_N=%s -D PROCESS_ELEMS_N=%s%s",
571 dcn, ocl::memopTypeToStr(depth), srcargs.c_str(),
572 indexdecl.c_str(), processelem.c_str(), cndecl.c_str()));
576 _dst.create(size, CV_MAKE_TYPE(depth, dcn));
577 UMat dst = _dst.getUMat();
580 for (int i = 0; i < dcn; ++i)
581 argidx = k.set(argidx, ocl::KernelArg::ReadOnlyNoSize(ksrc[i]));
582 argidx = k.set(argidx, ocl::KernelArg::WriteOnly(dst));
583 k.set(argidx, rowsPerWI);
585 size_t globalsize[2] = { (size_t)dst.cols, ((size_t)dst.rows + rowsPerWI - 1) / rowsPerWI };
586 return k.run(2, globalsize, NULL, false);
593 void cv::merge(InputArrayOfArrays _mv, OutputArray _dst)
595 CV_INSTRUMENT_REGION()
597 CV_OCL_RUN(_mv.isUMatVector() && _dst.isUMat(),
598 ocl_merge(_mv, _dst))
601 _mv.getMatVector(mv);
602 merge(!mv.empty() ? &mv[0] : 0, mv.size(), _dst);