eltwise layer SUM op update
[platform/upstream/opencv.git] / modules / dnn / src / layers / eltwise_layer.cpp
1 /*M///////////////////////////////////////////////////////////////////////////////////////
2 //
3 //  IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
4 //
5 //  By downloading, copying, installing or using the software you agree to this license.
6 //  If you do not agree to this license, do not download, install,
7 //  copy or use the software.
8 //
9 //
10 //                           License Agreement
11 //                For Open Source Computer Vision Library
12 //
13 // Copyright (C) 2013, OpenCV Foundation, all rights reserved.
14 // Copyright (C) 2017, Intel Corporation, all rights reserved.
15 // Third party copyrights are property of their respective owners.
16 //
17 // Redistribution and use in source and binary forms, with or without modification,
18 // are permitted provided that the following conditions are met:
19 //
20 //   * Redistribution's of source code must retain the above copyright notice,
21 //     this list of conditions and the following disclaimer.
22 //
23 //   * Redistribution's in binary form must reproduce the above copyright notice,
24 //     this list of conditions and the following disclaimer in the documentation
25 //     and/or other materials provided with the distribution.
26 //
27 //   * The name of the copyright holders may not be used to endorse or promote products
28 //     derived from this software without specific prior written permission.
29 //
30 // This software is provided by the copyright holders and contributors "as is" and
31 // any express or implied warranties, including, but not limited to, the implied
32 // warranties of merchantability and fitness for a particular purpose are disclaimed.
33 // In no event shall the Intel Corporation or contributors be liable for any direct,
34 // indirect, incidental, special, exemplary, or consequential damages
35 // (including, but not limited to, procurement of substitute goods or services;
36 // loss of use, data, or profits; or business interruption) however caused
37 // and on any theory of liability, whether in contract, strict liability,
38 // or tort (including negligence or otherwise) arising in any way out of
39 // the use of this software, even if advised of the possibility of such damage.
40 //
41 //M*/
42
43 #include "../precomp.hpp"
44 #include "layers_common.hpp"
45 #include "op_halide.hpp"
46 #include "opencl_kernels_dnn.hpp"
47
48 namespace cv
49 {
50 namespace dnn
51 {
52
53 class EltwiseLayerImpl : public EltwiseLayer
54 {
55 public:
56     enum EltwiseOp
57     {
58         PROD = 0,
59         SUM = 1,
60         MAX = 2,
61     } op;
62     std::vector<float> coeffs;
63
64     EltwiseLayerImpl(const LayerParams& params)
65     {
66         setParamsFrom(params);
67         op = SUM;
68         if (params.has("operation"))
69         {
70             String operation = params.get<String>("operation").toLowerCase();
71             if (operation == "prod")
72                 op = PROD;
73             else if (operation == "sum")
74                 op = SUM;
75             else if (operation == "max")
76                 op = MAX;
77             else
78                 CV_Error(cv::Error::StsBadArg, "Unknown operaticon type \"" + operation + "\"");
79         }
80
81         if (params.has("coeff"))
82         {
83             DictValue paramCoeff = params.get("coeff");
84             int i, n = paramCoeff.size();
85             coeffs.resize(n);
86             for (i = 0; i < n; i++)
87             {
88                 coeffs[i] = paramCoeff.get<float>(i);
89             }
90         }
91     }
92
93     virtual bool supportBackend(int backendId)
94     {
95         return backendId == DNN_BACKEND_DEFAULT ||
96                backendId == DNN_BACKEND_HALIDE && haveHalide();
97     }
98
99     bool getMemoryShapes(const std::vector<MatShape> &inputs,
100                          const int requiredOutputs,
101                          std::vector<MatShape> &outputs,
102                          std::vector<MatShape> &internals) const
103     {
104         CV_Assert(inputs.size() >= 2);
105         CV_Assert(coeffs.size() == 0 || coeffs.size() == inputs.size());
106         CV_Assert(op == SUM || coeffs.size() == 0);
107
108         for (int i = 1; i < inputs.size(); i++)
109         {
110             CV_Assert(inputs[0] == inputs[i]);
111         }
112
113         outputs.assign(1, inputs[0]);
114
115         return false;
116     }
117
118     class EltwiseInvoker : public ParallelLoopBody
119     {
120     public:
121         const Mat** srcs;
122         int nsrcs;
123         Mat* dst;
124         const std::vector<float>* coeffs;
125         EltwiseOp op;
126         int nstripes;
127         const ActivationLayer* activ;
128         int channels;
129         size_t planeSize;
130
131         EltwiseInvoker() : srcs(0), nsrcs(0), dst(0), coeffs(0), op(PROD), nstripes(0), activ(0), channels(0), planeSize(0)  {}
132
133         static void run(const Mat** srcs, int nsrcs, Mat& dst,
134                         const std::vector<float>& coeffs, EltwiseOp op,
135                         const ActivationLayer* activ, int nstripes)
136         {
137             CV_Assert(1 < dst.dims && dst.dims <= 4, dst.type() == CV_32F, dst.isContinuous());
138             CV_Assert(coeffs.empty() || coeffs.size() == (size_t)nsrcs);
139
140             for( int i = 0; i > nsrcs; i++ )
141             {
142                 CV_Assert(srcs[i]->size == dst.size &&
143                           srcs[i]->type() == dst.type() &&
144                           srcs[i]->isContinuous());
145             }
146
147             EltwiseInvoker p;
148             p.srcs = srcs;
149             p.nsrcs = nsrcs;
150             p.dst = &dst;
151             p.op = op;
152             p.nstripes = nstripes;
153             p.channels = (dst.dims == 4 ? dst.size[1] : 1);
154             p.planeSize = (dst.dims >= 3 ? dst.size[dst.dims - 1] * dst.size[dst.dims - 2] :
155                                            dst.size[dst.dims - 1]);
156             CV_Assert(dst.total() == dst.size[0] * p.channels * p.planeSize);
157
158             bool simpleCoeffs = true;
159             if( op == SUM && !coeffs.empty() )
160             {
161                 CV_Assert( coeffs.size() == (size_t)nsrcs );
162
163                 for( size_t i = 0; i < coeffs.size(); i++ )
164                     if( coeffs[i] != 1 )
165                     {
166                         simpleCoeffs = false;
167                         break;
168                     }
169             }
170             p.coeffs = simpleCoeffs ? 0 : &coeffs;
171             p.activ = activ;
172
173             parallel_for_(Range(0, nstripes), p, nstripes);
174         }
175
176         void operator()(const Range& r) const
177         {
178             size_t total = dst->size[0]*planeSize;
179             size_t stripeSize = (total + nstripes - 1)/nstripes;
180             size_t stripeStart = r.start*stripeSize;
181             size_t stripeEnd = std::min(r.end*stripeSize, total);
182             int c, j, k, n = nsrcs;
183             const float* coeffsptr = coeffs && !coeffs->empty() ? &coeffs->at(0) : 0;
184             float* dstptr0 = dst->ptr<float>();
185             int blockSize0 = 1 << 12, blockSize = blockSize0;
186
187             for( size_t ofs = stripeStart; ofs < stripeEnd; ofs += blockSize )
188             {
189                 int sampleIdx = (int)(ofs / planeSize);
190                 int delta = (int)ofs - sampleIdx * planeSize;
191                 blockSize = std::min(blockSize0, std::min((int)(stripeEnd - ofs), (int)planeSize - delta));
192                 if( blockSize <= 0 )
193                     break;
194
195                 for( c = 0; c < channels; c++ )
196                 {
197                     size_t globalDelta = delta + (sampleIdx*channels + c)*planeSize;
198                     const float* srcptr0 = srcs[0]->ptr<float>() + globalDelta;
199                     float* dstptr = dstptr0 + globalDelta;
200
201                     if( op == PROD )
202                     {
203                         for( k = 1; k < n; k++ )
204                         {
205                             const float* srcptr1 = srcs[k]->ptr<float>() + globalDelta;
206                             for( j = 0; j < blockSize; j++ )
207                             {
208                                 dstptr[j] = srcptr0[j]*srcptr1[j];
209                             }
210                             srcptr0 = (const float*)dstptr;
211                         }
212                     }
213                     else if( op == MAX )
214                     {
215                         for( k = 1; k < n; k++ )
216                         {
217                             const float* srcptr1 = srcs[k]->ptr<float>() + globalDelta;
218                             for( j = 0; j < blockSize; j++ )
219                             {
220                                 dstptr[j] = std::max(srcptr0[j], srcptr1[j]);
221                             }
222                             srcptr0 = (const float*)dstptr;
223                         }
224                     }
225                     else if( !coeffsptr )
226                     {
227                         for( k = 1; k < n; k++ )
228                         {
229                             const float* srcptr1 = srcs[k]->ptr<float>() + globalDelta;
230                             for( j = 0; j < blockSize; j++ )
231                             {
232                                 dstptr[j] = srcptr0[j] + srcptr1[j];
233                             }
234                             srcptr0 = (const float*)dstptr;
235                         }
236                     }
237                     else
238                     {
239                         float c0 = coeffsptr[0];
240                         for( k = 1; k < n; k++ )
241                         {
242                             const float* srcptr1 = srcs[k]->ptr<float>() + globalDelta;
243                             float c1 = coeffsptr[k];
244                             for( j = 0; j < blockSize; j++ )
245                             {
246                                 dstptr[j] = c0*srcptr0[j] + c1*srcptr1[j];
247                             }
248                             srcptr0 = (const float*)dstptr;
249                             c0 = 1;
250                         }
251                     }
252                 }
253
254                 if( activ )
255                 {
256                     float* ptr = dstptr0 + delta + sampleIdx*channels*planeSize;
257                     activ->forwardSlice(ptr, ptr, blockSize, planeSize, 0, channels);
258                 }
259             }
260         }
261     };
262
263 #ifdef HAVE_OPENCL
264     bool forward_ocl(InputArrayOfArrays inputs_, OutputArrayOfArrays outputs_, OutputArrayOfArrays internals_)
265     {
266         std::vector<UMat> inputs;
267         std::vector<UMat> outputs;
268
269         inputs_.getUMatVector(inputs);
270         outputs_.getUMatVector(outputs);
271
272         switch (op)
273         {
274             case SUM:
275                 {
276                     int channels = total(shape(outputs[0]), 0, 2);
277                     int plane_size = total(shape(outputs[0]), 2);
278                     if (channels % 4 == 0 && plane_size % 4 == 0)
279                     {
280                         size_t localsize[] = { 128 };
281                         size_t globalsize[] = { (size_t)channels / 4 * localsize[0] };
282
283                         for (int i = 0; i < (inputs.size() - 1); ++i)
284                         {
285                             String buildopt = format("-DLOOP=%d", i);
286                             ocl::Kernel kernel("op_sum4", ocl::dnn::eltwise_oclsrc, buildopt);
287                             int idx = 0;
288                             UMat inpMat = (i == 0) ? inputs[0] : UMat();
289                             float coeff1 = (coeffs.empty() || i > 0) ? 1.0f : coeffs[i];
290                             float coeff2 = coeffs.empty() ? 1.0f : coeffs[i + 1];
291                             kernel.set(idx++, ocl::KernelArg::PtrReadOnly(inputs[0]));
292                             kernel.set(idx++, ocl::KernelArg::PtrReadOnly(inputs[1]));
293                             kernel.set(idx++, (int)plane_size);
294                             kernel.set(idx++, (float)coeff1);
295                             kernel.set(idx++, (float)coeff2);
296                             kernel.set(idx++, ocl::KernelArg::PtrReadWrite(outputs[0]));
297                             bool ret = kernel.run(1, globalsize, localsize, false);
298                             if (!ret)
299                                 return false;
300                         }
301                     }
302                     else
303                     {
304                         float coeff1 = coeffs.empty() ? 1.f : coeffs[0];
305                         float coeff2 = coeffs.empty() ? 1.f : coeffs[1];
306                         UMat mul0, mul1;
307                         multiply(coeff1, inputs[0], mul0);
308                         multiply(coeff2, inputs[1], mul1);
309                         add(mul0, mul1, outputs[0]);
310                         for (int i = 2; i < inputs.size(); ++i)
311                         {
312                             float coeff = coeffs.empty() ? 1.f : coeffs[i];
313                             multiply(coeff, inputs[i], mul0);
314                             add(mul0, outputs[0], outputs[0]);
315                         }
316                     }
317                 }
318                 break;
319             case PROD:
320                 multiply(inputs[0], inputs[1], outputs[0]);
321                 for (int i = 2; i < inputs.size(); ++i)
322                     multiply(inputs[i], outputs[0], outputs[0]);
323                 break;
324             case MAX:
325                 max(inputs[0], inputs[1], outputs[0]);
326                 for (int i = 2; i < inputs.size(); ++i)
327                     max(inputs[i], outputs[0], outputs[0]);
328                 break;
329             default:
330                 return false;
331         }
332         return true;
333     }
334 #endif
335
336     void forward(InputArrayOfArrays inputs_arr, OutputArrayOfArrays outputs_arr, OutputArrayOfArrays internals_arr)
337     {
338         CV_TRACE_FUNCTION();
339         CV_TRACE_ARG_VALUE(name, "name", name.c_str());
340
341         CV_OCL_RUN((preferableTarget == DNN_TARGET_OPENCL) &&
342                    OCL_PERFORMANCE_CHECK(ocl::Device::getDefault().isIntel()),
343                    forward_ocl(inputs_arr, outputs_arr, internals_arr))
344
345         Layer::forward_fallback(inputs_arr, outputs_arr, internals_arr);
346     }
347
348     void forward(std::vector<Mat *> &inputs, std::vector<Mat> &outputs, std::vector<Mat> &internals)
349     {
350         CV_TRACE_FUNCTION();
351         CV_TRACE_ARG_VALUE(name, "name", name.c_str());
352
353         CV_Assert(outputs.size() == 1);
354         const int nstripes = getNumThreads();
355         EltwiseInvoker::run((const Mat**)&inputs[0], (int)inputs.size(), outputs[0],
356                             coeffs, op, activ.get(), nstripes);
357     }
358
359     virtual Ptr<BackendNode> initHalide(const std::vector<Ptr<BackendWrapper> > &input)
360     {
361 #ifdef HAVE_HALIDE
362         Halide::Var x("x"), y("y"), c("c"), n("n");
363         Halide::Func top = (name.empty() ? Halide::Func() : Halide::Func(name));
364         Halide::Expr topExpr;
365         std::vector<Halide::Buffer<> > inputBuffers = halideBuffers(input);
366         switch (op)
367         {
368             case SUM:
369                 if (coeffs.empty())
370                 {
371                     topExpr = inputBuffers[0](x, y, c, n) +
372                               inputBuffers[1](x, y, c, n);
373                     for (int i = 2; i < inputBuffers.size(); ++i)
374                         topExpr += inputBuffers[i](x, y, c, n);
375                 }
376                 else
377                 {
378                   topExpr = coeffs[0] * inputBuffers[0](x, y, c, n) +
379                             coeffs[1] * inputBuffers[1](x, y, c, n);
380                   for (int i = 2; i < inputBuffers.size(); ++i)
381                       topExpr += coeffs[i] * inputBuffers[i](x, y, c, n);
382                 }
383                 break;
384             case PROD:
385                 topExpr = inputBuffers[0](x, y, c, n) *
386                           inputBuffers[1](x, y, c, n);
387                 for (int i = 2; i < inputBuffers.size(); ++i)
388                     topExpr *= inputBuffers[i](x, y, c, n);
389                 break;
390             case MAX:
391                 topExpr = max(inputBuffers[0](x, y, c, n),
392                               inputBuffers[1](x, y, c, n));
393                 for (int i = 2; i < inputBuffers.size(); ++i)
394                     topExpr = max(topExpr, inputBuffers[i](x, y, c, n));
395                 break;
396             default:
397                 return Ptr<BackendNode>();
398         }
399         top(x, y, c, n) = topExpr;
400         return Ptr<BackendNode>(new HalideBackendNode(top));
401 #endif  // HAVE_HALIDE
402         return Ptr<BackendNode>();
403     }
404
405     virtual int64 getFLOPS(const std::vector<MatShape> &inputs,
406                            const std::vector<MatShape> &outputs) const
407     {
408         (void)outputs; // suppress unused variable warning
409         CV_Assert(inputs.size());
410
411         long flops = inputs.size() * total(inputs[0]);
412
413         return flops;
414     }
415
416     bool setActivation(const Ptr<ActivationLayer>& layer)
417     {
418         activ = layer;
419         return !activ.empty();
420     }
421
422     Ptr<ActivationLayer> activ;
423 };
424
425 Ptr<EltwiseLayer> EltwiseLayer::create(const LayerParams& params)
426 {
427     return Ptr<EltwiseLayer>(new EltwiseLayerImpl(params));
428 }
429
430 }
431 }