1 // This file is part of OpenCV project.
2 // It is subject to the license terms in the LICENSE file found in the top-level directory
3 // of this distribution and at http://opencv.org/license.html.
5 // Copyright (C) 2018, Intel Corporation, all rights reserved.
6 // Third party copyrights are property of their respective owners.
8 #include "../precomp.hpp"
9 #include "../op_inf_engine.hpp"
10 #include "../op_cuda.hpp"
11 #include "layers_common.hpp"
14 #include "opencl_kernels_dnn.hpp"
18 #include "../cuda4dnn/primitives/const.hpp"
19 using namespace cv::dnn::cuda4dnn;
22 namespace cv { namespace dnn {
24 class ConstLayerImpl CV_FINAL : public ConstLayer
27 ConstLayerImpl(const LayerParams& params)
29 setParamsFrom(params);
30 CV_Assert(blobs.size() == 1);
33 virtual bool supportBackend(int backendId) CV_OVERRIDE
35 return backendId == DNN_BACKEND_OPENCV ||
36 backendId == DNN_BACKEND_INFERENCE_ENGINE ||
37 backendId == DNN_BACKEND_CUDA;
40 virtual bool getMemoryShapes(const std::vector<MatShape> &inputs,
41 const int requiredOutputs,
42 std::vector<MatShape> &outputs,
43 std::vector<MatShape> &internals) const CV_OVERRIDE
45 CV_Assert(inputs.empty());
46 outputs.assign(1, shape(blobs[0]));
51 bool forward_ocl(InputArrayOfArrays inps, OutputArrayOfArrays outs, OutputArrayOfArrays internals)
53 std::vector<UMat> outputs;
54 outs.getUMatVector(outputs);
55 if (outs.depth() == CV_16S)
56 convertFp16(blobs[0], outputs[0]);
58 blobs[0].copyTo(outputs[0]);
63 void forward(InputArrayOfArrays inputs_arr, OutputArrayOfArrays outputs_arr, OutputArrayOfArrays internals_arr) CV_OVERRIDE
66 CV_TRACE_ARG_VALUE(name, "name", name.c_str());
68 CV_OCL_RUN(IS_DNN_OPENCL_TARGET(preferableTarget),
69 forward_ocl(inputs_arr, outputs_arr, internals_arr))
71 std::vector<Mat> outputs;
72 outputs_arr.getMatVector(outputs);
73 blobs[0].copyTo(outputs[0]);
76 #ifdef HAVE_INF_ENGINE
77 virtual Ptr<BackendNode> initInfEngine(const std::vector<Ptr<BackendWrapper> >&) CV_OVERRIDE
79 InferenceEngine::Builder::ConstLayer ieLayer(name);
80 ieLayer.setData(wrapToInfEngineBlob(blobs[0]));
81 return Ptr<BackendNode>(new InfEngineBackendNode(ieLayer));
83 #endif // HAVE_INF_ENGINE
86 Ptr<BackendNode> initCUDA(
88 const std::vector<Ptr<BackendWrapper>>& inputs,
89 const std::vector<Ptr<BackendWrapper>>& outputs
92 auto context = reinterpret_cast<csl::CSLContext*>(context_);
94 CV_Assert(blobs.size() == 1);
95 return make_cuda_node<cuda4dnn::ConstOp>(preferableTarget, std::move(context->stream), blobs[0]);
101 Ptr<Layer> ConstLayer::create(const LayerParams& params)
103 return Ptr<Layer>(new ConstLayerImpl(params));
106 }} // namespace cv::dnn