1 // Copyright (C) 2018-2019 Intel Corporation
2 // SPDX-License-Identifier: Apache-2.0
5 #include <gtest/gtest.h>
6 #include <gmock/gmock-spec-builders.h>
7 #include "mkldnn_plugin/mkldnn_graph.h"
9 #include "test_graph.hpp"
11 #include <mkldnn_plugin/mkldnn_extension_utils.h>
12 #include <mkldnn_plugin/mkldnn_plugin.h>
13 #include <mkldnn_plugin/config.h>
16 using namespace mkldnn;
18 class MKLDNNTestExecNetwork: public MKLDNNPlugin::MKLDNNExecNetwork {
20 MKLDNNTestExecNetwork(InferenceEngine::ICNNNetwork &network, const MKLDNNPlugin::Config &cfg)
21 : MKLDNNExecNetwork(network, cfg, {}) {}
22 MKLDNNPlugin::MKLDNNGraph& getGraph() {
27 class MKLDNNTestEngine: public MKLDNNPlugin::Engine {
29 MKLDNNPlugin::MKLDNNGraph& getGraph() {
30 auto * execNetworkInt =
31 dynamic_cast<InferenceEngine::ExecutableNetworkBase<InferenceEngine::ExecutableNetworkInternal> *>(_loadedNetwork.get());
33 THROW_IE_EXCEPTION << "Cannot find loaded network!";
35 auto * network = reinterpret_cast<MKLDNNTestExecNetwork *>(execNetworkInt->getImpl().get());
37 THROW_IE_EXCEPTION << "Cannot get mkldnn graph!";
38 return network->getGraph();
42 class MKLDNNGraphLeaksTests: public ::testing::Test {
44 void addOutputToEachNode(InferenceEngine::CNNNetReader& net_reader, std::vector<std::string>& new_outputs,
45 InferenceEngine::CNNLayerPtr cnnLayer) {
46 auto outputs = net_reader.getNetwork().getOutputsInfo();
47 if (outputs.find(cnnLayer->name) != outputs.end())
50 net_reader.getNetwork().addOutput(cnnLayer->name);
51 new_outputs.push_back(cnnLayer->name);
53 for (const auto &layer : cnnLayer->outData) {
54 for (const auto &data : layer->getInputTo()) {
55 addOutputToEachNode(net_reader, new_outputs, data.second);
60 void fill_data(float *data, size_t size, size_t duty_ratio = 10) {
61 for (size_t i = 0; i < size; i++) {
62 if ( ( i / duty_ratio)%2 == 1) {
65 data[i] = (float) sin((float)i);
71 TEST_F(MKLDNNGraphLeaksTests, MKLDNN_not_release_outputs_fp32) {
73 InferenceEngine::CNNNetReader net_reader;
74 std::string model = "<net name=\"LeNet\" version=\"2\" batch=\"1\">\n"
76 " <layer name=\"data\" type=\"Input\" precision=\"FP32\" id=\"0\">\n"
86 " <layer name=\"conv1\" type=\"Convolution\" precision=\"FP32\" id=\"1\">\n"
87 " <convolution_data stride-x=\"1\" stride-y=\"1\" pad-x=\"0\" pad-y=\"0\" kernel-x=\"5\" kernel-y=\"5\" output=\"20\" group=\"1\"/>\n"
104 " <weights offset=\"0\" size=\"2000\"/>\n"
105 " <biases offset=\"2000\" size=\"80\"/>\n"
107 " <layer name=\"pool1\" type=\"Pooling\" precision=\"FP32\" id=\"2\">\n"
108 " <pooling_data kernel-x=\"2\" kernel-y=\"2\" pad-x=\"0\" pad-y=\"0\" stride-x=\"2\" stride-y=\"2\" rounding-type=\"ceil\" pool-method=\"max\"/>\n"
126 " <layer name=\"conv2\" type=\"Convolution\" precision=\"FP32\" id=\"3\">\n"
127 " <convolution_data stride-x=\"1\" stride-y=\"1\" pad-x=\"0\" pad-y=\"0\" kernel-x=\"5\" kernel-y=\"5\" output=\"50\" group=\"1\"/>\n"
144 " <weights offset=\"2080\" size=\"100000\"/>\n"
145 " <biases offset=\"102080\" size=\"200\"/>\n"
147 " <layer name=\"pool2\" type=\"Pooling\" precision=\"FP32\" id=\"4\">\n"
148 " <pooling_data kernel-x=\"2\" kernel-y=\"2\" pad-x=\"0\" pad-y=\"0\" stride-x=\"2\" stride-y=\"2\" rounding-type=\"ceil\" pool-method=\"max\"/>\n"
166 " <layer name=\"ip1\" type=\"FullyConnected\" precision=\"FP32\" id=\"5\">\n"
167 " <fc_data out-size=\"500\"/>\n"
177 " <port id=\"10\">\n"
182 " <weights offset=\"102280\" size=\"1600000\"/>\n"
183 " <biases offset=\"1702280\" size=\"2000\"/>\n"
185 " <layer name=\"relu1\" type=\"ReLU\" precision=\"FP32\" id=\"6\">\n"
187 " <port id=\"11\">\n"
193 " <port id=\"12\">\n"
199 " <layer name=\"ip2\" type=\"FullyConnected\" precision=\"FP32\" id=\"7\">\n"
200 " <fc_data out-size=\"10\"/>\n"
202 " <port id=\"13\">\n"
208 " <port id=\"14\">\n"
213 " <weights offset=\"1704280\" size=\"20000\"/>\n"
214 " <biases offset=\"1724280\" size=\"40\"/>\n"
216 " <layer name=\"prob\" type=\"SoftMax\" precision=\"FP32\" id=\"8\">\n"
218 " <port id=\"15\">\n"
224 " <port id=\"16\">\n"
232 " <edge from-layer=\"0\" from-port=\"0\" to-layer=\"1\" to-port=\"1\"/>\n"
233 " <edge from-layer=\"1\" from-port=\"2\" to-layer=\"2\" to-port=\"3\"/>\n"
234 " <edge from-layer=\"2\" from-port=\"4\" to-layer=\"3\" to-port=\"5\"/>\n"
235 " <edge from-layer=\"3\" from-port=\"6\" to-layer=\"4\" to-port=\"7\"/>\n"
236 " <edge from-layer=\"4\" from-port=\"8\" to-layer=\"5\" to-port=\"9\"/>\n"
237 " <edge from-layer=\"5\" from-port=\"10\" to-layer=\"6\" to-port=\"11\"/>\n"
238 " <edge from-layer=\"6\" from-port=\"12\" to-layer=\"7\" to-port=\"13\"/>\n"
239 " <edge from-layer=\"7\" from-port=\"14\" to-layer=\"8\" to-port=\"15\"/>\n"
243 size_t weights_size = 1724320;
244 net_reader.ReadNetwork(model.c_str(), model.size());
246 InferenceEngine::TBlob<uint8_t> *weights = new InferenceEngine::TBlob<uint8_t>(InferenceEngine::Precision::U8, InferenceEngine::C, {weights_size});
248 fill_data((float *) weights->buffer(), weights->size() / sizeof(float));
249 InferenceEngine::TBlob<uint8_t>::Ptr weights_ptr = InferenceEngine::TBlob<uint8_t>::Ptr(weights);
251 net_reader.SetWeights(weights_ptr);
253 auto outputs = net_reader.getNetwork().getOutputsInfo();
254 std::vector<std::string> new_outputs;
256 for (auto input : net_reader.getNetwork().getInputsInfo()) {
257 for (const auto &layer : input.second->getInputData()->getInputTo()) {
258 addOutputToEachNode(net_reader, new_outputs, layer.second);
262 ASSERT_NE(1, net_reader.getNetwork().getOutputsInfo().size());
264 std::shared_ptr<MKLDNNTestEngine> score_engine(new MKLDNNTestEngine());
265 ASSERT_NO_THROW(score_engine->LoadNetwork(net_reader.getNetwork()));
267 size_t modified_outputs_size = score_engine->getGraph().GetOutputNodes().size();
269 InferenceEngine::CNNNetReader net_reader2;
270 net_reader2.ReadNetwork(model.c_str(), model.size());
271 net_reader2.SetWeights(weights_ptr);
272 ASSERT_EQ(1, net_reader2.getNetwork().getOutputsInfo().size());
274 ASSERT_NO_THROW(score_engine->LoadNetwork(net_reader2.getNetwork()));
276 size_t original_outputs_size = score_engine->getGraph().GetOutputNodes().size();
278 ASSERT_NE(modified_outputs_size, original_outputs_size);
279 ASSERT_EQ(1, original_outputs_size);