Backend backendId = get<0>(get<2>(GetParam()));
Target targetId = get<1>(get<2>(GetParam()));
- // bug: https://github.com/opencv/opencv/issues/17964
- if (actType == "Power" && backendId == DNN_BACKEND_OPENCV && (targetId == DNN_TARGET_OPENCL || targetId == DNN_TARGET_OPENCL_FP16))
- applyTestTag(CV_TEST_TAG_DNN_SKIP_OPENCL);
-
Net net;
int convId = net.addLayer(convParams.name, convParams.type, convParams);
int activId = net.addLayerToPrev(activationParams.name, activationParams.type, activationParams);
expectedFusedLayers.push_back(activId); // all activations are fused
else if (targetId == DNN_TARGET_OPENCL || targetId == DNN_TARGET_OPENCL_FP16)
{
- if (actType == "ReLU" || actType == "ChannelsPReLU" || actType == "ReLU6" || actType == "TanH" || actType == "Power")
+ if (actType == "ReLU" || actType == "ChannelsPReLU" || actType == "ReLU6" || actType == "TanH" /*|| actType == "Power"*/)
expectedFusedLayers.push_back(activId);
}
}
Backend backendId = get<0>(get<4>(GetParam()));
Target targetId = get<1>(get<4>(GetParam()));
- // bug: https://github.com/opencv/opencv/issues/17945
- if ((eltwiseOp != "sum" || weightedEltwise) && backendId == DNN_BACKEND_OPENCV && (targetId == DNN_TARGET_OPENCL || targetId == DNN_TARGET_OPENCL_FP16))
- applyTestTag(CV_TEST_TAG_DNN_SKIP_OPENCL);
-
- // bug: https://github.com/opencv/opencv/issues/17953
- if (eltwiseOp == "sum" && actType == "ChannelsPReLU" && bias_term == false &&
- backendId == DNN_BACKEND_OPENCV && (targetId == DNN_TARGET_OPENCL || targetId == DNN_TARGET_OPENCL_FP16))
- {
- applyTestTag(CV_TEST_TAG_DNN_SKIP_OPENCL);
- }
-
- // bug: https://github.com/opencv/opencv/issues/17964
- if (actType == "Power" && backendId == DNN_BACKEND_OPENCV && (targetId == DNN_TARGET_OPENCL || targetId == DNN_TARGET_OPENCL_FP16))
- applyTestTag(CV_TEST_TAG_DNN_SKIP_OPENCL);
-
Net net;
int convId = net.addLayer(convParams.name, convParams.type, convParams);
int eltwiseId = net.addLayer(eltwiseParams.name, eltwiseParams.type, eltwiseParams);
expectedFusedLayers.push_back(activId); // activation is fused with eltwise layer
else if (targetId == DNN_TARGET_OPENCL || targetId == DNN_TARGET_OPENCL_FP16)
{
- if (actType == "ReLU" || actType == "ChannelsPReLU" || actType == "Power")
+ if (eltwiseOp == "sum" && !weightedEltwise &&
+ (actType == "ReLU" || actType == "ChannelsPReLU" /*|| actType == "Power"*/)
+ )
{
expectedFusedLayers.push_back(eltwiseId);
expectedFusedLayers.push_back(activId);
Backend backendId = get<0>(get<4>(GetParam()));
Target targetId = get<1>(get<4>(GetParam()));
- // bug: https://github.com/opencv/opencv/issues/17964
- if (actType == "Power" && backendId == DNN_BACKEND_OPENCV && (targetId == DNN_TARGET_OPENCL || targetId == DNN_TARGET_OPENCL_FP16))
- applyTestTag(CV_TEST_TAG_DNN_SKIP_OPENCL);
-
- // bug: https://github.com/opencv/opencv/issues/17953
- if (actType == "ChannelsPReLU" && bias_term == false &&
- backendId == DNN_BACKEND_OPENCV && (targetId == DNN_TARGET_OPENCL || targetId == DNN_TARGET_OPENCL_FP16))
- {
- applyTestTag(CV_TEST_TAG_DNN_SKIP_OPENCL);
- }
-
Net net;
int convId = net.addLayer(convParams.name, convParams.type, convParams);
int activId = net.addLayer(activationParams.name, activationParams.type, activationParams);
expectedFusedLayers.push_back(activId); // activation fused with convolution
else if (targetId == DNN_TARGET_OPENCL || targetId == DNN_TARGET_OPENCL_FP16)
{
- if (actType == "ReLU" || actType == "ChannelsPReLU" || actType == "ReLU6" || actType == "TanH" || actType == "Power")
+ if (actType == "ReLU" || actType == "ChannelsPReLU" || actType == "ReLU6" || actType == "TanH" /*|| actType == "Power"*/)
expectedFusedLayers.push_back(activId); // activation fused with convolution
}
}