step_pyr[0] = data_cost.step / type_size;\r
\r
Size temp_size = data_cost_size;\r
- if (data_cost.step * data_cost_size.height < static_cast<size_t>(step_pyr[levels - 1]) * rows_pyr[levels - 1] * ndisp)\r
+ if (data_cost_size.width * data_cost_size.height < static_cast<size_t>(step_pyr[levels - 1]) * rows_pyr[levels - 1] * ndisp)\r
{\r
- temp_size = Size(step_pyr[levels - 1], rows_pyr[levels - 1] * nr_plane);\r
+ temp_size = Size(step_pyr[levels - 1], rows_pyr[levels - 1] * ndisp);\r
}\r
\r
temp1.create(temp_size, msg_type);\r
const int threadsNum = 256;\r
const size_t smem_size = threadsNum * sizeof(float);\r
\r
- dim3 threads(winsz, 1, threadsNum/winsz);\r
+ dim3 threads(winsz, 1, threadsNum / winsz);\r
dim3 grid(w, h, 1); \r
grid.y *= divUp(ndisp, threads.z);\r
\r
case 1: csbp_kernels::data_init<T, winsz, 1><<<grid, threads, smem_size, stream>>>(level, rows, cols, h); break;\r
case 3: csbp_kernels::data_init<T, winsz, 3><<<grid, threads, smem_size, stream>>>(level, rows, cols, h); break;\r
default: cv::gpu::error("Unsupported channels count", __FILE__, __LINE__);\r
- } \r
+ }\r
}\r
\r
typedef void (*DataInitCaller)(int cols, int rows, int w, int h, int level, int ndisp, int channels, const cudaStream_t& stream);\r
\r
typedef void (*ComputeDataCostCaller)(const DevMem2D& disp_selected_pyr, const DevMem2D& data_cost, \r
int h, int w, int level, int nr_plane, int channels, const cudaStream_t& stream);\r
-\r
+ \r
void compute_data_cost(const DevMem2D& disp_selected_pyr, const DevMem2D& data_cost, size_t msg_step1, size_t msg_step2, int msg_type,\r
int h, int w, int h2, int level, int nr_plane, int channels, const cudaStream_t& stream)\r
{\r