{
const int dir[][2] = { {-1,-1}, {-1,0}, {-1,1}, {0,1}, {1,1}, {1,0}, {1,-1}, {0,-1} };
coordinate_t next;
- coordinate_t next_temp;
unsigned char ptr;
assert (direction==-1 || !coords.empty());
CV_Assert( std::abs(dsize.width - ssize.width*2) == dsize.width % 2 &&
std::abs(dsize.height - ssize.height*2) == dsize.height % 2);
- int k, x, sy0 = -PU_SZ/2, sy = sy0, width0 = ssize.width - 1;
+ int k, x, sy0 = -PU_SZ/2, sy = sy0;
ssize.width *= cn;
dsize.width *= cn;
- width0 *= cn;
for( x = 0; x < ssize.width; x++ )
dtab[x] = (x/cn)*2*cn + x % cn;
const int n, const int bx, const int by,
const int threadsNum, int *kLevels, int **processingLevels)
{
- int rootFilterDim, sumPartFiltersDim, i, numLevels, dbx, dby, numDotProducts;
+ int rootFilterDim, sumPartFiltersDim, i, numLevels, dbx, dby;
int j, minValue, argMin, lambda, maxValue, k;
int *dotProd, *weights, *disp;
if (H == NULL || all_F == NULL)
// of feature map with part filter
dbx = 2 * bx;
dby = 2 * by;
- // Total number of dot products for all levels
- numDotProducts = 0;
lambda = LAMBDA;
for (i = 0; i < numLevels; i++)
{
H->pyramid[i + lambda]->sizeY * rootFilterDim +
(H->pyramid[i]->sizeX + dbx) *
(H->pyramid[i]->sizeY + dby) * sumPartFiltersDim;
- numDotProducts += dotProd[i];
}
// Allocation memory for saving dot product number performed by each thread
weights = (int *)malloc(sizeof(int) * threadsNum);