use NEON as same as SSE
authorTomoaki Teshima <tomoaki.teshima@gmail.com>
Thu, 23 Jun 2016 04:22:46 +0000 (13:22 +0900)
committerTomoaki Teshima <tomoaki.teshima@gmail.com>
Sat, 9 Jul 2016 07:53:01 +0000 (16:53 +0900)
modules/objdetect/src/hog.cpp

index 77dd712..1055525 100644 (file)
@@ -222,6 +222,17 @@ void HOGDescriptor::copyTo(HOGDescriptor& c) const
     c.signedGradient = signedGradient;
 }
 
+#if CV_NEON
+// replace of _mm_set_ps
+inline float32x4_t vsetq_f32(float f0, float f1, float f2, float f3)
+{
+    float32x4_t a = vdupq_n_f32(f0);
+    a = vsetq_lane_f32(f1, a, 1);
+    a = vsetq_lane_f32(f2, a, 2);
+    a = vsetq_lane_f32(f3, a, 3);
+    return a;
+}
+#endif
 void HOGDescriptor::computeGradient(const Mat& img, Mat& grad, Mat& qangle,
     Size paddingTL, Size paddingBR) const
 {
@@ -259,6 +270,21 @@ void HOGDescriptor::computeGradient(const Mat& img, Mat& grad, Mat& qangle,
             _mm_storeu_ps(_data + i, _mm_cvtepi32_ps(idx));
             idx = _mm_add_epi32(idx, ifour);
         }
+#elif CV_NEON
+    const int indeces[] = { 0, 1, 2, 3 };
+    uint32x4_t idx = *(uint32x4_t*)indeces;
+    uint32x4_t ifour = vdupq_n_u32(4);
+
+    float* const _data = &_lut(0, 0);
+    if( gammaCorrection )
+        for( i = 0; i < 256; i++ )
+            _lut(0,i) = std::sqrt((float)i);
+    else
+        for( i = 0; i < 256; i += 4 )
+        {
+            vst1q_f32(_data + i, vcvtq_f32_u32(idx));
+            idx = vaddq_u32 (idx, ifour);
+        }
 #else
     if( gammaCorrection )
         for( i = 0; i < 256; i++ )
@@ -299,6 +325,10 @@ void HOGDescriptor::computeGradient(const Mat& img, Mat& grad, Mat& qangle,
         for ( ; x <= end - 4; x += 4)
             _mm_storeu_si128((__m128i*)(xmap + x), _mm_mullo_epi16(ithree,
                 _mm_loadu_si128((const __m128i*)(xmap + x))));
+#elif CV_NEON
+        int32x4_t ithree = vdupq_n_s32(3);
+        for ( ; x <= end - 4; x += 4)
+            vst1q_s32(xmap + x, vmulq_s32(ithree, vld1q_s32(xmap + x)));
 #endif
         for ( ; x < end; ++x)
             xmap[x] *= 3;
@@ -368,6 +398,45 @@ void HOGDescriptor::computeGradient(const Mat& img, Mat& grad, Mat& qangle,
                 _mm_storeu_ps(dbuf + x, _dx2);
                 _mm_storeu_ps(dbuf + x + width, _dy2);
             }
+#elif CV_NEON
+            for( ; x <= width - 4; x += 4 )
+            {
+                int x0 = xmap[x], x1 = xmap[x+1], x2 = xmap[x+2], x3 = xmap[x+3];
+                typedef const uchar* const T;
+                T p02 = imgPtr + xmap[x+1], p00 = imgPtr + xmap[x-1];
+                T p12 = imgPtr + xmap[x+2], p10 = imgPtr + xmap[x];
+                T p22 = imgPtr + xmap[x+3], p20 = p02;
+                T p32 = imgPtr + xmap[x+4], p30 = p12;
+
+                float32x4_t _dx0 = vsubq_f32(vsetq_f32(lut[p02[0]], lut[p12[0]], lut[p22[0]], lut[p32[0]]),
+                                             vsetq_f32(lut[p00[0]], lut[p10[0]], lut[p20[0]], lut[p30[0]]));
+                float32x4_t _dx1 = vsubq_f32(vsetq_f32(lut[p02[1]], lut[p12[1]], lut[p22[1]], lut[p32[1]]),
+                                             vsetq_f32(lut[p00[1]], lut[p10[1]], lut[p20[1]], lut[p30[1]]));
+                float32x4_t _dx2 = vsubq_f32(vsetq_f32(lut[p02[2]], lut[p12[2]], lut[p22[2]], lut[p32[2]]),
+                                             vsetq_f32(lut[p00[2]], lut[p10[2]], lut[p20[2]], lut[p30[2]]));
+
+                float32x4_t _dy0 = vsubq_f32(vsetq_f32(lut[nextPtr[x0]], lut[nextPtr[x1]], lut[nextPtr[x2]], lut[nextPtr[x3]]),
+                                             vsetq_f32(lut[prevPtr[x0]], lut[prevPtr[x1]], lut[prevPtr[x2]], lut[prevPtr[x3]]));
+                float32x4_t _dy1 = vsubq_f32(vsetq_f32(lut[nextPtr[x0+1]], lut[nextPtr[x1+1]], lut[nextPtr[x2+1]], lut[nextPtr[x3+1]]),
+                                             vsetq_f32(lut[prevPtr[x0+1]], lut[prevPtr[x1+1]], lut[prevPtr[x2+1]], lut[prevPtr[x3+1]]));
+                float32x4_t _dy2 = vsubq_f32(vsetq_f32(lut[nextPtr[x0+2]], lut[nextPtr[x1+2]], lut[nextPtr[x2+2]], lut[nextPtr[x3+2]]),
+                                             vsetq_f32(lut[prevPtr[x0+2]], lut[prevPtr[x1+2]], lut[prevPtr[x2+2]], lut[prevPtr[x3+2]]));
+
+                float32x4_t _mag0 = vaddq_f32(vmulq_f32(_dx0, _dx0), vmulq_f32(_dy0, _dy0));
+                float32x4_t _mag1 = vaddq_f32(vmulq_f32(_dx1, _dx1), vmulq_f32(_dy1, _dy1));
+                float32x4_t _mag2 = vaddq_f32(vmulq_f32(_dx2, _dx2), vmulq_f32(_dy2, _dy2));
+
+                uint32x4_t mask = vcgtq_f32(_mag2, _mag1);
+                _dx2 = vbslq_f32(mask, _dx2, _dx1);
+                _dy2 = vbslq_f32(mask, _dy2, _dy1);
+
+                mask = vcgtq_f32(vmaxq_f32(_mag2, _mag1), _mag0);
+                _dx2 = vbslq_f32(mask, _dx2, _dx0);
+                _dy2 = vbslq_f32(mask, _dy2, _dy0);
+
+                vst1q_f32(dbuf + x, _dx2);
+                vst1q_f32(dbuf + x + width, _dy2);
+            }
 #endif
             for( ; x < width; x++ )
             {
@@ -600,6 +669,19 @@ void HOGCache::init(const HOGDescriptor* _descriptor,
             idx = _mm_add_epi32(idx, ifour);
             _mm_storeu_ps(_di + i, t);
         }
+    #elif CV_NEON
+        const int a[] = { 0, 1, 2, 3 };
+        int32x4_t idx = vld1q_s32(a);
+        float32x4_t _bw = vdupq_n_f32(bw), _bh = vdupq_n_f32(bh);
+        int32x4_t ifour = vdupq_n_s32(4);
+
+        for (; i <= blockSize.height - 4; i += 4)
+        {
+            float32x4_t t = vsubq_f32(vcvtq_f32_s32(idx), _bh);
+            t = vmulq_f32(t, t);
+            idx = vaddq_s32(idx, ifour);
+            vst1q_f32(_di + i, t);
+        }
     #endif
         for ( ; i < blockSize.height; ++i)
         {
@@ -617,6 +699,15 @@ void HOGCache::init(const HOGDescriptor* _descriptor,
             idx = _mm_add_epi32(idx, ifour);
             _mm_storeu_ps(_dj + j, t);
         }
+    #elif CV_NEON
+        idx = vld1q_s32(a);
+        for (; j <= blockSize.width - 4; j += 4)
+        {
+            float32x4_t t = vsubq_f32(vcvtq_f32_s32(idx), _bw);
+            t = vmulq_f32(t, t);
+            idx = vaddq_s32(idx, ifour);
+            vst1q_f32(_dj + j, t);
+        }
     #endif
         for ( ; j < blockSize.width; ++j)
         {
@@ -839,6 +930,31 @@ const float* HOGCache::getBlock(Point pt, float* buf)
         t1 = hist[h1] + hist1[1];
         hist[h0] = t0; hist[h1] = t1;
     }
+#elif CV_NEON
+    float hist0[4], hist1[4];
+    for( ; k < C2; k++ )
+    {
+        const PixData& pk = _pixData[k];
+        const float* const a = gradPtr + pk.gradOfs;
+        const uchar* const h = qanglePtr + pk.qangleOfs;
+        int h0 = h[0], h1 = h[1];
+
+        float32x4_t _a0 = vdupq_n_f32(a[0]), _a1 = vdupq_n_f32(a[1]);
+        float32x4_t _w = vmulq_f32(vdupq_n_f32(pk.gradWeight), vld1q_f32(pk.histWeights));
+
+        float32x4_t _h0 = vsetq_f32((blockHist + pk.histOfs[0])[h0], (blockHist + pk.histOfs[1])[h0], 0,  0);
+        float32x4_t _h1 = vsetq_f32((blockHist + pk.histOfs[0])[h1], (blockHist + pk.histOfs[1])[h1], 0,  0);
+
+        float32x4_t _t0 = vmlaq_f32(_h0, _a0, _w), _t1 = vmlaq_f32(_h1, _a1, _w);
+        vst1q_f32(hist0, _t0);
+        vst1q_f32(hist1, _t1);
+
+        (blockHist + pk.histOfs[0])[h0] = hist0[0];
+        (blockHist + pk.histOfs[1])[h0] = hist0[1];
+
+        (blockHist + pk.histOfs[0])[h1] = hist1[0];
+        (blockHist + pk.histOfs[1])[h1] = hist1[1];
+    }
 #else
     for( ; k < C2; k++ )
     {
@@ -918,6 +1034,41 @@ const float* HOGCache::getBlock(Point pt, float* buf)
 //        (pk.histOfs[2] + blockHist)[h1] = hist1[2];
 //        (pk.histOfs[3] + blockHist)[h1] = hist1[3];
     }
+#elif CV_NEON
+    for( ; k < C4; k++ )
+    {
+        const PixData& pk = _pixData[k];
+        const float* const a = gradPtr + pk.gradOfs;
+        const uchar* const h = qanglePtr + pk.qangleOfs;
+        int h0 = h[0], h1 = h[1];
+
+        float32x4_t _a0 = vdupq_n_f32(a[0]), _a1 = vdupq_n_f32(a[1]);
+        float32x4_t _w = vmulq_f32(vdupq_n_f32(pk.gradWeight), vld1q_f32(pk.histWeights));
+
+        float32x4_t _h0 = vsetq_f32((blockHist + pk.histOfs[0])[h0],
+                                    (blockHist + pk.histOfs[1])[h0],
+                                    (blockHist + pk.histOfs[2])[h0],
+                                    (blockHist + pk.histOfs[3])[h0]);
+        float32x4_t _h1 = vsetq_f32((blockHist + pk.histOfs[0])[h1],
+                                    (blockHist + pk.histOfs[1])[h1],
+                                    (blockHist + pk.histOfs[2])[h1],
+                                    (blockHist + pk.histOfs[3])[h1]);
+
+
+        float32x4_t _t0 = vmlaq_f32(_h0, _a0, _w), _t1 = vmlaq_f32(_h1, _a1, _w);
+        vst1q_f32(hist0, _t0);
+        vst1q_f32(hist1, _t1);
+
+        (blockHist + pk.histOfs[0])[h0] = hist0[0];
+        (blockHist + pk.histOfs[1])[h0] = hist0[1];
+        (blockHist + pk.histOfs[2])[h0] = hist0[2];
+        (blockHist + pk.histOfs[3])[h0] = hist0[3];
+
+        (blockHist + pk.histOfs[0])[h1] = hist1[0];
+        (blockHist + pk.histOfs[1])[h1] = hist1[1];
+        (blockHist + pk.histOfs[2])[h1] = hist1[2];
+        (blockHist + pk.histOfs[3])[h1] = hist1[3];
+    }
 #else
     for( ; k < C4; k++ )
     {
@@ -973,6 +1124,16 @@ void HOGCache::normalizeBlockHistogram(float* _hist) const
         s = _mm_add_ps(s, _mm_mul_ps(p0, p0));
     }
     _mm_storeu_ps(partSum, s);
+#elif CV_NEON
+    float32x4_t p0 = vld1q_f32(hist);
+    float32x4_t s = vmulq_f32(p0, p0);
+
+    for (i = 4; i <= sz - 4; i += 4)
+    {
+        p0 = vld1q_f32(hist + i);
+        s = vaddq_f32(s, vmulq_f32(p0, p0));
+    }
+    vst1q_f32(partSum, s);
 #else
     partSum[0] = 0.0f;
     partSum[1] = 0.0f;
@@ -1014,6 +1175,25 @@ void HOGCache::normalizeBlockHistogram(float* _hist) const
     }
 
     _mm_storeu_ps(partSum, s);
+#elif CV_NEON
+    float32x4_t _scale = vdupq_n_f32(scale);
+    static float32x4_t _threshold = vdupq_n_f32(thresh);
+
+    float32x4_t p = vmulq_f32(_scale, vld1q_f32(hist));
+    p = vminq_f32(p, _threshold);
+    s = vmulq_f32(p, p);
+    vst1q_f32(hist, p);
+
+    for(i = 4 ; i <= sz - 4; i += 4)
+    {
+        p = vld1q_f32(hist + i);
+        p = vmulq_f32(p, _scale);
+        p = vminq_f32(p, _threshold);
+        s = vaddq_f32(s, vmulq_f32(p, p));
+        vst1q_f32(hist + i, p);
+    }
+
+    vst1q_f32(partSum, s);
 #else
     partSum[0] = 0.0f;
     partSum[1] = 0.0f;
@@ -1048,6 +1228,13 @@ void HOGCache::normalizeBlockHistogram(float* _hist) const
         __m128 t = _mm_mul_ps(_scale2, _mm_loadu_ps(hist + i));
         _mm_storeu_ps(hist + i, t);
     }
+#elif CV_NEON
+    float32x4_t _scale2 = vdupq_n_f32(scale);
+    for ( ; i <= sz - 4; i += 4)
+    {
+        float32x4_t t = vmulq_f32(_scale2, vld1q_f32(hist + i));
+        vst1q_f32(hist + i, t);
+    }
 #endif
     for ( ; i < sz; ++i)
         hist[i] *= scale;
@@ -1489,7 +1676,7 @@ void HOGDescriptor::detect(const Mat& img,
     double rho = svmDetector.size() > dsize ? svmDetector[dsize] : 0;
     std::vector<float> blockHist(blockHistogramSize);
 
-#if CV_SSE2
+#if CV_SSE2 || CV_NEON
     float partSum[4];
 #endif
 
@@ -1535,6 +1722,23 @@ void HOGDescriptor::detect(const Mat& img,
             double t0 = partSum[0] + partSum[1];
             double t1 = partSum[2] + partSum[3];
             s += t0 + t1;
+#elif CV_NEON
+            float32x4_t _vec = vld1q_f32(vec);
+            float32x4_t _svmVec = vld1q_f32(svmVec);
+            float32x4_t sum = vmulq_f32(_svmVec, _vec);
+
+            for( k = 4; k <= blockHistogramSize - 4; k += 4 )
+            {
+                _vec = vld1q_f32(vec + k);
+                _svmVec = vld1q_f32(svmVec + k);
+
+                sum = vaddq_f32(sum, vmulq_f32(_vec, _svmVec));
+            }
+
+            vst1q_f32(partSum, sum);
+            double t0 = partSum[0] + partSum[1];
+            double t1 = partSum[2] + partSum[3];
+            s += t0 + t1;
 #else
             for( k = 0; k <= blockHistogramSize - 4; k += 4 )
                 s += vec[k]*svmVec[k] + vec[k+1]*svmVec[k+1] +
@@ -3357,7 +3561,7 @@ void HOGDescriptor::detectROI(const cv::Mat& img, const std::vector<cv::Point> &
     double rho = svmDetector.size() > dsize ? svmDetector[dsize] : 0;
     std::vector<float> blockHist(blockHistogramSize);
 
-#if CV_SSE2
+#if CV_SSE2 || CV_NEON
     float partSum[4];
 #endif
 
@@ -3401,6 +3605,23 @@ void HOGDescriptor::detectROI(const cv::Mat& img, const std::vector<cv::Point> &
             double t0 = partSum[0] + partSum[1];
             double t1 = partSum[2] + partSum[3];
             s += t0 + t1;
+#elif CV_NEON
+            float32x4_t _vec = vld1q_f32(vec);
+            float32x4_t _svmVec = vld1q_f32(svmVec);
+            float32x4_t sum = vmulq_f32(_svmVec, _vec);
+
+            for( k = 4; k <= blockHistogramSize - 4; k += 4 )
+            {
+                _vec = vld1q_f32(vec + k);
+                _svmVec = vld1q_f32(svmVec + k);
+
+                sum = vaddq_f32(sum, vmulq_f32(_vec, _svmVec));
+            }
+
+            vst1q_f32(partSum, sum);
+            double t0 = partSum[0] + partSum[1];
+            double t1 = partSum[2] + partSum[3];
+            s += t0 + t1;
 #else
             for( k = 0; k <= blockHistogramSize - 4; k += 4 )
                 s += vec[k]*svmVec[k] + vec[k+1]*svmVec[k+1] +