Update to 2.0.0 tree from current Fremantle build
[opencv] / src / cv / cvthresh.cpp
diff --git a/src/cv/cvthresh.cpp b/src/cv/cvthresh.cpp
new file mode 100644 (file)
index 0000000..ca34ca1
--- /dev/null
@@ -0,0 +1,579 @@
+/*M///////////////////////////////////////////////////////////////////////////////////////
+//
+//  IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
+//
+//  By downloading, copying, installing or using the software you agree to this license.
+//  If you do not agree to this license, do not download, install,
+//  copy or use the software.
+//
+//
+//                           License Agreement
+//                For Open Source Computer Vision Library
+//
+// Copyright (C) 2000-2008, Intel Corporation, all rights reserved.
+// Copyright (C) 2009, Willow Garage Inc., all rights reserved.
+// Third party copyrights are property of their respective owners.
+//
+// Redistribution and use in source and binary forms, with or without modification,
+// are permitted provided that the following conditions are met:
+//
+//   * Redistribution's of source code must retain the above copyright notice,
+//     this list of conditions and the following disclaimer.
+//
+//   * Redistribution's in binary form must reproduce the above copyright notice,
+//     this list of conditions and the following disclaimer in the documentation
+//     and/or other materials provided with the distribution.
+//
+//   * The name of the copyright holders may not be used to endorse or promote products
+//     derived from this software without specific prior written permission.
+//
+// This software is provided by the copyright holders and contributors "as is" and
+// any express or implied warranties, including, but not limited to, the implied
+// warranties of merchantability and fitness for a particular purpose are disclaimed.
+// In no event shall the Intel Corporation or contributors be liable for any direct,
+// indirect, incidental, special, exemplary, or consequential damages
+// (including, but not limited to, procurement of substitute goods or services;
+// loss of use, data, or profits; or business interruption) however caused
+// and on any theory of liability, whether in contract, strict liability,
+// or tort (including negligence or otherwise) arising in any way out of
+// the use of this software, even if advised of the possibility of such damage.
+//
+//M*/
+
+#include "_cv.h"
+
+namespace cv
+{
+
+static void
+thresh_8u( const Mat& _src, Mat& _dst, uchar thresh, uchar maxval, int type )
+{
+    int i, j;
+    uchar tab[256];
+    Size roi = _src.size();
+    roi.width *= _src.channels();
+#if CV_SSE2
+    __m128i _x80 = _mm_set1_epi8('\x80');
+    __m128i thresh_u = _mm_set1_epi8(thresh);
+    __m128i thresh_s = _mm_set1_epi8(thresh ^ 0x80);
+    __m128i maxval_ = _mm_set1_epi8(maxval);
+#endif
+
+    if( _src.isContinuous() && _dst.isContinuous() )
+    {
+        roi.width *= roi.height;
+        roi.height = 1;
+    }
+
+    switch( type )
+    {
+    case THRESH_BINARY:
+        for( i = 0; i <= thresh; i++ )
+            tab[i] = 0;
+        for( ; i < 256; i++ )
+            tab[i] = maxval;
+        break;
+    case THRESH_BINARY_INV:
+        for( i = 0; i <= thresh; i++ )
+            tab[i] = maxval;
+        for( ; i < 256; i++ )
+            tab[i] = 0;
+        break;
+    case THRESH_TRUNC:
+        for( i = 0; i <= thresh; i++ )
+            tab[i] = (uchar)i;
+        for( ; i < 256; i++ )
+            tab[i] = thresh;
+        break;
+    case THRESH_TOZERO:
+        for( i = 0; i <= thresh; i++ )
+            tab[i] = 0;
+        for( ; i < 256; i++ )
+            tab[i] = (uchar)i;
+        break;
+    case THRESH_TOZERO_INV:
+        for( i = 0; i <= thresh; i++ )
+            tab[i] = (uchar)i;
+        for( ; i < 256; i++ )
+            tab[i] = 0;
+        break;
+    default:
+        CV_Error( CV_StsBadArg, "Unknown threshold type" );
+    }
+
+    for( i = 0; i < roi.height; i++ )
+    {
+        const uchar* src = (const uchar*)(_src.data + _src.step*i);
+        uchar* dst = (uchar*)(_dst.data + _dst.step*i);
+        j = 0;
+
+    #if CV_SSE2
+        switch( type )
+        {
+        case THRESH_BINARY:
+            for( ; j <= roi.width - 32; j += 32 )
+            {
+                __m128i v0, v1;
+                v0 = _mm_loadu_si128( (const __m128i*)(src + j) );
+                v1 = _mm_loadu_si128( (const __m128i*)(src + j + 16) );
+                v0 = _mm_cmpgt_epi8( _mm_xor_si128(v0, _x80), thresh_s );
+                v1 = _mm_cmpgt_epi8( _mm_xor_si128(v1, _x80), thresh_s );
+                v0 = _mm_and_si128( v0, maxval_ );
+                v1 = _mm_and_si128( v1, maxval_ );
+                _mm_storeu_si128( (__m128i*)(dst + j), v0 );
+                _mm_storeu_si128( (__m128i*)(dst + j + 16), v1 );
+            }
+
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128i v0 = _mm_loadl_epi64( (const __m128i*)(src + j) );
+                v0 = _mm_cmpgt_epi8( _mm_xor_si128(v0, _x80), thresh_s );
+                v0 = _mm_and_si128( v0, maxval_ );
+                _mm_storel_epi64( (__m128i*)(dst + j), v0 );
+            }
+            break;
+
+        case THRESH_BINARY_INV:
+            for( ; j <= roi.width - 32; j += 32 )
+            {
+                __m128i v0, v1;
+                v0 = _mm_loadu_si128( (const __m128i*)(src + j) );
+                v1 = _mm_loadu_si128( (const __m128i*)(src + j + 16) );
+                v0 = _mm_cmpgt_epi8( _mm_xor_si128(v0, _x80), thresh_s );
+                v1 = _mm_cmpgt_epi8( _mm_xor_si128(v1, _x80), thresh_s );
+                v0 = _mm_andnot_si128( v0, maxval_ );
+                v1 = _mm_andnot_si128( v1, maxval_ );
+                _mm_storeu_si128( (__m128i*)(dst + j), v0 );
+                _mm_storeu_si128( (__m128i*)(dst + j + 16), v1 );
+            }
+
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128i v0 = _mm_loadl_epi64( (const __m128i*)(src + j) );
+                v0 = _mm_cmpgt_epi8( _mm_xor_si128(v0, _x80), thresh_s );
+                v0 = _mm_andnot_si128( v0, maxval_ );
+                _mm_storel_epi64( (__m128i*)(dst + j), v0 );
+            }
+            break;
+
+        case THRESH_TRUNC:
+            for( ; j <= roi.width - 32; j += 32 )
+            {
+                __m128i v0, v1;
+                v0 = _mm_loadu_si128( (const __m128i*)(src + j) );
+                v1 = _mm_loadu_si128( (const __m128i*)(src + j + 16) );
+                v0 = _mm_subs_epu8( v0, _mm_subs_epu8( v0, thresh_u ));
+                v1 = _mm_subs_epu8( v1, _mm_subs_epu8( v1, thresh_u ));
+                _mm_storeu_si128( (__m128i*)(dst + j), v0 );
+                _mm_storeu_si128( (__m128i*)(dst + j + 16), v1 );
+            }
+
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128i v0 = _mm_loadl_epi64( (const __m128i*)(src + j) );
+                v0 = _mm_subs_epu8( v0, _mm_subs_epu8( v0, thresh_u ));
+                _mm_storel_epi64( (__m128i*)(dst + j), v0 );
+            }
+            break;
+
+        case THRESH_TOZERO:
+            for( ; j <= roi.width - 32; j += 32 )
+            {
+                __m128i v0, v1;
+                v0 = _mm_loadu_si128( (const __m128i*)(src + j) );
+                v1 = _mm_loadu_si128( (const __m128i*)(src + j + 16) );
+                v0 = _mm_and_si128( v0, _mm_cmpgt_epi8(_mm_xor_si128(v0, _x80), thresh_s ));
+                v1 = _mm_and_si128( v1, _mm_cmpgt_epi8(_mm_xor_si128(v1, _x80), thresh_s ));
+                _mm_storeu_si128( (__m128i*)(dst + j), v0 );
+                _mm_storeu_si128( (__m128i*)(dst + j + 16), v1 );
+            }
+
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128i v0 = _mm_loadl_epi64( (const __m128i*)(src + j) );
+                v0 = _mm_and_si128( v0, _mm_cmpgt_epi8(_mm_xor_si128(v0, _x80), thresh_s ));
+                _mm_storel_epi64( (__m128i*)(dst + j), v0 );
+            }
+            break;
+
+        case THRESH_TOZERO_INV:
+            for( ; j <= roi.width - 32; j += 32 )
+            {
+                __m128i v0, v1;
+                v0 = _mm_loadu_si128( (const __m128i*)(src + j) );
+                v1 = _mm_loadu_si128( (const __m128i*)(src + j + 16) );
+                v0 = _mm_andnot_si128( _mm_cmpgt_epi8(_mm_xor_si128(v0, _x80), thresh_s ), v0 );
+                v1 = _mm_andnot_si128( _mm_cmpgt_epi8(_mm_xor_si128(v1, _x80), thresh_s ), v1 );
+                _mm_storeu_si128( (__m128i*)(dst + j), v0 );
+                _mm_storeu_si128( (__m128i*)(dst + j + 16), v1 );
+            }
+
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128i v0 = _mm_loadl_epi64( (const __m128i*)(src + j) );
+                v0 = _mm_andnot_si128( _mm_cmpgt_epi8(_mm_xor_si128(v0, _x80), thresh_s ), v0 );
+                _mm_storel_epi64( (__m128i*)(dst + j), v0 );
+            }
+            break;
+        }
+    #endif        
+
+        for( ; j <= roi.width - 4; j += 4 )
+        {
+            uchar t0 = tab[src[j]];
+            uchar t1 = tab[src[j+1]];
+
+            dst[j] = t0;
+            dst[j+1] = t1;
+
+            t0 = tab[src[j+2]];
+            t1 = tab[src[j+3]];
+
+            dst[j+2] = t0;
+            dst[j+3] = t1;
+        }
+
+        for( ; j < roi.width; j++ )
+            dst[j] = tab[src[j]];
+    }
+}
+
+
+static void
+thresh_32f( const Mat& _src, Mat& _dst, float thresh, float maxval, int type )
+{
+    int i, j;
+    Size roi = _src.size();
+    roi.width *= _src.channels();
+    const float* src = (const float*)_src.data;
+    float* dst = (float*)_dst.data;
+    size_t src_step = _src.step/sizeof(src[0]);
+    size_t dst_step = _dst.step/sizeof(dst[0]);
+#if CV_SSE2
+    __m128 thresh4 = _mm_set1_ps(thresh), maxval4 = _mm_set1_ps(maxval);
+#endif
+
+    if( _src.isContinuous() && _dst.isContinuous() )
+    {
+        roi.width *= roi.height;
+        roi.height = 1;
+    }
+
+    switch( type )
+    {
+    case THRESH_BINARY:
+        for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
+        {
+            j = 0;
+        #if CV_SSE2
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128 v0, v1;
+                v0 = _mm_loadu_ps( src + j );
+                v1 = _mm_loadu_ps( src + j + 4 );
+                v0 = _mm_cmpgt_ps( v0, thresh4 );
+                v1 = _mm_cmpgt_ps( v1, thresh4 );
+                v0 = _mm_and_ps( v0, maxval4 );
+                v1 = _mm_and_ps( v1, maxval4 );
+                _mm_storeu_ps( dst + j, v0 );
+                _mm_storeu_ps( dst + j + 4, v1 );
+            }
+        #endif
+
+            for( ; j < roi.width; j++ )
+                dst[j] = src[j] > thresh ? maxval : 0;
+        }
+        break;
+
+    case THRESH_BINARY_INV:
+        for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
+        {
+            j = 0;
+        #if CV_SSE2
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128 v0, v1;
+                v0 = _mm_loadu_ps( src + j );
+                v1 = _mm_loadu_ps( src + j + 4 );
+                v0 = _mm_cmple_ps( v0, thresh4 );
+                v1 = _mm_cmple_ps( v1, thresh4 );
+                v0 = _mm_and_ps( v0, maxval4 );
+                v1 = _mm_and_ps( v1, maxval4 );
+                _mm_storeu_ps( dst + j, v0 );
+                _mm_storeu_ps( dst + j + 4, v1 );
+            }
+        #endif            
+            
+            for( ; j < roi.width; j++ )
+                dst[j] = src[j] <= thresh ? maxval : 0;
+        }
+        break;
+
+    case THRESH_TRUNC:
+        for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
+        {
+            j = 0;
+        #if CV_SSE2
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128 v0, v1;
+                v0 = _mm_loadu_ps( src + j );
+                v1 = _mm_loadu_ps( src + j + 4 );
+                v0 = _mm_min_ps( v0, thresh4 );
+                v1 = _mm_min_ps( v1, thresh4 );
+                _mm_storeu_ps( dst + j, v0 );
+                _mm_storeu_ps( dst + j + 4, v1 );
+            }
+        #endif            
+            
+            for( ; j < roi.width; j++ )
+                dst[j] = std::min(src[j], thresh);
+        }
+        break;
+
+    case THRESH_TOZERO:
+        for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
+        {
+            j = 0;
+        #if CV_SSE2
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128 v0, v1;
+                v0 = _mm_loadu_ps( src + j );
+                v1 = _mm_loadu_ps( src + j + 4 );
+                v0 = _mm_and_ps(v0, _mm_cmpgt_ps(v0, thresh4));
+                v1 = _mm_and_ps(v1, _mm_cmpgt_ps(v1, thresh4));
+                _mm_storeu_ps( dst + j, v0 );
+                _mm_storeu_ps( dst + j + 4, v1 );
+            }
+        #endif
+            
+            for( ; j < roi.width; j++ )
+            {
+                float v = src[j];
+                dst[j] = v > thresh ? v : 0;
+            }
+        }
+        break;
+
+    case THRESH_TOZERO_INV:
+        for( i = 0; i < roi.height; i++, src += src_step, dst += dst_step )
+        {
+            j = 0;
+        #if CV_SSE2
+            for( ; j <= roi.width - 8; j += 8 )
+            {
+                __m128 v0, v1;
+                v0 = _mm_loadu_ps( src + j );
+                v1 = _mm_loadu_ps( src + j + 4 );
+                v0 = _mm_and_ps(v0, _mm_cmple_ps(v0, thresh4));
+                v1 = _mm_and_ps(v1, _mm_cmple_ps(v1, thresh4));
+                _mm_storeu_ps( dst + j, v0 );
+                _mm_storeu_ps( dst + j + 4, v1 );
+            }
+        #endif
+            for( ; j < roi.width; j++ )
+            {
+                float v = src[j];
+                dst[j] = v <= thresh ? v : 0;
+            }
+        }
+        break;
+    default:
+        return CV_Error( CV_StsBadArg, "" );
+    }
+}
+
+
+static double
+getThreshVal_Otsu_8u( const Mat& _src )
+{
+    Size size = _src.size();
+    if( _src.isContinuous() )
+    {
+        size.width *= size.height;
+        size.height = 1;
+    }
+    const int N = 256;
+    int i, j, h[N] = {0};
+    for( i = 0; i < size.height; i++ )
+    {
+        const uchar* src = _src.data + _src.step*i;
+        for( j = 0; j <= size.width - 4; j += 4 )
+        {
+            int v0 = src[j], v1 = src[j+1];
+            h[v0]++; h[v1]++;
+            v0 = src[j+2]; v1 = src[j+3];
+            h[v0]++; h[v1]++;
+        }
+        for( ; j < size.width; j++ )
+            h[src[j]]++;
+    }
+
+    double mu = 0, scale = 1./(size.width*size.height);
+    for( i = 0; i < N; i++ )
+        mu += i*h[i];
+    
+    mu *= scale;
+    double mu1 = 0, q1 = 0;
+    double max_sigma = 0, max_val = 0;
+
+    for( i = 0; i < N; i++ )
+    {
+        double p_i, q2, mu2, sigma;
+
+        p_i = h[i]*scale;
+        mu1 *= q1;
+        q1 += p_i;
+        q2 = 1. - q1;
+
+        if( std::min(q1,q2) < FLT_EPSILON || std::max(q1,q2) > 1. - FLT_EPSILON )
+            continue;
+
+        mu1 = (mu1 + i*p_i)/q1;
+        mu2 = (mu - q1*mu1)/q2;
+        sigma = q1*q2*(mu1 - mu2)*(mu1 - mu2);
+        if( sigma > max_sigma )
+        {
+            max_sigma = sigma;
+            max_val = i;
+        }
+    }
+
+    return max_val;
+}
+
+
+double threshold( const Mat& _src, Mat& _dst, double thresh, double maxval, int type )
+{
+    bool use_otsu = (type & THRESH_OTSU) != 0;
+    type &= THRESH_MASK;
+
+    if( use_otsu )
+    {
+        CV_Assert( _src.type() == CV_8UC1 );
+        thresh = getThreshVal_Otsu_8u(_src);
+    }
+  
+    _dst.create( _src.size(), _src.type() );
+    if( _src.depth() == CV_8U )
+    {
+        int ithresh = cvFloor(thresh);
+        thresh = ithresh;
+        int imaxval = cvRound(maxval);
+        if( type == THRESH_TRUNC )
+            imaxval = ithresh;
+        imaxval = saturate_cast<uchar>(imaxval);
+
+        if( ithresh < 0 || ithresh >= 255 )
+        {
+            if( type == THRESH_BINARY || type == THRESH_BINARY_INV ||
+                ((type == THRESH_TRUNC || type == THRESH_TOZERO_INV) && ithresh < 0) ||
+                (type == THRESH_TOZERO && ithresh >= 255) )
+            {
+                int v = type == THRESH_BINARY ? (ithresh >= 255 ? 0 : imaxval) :
+                        type == THRESH_BINARY_INV ? (ithresh >= 255 ? imaxval : 0) :
+                        type == THRESH_TRUNC ? imaxval : 0;
+                _dst = Scalar::all(v);
+            }
+            else
+                _src.copyTo(_dst);
+        }
+        else
+            thresh_8u( _src, _dst, (uchar)ithresh, (uchar)imaxval, type );
+    }
+    else if( _src.depth() == CV_32F )
+        thresh_32f( _src, _dst, (float)thresh, (float)maxval, type );
+    else
+        CV_Error( CV_StsUnsupportedFormat, "" );
+
+    return thresh;
+}
+
+
+void adaptiveThreshold( const Mat& _src, Mat& _dst, double maxValue,
+                        int method, int type, int blockSize, double delta )
+{
+    CV_Assert( _src.type() == CV_8UC1 );
+    CV_Assert( blockSize % 2 == 1 && blockSize > 1 );
+    Size size = _src.size();
+
+    _dst.create( size, _src.type() );
+
+    if( maxValue < 0 )
+    {
+        _dst = Scalar(0);
+        return;
+    }
+    
+    Mat _mean;
+
+    if( _src.data != _dst.data )
+        _mean = _dst;
+
+    if( method == ADAPTIVE_THRESH_MEAN_C )
+        boxFilter( _src, _mean, _src.type(), Size(blockSize, blockSize),
+                   Point(-1,-1), true, BORDER_REPLICATE );
+    else if( method == ADAPTIVE_THRESH_GAUSSIAN_C )
+        GaussianBlur( _src, _mean, Size(blockSize, blockSize), 0, 0, BORDER_REPLICATE );
+    else
+        CV_Error( CV_StsBadFlag, "Unknown/unsupported adaptive threshold method" );
+
+    int i, j;
+    uchar imaxval = saturate_cast<uchar>(maxValue);
+    int idelta = type == THRESH_BINARY ? cvCeil(delta) : cvFloor(delta);
+    uchar tab[768];    
+
+    if( type == CV_THRESH_BINARY )
+        for( i = 0; i < 768; i++ )
+            tab[i] = (uchar)(i - 255 > -idelta ? imaxval : 0);
+    else if( type == CV_THRESH_BINARY_INV )
+        for( i = 0; i < 768; i++ )
+            tab[i] = (uchar)(i - 255 <= -idelta ? imaxval : 0);
+    else
+        CV_Error( CV_StsBadFlag, "Unknown/unsupported threshold type" );
+
+    if( _src.isContinuous() && _mean.isContinuous() && _dst.isContinuous() )
+    {
+        size.width *= size.height;
+        size.height = 1;
+    }
+
+    for( i = 0; i < size.height; i++ )
+    {
+        const uchar* src = _src.data + _src.step*i;
+        const uchar* mean = _mean.data + _mean.step*i;
+        uchar* dst = _dst.data + _dst.step*i;
+
+        for( j = 0; j < size.width; j++ )
+            dst[j] = tab[src[j] - mean[j] + 255];
+    }
+}
+
+}
+
+CV_IMPL double
+cvThreshold( const void* srcarr, void* dstarr, double thresh, double maxval, int type )
+{
+    cv::Mat src = cv::cvarrToMat(srcarr), dst = cv::cvarrToMat(dstarr), dst0 = dst;
+
+    CV_Assert( src.size() == dst.size() && src.channels() == dst.channels() &&
+        (src.depth() == dst.depth() || dst.depth() == CV_8U));
+
+    thresh = cv::threshold( src, dst, thresh, maxval, type );
+    if( dst0.data != dst.data )
+        dst.convertTo( dst0, dst0.depth() );
+    return thresh;
+}
+
+
+CV_IMPL void
+cvAdaptiveThreshold( const void *srcIm, void *dstIm, double maxValue,
+                     int method, int type, int blockSize, double delta )
+{
+    cv::Mat src = cv::cvarrToMat(srcIm), dst = cv::cvarrToMat(dstIm);
+    CV_Assert( src.size() == dst.size() && src.type() == dst.type() );
+    cv::adaptiveThreshold( src, dst, maxValue, method, type, blockSize, delta );
+}
+
+/* End of file. */