From ae02ecec6003af181abdee97e76dfdaf9d902248 Mon Sep 17 00:00:00 2001 From: Rahul Kavi Date: Mon, 11 Nov 2013 15:29:19 -0500 Subject: [PATCH] removed a couple of unnecessary comments in Logistic Regression training method --- modules/ml/src/lr.cpp | 2 -- 1 file changed, 2 deletions(-) diff --git a/modules/ml/src/lr.cpp b/modules/ml/src/lr.cpp index 09acf9d..2411ea3 100644 --- a/modules/ml/src/lr.cpp +++ b/modules/ml/src/lr.cpp @@ -159,7 +159,6 @@ bool LogisticRegression::train(cv::InputArray data_ip, cv::InputArray labels_ip) if(num_classes == 2) { labels_l.convertTo(labels, CV_32F); - // new_theta = compute_batch_gradient(data_t, labels, init_theta); //currently supported training methods LogisticRegression::BATCH and LogisticRegression::MINI_BATCH if(this->params.train_method == LogisticRegression::BATCH) new_theta = compute_batch_gradient(data_t, labels, init_theta); @@ -177,7 +176,6 @@ bool LogisticRegression::train(cv::InputArray data_ip, cv::InputArray labels_ip) { new_local_labels = (labels_l == it->second)/255; new_local_labels.convertTo(labels, CV_32F); - // new_theta = compute_batch_gradient(data_t, labels, init_theta); // currently supported training methods LogisticRegression::BATCH and LogisticRegression::MINI_BATCH if(this->params.train_method == LogisticRegression::BATCH) new_theta = compute_batch_gradient(data_t, labels, init_theta); -- 2.7.4