From 8c58e415b40e16e5f13b23cf2790cc63ca59ca12 Mon Sep 17 00:00:00 2001 From: Jeff Donahue Date: Tue, 25 Feb 2014 20:39:01 -0800 Subject: [PATCH] fix linter errors in examples --- examples/cifar/convert_cifar_data.cpp | 47 +++++++++++++++++------------------ examples/lenet/convert_mnist_data.cpp | 28 ++++++++++----------- tools/compute_image_mean.cpp | 3 ++- tools/convert_imageset.cpp | 23 +++++++++-------- tools/device_query.cpp | 2 +- tools/dump_network.cpp | 20 +++++++++------ tools/finetune_net.cpp | 4 +-- tools/net_speed_benchmark.cpp | 37 +++++++++++++++++---------- tools/test_net.cpp | 6 +++-- tools/train_net.cpp | 2 +- 10 files changed, 96 insertions(+), 76 deletions(-) diff --git a/examples/cifar/convert_cifar_data.cpp b/examples/cifar/convert_cifar_data.cpp index 083ea9e..648dd37 100644 --- a/examples/cifar/convert_cifar_data.cpp +++ b/examples/cifar/convert_cifar_data.cpp @@ -12,25 +12,23 @@ #include #include -#include -#include +#include // NOLINT(readability/streams) #include #include "caffe/proto/caffe.pb.h" using std::string; +const int kCIFARSize = 32; +const int kCIFARImageNBytes = 3072; +const int kCIFARBatchSize = 10000; +const int kCIFARTrainBatches = 5; -const int kCIFAR_SIZE=32; -const int kCIFAR_IMAGE_NBYTES=3072; -const int kCIFAR_BATCHSIZE=10000; -const int kCIFAR_TRAIN_BATCHES=5; - -void read_image(std::ifstream& file, int* label, char* buffer) { +void read_image(std::ifstream* file, int* label, char* buffer) { char label_char; - file.read(&label_char, 1); + file->read(&label_char, 1); *label = label_char; - file.read(buffer, kCIFAR_IMAGE_NBYTES); + file->read(buffer, kCIFARImageNBytes); return; } @@ -41,12 +39,12 @@ void convert_dataset(const string& input_folder, const string& output_folder) { options.error_if_exists = true; // Data buffer int label; - char str_buffer[kCIFAR_IMAGE_NBYTES]; + char str_buffer[kCIFARImageNBytes]; string value; caffe::Datum datum; datum.set_channels(3); - datum.set_height(kCIFAR_SIZE); - datum.set_width(kCIFAR_SIZE); + datum.set_height(kCIFARSize); + datum.set_width(kCIFARSize); LOG(INFO) << "Writing Training data"; leveldb::DB* train_db; @@ -54,19 +52,20 @@ void convert_dataset(const string& input_folder, const string& output_folder) { status = leveldb::DB::Open(options, output_folder + "/cifar-train-leveldb", &train_db); CHECK(status.ok()) << "Failed to open leveldb."; - for (int fileid = 0; fileid < kCIFAR_TRAIN_BATCHES; ++fileid) { + for (int fileid = 0; fileid < kCIFARTrainBatches; ++fileid) { // Open files LOG(INFO) << "Training Batch " << fileid + 1; - sprintf(str_buffer, "/data_batch_%d.bin", fileid + 1); + snprintf(str_buffer, kCIFARImageNBytes, "/data_batch_%d.bin", fileid + 1); std::ifstream data_file((input_folder + str_buffer).c_str(), std::ios::in | std::ios::binary); CHECK(data_file) << "Unable to open train file #" << fileid + 1; - for (int itemid = 0; itemid < kCIFAR_BATCHSIZE; ++itemid) { - read_image(data_file, &label, str_buffer); + for (int itemid = 0; itemid < kCIFARBatchSize; ++itemid) { + read_image(&data_file, &label, str_buffer); datum.set_label(label); - datum.set_data(str_buffer, kCIFAR_IMAGE_NBYTES); + datum.set_data(str_buffer, kCIFARImageNBytes); datum.SerializeToString(&value); - sprintf(str_buffer, "%05d", fileid * kCIFAR_BATCHSIZE + itemid); + snprintf(str_buffer, kCIFARImageNBytes, "%05d", + fileid * kCIFARBatchSize + itemid); train_db->Put(leveldb::WriteOptions(), string(str_buffer), value); } } @@ -79,12 +78,12 @@ void convert_dataset(const string& input_folder, const string& output_folder) { std::ifstream data_file((input_folder + "/test_batch.bin").c_str(), std::ios::in | std::ios::binary); CHECK(data_file) << "Unable to open test file."; - for (int itemid = 0; itemid < kCIFAR_BATCHSIZE; ++itemid) { - read_image(data_file, &label, str_buffer); + for (int itemid = 0; itemid < kCIFARBatchSize; ++itemid) { + read_image(&data_file, &label, str_buffer); datum.set_label(label); - datum.set_data(str_buffer, kCIFAR_IMAGE_NBYTES); + datum.set_data(str_buffer, kCIFARImageNBytes); datum.SerializeToString(&value); - sprintf(str_buffer, "%05d", itemid); + snprintf(str_buffer, kCIFARImageNBytes, "%05d", itemid); test_db->Put(leveldb::WriteOptions(), string(str_buffer), value); } @@ -92,7 +91,7 @@ void convert_dataset(const string& input_folder, const string& output_folder) { delete test_db; } -int main (int argc, char** argv) { +int main(int argc, char** argv) { if (argc != 3) { printf("This script converts the CIFAR dataset to the leveldb format used\n" "by caffe to perform classification.\n" diff --git a/examples/lenet/convert_mnist_data.cpp b/examples/lenet/convert_mnist_data.cpp index 32d9b9d..1bf1d66 100644 --- a/examples/lenet/convert_mnist_data.cpp +++ b/examples/lenet/convert_mnist_data.cpp @@ -12,14 +12,13 @@ #include #include -#include -#include +#include // NOLINT(readability/streams) +#include #include "caffe/proto/caffe.pb.h" -uint32_t swap_endian( uint32_t val ) -{ - val = ((val << 8) & 0xFF00FF00 ) | ((val >> 8) & 0xFF00FF ); +uint32_t swap_endian(uint32_t val) { + val = ((val << 8) & 0xFF00FF00) | ((val >> 8) & 0xFF00FF); return (val << 16) | (val >> 16); } @@ -37,20 +36,20 @@ void convert_dataset(const char* image_filename, const char* label_filename, uint32_t rows; uint32_t cols; - image_file.read((char*)(&magic), 4); + image_file.read(reinterpret_cast(&magic), 4); magic = swap_endian(magic); CHECK_EQ(magic, 2051) << "Incorrect image file magic."; - label_file.read((char*)(&magic), 4); + label_file.read(reinterpret_cast(&magic), 4); magic = swap_endian(magic); CHECK_EQ(magic, 2049) << "Incorrect label file magic."; - image_file.read((char*)(&num_items), 4); + image_file.read(reinterpret_cast(&num_items), 4); num_items = swap_endian(num_items); - label_file.read((char*)(&num_labels), 4); + label_file.read(reinterpret_cast(&num_labels), 4); num_labels = swap_endian(num_labels); CHECK_EQ(num_items, num_labels); - image_file.read((char*)(&rows), 4); + image_file.read(reinterpret_cast(&rows), 4); rows = swap_endian(rows); - image_file.read((char*)(&cols), 4); + image_file.read(reinterpret_cast(&cols), 4); cols = swap_endian(cols); // Open leveldb @@ -65,7 +64,8 @@ void convert_dataset(const char* image_filename, const char* label_filename, char label; char* pixels = new char[rows * cols]; - char key[10]; + const int kMaxKeyLength = 10; + char key[kMaxKeyLength]; std::string value; caffe::Datum datum; @@ -80,7 +80,7 @@ void convert_dataset(const char* image_filename, const char* label_filename, datum.set_data(pixels, rows*cols); datum.set_label(label); datum.SerializeToString(&value); - sprintf(key, "%08d", itemid); + snprintf(key, kMaxKeyLength, "%08d", itemid); db->Put(leveldb::WriteOptions(), std::string(key), value); } @@ -88,7 +88,7 @@ void convert_dataset(const char* image_filename, const char* label_filename, delete pixels; } -int main (int argc, char** argv) { +int main(int argc, char** argv) { if (argc != 4) { printf("This script converts the MNIST dataset to the leveldb format used\n" "by caffe to perform classification.\n" diff --git a/tools/compute_image_mean.cpp b/tools/compute_image_mean.cpp index 3c243d6..65ff538 100644 --- a/tools/compute_image_mean.cpp +++ b/tools/compute_image_mean.cpp @@ -48,7 +48,8 @@ int main(int argc, char** argv) { // just a dummy operation datum.ParseFromString(it->value().ToString()); const string& data = datum.data(); - CHECK_EQ(data.size(), data_size) << "Incorrect data field size " << data.size(); + CHECK_EQ(data.size(), data_size) << "Incorrect data field size " + << data.size(); for (int i = 0; i < data.size(); ++i) { sum_blob.set_data(i, sum_blob.data(i) + (uint8_t)data[i]); } diff --git a/tools/convert_imageset.cpp b/tools/convert_imageset.cpp index e739716..50b53e6 100644 --- a/tools/convert_imageset.cpp +++ b/tools/convert_imageset.cpp @@ -16,14 +16,15 @@ #include #include +#include // NOLINT(readability/streams) #include -#include -#include +#include +#include #include "caffe/proto/caffe.pb.h" #include "caffe/util/io.hpp" -using namespace caffe; +using namespace caffe; // NOLINT(build/namespaces) using std::pair; using std::string; @@ -66,25 +67,27 @@ int main(int argc, char** argv) { string root_folder(argv[1]); Datum datum; int count = 0; - const int maxKeyLength = 256; - char key_cstr[maxKeyLength]; + const int kMaxKeyLength = 256; + char key_cstr[kMaxKeyLength]; leveldb::WriteBatch* batch = new leveldb::WriteBatch(); int data_size; bool data_size_initialized = false; for (int line_id = 0; line_id < lines.size(); ++line_id) { - if (!ReadImageToDatum(root_folder + lines[line_id].first, lines[line_id].second, - &datum)) { + if (!ReadImageToDatum(root_folder + lines[line_id].first, + lines[line_id].second, &datum)) { continue; - }; + } if (!data_size_initialized) { data_size = datum.channels() * datum.height() * datum.width(); data_size_initialized = true; } else { const string& data = datum.data(); - CHECK_EQ(data.size(), data_size) << "Incorrect data field size " << data.size(); + CHECK_EQ(data.size(), data_size) << "Incorrect data field size " + << data.size(); } // sequential - snprintf(key_cstr, maxKeyLength, "%08d_%s", line_id, lines[line_id].first.c_str()); + snprintf(key_cstr, kMaxKeyLength, "%08d_%s", line_id, + lines[line_id].first.c_str()); string value; // get the value datum.SerializeToString(&value); diff --git a/tools/device_query.cpp b/tools/device_query.cpp index 88bf5aa..920e81b 100644 --- a/tools/device_query.cpp +++ b/tools/device_query.cpp @@ -5,7 +5,7 @@ #include "caffe/net.hpp" -using namespace caffe; +using namespace caffe; // NOLINT(build/namespaces) int main(int argc, char** argv) { if (argc > 2) { diff --git a/tools/dump_network.cpp b/tools/dump_network.cpp index 0d6e2d0..48804f2 100644 --- a/tools/dump_network.cpp +++ b/tools/dump_network.cpp @@ -4,16 +4,18 @@ // all the intermediate blobs produced by the net to individual binary // files stored in protobuffer binary formats. // Usage: -// dump_network input_net_param trained_net_param input_blob output_prefix 0/1 +// dump_network input_net_param trained_net_param \ +// input_blob output_prefix 0/1 // if input_net_param is 'none', we will directly load the network from // trained_net_param. If the last argv is 1, we will do a forward-backward pass // before dumping everyting, and also dump the who network. -#include -#include -#include +#include +#include -#include +#include "cuda_runtime.h" +#include "fcntl.h" +#include "google/protobuf/text_format.h" #include "caffe/blob.hpp" #include "caffe/common.hpp" @@ -23,7 +25,7 @@ #include "caffe/util/io.hpp" #include "caffe/solver.hpp" -using namespace caffe; +using namespace caffe; // NOLINT(build/namespaces) int main(int argc, char** argv) { cudaSetDevice(1); @@ -63,7 +65,8 @@ int main(int argc, char** argv) { // Dump the network NetParameter output_net_param; caffe_net->ToProto(&output_net_param, true); - WriteProtoToBinaryFile(output_net_param, output_prefix + output_net_param.name()); + WriteProtoToBinaryFile(output_net_param, + output_prefix + output_net_param.name()); } // Now, let's dump all the layers @@ -74,7 +77,8 @@ int main(int argc, char** argv) { LOG(ERROR) << "Dumping " << blob_names[blobid]; BlobProto output_blob_proto; blobs[blobid]->ToProto(&output_blob_proto); - WriteProtoToBinaryFile(output_blob_proto, output_prefix + blob_names[blobid]); + WriteProtoToBinaryFile(output_blob_proto, + output_prefix + blob_names[blobid]); } return 0; diff --git a/tools/finetune_net.cpp b/tools/finetune_net.cpp index 559715d..2aad738 100644 --- a/tools/finetune_net.cpp +++ b/tools/finetune_net.cpp @@ -6,11 +6,11 @@ #include -#include +#include #include "caffe/caffe.hpp" -using namespace caffe; +using namespace caffe; // NOLINT(build/namespaces) int main(int argc, char** argv) { ::google::InitGoogleLogging(argv[0]); diff --git a/tools/net_speed_benchmark.cpp b/tools/net_speed_benchmark.cpp index dd6f3ed..9976744 100644 --- a/tools/net_speed_benchmark.cpp +++ b/tools/net_speed_benchmark.cpp @@ -1,11 +1,12 @@ // Copyright 2013 Yangqing Jia -#include -#include -#include - -#include #include +#include +#include + +#include "cuda_runtime.h" +#include "fcntl.h" +#include "google/protobuf/text_format.h" #include "caffe/blob.hpp" #include "caffe/common.hpp" @@ -15,14 +16,16 @@ #include "caffe/util/io.hpp" #include "caffe/solver.hpp" -using namespace caffe; +using boost::shared_ptr; -int main(int argc, char** argv) { +using namespace caffe; // NOLINT(build/namespaces) +int main(int argc, char** argv) { int total_iter = 50; if (argc < 2) { - LOG(ERROR) << "net_speed_benchmark net_proto [iterations=50] [CPU/GPU] [Device_id=0]"; + LOG(ERROR) << "net_speed_benchmark net_proto [iterations=50] [CPU/GPU] " + << "[Device_id=0]"; return 0; } @@ -72,9 +75,12 @@ int main(int argc, char** argv) { layers[i]->Forward(bottom_vecs[i], &top_vecs[i]); } LOG(ERROR) << layername << "\tforward: " - << float(clock() - start) / CLOCKS_PER_SEC << " seconds."; + << static_cast(clock() - start) / CLOCKS_PER_SEC + << " seconds."; } - LOG(ERROR) << "Forward pass: " << float(clock() - forward_start) / CLOCKS_PER_SEC << " seconds."; + LOG(ERROR) << "Forward pass: " + << static_cast(clock() - forward_start) / CLOCKS_PER_SEC + << " seconds."; clock_t backward_start = clock(); for (int i = layers.size() - 1; i >= 0; --i) { const string& layername = layers[i]->layer_param().name(); @@ -83,10 +89,15 @@ int main(int argc, char** argv) { layers[i]->Backward(top_vecs[i], true, &bottom_vecs[i]); } LOG(ERROR) << layername << "\tbackward: " - << float(clock() - start) / CLOCKS_PER_SEC << " seconds."; + << static_cast(clock() - start) / CLOCKS_PER_SEC + << " seconds."; } - LOG(ERROR) << "Backward pass: " << float(clock() - backward_start) / CLOCKS_PER_SEC << " seconds."; - LOG(ERROR) << "Total Time: " << float(clock() - forward_start) / CLOCKS_PER_SEC << " seconds."; + LOG(ERROR) << "Backward pass: " + << static_cast(clock() - backward_start) / CLOCKS_PER_SEC + << " seconds."; + LOG(ERROR) << "Total Time: " + << static_cast(clock() - forward_start) / CLOCKS_PER_SEC + << " seconds."; LOG(ERROR) << "*** Benchmark ends ***"; return 0; } diff --git a/tools/test_net.cpp b/tools/test_net.cpp index 5b8305a..c4c992a 100644 --- a/tools/test_net.cpp +++ b/tools/test_net.cpp @@ -10,14 +10,16 @@ #include #include +#include #include "caffe/caffe.hpp" -using namespace caffe; +using namespace caffe; // NOLINT(build/namespaces) int main(int argc, char** argv) { if (argc < 4) { - LOG(ERROR) << "test_net net_proto pretrained_net_proto iterations [CPU/GPU]"; + LOG(ERROR) << "test_net net_proto pretrained_net_proto iterations " + << "[CPU/GPU]"; return 0; } diff --git a/tools/train_net.cpp b/tools/train_net.cpp index ce62616..3bd4f87 100644 --- a/tools/train_net.cpp +++ b/tools/train_net.cpp @@ -11,7 +11,7 @@ #include "caffe/caffe.hpp" -using namespace caffe; +using namespace caffe; // NOLINT(build/namespaces) int main(int argc, char** argv) { ::google::InitGoogleLogging(argv[0]); -- 2.7.4