From: Evan Shelhamer Date: Sun, 19 Jan 2014 22:34:12 +0000 (-0800) Subject: cleanup whitespace X-Git-Url: http://review.tizen.org/git/?a=commitdiff_plain;h=e54bd1be09e02c2926663d7c9a701f50a01babd7;p=platform%2Fupstream%2Fcaffe.git cleanup whitespace --- diff --git a/Makefile b/Makefile index f9caee1..9c97ce7 100644 --- a/Makefile +++ b/Makefile @@ -90,12 +90,12 @@ PYTHON_LDFLAGS := $(LDFLAGS) $(foreach library,$(PYTHON_LIBRARIES),-l$(library)) all: init $(NAME) $(STATIC_NAME) examples @echo $(CXX_OBJS) - -init: + +init: @ mkdir -p $(foreach obj,$(OBJS),$(dir $(obj))) @ mkdir -p $(foreach obj,$(EXAMPLE_OBJS),$(dir $(obj))) @ mkdir -p $(foreach obj,$(TEST_OBJS),$(dir $(obj))) - @ mkdir -p $(foreach obj,$(GTEST_OBJ),$(dir $(obj))) + @ mkdir -p $(foreach obj,$(GTEST_OBJ),$(dir $(obj))) linecount: clean cloc --read-lang-def=$(PROJECT).cloc src/$(PROJECT)/ @@ -110,7 +110,7 @@ py: init $(STATIC_NAME) $(PY$(PROJECT)_SRC) $(PROTO_GEN_PY) $(CXX) -shared -o $(PY$(PROJECT)_SO) $(PY$(PROJECT)_SRC) \ $(STATIC_NAME) $(CXXFLAGS) $(PYTHON_LDFLAGS) @echo - + mat$(PROJECT): mat mat: init $(STATIC_NAME) $(MAT$(PROJECT)_SRC) @@ -119,7 +119,7 @@ mat: init $(STATIC_NAME) $(MAT$(PROJECT)_SRC) CXXLIBS="\$$CXXLIBS $(LDFLAGS)" \ -o $(MAT$(PROJECT)_SO) @echo - + $(NAME): init $(PROTO_OBJS) $(OBJS) $(CXX) -shared -o $(NAME) $(OBJS) $(LDFLAGS) $(WARNINGS) @echo @@ -143,35 +143,35 @@ $(OBJS): $(PROTO_GEN_CC) $(HXX_SRCS) $(BUILD_DIR)/src/$(PROJECT)/%.o: src/$(PROJECT)/%.cpp $(CXX) $< $(CXXFLAGS) -c -o $@ $(LDFLAGS) @echo - + $(BUILD_DIR)/src/$(PROJECT)/layers/%.o: src/$(PROJECT)/layers/%.cpp $(CXX) $< $(CXXFLAGS) -c -o $@ $(LDFLAGS) @echo - + $(BUILD_DIR)/src/$(PROJECT)/proto/%.o: src/$(PROJECT)/proto/%.cc $(CXX) $< $(CXXFLAGS) -c -o $@ $(LDFLAGS) @echo - + $(BUILD_DIR)/src/$(PROJECT)/test/%.o: src/test/%.cpp $(CXX) $< $(CXXFLAGS) -c -o $@ $(LDFLAGS) @echo - + $(BUILD_DIR)/src/$(PROJECT)/util/%.o: src/$(PROJECT)/util/%.cpp $(CXX) $< $(CXXFLAGS) -c -o $@ $(LDFLAGS) @echo - + $(BUILD_DIR)/src/gtest/%.o: src/gtest/%.cpp $(CXX) $< $(CXXFLAGS) -c -o $@ $(LDFLAGS) @echo - + $(BUILD_DIR)/src/$(PROJECT)/layers/%.cuo: src/$(PROJECT)/layers/%.cu $(CUDA_DIR)/bin/nvcc $(NVCCFLAGS) $(CUDA_ARCH) -c $< -o $@ @echo - + $(BUILD_DIR)/src/$(PROJECT)/util/%.cuo: src/$(PROJECT)/util/%.cu $(CUDA_DIR)/bin/nvcc $(NVCCFLAGS) $(CUDA_ARCH) -c $< -o $@ @echo - + $(BUILD_DIR)/examples/%.o: examples/%.cpp $(CXX) $< $(CXXFLAGS) -c -o $@ $(LDFLAGS) @echo @@ -187,7 +187,7 @@ $(PROTO_GEN_CC): $(PROTO_SRCS) mkdir -p include/$(PROJECT)/proto cp $(PROTO_GEN_HEADER) include/$(PROJECT)/proto/ @echo - + clean: @- $(RM) $(NAME) $(STATIC_NAME) @- $(RM) $(PROTO_GEN_HEADER) $(PROTO_GEN_CC) $(PROTO_GEN_PY) diff --git a/Makefile.config b/Makefile.config index 356493a..111bee9 100644 --- a/Makefile.config +++ b/Makefile.config @@ -20,4 +20,4 @@ LIBRARY_DIRS := /usr/lib /usr/local/lib CXX=/usr/bin/g++ BUILD_DIR=build -DISTRIBUTE_DIR=distribute \ No newline at end of file +DISTRIBUTE_DIR=distribute diff --git a/examples/synset_words.txt b/examples/synset_words.txt index 9243b00..a9e8c7f 100644 --- a/examples/synset_words.txt +++ b/examples/synset_words.txt @@ -997,4 +997,4 @@ n13044778 earthstar n13052670 hen-of-the-woods, hen of the woods, Polyporus frondosus, Grifola frondosa n13054560 bolete n13133613 ear, spike, capitulum -n15075141 toilet tissue, toilet paper, bathroom tissue \ No newline at end of file +n15075141 toilet tissue, toilet paper, bathroom tissue diff --git a/include/caffe/syncedmem.hpp b/include/caffe/syncedmem.hpp index 4ffa6c1..97688cb 100644 --- a/include/caffe/syncedmem.hpp +++ b/include/caffe/syncedmem.hpp @@ -17,7 +17,7 @@ namespace caffe { // problem when running on a machine without GPU. Thus, we simply define // these two functions for safety and possible future change if the problem // of calling cuda functions disappears in a future version. -// +// // In practice, although we are creating unpinned memory here, as long as we // are constantly accessing them the memory pages almost always stays in // the physical memory (assuming we have large enough memory installed), and diff --git a/src/caffe/test/test_blob.cpp b/src/caffe/test/test_blob.cpp index ba76ed1..7c3084e 100644 --- a/src/caffe/test/test_blob.cpp +++ b/src/caffe/test/test_blob.cpp @@ -11,7 +11,7 @@ #include "caffe/test/test_caffe_main.hpp" namespace caffe { - + template class BlobSimpleTest : public ::testing::Test { protected: diff --git a/src/caffe/test/test_flatten_layer.cpp b/src/caffe/test/test_flatten_layer.cpp index b97e56a..805fd72 100644 --- a/src/caffe/test/test_flatten_layer.cpp +++ b/src/caffe/test/test_flatten_layer.cpp @@ -15,7 +15,7 @@ namespace caffe { extern cudaDeviceProp CAFFE_TEST_CUDA_PROP; - + template class FlattenLayerTest : public ::testing::Test { protected: diff --git a/src/caffe/test/test_im2col_layer.cpp b/src/caffe/test/test_im2col_layer.cpp index 65d9153..dc6445d 100644 --- a/src/caffe/test/test_im2col_layer.cpp +++ b/src/caffe/test/test_im2col_layer.cpp @@ -15,7 +15,7 @@ namespace caffe { extern cudaDeviceProp CAFFE_TEST_CUDA_PROP; - + template class Im2colLayerTest : public ::testing::Test { protected: diff --git a/src/caffe/test/test_lrn_layer.cpp b/src/caffe/test/test_lrn_layer.cpp index f3fc1bd..757bac3 100644 --- a/src/caffe/test/test_lrn_layer.cpp +++ b/src/caffe/test/test_lrn_layer.cpp @@ -20,7 +20,7 @@ using std::max; namespace caffe { extern cudaDeviceProp CAFFE_TEST_CUDA_PROP; - + template class LRNLayerTest : public ::testing::Test { protected: @@ -48,7 +48,7 @@ class LRNLayerTest : public ::testing::Test { template void LRNLayerTest::ReferenceLRNForward( - const Blob& blob_bottom, const LayerParameter& layer_param, + const Blob& blob_bottom, const LayerParameter& layer_param, Blob* blob_top) { blob_top->Reshape(blob_bottom.num(), blob_bottom.channels(), blob_bottom.height(), blob_bottom.width()); @@ -139,7 +139,7 @@ TYPED_TEST(LRNLayerTest, TestCPUGradient) { // std::cout << "CPU diff " << this->blob_bottom_->cpu_diff()[i] << std::endl; //} checker.CheckGradientExhaustive(layer, this->blob_bottom_vec_, this->blob_top_vec_); -} +} TYPED_TEST(LRNLayerTest, TestGPUGradient) { LayerParameter layer_param; diff --git a/src/caffe/test/test_neuron_layer.cpp b/src/caffe/test/test_neuron_layer.cpp index 0bca34d..1abee7c 100644 --- a/src/caffe/test/test_neuron_layer.cpp +++ b/src/caffe/test/test_neuron_layer.cpp @@ -15,7 +15,7 @@ namespace caffe { extern cudaDeviceProp CAFFE_TEST_CUDA_PROP; - + template class NeuronLayerTest : public ::testing::Test { protected: diff --git a/src/caffe/test/test_padding_layer.cpp b/src/caffe/test/test_padding_layer.cpp index bb47e35..da48111 100644 --- a/src/caffe/test/test_padding_layer.cpp +++ b/src/caffe/test/test_padding_layer.cpp @@ -15,7 +15,7 @@ namespace caffe { extern cudaDeviceProp CAFFE_TEST_CUDA_PROP; - + template class PaddingLayerTest : public ::testing::Test { protected: diff --git a/src/caffe/test/test_syncedmem.cpp b/src/caffe/test/test_syncedmem.cpp index 216dc87..b834710 100644 --- a/src/caffe/test/test_syncedmem.cpp +++ b/src/caffe/test/test_syncedmem.cpp @@ -46,7 +46,7 @@ TEST_F(SyncedMemoryTest, TestCPUWrite) { for (int i = 0; i < mem.size(); ++i) { EXPECT_EQ(((char*)recovered_value)[i], 1); } - // do another round + // do another round cpu_data = mem.mutable_cpu_data(); EXPECT_EQ(mem.head(), SyncedMemory::HEAD_AT_CPU); memset(cpu_data, 2, mem.size()); @@ -73,7 +73,7 @@ TEST_F(SyncedMemoryTest, TestGPUWrite) { EXPECT_EQ(((char*)cpu_data)[i], 1); } EXPECT_EQ(mem.head(), SyncedMemory::SYNCED); - + gpu_data = mem.mutable_gpu_data(); EXPECT_EQ(mem.head(), SyncedMemory::HEAD_AT_GPU); CUDA_CHECK(cudaMemset(gpu_data, 2, mem.size()));