CHECK_GE(top->size(), 1) << "Split Layer takes at least one blob as output.";
count_ = bottom[0]->count();
for (int i = 0; i < top->size(); ++i) {
+ // Allow the 0th top blob to be 'in-place', but no others.
+ if (i == 0 && (*top)[i] == bottom[0]) {
+ continue;
+ } else {
+ CHECK_NE((*top)[i], bottom[0]) << "Only 0th top blob may be in place.";
+ }
(*top)[i]->Reshape(bottom[0]->num(), bottom[0]->channels(),
bottom[0]->height(), bottom[0]->width());
CHECK_EQ(count_, (*top)[i]->count());
vector<Blob<Dtype>*>* top) {
const Dtype* bottom_data = bottom[0]->cpu_data();
for (int i = 0; i < top->size(); ++i) {
+ if (i == 0 && (*top)[i] == bottom[0]) {
+ continue;
+ }
Dtype* top_data = (*top)[i]->mutable_cpu_data();
caffe_copy(count_, bottom_data, top_data);
}
vector<Blob<Dtype>*>* top) {
const Dtype* bottom_data = bottom[0]->gpu_data();
for (int i = 0; i < top->size(); ++i) {
+ if (i == 0 && (*top)[i] == bottom[0]) {
+ continue;
+ }
Dtype* top_data = (*top)[i]->mutable_gpu_data();
caffe_gpu_copy(count_, bottom_data, top_data);
}
if (propagate_down) {
const Dtype* top_diff = top[0]->cpu_diff();
Dtype* bottom_diff = (*bottom)[0]->mutable_cpu_diff();
- caffe_copy(count_, top_diff, bottom_diff);
+ // Initialize by copying first top blob diff to our diff, unless we're
+ // doing in-place computation for the first blob, in which case the diff is
+ // already initialized.
+ if (top[0] != (*bottom)[0]) {
+ caffe_copy(count_, top_diff, bottom_diff);
+ }
+ // Add remaining top blob diffs.
for (int i = 1; i < top.size(); ++i) {
top_diff = top[i]->cpu_diff();
caffe_axpy(count_, Dtype(1.), top_diff, bottom_diff);
if (propagate_down) {
const Dtype* top_diff = top[0]->gpu_diff();
Dtype* bottom_diff = (*bottom)[0]->mutable_gpu_diff();
- caffe_gpu_copy(count_, top_diff, bottom_diff);
+ // Initialize by copying first top blob diff to our diff, unless we're
+ // doing in-place computation for the first blob, in which case the diff is
+ // already initialized.
+ if (top[0] != (*bottom)[0]) {
+ caffe_gpu_copy(count_, top_diff, bottom_diff);
+ }
+ // Add remaining top blob diffs.
for (int i = 1; i < top.size(); ++i) {
top_diff = top[i]->gpu_diff();
caffe_gpu_axpy(count_, Dtype(1.), top_diff, bottom_diff);
}
}
+TYPED_TEST(SplitLayerTest, TestCPUInPlace) {
+ LayerParameter layer_param;
+ SplitLayer<TypeParam> layer(layer_param);
+ Caffe::set_mode(Caffe::CPU);
+ this->blob_top_vec_[0] = this->blob_bottom_vec_[0];
+ layer.SetUp(this->blob_bottom_vec_, &(this->blob_top_vec_));
+ layer.Forward(this->blob_bottom_vec_, &(this->blob_top_vec_));
+ for (int i = 0; i < this->blob_bottom_->count(); ++i) {
+ TypeParam bottom_value = this->blob_bottom_->cpu_data()[i];
+ EXPECT_EQ(bottom_value, this->blob_top_b_->cpu_data()[i]);
+ }
+}
+
+TYPED_TEST(SplitLayerTest, TestGPUInPlace) {
+ LayerParameter layer_param;
+ SplitLayer<TypeParam> layer(layer_param);
+ Caffe::set_mode(Caffe::GPU);
+ this->blob_top_vec_[0] = this->blob_bottom_vec_[0];
+ layer.SetUp(this->blob_bottom_vec_, &(this->blob_top_vec_));
+ layer.Forward(this->blob_bottom_vec_, &(this->blob_top_vec_));
+ for (int i = 0; i < this->blob_bottom_->count(); ++i) {
+ TypeParam bottom_value = this->blob_bottom_->cpu_data()[i];
+ EXPECT_EQ(bottom_value, this->blob_top_b_->cpu_data()[i]);
+ }
+}
+
TYPED_TEST(SplitLayerTest, TestCPUGradient) {
LayerParameter layer_param;
Caffe::set_mode(Caffe::CPU);
this->blob_top_vec_);
}
+TYPED_TEST(SplitLayerTest, TestCPUGradientInPlace) {
+ LayerParameter layer_param;
+ Caffe::set_mode(Caffe::CPU);
+ SplitLayer<TypeParam> layer(layer_param);
+ GradientChecker<TypeParam> checker(1e-2, 1e-2);
+ this->blob_top_vec_[0] = this->blob_bottom_vec_[0];
+ checker.CheckGradientExhaustive(layer, this->blob_bottom_vec_, this->blob_top_vec_);
+ checker.CheckGradientExhaustive(layer, this->blob_bottom_vec_,
+ this->blob_top_vec_);
+}
+
+TYPED_TEST(SplitLayerTest, TestGPUGradientInPlace) {
+ LayerParameter layer_param;
+ Caffe::set_mode(Caffe::GPU);
+ SplitLayer<TypeParam> layer(layer_param);
+ GradientChecker<TypeParam> checker(1e-2, 1e-2);
+ this->blob_top_vec_[0] = this->blob_bottom_vec_[0];
+ checker.CheckGradientExhaustive(layer, this->blob_bottom_vec_, this->blob_top_vec_);
+ checker.CheckGradientExhaustive(layer, this->blob_bottom_vec_,
+ this->blob_top_vec_);
+}
+
template <typename Dtype>
class SplitLayerInsertionTest : public ::testing::Test {