virtual Dtype Backward_gpu(const vector<Blob<Dtype>*>& top,
const bool propagate_down, vector<Blob<Dtype>*>* bottom);
};
-REGISTER_LAYER("relu", ReLULayer);
template <typename Dtype>
float scale_;
unsigned int uint_thres_;
};
-REGISTER_LAYER("dropout", DropoutLayer);
template <typename Dtype>
bool biasterm_;
shared_ptr<SyncedMemory> bias_multiplier_;
};
-REGISTER_LAYER("innerproduct", InnerProductLayer);
template <typename Dtype>
int HEIGHT_OUT_;
int WIDTH_OUT_;
};
-REGISTER_LAYER("padding", PaddingLayer);
template <typename Dtype>
int height_;
int width_;
};
-REGISTER_LAYER("lrn", LRNLayer);
template <typename Dtype>
int HEIGHT_;
int WIDTH_;
};
-REGISTER_LAYER("im2col", Im2colLayer);
template <typename Dtype>
int POOLED_HEIGHT_;
int POOLED_WIDTH_;
};
-REGISTER_LAYER("pool", PoolingLayer);
template <typename Dtype>
int K_;
int N_;
};
-REGISTER_LAYER("conv", ConvolutionLayer);
+
// This function is used to create a pthread that prefetches the data.
template <typename Dtype>
shared_ptr<Blob<Dtype> > prefetch_label_;
Blob<Dtype> data_mean_;
};
-REGISTER_LAYER("data", DataLayer);
template <typename Dtype>
// scale is an intermediate blob to hold temporary results.
Blob<Dtype> scale_;
};
-REGISTER_LAYER("softmax", SoftmaxLayer);
template <typename Dtype>
// virtual Dtype Backward_gpu(const vector<Blob<Dtype>*>& top,
// const bool propagate_down, vector<Blob<Dtype>*>* bottom);
};
-REGISTER_LAYER("multinomial_logistic_loss", MultinomialLogisticLossLayer);
// SoftmaxWithLossLayer is a layer that implements softmax and then computes
vector<Blob<Dtype>*> softmax_bottom_vec_;
vector<Blob<Dtype>*> softmax_top_vec_;
};
-REGISTER_LAYER("softmax_loss", SoftmaxWithLossLayer);
template <typename Dtype>
// const bool propagate_down, vector<Blob<Dtype>*>* bottom);
Blob<Dtype> difference_;
};
-REGISTER_LAYER("euclidean_loss", EuclideanLossLayer);
template <typename Dtype>
return Dtype(0.);
}
};
-REGISTER_LAYER("accuracy", AccuracyLayer);
} // namespace caffe