void caffe_set(const int N, const Dtype alpha, Dtype *X);
inline void caffe_memset(const size_t N, const int alpha, void* X) {
- memset(X, alpha, N);
+ memset(X, alpha, N); // NOLINT(caffe/alt_fn)
}
template <typename Dtype>
}
current_row_ = 0;
}
- memcpy(&(*top)[0]->mutable_cpu_data()[i * data_count],
- &data_blob_.cpu_data()[current_row_ * data_count],
- sizeof(Dtype) * data_count);
- memcpy(&(*top)[1]->mutable_cpu_data()[i * label_data_count],
- &label_blob_.cpu_data()[current_row_ * label_data_count],
- sizeof(Dtype) * label_data_count);
+ caffe_copy(data_count, &data_blob_.cpu_data()[current_row_ * data_count],
+ &(*top)[0]->mutable_cpu_data()[i * data_count]);
+ caffe_copy(label_data_count,
+ &label_blob_.cpu_data()[current_row_ * label_data_count],
+ &(*top)[1]->mutable_cpu_data()[i * label_data_count]);
}
return Dtype(0.);
}
switch (head_) {
case UNINITIALIZED:
CaffeMallocHost(&cpu_ptr_, size_);
- memset(cpu_ptr_, 0, size_);
+ caffe_memset(size_, 0, cpu_ptr_);
head_ = HEAD_AT_CPU;
own_cpu_data_ = true;
break;
switch (head_) {
case UNINITIALIZED:
CUDA_CHECK(cudaMalloc(&gpu_ptr_, size_));
- CUDA_CHECK(cudaMemset(gpu_ptr_, 0, size_));
+ caffe_gpu_memset(size_, 0, gpu_ptr_);
head_ = HEAD_AT_GPU;
break;
case HEAD_AT_CPU:
SyncedMemory mem(10);
void* cpu_data = mem.mutable_cpu_data();
EXPECT_EQ(mem.head(), SyncedMemory::HEAD_AT_CPU);
- memset(cpu_data, 1, mem.size());
+ caffe_memset(mem.size(), 1, cpu_data);
for (int i = 0; i < mem.size(); ++i) {
EXPECT_EQ((static_cast<char*>(cpu_data))[i], 1);
}
// do another round
cpu_data = mem.mutable_cpu_data();
EXPECT_EQ(mem.head(), SyncedMemory::HEAD_AT_CPU);
- memset(cpu_data, 2, mem.size());
+ caffe_memset(mem.size(), 2, cpu_data);
for (int i = 0; i < mem.size(); ++i) {
EXPECT_EQ((static_cast<char*>(cpu_data))[i], 2);
}
SyncedMemory mem(10);
void* cpu_data = mem.mutable_cpu_data();
EXPECT_EQ(mem.head(), SyncedMemory::HEAD_AT_CPU);
- memset(cpu_data, 1, mem.size());
+ caffe_memset(mem.size(), 1, cpu_data);
const void* gpu_data = mem.gpu_data();
EXPECT_EQ(mem.head(), SyncedMemory::SYNCED);
// check if values are the same
// do another round
cpu_data = mem.mutable_cpu_data();
EXPECT_EQ(mem.head(), SyncedMemory::HEAD_AT_CPU);
- memset(cpu_data, 2, mem.size());
+ caffe_memset(mem.size(), 2, cpu_data);
for (int i = 0; i < mem.size(); ++i) {
EXPECT_EQ((static_cast<char*>(cpu_data))[i], 2);
}
SyncedMemory mem(10);
void* gpu_data = mem.mutable_gpu_data();
EXPECT_EQ(mem.head(), SyncedMemory::HEAD_AT_GPU);
- CUDA_CHECK(cudaMemset(gpu_data, 1, mem.size()));
+ caffe_gpu_memset(mem.size(), 1, gpu_data);
const void* cpu_data = mem.cpu_data();
for (int i = 0; i < mem.size(); ++i) {
EXPECT_EQ((static_cast<const char*>(cpu_data))[i], 1);
gpu_data = mem.mutable_gpu_data();
EXPECT_EQ(mem.head(), SyncedMemory::HEAD_AT_GPU);
- CUDA_CHECK(cudaMemset(gpu_data, 2, mem.size()));
+ caffe_gpu_memset(mem.size(), 2, gpu_data);
cpu_data = mem.cpu_data();
for (int i = 0; i < mem.size(); ++i) {
EXPECT_EQ((static_cast<const char*>(cpu_data))[i], 2);
template <typename Dtype>
void caffe_set(const int N, const Dtype alpha, Dtype* Y) {
if (alpha == 0) {
- memset(Y, 0, sizeof(Dtype) * N);
+ memset(Y, 0, sizeof(Dtype) * N); // NOLINT(caffe/alt_fn)
return;
}
for (int i = 0; i < N; ++i) {
if (X != Y) {
if (Caffe::mode() == Caffe::GPU) {
#ifndef CPU_ONLY
+ // NOLINT_NEXT_LINE(caffe/alt_fn)
CUDA_CHECK(cudaMemcpy(Y, X, sizeof(Dtype) * N, cudaMemcpyDefault));
#else
NO_GPU;
#endif
} else {
- memcpy(Y, X, sizeof(Dtype) * N);
+ memcpy(Y, X, sizeof(Dtype) * N); // NOLINT(caffe/alt_fn)
}
}
}
void caffe_gpu_memcpy(const size_t N, const void* X, void* Y) {
if (X != Y) {
- CUDA_CHECK(cudaMemcpy(Y, X, N, cudaMemcpyDefault));
+ CUDA_CHECK(cudaMemcpy(Y, X, N, cudaMemcpyDefault)); // NOLINT(caffe/alt_fn)
}
}
template <typename Dtype>
void caffe_gpu_set(const int N, const Dtype alpha, Dtype* Y) {
if (alpha == 0) {
- CUDA_CHECK(cudaMemset(Y, 0, sizeof(Dtype) * N));
+ CUDA_CHECK(cudaMemset(Y, 0, sizeof(Dtype) * N)); // NOLINT(caffe/alt_fn)
return;
}
// NOLINT_NEXT_LINE(whitespace/operators)