layers.insert(layers.end(), block.begin(), block.end());
}
- for (auto layer : layers) {
+ for (auto &layer : layers) {
model->addLayer(layer);
}
NNTrainer::InputTensorsInfo::~InputTensorsInfo() {
g_print("%s:%d:%s: <called>\n", __FILE__, __LINE__, __func__);
- for (auto data : tensor_data) {
+ for (auto &data : tensor_data) {
for (auto inputs : data.inputs) {
ml_logd("free: ##I addr:%p", inputs);
delete inputs;
ml_logd("<called>");
if (!nntrainer) {
ml_loge("Failed get nntrainer");
+ return -1;
}
if (!notifier) {
ml_loge("Failed get notify");
+ return -1;
}
nntrainer->notifier = notifier;
training_loss(0),
validation_loss(0),
num_push_data(0),
- model_config(_model_config) {
+ model_config(_model_config),
+ notifier(nullptr) {
ml_logd("<called>");
getNNStreamerProperties(prop);
createModel();
createDataset();
- // properties = prop;
ml_logd("<leave>");
}
**/
/// assume that loss layers have single output
if (layer.getNumOutputConnections() == 1) {
- for (auto loss : loss_type) {
+ for (auto &loss : loss_type) {
if (layer.getOutputConnections()[0].find(loss) != std::string::npos) {
is_output = true;
}
size_t id = 0;
size_t num_data = 0;
- for (auto c_name : class_names) {
+ for (auto &c_name : class_names) {
num_data = 0;
std::filesystem::directory_iterator itr(c_name);
while (itr != std::filesystem::end(itr)) {