// Resize the output tensor if the output tensor is dynamic.
if (IsDynamicTensor(op_context.output)) {
TF_LITE_ENSURE_OK(context, ResizeOutputTensor(context, &op_context));
- TfLiteTensorRealloc(op_context.output->bytes, op_context.output);
}
#define TF_LITE_BATCH_TO_SPACE_ND(type, scalar) \
TF_LITE_ENSURE_OK(context,
ResizeTempAxis(context, &op_context, resolved_axis));
TF_LITE_ENSURE_OK(context, ResizeOutputTensor(context, &op_context));
- TfLiteTensorRealloc(resolved_axis->bytes, resolved_axis);
- TfLiteTensorRealloc(op_context.output->bytes, op_context.output);
}
#define TF_LITE_MEAN(kernel_type, data_type) \
// Resize the output tensor if the output tensor is dynamic.
if (IsDynamicTensor(op_context.output)) {
TF_LITE_ENSURE_OK(context, ResizeOutputTensor(context, &op_context));
- TfLiteTensorRealloc(op_context.output->bytes, op_context.output);
}
// TODO(nupurgarg): Change kernel implementation to take in int* instead of
if (IsDynamicTensor(output)) {
TF_LITE_ENSURE_OK(context,
ResizeOutputTensor(context, input, size, output));
- TfLiteTensorRealloc(output->bytes, output);
}
if (output->type == kTfLiteFloat32) {
// Resize the output tensor if the output tensor is dynamic.
if (IsDynamicTensor(op_context.output)) {
TF_LITE_ENSURE_OK(context, ResizeOutputTensor(context, &op_context));
- TfLiteTensorRealloc(op_context.output->bytes, op_context.output);
}
#define TF_LITE_SPACE_TO_BATCH_ND(type, scalar) \
if (IsDynamicTensor(op_context.output)) {
TF_LITE_ENSURE_OK(context, ResizeOutputTensor(context, &op_context));
- TfLiteTensorRealloc(op_context.output->bytes, op_context.output);
}
std::vector<int32_t> starts;
// Resize the output tensor if the output tensor is dynamic.
if (IsDynamicTensor(op_context.output)) {
TF_LITE_ENSURE_OK(context, ResizeOutputTensor(context, &op_context));
- TfLiteTensorRealloc(op_context.output->bytes, op_context.output);
}
// Reverse the permuted axes and convert to 4D due to the way Dims are