/* thresholds for memory usage. RCU-protected */
struct mem_cgroup_threshold_ary *thresholds;
+ /*
+ * Preallocated buffer to be used in mem_cgroup_unregister_event()
+ * to make it "never fail".
+ * It must be able to store at least thresholds->size - 1 entries.
+ */
+ struct mem_cgroup_threshold_ary *__thresholds;
+
/* thresholds for mem+swap usage. RCU-protected */
struct mem_cgroup_threshold_ary *memsw_thresholds;
+ /* the same as __thresholds, but for memsw_thresholds */
+ struct mem_cgroup_threshold_ary *__memsw_thresholds;
+
/* For oom notifier event fd */
struct list_head oom_notify;
else
rcu_assign_pointer(memcg->memsw_thresholds, thresholds_new);
- /* To be sure that nobody uses thresholds before freeing it */
+ /* To be sure that nobody uses thresholds */
synchronize_rcu();
- kfree(thresholds);
+ /*
+ * Free old preallocated buffer and use thresholds as new
+ * preallocated buffer.
+ */
+ if (type == _MEM) {
+ kfree(memcg->__thresholds);
+ memcg->__thresholds = thresholds;
+ } else {
+ kfree(memcg->__memsw_thresholds);
+ memcg->__memsw_thresholds = thresholds;
+ }
unlock:
mutex_unlock(&memcg->thresholds_lock);
return ret;
}
-static int mem_cgroup_usage_unregister_event(struct cgroup *cgrp,
+static void mem_cgroup_usage_unregister_event(struct cgroup *cgrp,
struct cftype *cft, struct eventfd_ctx *eventfd)
{
struct mem_cgroup *memcg = mem_cgroup_from_cont(cgrp);
int type = MEMFILE_TYPE(cft->private);
u64 usage;
int size = 0;
- int i, j, ret = 0;
+ int i, j;
mutex_lock(&memcg->thresholds_lock);
if (type == _MEM)
size++;
}
+ /* Use preallocated buffer for new array of thresholds */
+ if (type == _MEM)
+ thresholds_new = memcg->__thresholds;
+ else
+ thresholds_new = memcg->__memsw_thresholds;
+
/* Set thresholds array to NULL if we don't have thresholds */
if (!size) {
+ kfree(thresholds_new);
thresholds_new = NULL;
- goto assign;
+ goto swap_buffers;
}
- /* Allocate memory for new array of thresholds */
- thresholds_new = kmalloc(sizeof(*thresholds_new) +
- size * sizeof(struct mem_cgroup_threshold),
- GFP_KERNEL);
- if (!thresholds_new) {
- ret = -ENOMEM;
- goto unlock;
- }
thresholds_new->size = size;
/* Copy thresholds and find current threshold */
j++;
}
-assign:
- if (type == _MEM)
+swap_buffers:
+ /* Swap thresholds array and preallocated buffer */
+ if (type == _MEM) {
+ memcg->__thresholds = thresholds;
rcu_assign_pointer(memcg->thresholds, thresholds_new);
- else
+ } else {
+ memcg->__memsw_thresholds = thresholds;
rcu_assign_pointer(memcg->memsw_thresholds, thresholds_new);
+ }
- /* To be sure that nobody uses thresholds before freeing it */
+ /* To be sure that nobody uses thresholds */
synchronize_rcu();
- kfree(thresholds);
-unlock:
mutex_unlock(&memcg->thresholds_lock);
-
- return ret;
}
static int mem_cgroup_oom_register_event(struct cgroup *cgrp,
return 0;
}
-static int mem_cgroup_oom_unregister_event(struct cgroup *cgrp,
+static void mem_cgroup_oom_unregister_event(struct cgroup *cgrp,
struct cftype *cft, struct eventfd_ctx *eventfd)
{
struct mem_cgroup *mem = mem_cgroup_from_cont(cgrp);
}
mutex_unlock(&memcg_oom_mutex);
-
- return 0;
}
static int mem_cgroup_oom_control_read(struct cgroup *cgrp,