*/
static void blkcg_destroy_blkgs(struct blkcg *blkcg)
{
- int cpu;
-
might_sleep();
- css_get(&blkcg->css);
spin_lock_irq(&blkcg->lock);
+
while (!hlist_empty(&blkcg->blkg_list)) {
struct blkcg_gq *blkg = hlist_entry(blkcg->blkg_list.first,
struct blkcg_gq, blkcg_node);
}
spin_unlock_irq(&blkcg->lock);
-
- /*
- * Flush all the non-empty percpu lockless lists.
- */
- for_each_possible_cpu(cpu) {
- struct llist_head *lhead = per_cpu_ptr(blkcg->lhead, cpu);
-
- if (!llist_empty(lhead))
- cgroup_rstat_css_cpu_flush(&blkcg->css, cpu);
- }
- css_put(&blkcg->css);
}
/**
void cgroup_rstat_flush_irqsafe(struct cgroup *cgrp);
void cgroup_rstat_flush_hold(struct cgroup *cgrp);
void cgroup_rstat_flush_release(void);
-void cgroup_rstat_css_cpu_flush(struct cgroup_subsys_state *css, int cpu);
/*
* Basic resource stats.
spin_unlock_irq(&cgroup_rstat_lock);
}
-/**
- * cgroup_rstat_css_cpu_flush - flush stats for the given css and cpu
- * @css: target css to be flush
- * @cpu: the cpu that holds the stats to be flush
- *
- * A lightweight rstat flush operation for a given css and cpu.
- * Only the cpu_lock is being held for mutual exclusion, the cgroup_rstat_lock
- * isn't used.
- */
-void cgroup_rstat_css_cpu_flush(struct cgroup_subsys_state *css, int cpu)
-{
- raw_spinlock_t *cpu_lock = per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu);
-
- raw_spin_lock_irq(cpu_lock);
- rcu_read_lock();
- css->ss->css_rstat_flush(css, cpu);
- rcu_read_unlock();
- raw_spin_unlock_irq(cpu_lock);
-}
-
int cgroup_rstat_init(struct cgroup *cgrp)
{
int cpu;