if (__split_bio(md, c))
bio_io_error(c);
}
+
+ clear_bit(DMF_BLOCK_IO, &md->flags);
}
static void __merge_pushback_list(struct mapped_device *md)
{
struct dm_table *map = NULL;
DECLARE_WAITQUEUE(wait, current);
- int r = -EINVAL;
+ int pending, r = 0;
int do_lockfs = suspend_flags & DM_SUSPEND_LOCKFS_FLAG ? 1 : 0;
int noflush = suspend_flags & DM_SUSPEND_NOFLUSH_FLAG ? 1 : 0;
mutex_lock(&md->suspend_lock);
- if (dm_suspended(md))
+ if (dm_suspended(md)) {
+ r = -EINVAL;
goto out_unlock;
+ }
map = dm_get_table(md);
set_current_state(TASK_INTERRUPTIBLE);
smp_mb();
- if (!atomic_read(&md->pending) || signal_pending(current))
+ pending = atomic_read(&md->pending);
+ if (!pending || signal_pending(current))
break;
io_schedule();
__merge_pushback_list(md);
/* were we interrupted ? */
- r = -EINTR;
- if (atomic_read(&md->pending)) {
- clear_bit(DMF_BLOCK_IO, &md->flags);
+ if (pending) {
__flush_deferred_io(md);
up_write(&md->io_lock);
+
unlock_fs(md);
+ r = -EINTR;
goto out; /* pushback list is already flushed, so skip flush */
}
up_write(&md->io_lock);
set_bit(DMF_SUSPENDED, &md->flags);
- r = 0;
-
flush_and_out:
if (r && noflush) {
/*
goto out;
down_write(&md->io_lock);
- clear_bit(DMF_BLOCK_IO, &md->flags);
-
__flush_deferred_io(md);
up_write(&md->io_lock);