return 1;
}
-static void fd_emulate_sync_cache(struct se_cmd *cmd)
+static int fd_execute_sync_cache(struct se_cmd *cmd)
{
struct se_device *dev = cmd->se_dev;
struct fd_dev *fd_dev = dev->dev_ptr;
pr_err("FILEIO: vfs_fsync_range() failed: %d\n", ret);
if (immed)
- return;
+ return 0;
if (ret) {
cmd->scsi_sense_reason = TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
} else {
target_complete_cmd(cmd, SAM_STAT_GOOD);
}
+
+ return 0;
}
static int fd_execute_rw(struct se_cmd *cmd)
static struct spc_ops fd_spc_ops = {
.execute_rw = fd_execute_rw,
+ .execute_sync_cache = fd_execute_sync_cache,
};
static int fd_parse_cdb(struct se_cmd *cmd)
.create_virtdevice = fd_create_virtdevice,
.free_device = fd_free_device,
.parse_cdb = fd_parse_cdb,
- .do_sync_cache = fd_emulate_sync_cache,
.check_configfs_dev_params = fd_check_configfs_dev_params,
.set_configfs_dev_params = fd_set_configfs_dev_params,
.show_configfs_dev_params = fd_show_configfs_dev_params,
* Implement SYCHRONIZE CACHE. Note that we can't handle lba ranges and must
* always flush the whole cache.
*/
-static void iblock_emulate_sync_cache(struct se_cmd *cmd)
+static int iblock_execute_sync_cache(struct se_cmd *cmd)
{
struct iblock_dev *ib_dev = cmd->se_dev->dev_ptr;
int immed = (cmd->t_task_cdb[1] & 0x2);
if (!immed)
bio->bi_private = cmd;
submit_bio(WRITE_FLUSH, bio);
+ return 0;
}
static int iblock_do_discard(struct se_device *dev, sector_t lba, u32 range)
static struct spc_ops iblock_spc_ops = {
.execute_rw = iblock_execute_rw,
+ .execute_sync_cache = iblock_execute_sync_cache,
};
static int iblock_parse_cdb(struct se_cmd *cmd)
.free_device = iblock_free_device,
.parse_cdb = iblock_parse_cdb,
.do_discard = iblock_do_discard,
- .do_sync_cache = iblock_emulate_sync_cache,
.check_configfs_dev_params = iblock_check_configfs_dev_params,
.set_configfs_dev_params = iblock_set_configfs_dev_params,
.show_configfs_dev_params = iblock_show_configfs_dev_params,
return 0;
}
-static int sbc_emulate_synchronize_cache(struct se_cmd *cmd)
-{
- if (!cmd->se_dev->transport->do_sync_cache) {
- pr_err("SYNCHRONIZE_CACHE emulation not supported"
- " for: %s\n", cmd->se_dev->transport->name);
- cmd->scsi_sense_reason = TCM_UNSUPPORTED_SCSI_OPCODE;
- return -ENOSYS;
- }
-
- cmd->se_dev->transport->do_sync_cache(cmd);
- return 0;
-}
-
static int sbc_emulate_verify(struct se_cmd *cmd)
{
target_complete_cmd(cmd, GOOD);
break;
case SYNCHRONIZE_CACHE:
case SYNCHRONIZE_CACHE_16:
+ if (!ops->execute_sync_cache)
+ goto out_unsupported_cdb;
+
/*
* Extract LBA and range to be flushed for emulated SYNCHRONIZE_CACHE
*/
if (sbc_check_valid_sectors(cmd) < 0)
goto out_invalid_cdb_field;
}
- cmd->execute_cmd = sbc_emulate_synchronize_cache;
+ cmd->execute_cmd = ops->execute_sync_cache;
break;
case UNMAP:
size = get_unaligned_be16(&cdb[7]);
int (*parse_cdb)(struct se_cmd *cmd);
int (*do_discard)(struct se_device *, sector_t, u32);
- void (*do_sync_cache)(struct se_cmd *);
ssize_t (*check_configfs_dev_params)(struct se_hba *,
struct se_subsystem_dev *);
ssize_t (*set_configfs_dev_params)(struct se_hba *,
struct spc_ops {
int (*execute_rw)(struct se_cmd *cmd);
+ int (*execute_sync_cache)(struct se_cmd *cmd);
};
int transport_subsystem_register(struct se_subsystem_api *);