}
pci_set_master(pdev);
/* Completion for VF2PF request/response message exchange */
- init_completion(&accel_dev->vf.iov_msg_completion);
+ init_completion(&accel_dev->vf.msg_received);
ret = qat_crypto_dev_config(accel_dev);
if (ret)
}
pci_set_master(pdev);
/* Completion for VF2PF request/response message exchange */
- init_completion(&accel_dev->vf.iov_msg_completion);
+ init_completion(&accel_dev->vf.msg_received);
ret = qat_crypto_dev_config(accel_dev);
if (ret)
char irq_name[ADF_MAX_MSIX_VECTOR_NAME];
struct tasklet_struct pf2vf_bh_tasklet;
struct mutex vf2pf_lock; /* protect CSR access */
- struct completion iov_msg_completion;
- u8 compatible;
+ struct completion msg_received;
+ u32 response; /* temp field holding pf2vf response */
u8 pf_version;
} vf;
};
int adf_vf2pf_request_version(struct adf_accel_dev *accel_dev)
{
struct adf_hw_device_data *hw_data = accel_dev->hw_device;
+ u8 pf_version;
u32 msg = 0;
+ int compat;
+ u32 resp;
int ret;
msg = ADF_VF2PF_MSGORIGIN_SYSTEM;
msg |= ADF_PFVF_COMPAT_THIS_VERSION << ADF_VF2PF_COMPAT_VER_REQ_SHIFT;
BUILD_BUG_ON(ADF_PFVF_COMPAT_THIS_VERSION > 255);
- ret = adf_send_vf2pf_req(accel_dev, msg);
+ ret = adf_send_vf2pf_req(accel_dev, msg, &resp);
if (ret) {
dev_err(&GET_DEV(accel_dev),
"Failed to send Compatibility Version Request.\n");
return ret;
}
+ pf_version = (resp & ADF_PF2VF_VERSION_RESP_VERS_MASK)
+ >> ADF_PF2VF_VERSION_RESP_VERS_SHIFT;
+ compat = (resp & ADF_PF2VF_VERSION_RESP_RESULT_MASK)
+ >> ADF_PF2VF_VERSION_RESP_RESULT_SHIFT;
+
/* Response from PF received, check compatibility */
- switch (accel_dev->vf.compatible) {
+ switch (compat) {
case ADF_PF2VF_VF_COMPATIBLE:
break;
case ADF_PF2VF_VF_COMPAT_UNKNOWN:
/* VF is newer than PF and decides whether it is compatible */
- if (accel_dev->vf.pf_version >= hw_data->min_iov_compat_ver) {
- accel_dev->vf.compatible = ADF_PF2VF_VF_COMPATIBLE;
+ if (pf_version >= hw_data->min_iov_compat_ver)
break;
- }
fallthrough;
case ADF_PF2VF_VF_INCOMPATIBLE:
dev_err(&GET_DEV(accel_dev),
"PF (vers %d) and VF (vers %d) are not compatible\n",
- accel_dev->vf.pf_version,
- ADF_PFVF_COMPAT_THIS_VERSION);
+ pf_version, ADF_PFVF_COMPAT_THIS_VERSION);
return -EINVAL;
default:
dev_err(&GET_DEV(accel_dev),
"Invalid response from PF; assume not compatible\n");
return -EINVAL;
}
- return ret;
+
+ accel_dev->vf.pf_version = pf_version;
+ return 0;
}
* adf_send_vf2pf_req() - send VF2PF request message
* @accel_dev: Pointer to acceleration device.
* @msg: Request message to send
+ * @resp: Returned PF response
*
* This function sends a message that requires a response from the VF to the PF
* and waits for a reply.
*
* Return: 0 on success, error code otherwise.
*/
-int adf_send_vf2pf_req(struct adf_accel_dev *accel_dev, u32 msg)
+int adf_send_vf2pf_req(struct adf_accel_dev *accel_dev, u32 msg, u32 *resp)
{
unsigned long timeout = msecs_to_jiffies(ADF_PFVF_MSG_RESP_TIMEOUT);
int ret;
- reinit_completion(&accel_dev->vf.iov_msg_completion);
+ reinit_completion(&accel_dev->vf.msg_received);
/* Send request from VF to PF */
ret = adf_send_vf2pf_msg(accel_dev, msg);
}
/* Wait for response */
- if (!wait_for_completion_timeout(&accel_dev->vf.iov_msg_completion,
+ if (!wait_for_completion_timeout(&accel_dev->vf.msg_received,
timeout)) {
dev_err(&GET_DEV(accel_dev),
"PFVF request/response message timeout expired\n");
return -EIO;
}
+ if (likely(resp))
+ *resp = accel_dev->vf.response;
+
+ /* Once copied, set to an invalid value */
+ accel_dev->vf.response = 0;
+
return 0;
}
adf_pf2vf_handle_pf_restarting(accel_dev);
return false;
case ADF_PF2VF_MSGTYPE_VERSION_RESP:
- dev_dbg(&GET_DEV(accel_dev),
- "Version resp received from PF 0x%x\n", msg);
- accel_dev->vf.pf_version =
- (msg & ADF_PF2VF_VERSION_RESP_VERS_MASK) >>
- ADF_PF2VF_VERSION_RESP_VERS_SHIFT;
- accel_dev->vf.compatible =
- (msg & ADF_PF2VF_VERSION_RESP_RESULT_MASK) >>
- ADF_PF2VF_VERSION_RESP_RESULT_SHIFT;
- complete(&accel_dev->vf.iov_msg_completion);
+ accel_dev->vf.response = msg;
+ complete(&accel_dev->vf.msg_received);
return true;
default:
dev_err(&GET_DEV(accel_dev),
#include "adf_accel_devices.h"
int adf_send_vf2pf_msg(struct adf_accel_dev *accel_dev, u32 msg);
-int adf_send_vf2pf_req(struct adf_accel_dev *accel_dev, u32 msg);
+int adf_send_vf2pf_req(struct adf_accel_dev *accel_dev, u32 msg, u32 *resp);
int adf_enable_vf2pf_comms(struct adf_accel_dev *accel_dev);
}
pci_set_master(pdev);
/* Completion for VF2PF request/response message exchange */
- init_completion(&accel_dev->vf.iov_msg_completion);
+ init_completion(&accel_dev->vf.msg_received);
ret = qat_crypto_dev_config(accel_dev);
if (ret)