1 // SPDX-License-Identifier: GPL-2.0+
2 // Copyright (c) 2016-2017 Hisilicon Limited.
4 #include <linux/etherdevice.h>
5 #include <linux/iopoll.h>
6 #include <net/rtnetlink.h>
7 #include "hclgevf_cmd.h"
8 #include "hclgevf_main.h"
9 #include "hclgevf_regs.h"
10 #include "hclge_mbx.h"
12 #include "hclgevf_devlink.h"
13 #include "hclge_comm_rss.h"
15 #define HCLGEVF_NAME "hclgevf"
17 #define HCLGEVF_RESET_MAX_FAIL_CNT 5
19 static int hclgevf_reset_hdev(struct hclgevf_dev *hdev);
20 static void hclgevf_task_schedule(struct hclgevf_dev *hdev,
23 static struct hnae3_ae_algo ae_algovf;
25 static struct workqueue_struct *hclgevf_wq;
27 static const struct pci_device_id ae_algovf_pci_tbl[] = {
28 {PCI_VDEVICE(HUAWEI, HNAE3_DEV_ID_VF), 0},
29 {PCI_VDEVICE(HUAWEI, HNAE3_DEV_ID_RDMA_DCB_PFC_VF),
30 HNAE3_DEV_SUPPORT_ROCE_DCB_BITS},
31 /* required last entry */
35 MODULE_DEVICE_TABLE(pci, ae_algovf_pci_tbl);
37 /* hclgevf_cmd_send - send command to command queue
38 * @hw: pointer to the hw struct
39 * @desc: prefilled descriptor for describing the command
40 * @num : the number of descriptors to be sent
42 * This is the main send command for command queue, it
43 * sends the queue, cleans the queue, etc
45 int hclgevf_cmd_send(struct hclgevf_hw *hw, struct hclge_desc *desc, int num)
47 return hclge_comm_cmd_send(&hw->hw, desc, num);
50 void hclgevf_arq_init(struct hclgevf_dev *hdev)
52 struct hclge_comm_cmq *cmdq = &hdev->hw.hw.cmq;
54 spin_lock(&cmdq->crq.lock);
55 /* initialize the pointers of async rx queue of mailbox */
56 hdev->arq.hdev = hdev;
59 atomic_set(&hdev->arq.count, 0);
60 spin_unlock(&cmdq->crq.lock);
63 struct hclgevf_dev *hclgevf_ae_get_hdev(struct hnae3_handle *handle)
66 return container_of(handle, struct hclgevf_dev, nic);
67 else if (handle->client->type == HNAE3_CLIENT_ROCE)
68 return container_of(handle, struct hclgevf_dev, roce);
70 return container_of(handle, struct hclgevf_dev, nic);
73 static void hclgevf_update_stats(struct hnae3_handle *handle)
75 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
78 status = hclge_comm_tqps_update_stats(handle, &hdev->hw.hw);
80 dev_err(&hdev->pdev->dev,
81 "VF update of TQPS stats fail, status = %d.\n",
85 static int hclgevf_get_sset_count(struct hnae3_handle *handle, int strset)
87 if (strset == ETH_SS_TEST)
89 else if (strset == ETH_SS_STATS)
90 return hclge_comm_tqps_get_sset_count(handle);
95 static void hclgevf_get_strings(struct hnae3_handle *handle, u32 strset,
100 if (strset == ETH_SS_STATS)
101 p = hclge_comm_tqps_get_strings(handle, p);
104 static void hclgevf_get_stats(struct hnae3_handle *handle, u64 *data)
106 hclge_comm_tqps_get_stats(handle, data);
109 static void hclgevf_build_send_msg(struct hclge_vf_to_pf_msg *msg, u8 code,
113 memset(msg, 0, sizeof(struct hclge_vf_to_pf_msg));
115 msg->subcode = subcode;
119 static int hclgevf_get_basic_info(struct hclgevf_dev *hdev)
121 struct hnae3_ae_dev *ae_dev = hdev->ae_dev;
122 u8 resp_msg[HCLGE_MBX_MAX_RESP_DATA_SIZE];
123 struct hclge_basic_info *basic_info;
124 struct hclge_vf_to_pf_msg send_msg;
128 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_GET_BASIC_INFO, 0);
129 status = hclgevf_send_mbx_msg(hdev, &send_msg, true, resp_msg,
132 dev_err(&hdev->pdev->dev,
133 "failed to get basic info from pf, ret = %d", status);
137 basic_info = (struct hclge_basic_info *)resp_msg;
139 hdev->hw_tc_map = basic_info->hw_tc_map;
140 hdev->mbx_api_version = le16_to_cpu(basic_info->mbx_api_version);
141 caps = le32_to_cpu(basic_info->pf_caps);
142 if (test_bit(HNAE3_PF_SUPPORT_VLAN_FLTR_MDF_B, &caps))
143 set_bit(HNAE3_DEV_SUPPORT_VLAN_FLTR_MDF_B, ae_dev->caps);
148 static int hclgevf_get_port_base_vlan_filter_state(struct hclgevf_dev *hdev)
150 struct hnae3_handle *nic = &hdev->nic;
151 struct hclge_vf_to_pf_msg send_msg;
155 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_SET_VLAN,
156 HCLGE_MBX_GET_PORT_BASE_VLAN_STATE);
157 ret = hclgevf_send_mbx_msg(hdev, &send_msg, true, &resp_msg,
160 dev_err(&hdev->pdev->dev,
161 "VF request to get port based vlan state failed %d",
166 nic->port_base_vlan_state = resp_msg;
171 static int hclgevf_get_queue_info(struct hclgevf_dev *hdev)
173 #define HCLGEVF_TQPS_RSS_INFO_LEN 6
175 struct hclge_mbx_vf_queue_info *queue_info;
176 u8 resp_msg[HCLGEVF_TQPS_RSS_INFO_LEN];
177 struct hclge_vf_to_pf_msg send_msg;
180 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_GET_QINFO, 0);
181 status = hclgevf_send_mbx_msg(hdev, &send_msg, true, resp_msg,
182 HCLGEVF_TQPS_RSS_INFO_LEN);
184 dev_err(&hdev->pdev->dev,
185 "VF request to get tqp info from PF failed %d",
190 queue_info = (struct hclge_mbx_vf_queue_info *)resp_msg;
191 hdev->num_tqps = le16_to_cpu(queue_info->num_tqps);
192 hdev->rss_size_max = le16_to_cpu(queue_info->rss_size);
193 hdev->rx_buf_len = le16_to_cpu(queue_info->rx_buf_len);
198 static int hclgevf_get_queue_depth(struct hclgevf_dev *hdev)
200 #define HCLGEVF_TQPS_DEPTH_INFO_LEN 4
202 struct hclge_mbx_vf_queue_depth *queue_depth;
203 u8 resp_msg[HCLGEVF_TQPS_DEPTH_INFO_LEN];
204 struct hclge_vf_to_pf_msg send_msg;
207 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_GET_QDEPTH, 0);
208 ret = hclgevf_send_mbx_msg(hdev, &send_msg, true, resp_msg,
209 HCLGEVF_TQPS_DEPTH_INFO_LEN);
211 dev_err(&hdev->pdev->dev,
212 "VF request to get tqp depth info from PF failed %d",
217 queue_depth = (struct hclge_mbx_vf_queue_depth *)resp_msg;
218 hdev->num_tx_desc = le16_to_cpu(queue_depth->num_tx_desc);
219 hdev->num_rx_desc = le16_to_cpu(queue_depth->num_rx_desc);
224 static u16 hclgevf_get_qid_global(struct hnae3_handle *handle, u16 queue_id)
226 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
227 struct hclge_vf_to_pf_msg send_msg;
232 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_GET_QID_IN_PF, 0);
233 *(__le16 *)send_msg.data = cpu_to_le16(queue_id);
234 ret = hclgevf_send_mbx_msg(hdev, &send_msg, true, resp_data,
237 qid_in_pf = le16_to_cpu(*(__le16 *)resp_data);
242 static int hclgevf_get_pf_media_type(struct hclgevf_dev *hdev)
244 struct hclge_vf_to_pf_msg send_msg;
248 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_GET_MEDIA_TYPE, 0);
249 ret = hclgevf_send_mbx_msg(hdev, &send_msg, true, resp_msg,
252 dev_err(&hdev->pdev->dev,
253 "VF request to get the pf port media type failed %d",
258 hdev->hw.mac.media_type = resp_msg[0];
259 hdev->hw.mac.module_type = resp_msg[1];
264 static int hclgevf_alloc_tqps(struct hclgevf_dev *hdev)
266 struct hnae3_ae_dev *ae_dev = pci_get_drvdata(hdev->pdev);
267 struct hclge_comm_tqp *tqp;
270 hdev->htqp = devm_kcalloc(&hdev->pdev->dev, hdev->num_tqps,
271 sizeof(struct hclge_comm_tqp), GFP_KERNEL);
277 for (i = 0; i < hdev->num_tqps; i++) {
278 tqp->dev = &hdev->pdev->dev;
281 tqp->q.ae_algo = &ae_algovf;
282 tqp->q.buf_size = hdev->rx_buf_len;
283 tqp->q.tx_desc_num = hdev->num_tx_desc;
284 tqp->q.rx_desc_num = hdev->num_rx_desc;
286 /* need an extended offset to configure queues >=
287 * HCLGEVF_TQP_MAX_SIZE_DEV_V2.
289 if (i < HCLGEVF_TQP_MAX_SIZE_DEV_V2)
290 tqp->q.io_base = hdev->hw.hw.io_base +
291 HCLGEVF_TQP_REG_OFFSET +
292 i * HCLGEVF_TQP_REG_SIZE;
294 tqp->q.io_base = hdev->hw.hw.io_base +
295 HCLGEVF_TQP_REG_OFFSET +
296 HCLGEVF_TQP_EXT_REG_OFFSET +
297 (i - HCLGEVF_TQP_MAX_SIZE_DEV_V2) *
298 HCLGEVF_TQP_REG_SIZE;
300 /* when device supports tx push and has device memory,
301 * the queue can execute push mode or doorbell mode on
304 if (test_bit(HNAE3_DEV_SUPPORT_TX_PUSH_B, ae_dev->caps))
305 tqp->q.mem_base = hdev->hw.hw.mem_base +
306 HCLGEVF_TQP_MEM_OFFSET(hdev, i);
314 static int hclgevf_knic_setup(struct hclgevf_dev *hdev)
316 struct hnae3_handle *nic = &hdev->nic;
317 struct hnae3_knic_private_info *kinfo;
318 u16 new_tqps = hdev->num_tqps;
323 kinfo->num_tx_desc = hdev->num_tx_desc;
324 kinfo->num_rx_desc = hdev->num_rx_desc;
325 kinfo->rx_buf_len = hdev->rx_buf_len;
326 for (i = 0; i < HCLGE_COMM_MAX_TC_NUM; i++)
327 if (hdev->hw_tc_map & BIT(i))
330 num_tc = num_tc ? num_tc : 1;
331 kinfo->tc_info.num_tc = num_tc;
332 kinfo->rss_size = min_t(u16, hdev->rss_size_max, new_tqps / num_tc);
333 new_tqps = kinfo->rss_size * num_tc;
334 kinfo->num_tqps = min(new_tqps, hdev->num_tqps);
336 kinfo->tqp = devm_kcalloc(&hdev->pdev->dev, kinfo->num_tqps,
337 sizeof(struct hnae3_queue *), GFP_KERNEL);
341 for (i = 0; i < kinfo->num_tqps; i++) {
342 hdev->htqp[i].q.handle = &hdev->nic;
343 hdev->htqp[i].q.tqp_index = i;
344 kinfo->tqp[i] = &hdev->htqp[i].q;
347 /* after init the max rss_size and tqps, adjust the default tqp numbers
348 * and rss size with the actual vector numbers
350 kinfo->num_tqps = min_t(u16, hdev->num_nic_msix - 1, kinfo->num_tqps);
351 kinfo->rss_size = min_t(u16, kinfo->num_tqps / num_tc,
357 static void hclgevf_request_link_info(struct hclgevf_dev *hdev)
359 struct hclge_vf_to_pf_msg send_msg;
362 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_GET_LINK_STATUS, 0);
363 status = hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
365 dev_err(&hdev->pdev->dev,
366 "VF failed to fetch link status(%d) from PF", status);
369 void hclgevf_update_link_status(struct hclgevf_dev *hdev, int link_state)
371 struct hnae3_handle *rhandle = &hdev->roce;
372 struct hnae3_handle *handle = &hdev->nic;
373 struct hnae3_client *rclient;
374 struct hnae3_client *client;
376 if (test_and_set_bit(HCLGEVF_STATE_LINK_UPDATING, &hdev->state))
379 client = handle->client;
380 rclient = hdev->roce_client;
383 test_bit(HCLGEVF_STATE_DOWN, &hdev->state) ? 0 : link_state;
384 if (link_state != hdev->hw.mac.link) {
385 hdev->hw.mac.link = link_state;
386 client->ops->link_status_change(handle, !!link_state);
387 if (rclient && rclient->ops->link_status_change)
388 rclient->ops->link_status_change(rhandle, !!link_state);
391 clear_bit(HCLGEVF_STATE_LINK_UPDATING, &hdev->state);
394 static void hclgevf_update_link_mode(struct hclgevf_dev *hdev)
396 #define HCLGEVF_ADVERTISING 0
397 #define HCLGEVF_SUPPORTED 1
399 struct hclge_vf_to_pf_msg send_msg;
401 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_GET_LINK_MODE, 0);
402 send_msg.data[0] = HCLGEVF_ADVERTISING;
403 hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
404 send_msg.data[0] = HCLGEVF_SUPPORTED;
405 hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
408 static int hclgevf_set_handle_info(struct hclgevf_dev *hdev)
410 struct hnae3_handle *nic = &hdev->nic;
413 nic->ae_algo = &ae_algovf;
414 nic->pdev = hdev->pdev;
415 nic->numa_node_mask = hdev->numa_node_mask;
416 nic->flags |= HNAE3_SUPPORT_VF;
417 nic->kinfo.io_base = hdev->hw.hw.io_base;
419 ret = hclgevf_knic_setup(hdev);
421 dev_err(&hdev->pdev->dev, "VF knic setup failed %d\n",
426 static void hclgevf_free_vector(struct hclgevf_dev *hdev, int vector_id)
428 if (hdev->vector_status[vector_id] == HCLGEVF_INVALID_VPORT) {
429 dev_warn(&hdev->pdev->dev,
430 "vector(vector_id %d) has been freed.\n", vector_id);
434 hdev->vector_status[vector_id] = HCLGEVF_INVALID_VPORT;
435 hdev->num_msi_left += 1;
436 hdev->num_msi_used -= 1;
439 static int hclgevf_get_vector(struct hnae3_handle *handle, u16 vector_num,
440 struct hnae3_vector_info *vector_info)
442 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
443 struct hnae3_vector_info *vector = vector_info;
447 vector_num = min_t(u16, hdev->num_nic_msix - 1, vector_num);
448 vector_num = min(hdev->num_msi_left, vector_num);
450 for (j = 0; j < vector_num; j++) {
451 for (i = HCLGEVF_MISC_VECTOR_NUM + 1; i < hdev->num_msi; i++) {
452 if (hdev->vector_status[i] == HCLGEVF_INVALID_VPORT) {
453 vector->vector = pci_irq_vector(hdev->pdev, i);
454 vector->io_addr = hdev->hw.hw.io_base +
455 HCLGEVF_VECTOR_REG_BASE +
456 (i - 1) * HCLGEVF_VECTOR_REG_OFFSET;
457 hdev->vector_status[i] = 0;
458 hdev->vector_irq[i] = vector->vector;
467 hdev->num_msi_left -= alloc;
468 hdev->num_msi_used += alloc;
473 static int hclgevf_get_vector_index(struct hclgevf_dev *hdev, int vector)
477 for (i = 0; i < hdev->num_msi; i++)
478 if (vector == hdev->vector_irq[i])
484 /* for revision 0x20, vf shared the same rss config with pf */
485 static int hclgevf_get_rss_hash_key(struct hclgevf_dev *hdev)
487 #define HCLGEVF_RSS_MBX_RESP_LEN 8
488 struct hclge_comm_rss_cfg *rss_cfg = &hdev->rss_cfg;
489 u8 resp_msg[HCLGEVF_RSS_MBX_RESP_LEN];
490 struct hclge_vf_to_pf_msg send_msg;
491 u16 msg_num, hash_key_index;
495 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_GET_RSS_KEY, 0);
496 msg_num = (HCLGE_COMM_RSS_KEY_SIZE + HCLGEVF_RSS_MBX_RESP_LEN - 1) /
497 HCLGEVF_RSS_MBX_RESP_LEN;
498 for (index = 0; index < msg_num; index++) {
499 send_msg.data[0] = index;
500 ret = hclgevf_send_mbx_msg(hdev, &send_msg, true, resp_msg,
501 HCLGEVF_RSS_MBX_RESP_LEN);
503 dev_err(&hdev->pdev->dev,
504 "VF get rss hash key from PF failed, ret=%d",
509 hash_key_index = HCLGEVF_RSS_MBX_RESP_LEN * index;
510 if (index == msg_num - 1)
511 memcpy(&rss_cfg->rss_hash_key[hash_key_index],
513 HCLGE_COMM_RSS_KEY_SIZE - hash_key_index);
515 memcpy(&rss_cfg->rss_hash_key[hash_key_index],
516 &resp_msg[0], HCLGEVF_RSS_MBX_RESP_LEN);
522 static int hclgevf_get_rss(struct hnae3_handle *handle, u32 *indir, u8 *key,
525 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
526 struct hclge_comm_rss_cfg *rss_cfg = &hdev->rss_cfg;
529 if (hdev->ae_dev->dev_version >= HNAE3_DEVICE_VERSION_V2) {
530 hclge_comm_get_rss_hash_info(rss_cfg, key, hfunc);
533 *hfunc = ETH_RSS_HASH_TOP;
535 ret = hclgevf_get_rss_hash_key(hdev);
538 memcpy(key, rss_cfg->rss_hash_key,
539 HCLGE_COMM_RSS_KEY_SIZE);
543 hclge_comm_get_rss_indir_tbl(rss_cfg, indir,
544 hdev->ae_dev->dev_specs.rss_ind_tbl_size);
549 static int hclgevf_set_rss(struct hnae3_handle *handle, const u32 *indir,
550 const u8 *key, const u8 hfunc)
552 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
553 struct hclge_comm_rss_cfg *rss_cfg = &hdev->rss_cfg;
556 if (hdev->ae_dev->dev_version >= HNAE3_DEVICE_VERSION_V2) {
557 ret = hclge_comm_set_rss_hash_key(rss_cfg, &hdev->hw.hw, key,
563 /* update the shadow RSS table with user specified qids */
564 for (i = 0; i < hdev->ae_dev->dev_specs.rss_ind_tbl_size; i++)
565 rss_cfg->rss_indirection_tbl[i] = indir[i];
567 /* update the hardware */
568 return hclge_comm_set_rss_indir_table(hdev->ae_dev, &hdev->hw.hw,
569 rss_cfg->rss_indirection_tbl);
572 static int hclgevf_set_rss_tuple(struct hnae3_handle *handle,
573 struct ethtool_rxnfc *nfc)
575 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
578 if (hdev->ae_dev->dev_version < HNAE3_DEVICE_VERSION_V2)
581 ret = hclge_comm_set_rss_tuple(hdev->ae_dev, &hdev->hw.hw,
582 &hdev->rss_cfg, nfc);
584 dev_err(&hdev->pdev->dev,
585 "failed to set rss tuple, ret = %d.\n", ret);
590 static int hclgevf_get_rss_tuple(struct hnae3_handle *handle,
591 struct ethtool_rxnfc *nfc)
593 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
597 if (hdev->ae_dev->dev_version < HNAE3_DEVICE_VERSION_V2)
602 ret = hclge_comm_get_rss_tuple(&hdev->rss_cfg, nfc->flow_type,
604 if (ret || !tuple_sets)
607 nfc->data = hclge_comm_convert_rss_tuple(tuple_sets);
612 static int hclgevf_get_tc_size(struct hnae3_handle *handle)
614 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
615 struct hclge_comm_rss_cfg *rss_cfg = &hdev->rss_cfg;
617 return rss_cfg->rss_size;
620 static int hclgevf_bind_ring_to_vector(struct hnae3_handle *handle, bool en,
622 struct hnae3_ring_chain_node *ring_chain)
624 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
625 struct hclge_vf_to_pf_msg send_msg;
626 struct hnae3_ring_chain_node *node;
630 memset(&send_msg, 0, sizeof(send_msg));
631 send_msg.code = en ? HCLGE_MBX_MAP_RING_TO_VECTOR :
632 HCLGE_MBX_UNMAP_RING_TO_VECTOR;
633 send_msg.vector_id = vector_id;
635 for (node = ring_chain; node; node = node->next) {
636 send_msg.param[i].ring_type =
637 hnae3_get_bit(node->flag, HNAE3_RING_TYPE_B);
639 send_msg.param[i].tqp_index = node->tqp_index;
640 send_msg.param[i].int_gl_index =
641 hnae3_get_field(node->int_gl_idx,
643 HNAE3_RING_GL_IDX_S);
646 if (i == HCLGE_MBX_MAX_RING_CHAIN_PARAM_NUM || !node->next) {
647 send_msg.ring_num = i;
649 status = hclgevf_send_mbx_msg(hdev, &send_msg, false,
652 dev_err(&hdev->pdev->dev,
653 "Map TQP fail, status is %d.\n",
664 static int hclgevf_map_ring_to_vector(struct hnae3_handle *handle, int vector,
665 struct hnae3_ring_chain_node *ring_chain)
667 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
670 vector_id = hclgevf_get_vector_index(hdev, vector);
672 dev_err(&handle->pdev->dev,
673 "Get vector index fail. ret =%d\n", vector_id);
677 return hclgevf_bind_ring_to_vector(handle, true, vector_id, ring_chain);
680 static int hclgevf_unmap_ring_from_vector(
681 struct hnae3_handle *handle,
683 struct hnae3_ring_chain_node *ring_chain)
685 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
688 if (test_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state))
691 vector_id = hclgevf_get_vector_index(hdev, vector);
693 dev_err(&handle->pdev->dev,
694 "Get vector index fail. ret =%d\n", vector_id);
698 ret = hclgevf_bind_ring_to_vector(handle, false, vector_id, ring_chain);
700 dev_err(&handle->pdev->dev,
701 "Unmap ring from vector fail. vector=%d, ret =%d\n",
708 static int hclgevf_put_vector(struct hnae3_handle *handle, int vector)
710 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
713 vector_id = hclgevf_get_vector_index(hdev, vector);
715 dev_err(&handle->pdev->dev,
716 "hclgevf_put_vector get vector index fail. ret =%d\n",
721 hclgevf_free_vector(hdev, vector_id);
726 static int hclgevf_cmd_set_promisc_mode(struct hclgevf_dev *hdev,
727 bool en_uc_pmc, bool en_mc_pmc,
730 struct hnae3_handle *handle = &hdev->nic;
731 struct hclge_vf_to_pf_msg send_msg;
734 memset(&send_msg, 0, sizeof(send_msg));
735 send_msg.code = HCLGE_MBX_SET_PROMISC_MODE;
736 send_msg.en_bc = en_bc_pmc ? 1 : 0;
737 send_msg.en_uc = en_uc_pmc ? 1 : 0;
738 send_msg.en_mc = en_mc_pmc ? 1 : 0;
739 send_msg.en_limit_promisc = test_bit(HNAE3_PFLAG_LIMIT_PROMISC,
740 &handle->priv_flags) ? 1 : 0;
742 ret = hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
744 dev_err(&hdev->pdev->dev,
745 "Set promisc mode fail, status is %d.\n", ret);
750 static int hclgevf_set_promisc_mode(struct hnae3_handle *handle, bool en_uc_pmc,
753 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
756 en_bc_pmc = hdev->ae_dev->dev_version >= HNAE3_DEVICE_VERSION_V2;
758 return hclgevf_cmd_set_promisc_mode(hdev, en_uc_pmc, en_mc_pmc,
762 static void hclgevf_request_update_promisc_mode(struct hnae3_handle *handle)
764 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
766 set_bit(HCLGEVF_STATE_PROMISC_CHANGED, &hdev->state);
767 hclgevf_task_schedule(hdev, 0);
770 static void hclgevf_sync_promisc_mode(struct hclgevf_dev *hdev)
772 struct hnae3_handle *handle = &hdev->nic;
773 bool en_uc_pmc = handle->netdev_flags & HNAE3_UPE;
774 bool en_mc_pmc = handle->netdev_flags & HNAE3_MPE;
777 if (test_bit(HCLGEVF_STATE_PROMISC_CHANGED, &hdev->state)) {
778 ret = hclgevf_set_promisc_mode(handle, en_uc_pmc, en_mc_pmc);
780 clear_bit(HCLGEVF_STATE_PROMISC_CHANGED, &hdev->state);
784 static int hclgevf_tqp_enable_cmd_send(struct hclgevf_dev *hdev, u16 tqp_id,
785 u16 stream_id, bool enable)
787 struct hclgevf_cfg_com_tqp_queue_cmd *req;
788 struct hclge_desc desc;
790 req = (struct hclgevf_cfg_com_tqp_queue_cmd *)desc.data;
792 hclgevf_cmd_setup_basic_desc(&desc, HCLGE_OPC_CFG_COM_TQP_QUEUE, false);
793 req->tqp_id = cpu_to_le16(tqp_id & HCLGEVF_RING_ID_MASK);
794 req->stream_id = cpu_to_le16(stream_id);
796 req->enable |= 1U << HCLGEVF_TQP_ENABLE_B;
798 return hclgevf_cmd_send(&hdev->hw, &desc, 1);
801 static int hclgevf_tqp_enable(struct hnae3_handle *handle, bool enable)
803 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
807 for (i = 0; i < handle->kinfo.num_tqps; i++) {
808 ret = hclgevf_tqp_enable_cmd_send(hdev, i, 0, enable);
816 static int hclgevf_get_host_mac_addr(struct hclgevf_dev *hdev, u8 *p)
818 struct hclge_vf_to_pf_msg send_msg;
819 u8 host_mac[ETH_ALEN];
822 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_GET_MAC_ADDR, 0);
823 status = hclgevf_send_mbx_msg(hdev, &send_msg, true, host_mac,
826 dev_err(&hdev->pdev->dev,
827 "fail to get VF MAC from host %d", status);
831 ether_addr_copy(p, host_mac);
836 static void hclgevf_get_mac_addr(struct hnae3_handle *handle, u8 *p)
838 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
839 u8 host_mac_addr[ETH_ALEN];
841 if (hclgevf_get_host_mac_addr(hdev, host_mac_addr))
844 hdev->has_pf_mac = !is_zero_ether_addr(host_mac_addr);
845 if (hdev->has_pf_mac)
846 ether_addr_copy(p, host_mac_addr);
848 ether_addr_copy(p, hdev->hw.mac.mac_addr);
851 static int hclgevf_set_mac_addr(struct hnae3_handle *handle, const void *p,
854 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
855 u8 *old_mac_addr = (u8 *)hdev->hw.mac.mac_addr;
856 struct hclge_vf_to_pf_msg send_msg;
857 u8 *new_mac_addr = (u8 *)p;
860 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_SET_UNICAST, 0);
861 send_msg.subcode = HCLGE_MBX_MAC_VLAN_UC_MODIFY;
862 ether_addr_copy(send_msg.data, new_mac_addr);
863 if (is_first && !hdev->has_pf_mac)
864 eth_zero_addr(&send_msg.data[ETH_ALEN]);
866 ether_addr_copy(&send_msg.data[ETH_ALEN], old_mac_addr);
867 status = hclgevf_send_mbx_msg(hdev, &send_msg, true, NULL, 0);
869 ether_addr_copy(hdev->hw.mac.mac_addr, new_mac_addr);
874 static struct hclgevf_mac_addr_node *
875 hclgevf_find_mac_node(struct list_head *list, const u8 *mac_addr)
877 struct hclgevf_mac_addr_node *mac_node, *tmp;
879 list_for_each_entry_safe(mac_node, tmp, list, node)
880 if (ether_addr_equal(mac_addr, mac_node->mac_addr))
886 static void hclgevf_update_mac_node(struct hclgevf_mac_addr_node *mac_node,
887 enum HCLGEVF_MAC_NODE_STATE state)
890 /* from set_rx_mode or tmp_add_list */
891 case HCLGEVF_MAC_TO_ADD:
892 if (mac_node->state == HCLGEVF_MAC_TO_DEL)
893 mac_node->state = HCLGEVF_MAC_ACTIVE;
895 /* only from set_rx_mode */
896 case HCLGEVF_MAC_TO_DEL:
897 if (mac_node->state == HCLGEVF_MAC_TO_ADD) {
898 list_del(&mac_node->node);
901 mac_node->state = HCLGEVF_MAC_TO_DEL;
904 /* only from tmp_add_list, the mac_node->state won't be
907 case HCLGEVF_MAC_ACTIVE:
908 if (mac_node->state == HCLGEVF_MAC_TO_ADD)
909 mac_node->state = HCLGEVF_MAC_ACTIVE;
914 static int hclgevf_update_mac_list(struct hnae3_handle *handle,
915 enum HCLGEVF_MAC_NODE_STATE state,
916 enum HCLGEVF_MAC_ADDR_TYPE mac_type,
917 const unsigned char *addr)
919 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
920 struct hclgevf_mac_addr_node *mac_node;
921 struct list_head *list;
923 list = (mac_type == HCLGEVF_MAC_ADDR_UC) ?
924 &hdev->mac_table.uc_mac_list : &hdev->mac_table.mc_mac_list;
926 spin_lock_bh(&hdev->mac_table.mac_list_lock);
928 /* if the mac addr is already in the mac list, no need to add a new
929 * one into it, just check the mac addr state, convert it to a new
930 * state, or just remove it, or do nothing.
932 mac_node = hclgevf_find_mac_node(list, addr);
934 hclgevf_update_mac_node(mac_node, state);
935 spin_unlock_bh(&hdev->mac_table.mac_list_lock);
938 /* if this address is never added, unnecessary to delete */
939 if (state == HCLGEVF_MAC_TO_DEL) {
940 spin_unlock_bh(&hdev->mac_table.mac_list_lock);
944 mac_node = kzalloc(sizeof(*mac_node), GFP_ATOMIC);
946 spin_unlock_bh(&hdev->mac_table.mac_list_lock);
950 mac_node->state = state;
951 ether_addr_copy(mac_node->mac_addr, addr);
952 list_add_tail(&mac_node->node, list);
954 spin_unlock_bh(&hdev->mac_table.mac_list_lock);
958 static int hclgevf_add_uc_addr(struct hnae3_handle *handle,
959 const unsigned char *addr)
961 return hclgevf_update_mac_list(handle, HCLGEVF_MAC_TO_ADD,
962 HCLGEVF_MAC_ADDR_UC, addr);
965 static int hclgevf_rm_uc_addr(struct hnae3_handle *handle,
966 const unsigned char *addr)
968 return hclgevf_update_mac_list(handle, HCLGEVF_MAC_TO_DEL,
969 HCLGEVF_MAC_ADDR_UC, addr);
972 static int hclgevf_add_mc_addr(struct hnae3_handle *handle,
973 const unsigned char *addr)
975 return hclgevf_update_mac_list(handle, HCLGEVF_MAC_TO_ADD,
976 HCLGEVF_MAC_ADDR_MC, addr);
979 static int hclgevf_rm_mc_addr(struct hnae3_handle *handle,
980 const unsigned char *addr)
982 return hclgevf_update_mac_list(handle, HCLGEVF_MAC_TO_DEL,
983 HCLGEVF_MAC_ADDR_MC, addr);
986 static int hclgevf_add_del_mac_addr(struct hclgevf_dev *hdev,
987 struct hclgevf_mac_addr_node *mac_node,
988 enum HCLGEVF_MAC_ADDR_TYPE mac_type)
990 struct hclge_vf_to_pf_msg send_msg;
993 if (mac_type == HCLGEVF_MAC_ADDR_UC) {
994 code = HCLGE_MBX_SET_UNICAST;
995 if (mac_node->state == HCLGEVF_MAC_TO_ADD)
996 subcode = HCLGE_MBX_MAC_VLAN_UC_ADD;
998 subcode = HCLGE_MBX_MAC_VLAN_UC_REMOVE;
1000 code = HCLGE_MBX_SET_MULTICAST;
1001 if (mac_node->state == HCLGEVF_MAC_TO_ADD)
1002 subcode = HCLGE_MBX_MAC_VLAN_MC_ADD;
1004 subcode = HCLGE_MBX_MAC_VLAN_MC_REMOVE;
1007 hclgevf_build_send_msg(&send_msg, code, subcode);
1008 ether_addr_copy(send_msg.data, mac_node->mac_addr);
1009 return hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
1012 static void hclgevf_config_mac_list(struct hclgevf_dev *hdev,
1013 struct list_head *list,
1014 enum HCLGEVF_MAC_ADDR_TYPE mac_type)
1016 char format_mac_addr[HNAE3_FORMAT_MAC_ADDR_LEN];
1017 struct hclgevf_mac_addr_node *mac_node, *tmp;
1020 list_for_each_entry_safe(mac_node, tmp, list, node) {
1021 ret = hclgevf_add_del_mac_addr(hdev, mac_node, mac_type);
1023 hnae3_format_mac_addr(format_mac_addr,
1024 mac_node->mac_addr);
1025 dev_err(&hdev->pdev->dev,
1026 "failed to configure mac %s, state = %d, ret = %d\n",
1027 format_mac_addr, mac_node->state, ret);
1030 if (mac_node->state == HCLGEVF_MAC_TO_ADD) {
1031 mac_node->state = HCLGEVF_MAC_ACTIVE;
1033 list_del(&mac_node->node);
1039 static void hclgevf_sync_from_add_list(struct list_head *add_list,
1040 struct list_head *mac_list)
1042 struct hclgevf_mac_addr_node *mac_node, *tmp, *new_node;
1044 list_for_each_entry_safe(mac_node, tmp, add_list, node) {
1045 /* if the mac address from tmp_add_list is not in the
1046 * uc/mc_mac_list, it means have received a TO_DEL request
1047 * during the time window of sending mac config request to PF
1048 * If mac_node state is ACTIVE, then change its state to TO_DEL,
1049 * then it will be removed at next time. If is TO_ADD, it means
1050 * send TO_ADD request failed, so just remove the mac node.
1052 new_node = hclgevf_find_mac_node(mac_list, mac_node->mac_addr);
1054 hclgevf_update_mac_node(new_node, mac_node->state);
1055 list_del(&mac_node->node);
1057 } else if (mac_node->state == HCLGEVF_MAC_ACTIVE) {
1058 mac_node->state = HCLGEVF_MAC_TO_DEL;
1059 list_move_tail(&mac_node->node, mac_list);
1061 list_del(&mac_node->node);
1067 static void hclgevf_sync_from_del_list(struct list_head *del_list,
1068 struct list_head *mac_list)
1070 struct hclgevf_mac_addr_node *mac_node, *tmp, *new_node;
1072 list_for_each_entry_safe(mac_node, tmp, del_list, node) {
1073 new_node = hclgevf_find_mac_node(mac_list, mac_node->mac_addr);
1075 /* If the mac addr is exist in the mac list, it means
1076 * received a new request TO_ADD during the time window
1077 * of sending mac addr configurrequest to PF, so just
1078 * change the mac state to ACTIVE.
1080 new_node->state = HCLGEVF_MAC_ACTIVE;
1081 list_del(&mac_node->node);
1084 list_move_tail(&mac_node->node, mac_list);
1089 static void hclgevf_clear_list(struct list_head *list)
1091 struct hclgevf_mac_addr_node *mac_node, *tmp;
1093 list_for_each_entry_safe(mac_node, tmp, list, node) {
1094 list_del(&mac_node->node);
1099 static void hclgevf_sync_mac_list(struct hclgevf_dev *hdev,
1100 enum HCLGEVF_MAC_ADDR_TYPE mac_type)
1102 struct hclgevf_mac_addr_node *mac_node, *tmp, *new_node;
1103 struct list_head tmp_add_list, tmp_del_list;
1104 struct list_head *list;
1106 INIT_LIST_HEAD(&tmp_add_list);
1107 INIT_LIST_HEAD(&tmp_del_list);
1109 /* move the mac addr to the tmp_add_list and tmp_del_list, then
1110 * we can add/delete these mac addr outside the spin lock
1112 list = (mac_type == HCLGEVF_MAC_ADDR_UC) ?
1113 &hdev->mac_table.uc_mac_list : &hdev->mac_table.mc_mac_list;
1115 spin_lock_bh(&hdev->mac_table.mac_list_lock);
1117 list_for_each_entry_safe(mac_node, tmp, list, node) {
1118 switch (mac_node->state) {
1119 case HCLGEVF_MAC_TO_DEL:
1120 list_move_tail(&mac_node->node, &tmp_del_list);
1122 case HCLGEVF_MAC_TO_ADD:
1123 new_node = kzalloc(sizeof(*new_node), GFP_ATOMIC);
1127 ether_addr_copy(new_node->mac_addr, mac_node->mac_addr);
1128 new_node->state = mac_node->state;
1129 list_add_tail(&new_node->node, &tmp_add_list);
1137 spin_unlock_bh(&hdev->mac_table.mac_list_lock);
1139 /* delete first, in order to get max mac table space for adding */
1140 hclgevf_config_mac_list(hdev, &tmp_del_list, mac_type);
1141 hclgevf_config_mac_list(hdev, &tmp_add_list, mac_type);
1143 /* if some mac addresses were added/deleted fail, move back to the
1144 * mac_list, and retry at next time.
1146 spin_lock_bh(&hdev->mac_table.mac_list_lock);
1148 hclgevf_sync_from_del_list(&tmp_del_list, list);
1149 hclgevf_sync_from_add_list(&tmp_add_list, list);
1151 spin_unlock_bh(&hdev->mac_table.mac_list_lock);
1154 static void hclgevf_sync_mac_table(struct hclgevf_dev *hdev)
1156 hclgevf_sync_mac_list(hdev, HCLGEVF_MAC_ADDR_UC);
1157 hclgevf_sync_mac_list(hdev, HCLGEVF_MAC_ADDR_MC);
1160 static void hclgevf_uninit_mac_list(struct hclgevf_dev *hdev)
1162 spin_lock_bh(&hdev->mac_table.mac_list_lock);
1164 hclgevf_clear_list(&hdev->mac_table.uc_mac_list);
1165 hclgevf_clear_list(&hdev->mac_table.mc_mac_list);
1167 spin_unlock_bh(&hdev->mac_table.mac_list_lock);
1170 static int hclgevf_enable_vlan_filter(struct hnae3_handle *handle, bool enable)
1172 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
1173 struct hnae3_ae_dev *ae_dev = hdev->ae_dev;
1174 struct hclge_vf_to_pf_msg send_msg;
1176 if (!test_bit(HNAE3_DEV_SUPPORT_VLAN_FLTR_MDF_B, ae_dev->caps))
1179 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_SET_VLAN,
1180 HCLGE_MBX_ENABLE_VLAN_FILTER);
1181 send_msg.data[0] = enable ? 1 : 0;
1183 return hclgevf_send_mbx_msg(hdev, &send_msg, true, NULL, 0);
1186 static int hclgevf_set_vlan_filter(struct hnae3_handle *handle,
1187 __be16 proto, u16 vlan_id,
1190 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
1191 struct hclge_mbx_vlan_filter *vlan_filter;
1192 struct hclge_vf_to_pf_msg send_msg;
1195 if (vlan_id > HCLGEVF_MAX_VLAN_ID)
1198 if (proto != htons(ETH_P_8021Q))
1199 return -EPROTONOSUPPORT;
1201 /* When device is resetting or reset failed, firmware is unable to
1202 * handle mailbox. Just record the vlan id, and remove it after
1205 if ((test_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state) ||
1206 test_bit(HCLGEVF_STATE_RST_FAIL, &hdev->state)) && is_kill) {
1207 set_bit(vlan_id, hdev->vlan_del_fail_bmap);
1211 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_SET_VLAN,
1212 HCLGE_MBX_VLAN_FILTER);
1213 vlan_filter = (struct hclge_mbx_vlan_filter *)send_msg.data;
1214 vlan_filter->is_kill = is_kill;
1215 vlan_filter->vlan_id = cpu_to_le16(vlan_id);
1216 vlan_filter->proto = cpu_to_le16(be16_to_cpu(proto));
1218 /* when remove hw vlan filter failed, record the vlan id,
1219 * and try to remove it from hw later, to be consistence
1222 ret = hclgevf_send_mbx_msg(hdev, &send_msg, true, NULL, 0);
1224 set_bit(vlan_id, hdev->vlan_del_fail_bmap);
1229 static void hclgevf_sync_vlan_filter(struct hclgevf_dev *hdev)
1231 #define HCLGEVF_MAX_SYNC_COUNT 60
1232 struct hnae3_handle *handle = &hdev->nic;
1233 int ret, sync_cnt = 0;
1236 vlan_id = find_first_bit(hdev->vlan_del_fail_bmap, VLAN_N_VID);
1237 while (vlan_id != VLAN_N_VID) {
1238 ret = hclgevf_set_vlan_filter(handle, htons(ETH_P_8021Q),
1243 clear_bit(vlan_id, hdev->vlan_del_fail_bmap);
1245 if (sync_cnt >= HCLGEVF_MAX_SYNC_COUNT)
1248 vlan_id = find_first_bit(hdev->vlan_del_fail_bmap, VLAN_N_VID);
1252 static int hclgevf_en_hw_strip_rxvtag(struct hnae3_handle *handle, bool enable)
1254 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
1255 struct hclge_vf_to_pf_msg send_msg;
1257 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_SET_VLAN,
1258 HCLGE_MBX_VLAN_RX_OFF_CFG);
1259 send_msg.data[0] = enable ? 1 : 0;
1260 return hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
1263 static int hclgevf_reset_tqp(struct hnae3_handle *handle)
1265 #define HCLGEVF_RESET_ALL_QUEUE_DONE 1U
1266 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
1267 struct hclge_vf_to_pf_msg send_msg;
1268 u8 return_status = 0;
1272 /* disable vf queue before send queue reset msg to PF */
1273 ret = hclgevf_tqp_enable(handle, false);
1275 dev_err(&hdev->pdev->dev, "failed to disable tqp, ret = %d\n",
1280 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_QUEUE_RESET, 0);
1282 ret = hclgevf_send_mbx_msg(hdev, &send_msg, true, &return_status,
1283 sizeof(return_status));
1284 if (ret || return_status == HCLGEVF_RESET_ALL_QUEUE_DONE)
1287 for (i = 1; i < handle->kinfo.num_tqps; i++) {
1288 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_QUEUE_RESET, 0);
1289 *(__le16 *)send_msg.data = cpu_to_le16(i);
1290 ret = hclgevf_send_mbx_msg(hdev, &send_msg, true, NULL, 0);
1298 static int hclgevf_set_mtu(struct hnae3_handle *handle, int new_mtu)
1300 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
1301 struct hclge_mbx_mtu_info *mtu_info;
1302 struct hclge_vf_to_pf_msg send_msg;
1304 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_SET_MTU, 0);
1305 mtu_info = (struct hclge_mbx_mtu_info *)send_msg.data;
1306 mtu_info->mtu = cpu_to_le32(new_mtu);
1308 return hclgevf_send_mbx_msg(hdev, &send_msg, true, NULL, 0);
1311 static int hclgevf_notify_client(struct hclgevf_dev *hdev,
1312 enum hnae3_reset_notify_type type)
1314 struct hnae3_client *client = hdev->nic_client;
1315 struct hnae3_handle *handle = &hdev->nic;
1318 if (!test_bit(HCLGEVF_STATE_NIC_REGISTERED, &hdev->state) ||
1322 if (!client->ops->reset_notify)
1325 ret = client->ops->reset_notify(handle, type);
1327 dev_err(&hdev->pdev->dev, "notify nic client failed %d(%d)\n",
1333 static int hclgevf_notify_roce_client(struct hclgevf_dev *hdev,
1334 enum hnae3_reset_notify_type type)
1336 struct hnae3_client *client = hdev->roce_client;
1337 struct hnae3_handle *handle = &hdev->roce;
1340 if (!test_bit(HCLGEVF_STATE_ROCE_REGISTERED, &hdev->state) || !client)
1343 if (!client->ops->reset_notify)
1346 ret = client->ops->reset_notify(handle, type);
1348 dev_err(&hdev->pdev->dev, "notify roce client failed %d(%d)",
1353 static int hclgevf_reset_wait(struct hclgevf_dev *hdev)
1355 #define HCLGEVF_RESET_WAIT_US 20000
1356 #define HCLGEVF_RESET_WAIT_CNT 2000
1357 #define HCLGEVF_RESET_WAIT_TIMEOUT_US \
1358 (HCLGEVF_RESET_WAIT_US * HCLGEVF_RESET_WAIT_CNT)
1363 if (hdev->reset_type == HNAE3_VF_RESET)
1364 ret = readl_poll_timeout(hdev->hw.hw.io_base +
1365 HCLGEVF_VF_RST_ING, val,
1366 !(val & HCLGEVF_VF_RST_ING_BIT),
1367 HCLGEVF_RESET_WAIT_US,
1368 HCLGEVF_RESET_WAIT_TIMEOUT_US);
1370 ret = readl_poll_timeout(hdev->hw.hw.io_base +
1371 HCLGEVF_RST_ING, val,
1372 !(val & HCLGEVF_RST_ING_BITS),
1373 HCLGEVF_RESET_WAIT_US,
1374 HCLGEVF_RESET_WAIT_TIMEOUT_US);
1376 /* hardware completion status should be available by this time */
1378 dev_err(&hdev->pdev->dev,
1379 "couldn't get reset done status from h/w, timeout!\n");
1383 /* we will wait a bit more to let reset of the stack to complete. This
1384 * might happen in case reset assertion was made by PF. Yes, this also
1385 * means we might end up waiting bit more even for VF reset.
1387 if (hdev->reset_type == HNAE3_VF_FULL_RESET)
1395 static void hclgevf_reset_handshake(struct hclgevf_dev *hdev, bool enable)
1399 reg_val = hclgevf_read_dev(&hdev->hw, HCLGE_COMM_NIC_CSQ_DEPTH_REG);
1401 reg_val |= HCLGEVF_NIC_SW_RST_RDY;
1403 reg_val &= ~HCLGEVF_NIC_SW_RST_RDY;
1405 hclgevf_write_dev(&hdev->hw, HCLGE_COMM_NIC_CSQ_DEPTH_REG,
1409 static int hclgevf_reset_stack(struct hclgevf_dev *hdev)
1413 /* uninitialize the nic client */
1414 ret = hclgevf_notify_client(hdev, HNAE3_UNINIT_CLIENT);
1418 /* re-initialize the hclge device */
1419 ret = hclgevf_reset_hdev(hdev);
1421 dev_err(&hdev->pdev->dev,
1422 "hclge device re-init failed, VF is disabled!\n");
1426 /* bring up the nic client again */
1427 ret = hclgevf_notify_client(hdev, HNAE3_INIT_CLIENT);
1431 /* clear handshake status with IMP */
1432 hclgevf_reset_handshake(hdev, false);
1434 /* bring up the nic to enable TX/RX again */
1435 return hclgevf_notify_client(hdev, HNAE3_UP_CLIENT);
1438 static int hclgevf_reset_prepare_wait(struct hclgevf_dev *hdev)
1440 #define HCLGEVF_RESET_SYNC_TIME 100
1442 if (hdev->reset_type == HNAE3_VF_FUNC_RESET) {
1443 struct hclge_vf_to_pf_msg send_msg;
1446 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_RESET, 0);
1447 ret = hclgevf_send_mbx_msg(hdev, &send_msg, true, NULL, 0);
1449 dev_err(&hdev->pdev->dev,
1450 "failed to assert VF reset, ret = %d\n", ret);
1453 hdev->rst_stats.vf_func_rst_cnt++;
1456 set_bit(HCLGE_COMM_STATE_CMD_DISABLE, &hdev->hw.hw.comm_state);
1457 /* inform hardware that preparatory work is done */
1458 msleep(HCLGEVF_RESET_SYNC_TIME);
1459 hclgevf_reset_handshake(hdev, true);
1460 dev_info(&hdev->pdev->dev, "prepare reset(%d) wait done\n",
1466 static void hclgevf_dump_rst_info(struct hclgevf_dev *hdev)
1468 dev_info(&hdev->pdev->dev, "VF function reset count: %u\n",
1469 hdev->rst_stats.vf_func_rst_cnt);
1470 dev_info(&hdev->pdev->dev, "FLR reset count: %u\n",
1471 hdev->rst_stats.flr_rst_cnt);
1472 dev_info(&hdev->pdev->dev, "VF reset count: %u\n",
1473 hdev->rst_stats.vf_rst_cnt);
1474 dev_info(&hdev->pdev->dev, "reset done count: %u\n",
1475 hdev->rst_stats.rst_done_cnt);
1476 dev_info(&hdev->pdev->dev, "HW reset done count: %u\n",
1477 hdev->rst_stats.hw_rst_done_cnt);
1478 dev_info(&hdev->pdev->dev, "reset count: %u\n",
1479 hdev->rst_stats.rst_cnt);
1480 dev_info(&hdev->pdev->dev, "reset fail count: %u\n",
1481 hdev->rst_stats.rst_fail_cnt);
1482 dev_info(&hdev->pdev->dev, "vector0 interrupt enable status: 0x%x\n",
1483 hclgevf_read_dev(&hdev->hw, HCLGEVF_MISC_VECTOR_REG_BASE));
1484 dev_info(&hdev->pdev->dev, "vector0 interrupt status: 0x%x\n",
1485 hclgevf_read_dev(&hdev->hw, HCLGE_COMM_VECTOR0_CMDQ_STATE_REG));
1486 dev_info(&hdev->pdev->dev, "handshake status: 0x%x\n",
1487 hclgevf_read_dev(&hdev->hw, HCLGE_COMM_NIC_CSQ_DEPTH_REG));
1488 dev_info(&hdev->pdev->dev, "function reset status: 0x%x\n",
1489 hclgevf_read_dev(&hdev->hw, HCLGEVF_RST_ING));
1490 dev_info(&hdev->pdev->dev, "hdev state: 0x%lx\n", hdev->state);
1493 static void hclgevf_reset_err_handle(struct hclgevf_dev *hdev)
1495 /* recover handshake status with IMP when reset fail */
1496 hclgevf_reset_handshake(hdev, true);
1497 hdev->rst_stats.rst_fail_cnt++;
1498 dev_err(&hdev->pdev->dev, "failed to reset VF(%u)\n",
1499 hdev->rst_stats.rst_fail_cnt);
1501 if (hdev->rst_stats.rst_fail_cnt < HCLGEVF_RESET_MAX_FAIL_CNT)
1502 set_bit(hdev->reset_type, &hdev->reset_pending);
1504 if (hclgevf_is_reset_pending(hdev)) {
1505 set_bit(HCLGEVF_RESET_PENDING, &hdev->reset_state);
1506 hclgevf_reset_task_schedule(hdev);
1508 set_bit(HCLGEVF_STATE_RST_FAIL, &hdev->state);
1509 hclgevf_dump_rst_info(hdev);
1513 static int hclgevf_reset_prepare(struct hclgevf_dev *hdev)
1517 hdev->rst_stats.rst_cnt++;
1519 /* perform reset of the stack & ae device for a client */
1520 ret = hclgevf_notify_roce_client(hdev, HNAE3_DOWN_CLIENT);
1525 /* bring down the nic to stop any ongoing TX/RX */
1526 ret = hclgevf_notify_client(hdev, HNAE3_DOWN_CLIENT);
1531 return hclgevf_reset_prepare_wait(hdev);
1534 static int hclgevf_reset_rebuild(struct hclgevf_dev *hdev)
1538 hdev->rst_stats.hw_rst_done_cnt++;
1539 ret = hclgevf_notify_roce_client(hdev, HNAE3_UNINIT_CLIENT);
1544 /* now, re-initialize the nic client and ae device */
1545 ret = hclgevf_reset_stack(hdev);
1548 dev_err(&hdev->pdev->dev, "failed to reset VF stack\n");
1552 ret = hclgevf_notify_roce_client(hdev, HNAE3_INIT_CLIENT);
1553 /* ignore RoCE notify error if it fails HCLGEVF_RESET_MAX_FAIL_CNT - 1
1557 hdev->rst_stats.rst_fail_cnt < HCLGEVF_RESET_MAX_FAIL_CNT - 1)
1560 ret = hclgevf_notify_roce_client(hdev, HNAE3_UP_CLIENT);
1564 hdev->last_reset_time = jiffies;
1565 hdev->rst_stats.rst_done_cnt++;
1566 hdev->rst_stats.rst_fail_cnt = 0;
1567 clear_bit(HCLGEVF_STATE_RST_FAIL, &hdev->state);
1572 static void hclgevf_reset(struct hclgevf_dev *hdev)
1574 if (hclgevf_reset_prepare(hdev))
1577 /* check if VF could successfully fetch the hardware reset completion
1578 * status from the hardware
1580 if (hclgevf_reset_wait(hdev)) {
1581 /* can't do much in this situation, will disable VF */
1582 dev_err(&hdev->pdev->dev,
1583 "failed to fetch H/W reset completion status\n");
1587 if (hclgevf_reset_rebuild(hdev))
1593 hclgevf_reset_err_handle(hdev);
1596 static enum hnae3_reset_type hclgevf_get_reset_level(unsigned long *addr)
1598 enum hnae3_reset_type rst_level = HNAE3_NONE_RESET;
1600 /* return the highest priority reset level amongst all */
1601 if (test_bit(HNAE3_VF_RESET, addr)) {
1602 rst_level = HNAE3_VF_RESET;
1603 clear_bit(HNAE3_VF_RESET, addr);
1604 clear_bit(HNAE3_VF_PF_FUNC_RESET, addr);
1605 clear_bit(HNAE3_VF_FUNC_RESET, addr);
1606 } else if (test_bit(HNAE3_VF_FULL_RESET, addr)) {
1607 rst_level = HNAE3_VF_FULL_RESET;
1608 clear_bit(HNAE3_VF_FULL_RESET, addr);
1609 clear_bit(HNAE3_VF_FUNC_RESET, addr);
1610 } else if (test_bit(HNAE3_VF_PF_FUNC_RESET, addr)) {
1611 rst_level = HNAE3_VF_PF_FUNC_RESET;
1612 clear_bit(HNAE3_VF_PF_FUNC_RESET, addr);
1613 clear_bit(HNAE3_VF_FUNC_RESET, addr);
1614 } else if (test_bit(HNAE3_VF_FUNC_RESET, addr)) {
1615 rst_level = HNAE3_VF_FUNC_RESET;
1616 clear_bit(HNAE3_VF_FUNC_RESET, addr);
1617 } else if (test_bit(HNAE3_FLR_RESET, addr)) {
1618 rst_level = HNAE3_FLR_RESET;
1619 clear_bit(HNAE3_FLR_RESET, addr);
1625 static void hclgevf_reset_event(struct pci_dev *pdev,
1626 struct hnae3_handle *handle)
1628 struct hnae3_ae_dev *ae_dev = pci_get_drvdata(pdev);
1629 struct hclgevf_dev *hdev = ae_dev->priv;
1631 dev_info(&hdev->pdev->dev, "received reset request from VF enet\n");
1633 if (hdev->default_reset_request)
1635 hclgevf_get_reset_level(&hdev->default_reset_request);
1637 hdev->reset_level = HNAE3_VF_FUNC_RESET;
1639 /* reset of this VF requested */
1640 set_bit(HCLGEVF_RESET_REQUESTED, &hdev->reset_state);
1641 hclgevf_reset_task_schedule(hdev);
1643 hdev->last_reset_time = jiffies;
1646 static void hclgevf_set_def_reset_request(struct hnae3_ae_dev *ae_dev,
1647 enum hnae3_reset_type rst_type)
1649 struct hclgevf_dev *hdev = ae_dev->priv;
1651 set_bit(rst_type, &hdev->default_reset_request);
1654 static void hclgevf_enable_vector(struct hclgevf_misc_vector *vector, bool en)
1656 writel(en ? 1 : 0, vector->addr);
1659 static void hclgevf_reset_prepare_general(struct hnae3_ae_dev *ae_dev,
1660 enum hnae3_reset_type rst_type)
1662 #define HCLGEVF_RESET_RETRY_WAIT_MS 500
1663 #define HCLGEVF_RESET_RETRY_CNT 5
1665 struct hclgevf_dev *hdev = ae_dev->priv;
1669 while (retry_cnt++ < HCLGEVF_RESET_RETRY_CNT) {
1670 down(&hdev->reset_sem);
1671 set_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state);
1672 hdev->reset_type = rst_type;
1673 ret = hclgevf_reset_prepare(hdev);
1674 if (!ret && !hdev->reset_pending)
1677 dev_err(&hdev->pdev->dev,
1678 "failed to prepare to reset, ret=%d, reset_pending:0x%lx, retry_cnt:%d\n",
1679 ret, hdev->reset_pending, retry_cnt);
1680 clear_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state);
1681 up(&hdev->reset_sem);
1682 msleep(HCLGEVF_RESET_RETRY_WAIT_MS);
1685 /* disable misc vector before reset done */
1686 hclgevf_enable_vector(&hdev->misc_vector, false);
1688 if (hdev->reset_type == HNAE3_FLR_RESET)
1689 hdev->rst_stats.flr_rst_cnt++;
1692 static void hclgevf_reset_done(struct hnae3_ae_dev *ae_dev)
1694 struct hclgevf_dev *hdev = ae_dev->priv;
1697 hclgevf_enable_vector(&hdev->misc_vector, true);
1699 ret = hclgevf_reset_rebuild(hdev);
1701 dev_warn(&hdev->pdev->dev, "fail to rebuild, ret=%d\n",
1704 hdev->reset_type = HNAE3_NONE_RESET;
1705 clear_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state);
1706 up(&hdev->reset_sem);
1709 static u32 hclgevf_get_fw_version(struct hnae3_handle *handle)
1711 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
1713 return hdev->fw_version;
1716 static void hclgevf_get_misc_vector(struct hclgevf_dev *hdev)
1718 struct hclgevf_misc_vector *vector = &hdev->misc_vector;
1720 vector->vector_irq = pci_irq_vector(hdev->pdev,
1721 HCLGEVF_MISC_VECTOR_NUM);
1722 vector->addr = hdev->hw.hw.io_base + HCLGEVF_MISC_VECTOR_REG_BASE;
1723 /* vector status always valid for Vector 0 */
1724 hdev->vector_status[HCLGEVF_MISC_VECTOR_NUM] = 0;
1725 hdev->vector_irq[HCLGEVF_MISC_VECTOR_NUM] = vector->vector_irq;
1727 hdev->num_msi_left -= 1;
1728 hdev->num_msi_used += 1;
1731 void hclgevf_reset_task_schedule(struct hclgevf_dev *hdev)
1733 if (!test_bit(HCLGEVF_STATE_REMOVING, &hdev->state) &&
1734 test_bit(HCLGEVF_STATE_SERVICE_INITED, &hdev->state) &&
1735 !test_and_set_bit(HCLGEVF_STATE_RST_SERVICE_SCHED,
1737 mod_delayed_work(hclgevf_wq, &hdev->service_task, 0);
1740 void hclgevf_mbx_task_schedule(struct hclgevf_dev *hdev)
1742 if (!test_bit(HCLGEVF_STATE_REMOVING, &hdev->state) &&
1743 !test_and_set_bit(HCLGEVF_STATE_MBX_SERVICE_SCHED,
1745 mod_delayed_work(hclgevf_wq, &hdev->service_task, 0);
1748 static void hclgevf_task_schedule(struct hclgevf_dev *hdev,
1749 unsigned long delay)
1751 if (!test_bit(HCLGEVF_STATE_REMOVING, &hdev->state) &&
1752 !test_bit(HCLGEVF_STATE_RST_FAIL, &hdev->state))
1753 mod_delayed_work(hclgevf_wq, &hdev->service_task, delay);
1756 static void hclgevf_reset_service_task(struct hclgevf_dev *hdev)
1758 #define HCLGEVF_MAX_RESET_ATTEMPTS_CNT 3
1760 if (!test_and_clear_bit(HCLGEVF_STATE_RST_SERVICE_SCHED, &hdev->state))
1763 down(&hdev->reset_sem);
1764 set_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state);
1766 if (test_and_clear_bit(HCLGEVF_RESET_PENDING,
1767 &hdev->reset_state)) {
1768 /* PF has intimated that it is about to reset the hardware.
1769 * We now have to poll & check if hardware has actually
1770 * completed the reset sequence. On hardware reset completion,
1771 * VF needs to reset the client and ae device.
1773 hdev->reset_attempts = 0;
1775 hdev->last_reset_time = jiffies;
1777 hclgevf_get_reset_level(&hdev->reset_pending);
1778 if (hdev->reset_type != HNAE3_NONE_RESET)
1779 hclgevf_reset(hdev);
1780 } else if (test_and_clear_bit(HCLGEVF_RESET_REQUESTED,
1781 &hdev->reset_state)) {
1782 /* we could be here when either of below happens:
1783 * 1. reset was initiated due to watchdog timeout caused by
1784 * a. IMP was earlier reset and our TX got choked down and
1785 * which resulted in watchdog reacting and inducing VF
1786 * reset. This also means our cmdq would be unreliable.
1787 * b. problem in TX due to other lower layer(example link
1788 * layer not functioning properly etc.)
1789 * 2. VF reset might have been initiated due to some config
1792 * NOTE: Theres no clear way to detect above cases than to react
1793 * to the response of PF for this reset request. PF will ack the
1794 * 1b and 2. cases but we will not get any intimation about 1a
1795 * from PF as cmdq would be in unreliable state i.e. mailbox
1796 * communication between PF and VF would be broken.
1798 * if we are never geting into pending state it means either:
1799 * 1. PF is not receiving our request which could be due to IMP
1802 * We cannot do much for 2. but to check first we can try reset
1803 * our PCIe + stack and see if it alleviates the problem.
1805 if (hdev->reset_attempts > HCLGEVF_MAX_RESET_ATTEMPTS_CNT) {
1806 /* prepare for full reset of stack + pcie interface */
1807 set_bit(HNAE3_VF_FULL_RESET, &hdev->reset_pending);
1809 /* "defer" schedule the reset task again */
1810 set_bit(HCLGEVF_RESET_PENDING, &hdev->reset_state);
1812 hdev->reset_attempts++;
1814 set_bit(hdev->reset_level, &hdev->reset_pending);
1815 set_bit(HCLGEVF_RESET_PENDING, &hdev->reset_state);
1817 hclgevf_reset_task_schedule(hdev);
1820 hdev->reset_type = HNAE3_NONE_RESET;
1821 clear_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state);
1822 up(&hdev->reset_sem);
1825 static void hclgevf_mailbox_service_task(struct hclgevf_dev *hdev)
1827 if (!test_and_clear_bit(HCLGEVF_STATE_MBX_SERVICE_SCHED, &hdev->state))
1830 if (test_and_set_bit(HCLGEVF_STATE_MBX_HANDLING, &hdev->state))
1833 hclgevf_mbx_async_handler(hdev);
1835 clear_bit(HCLGEVF_STATE_MBX_HANDLING, &hdev->state);
1838 static void hclgevf_keep_alive(struct hclgevf_dev *hdev)
1840 struct hclge_vf_to_pf_msg send_msg;
1843 if (test_bit(HCLGE_COMM_STATE_CMD_DISABLE, &hdev->hw.hw.comm_state))
1846 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_KEEP_ALIVE, 0);
1847 ret = hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
1849 dev_err(&hdev->pdev->dev,
1850 "VF sends keep alive cmd failed(=%d)\n", ret);
1853 static void hclgevf_periodic_service_task(struct hclgevf_dev *hdev)
1855 unsigned long delta = round_jiffies_relative(HZ);
1856 struct hnae3_handle *handle = &hdev->nic;
1858 if (test_bit(HCLGEVF_STATE_RST_FAIL, &hdev->state))
1861 if (time_is_after_jiffies(hdev->last_serv_processed + HZ)) {
1862 delta = jiffies - hdev->last_serv_processed;
1864 if (delta < round_jiffies_relative(HZ)) {
1865 delta = round_jiffies_relative(HZ) - delta;
1870 hdev->serv_processed_cnt++;
1871 if (!(hdev->serv_processed_cnt % HCLGEVF_KEEP_ALIVE_TASK_INTERVAL))
1872 hclgevf_keep_alive(hdev);
1874 if (test_bit(HCLGEVF_STATE_DOWN, &hdev->state)) {
1875 hdev->last_serv_processed = jiffies;
1879 if (!(hdev->serv_processed_cnt % HCLGEVF_STATS_TIMER_INTERVAL))
1880 hclge_comm_tqps_update_stats(handle, &hdev->hw.hw);
1882 /* VF does not need to request link status when this bit is set, because
1883 * PF will push its link status to VFs when link status changed.
1885 if (!test_bit(HCLGEVF_STATE_PF_PUSH_LINK_STATUS, &hdev->state))
1886 hclgevf_request_link_info(hdev);
1888 hclgevf_update_link_mode(hdev);
1890 hclgevf_sync_vlan_filter(hdev);
1892 hclgevf_sync_mac_table(hdev);
1894 hclgevf_sync_promisc_mode(hdev);
1896 hdev->last_serv_processed = jiffies;
1899 hclgevf_task_schedule(hdev, delta);
1902 static void hclgevf_service_task(struct work_struct *work)
1904 struct hclgevf_dev *hdev = container_of(work, struct hclgevf_dev,
1907 hclgevf_reset_service_task(hdev);
1908 hclgevf_mailbox_service_task(hdev);
1909 hclgevf_periodic_service_task(hdev);
1911 /* Handle reset and mbx again in case periodical task delays the
1912 * handling by calling hclgevf_task_schedule() in
1913 * hclgevf_periodic_service_task()
1915 hclgevf_reset_service_task(hdev);
1916 hclgevf_mailbox_service_task(hdev);
1919 static void hclgevf_clear_event_cause(struct hclgevf_dev *hdev, u32 regclr)
1921 hclgevf_write_dev(&hdev->hw, HCLGE_COMM_VECTOR0_CMDQ_SRC_REG, regclr);
1924 static enum hclgevf_evt_cause hclgevf_check_evt_cause(struct hclgevf_dev *hdev,
1927 u32 val, cmdq_stat_reg, rst_ing_reg;
1929 /* fetch the events from their corresponding regs */
1930 cmdq_stat_reg = hclgevf_read_dev(&hdev->hw,
1931 HCLGE_COMM_VECTOR0_CMDQ_STATE_REG);
1932 if (BIT(HCLGEVF_VECTOR0_RST_INT_B) & cmdq_stat_reg) {
1933 rst_ing_reg = hclgevf_read_dev(&hdev->hw, HCLGEVF_RST_ING);
1934 dev_info(&hdev->pdev->dev,
1935 "receive reset interrupt 0x%x!\n", rst_ing_reg);
1936 set_bit(HNAE3_VF_RESET, &hdev->reset_pending);
1937 set_bit(HCLGEVF_RESET_PENDING, &hdev->reset_state);
1938 set_bit(HCLGE_COMM_STATE_CMD_DISABLE, &hdev->hw.hw.comm_state);
1939 *clearval = ~(1U << HCLGEVF_VECTOR0_RST_INT_B);
1940 hdev->rst_stats.vf_rst_cnt++;
1941 /* set up VF hardware reset status, its PF will clear
1942 * this status when PF has initialized done.
1944 val = hclgevf_read_dev(&hdev->hw, HCLGEVF_VF_RST_ING);
1945 hclgevf_write_dev(&hdev->hw, HCLGEVF_VF_RST_ING,
1946 val | HCLGEVF_VF_RST_ING_BIT);
1947 return HCLGEVF_VECTOR0_EVENT_RST;
1950 /* check for vector0 mailbox(=CMDQ RX) event source */
1951 if (BIT(HCLGEVF_VECTOR0_RX_CMDQ_INT_B) & cmdq_stat_reg) {
1952 /* for revision 0x21, clearing interrupt is writing bit 0
1953 * to the clear register, writing bit 1 means to keep the
1955 * for revision 0x20, the clear register is a read & write
1956 * register, so we should just write 0 to the bit we are
1957 * handling, and keep other bits as cmdq_stat_reg.
1959 if (hdev->ae_dev->dev_version >= HNAE3_DEVICE_VERSION_V2)
1960 *clearval = ~(1U << HCLGEVF_VECTOR0_RX_CMDQ_INT_B);
1962 *clearval = cmdq_stat_reg &
1963 ~BIT(HCLGEVF_VECTOR0_RX_CMDQ_INT_B);
1965 return HCLGEVF_VECTOR0_EVENT_MBX;
1968 /* print other vector0 event source */
1969 dev_info(&hdev->pdev->dev,
1970 "vector 0 interrupt from unknown source, cmdq_src = %#x\n",
1973 return HCLGEVF_VECTOR0_EVENT_OTHER;
1976 static irqreturn_t hclgevf_misc_irq_handle(int irq, void *data)
1978 enum hclgevf_evt_cause event_cause;
1979 struct hclgevf_dev *hdev = data;
1982 hclgevf_enable_vector(&hdev->misc_vector, false);
1983 event_cause = hclgevf_check_evt_cause(hdev, &clearval);
1984 if (event_cause != HCLGEVF_VECTOR0_EVENT_OTHER)
1985 hclgevf_clear_event_cause(hdev, clearval);
1987 switch (event_cause) {
1988 case HCLGEVF_VECTOR0_EVENT_RST:
1989 hclgevf_reset_task_schedule(hdev);
1991 case HCLGEVF_VECTOR0_EVENT_MBX:
1992 hclgevf_mbx_handler(hdev);
1998 hclgevf_enable_vector(&hdev->misc_vector, true);
2003 static int hclgevf_configure(struct hclgevf_dev *hdev)
2007 hdev->gro_en = true;
2009 ret = hclgevf_get_basic_info(hdev);
2013 /* get current port based vlan state from PF */
2014 ret = hclgevf_get_port_base_vlan_filter_state(hdev);
2018 /* get queue configuration from PF */
2019 ret = hclgevf_get_queue_info(hdev);
2023 /* get queue depth info from PF */
2024 ret = hclgevf_get_queue_depth(hdev);
2028 return hclgevf_get_pf_media_type(hdev);
2031 static int hclgevf_alloc_hdev(struct hnae3_ae_dev *ae_dev)
2033 struct pci_dev *pdev = ae_dev->pdev;
2034 struct hclgevf_dev *hdev;
2036 hdev = devm_kzalloc(&pdev->dev, sizeof(*hdev), GFP_KERNEL);
2041 hdev->ae_dev = ae_dev;
2042 ae_dev->priv = hdev;
2047 static int hclgevf_init_roce_base_info(struct hclgevf_dev *hdev)
2049 struct hnae3_handle *roce = &hdev->roce;
2050 struct hnae3_handle *nic = &hdev->nic;
2052 roce->rinfo.num_vectors = hdev->num_roce_msix;
2054 if (hdev->num_msi_left < roce->rinfo.num_vectors ||
2055 hdev->num_msi_left == 0)
2058 roce->rinfo.base_vector = hdev->roce_base_msix_offset;
2060 roce->rinfo.netdev = nic->kinfo.netdev;
2061 roce->rinfo.roce_io_base = hdev->hw.hw.io_base;
2062 roce->rinfo.roce_mem_base = hdev->hw.hw.mem_base;
2064 roce->pdev = nic->pdev;
2065 roce->ae_algo = nic->ae_algo;
2066 roce->numa_node_mask = nic->numa_node_mask;
2071 static int hclgevf_config_gro(struct hclgevf_dev *hdev)
2073 struct hclgevf_cfg_gro_status_cmd *req;
2074 struct hclge_desc desc;
2077 if (!hnae3_ae_dev_gro_supported(hdev->ae_dev))
2080 hclgevf_cmd_setup_basic_desc(&desc, HCLGE_OPC_GRO_GENERIC_CONFIG,
2082 req = (struct hclgevf_cfg_gro_status_cmd *)desc.data;
2084 req->gro_en = hdev->gro_en ? 1 : 0;
2086 ret = hclgevf_cmd_send(&hdev->hw, &desc, 1);
2088 dev_err(&hdev->pdev->dev,
2089 "VF GRO hardware config cmd failed, ret = %d.\n", ret);
2094 static int hclgevf_rss_init_hw(struct hclgevf_dev *hdev)
2096 struct hclge_comm_rss_cfg *rss_cfg = &hdev->rss_cfg;
2097 u16 tc_offset[HCLGE_COMM_MAX_TC_NUM];
2098 u16 tc_valid[HCLGE_COMM_MAX_TC_NUM];
2099 u16 tc_size[HCLGE_COMM_MAX_TC_NUM];
2102 if (hdev->ae_dev->dev_version >= HNAE3_DEVICE_VERSION_V2) {
2103 ret = hclge_comm_set_rss_algo_key(&hdev->hw.hw,
2105 rss_cfg->rss_hash_key);
2109 ret = hclge_comm_set_rss_input_tuple(&hdev->hw.hw, rss_cfg);
2114 ret = hclge_comm_set_rss_indir_table(hdev->ae_dev, &hdev->hw.hw,
2115 rss_cfg->rss_indirection_tbl);
2119 hclge_comm_get_rss_tc_info(rss_cfg->rss_size, hdev->hw_tc_map,
2120 tc_offset, tc_valid, tc_size);
2122 return hclge_comm_set_rss_tc_mode(&hdev->hw.hw, tc_offset,
2126 static int hclgevf_init_vlan_config(struct hclgevf_dev *hdev)
2128 struct hnae3_handle *nic = &hdev->nic;
2131 ret = hclgevf_en_hw_strip_rxvtag(nic, true);
2133 dev_err(&hdev->pdev->dev,
2134 "failed to enable rx vlan offload, ret = %d\n", ret);
2138 return hclgevf_set_vlan_filter(&hdev->nic, htons(ETH_P_8021Q), 0,
2142 static void hclgevf_flush_link_update(struct hclgevf_dev *hdev)
2144 #define HCLGEVF_FLUSH_LINK_TIMEOUT 100000
2146 unsigned long last = hdev->serv_processed_cnt;
2149 while (test_bit(HCLGEVF_STATE_LINK_UPDATING, &hdev->state) &&
2150 i++ < HCLGEVF_FLUSH_LINK_TIMEOUT &&
2151 last == hdev->serv_processed_cnt)
2155 static void hclgevf_set_timer_task(struct hnae3_handle *handle, bool enable)
2157 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
2160 hclgevf_task_schedule(hdev, 0);
2162 set_bit(HCLGEVF_STATE_DOWN, &hdev->state);
2164 /* flush memory to make sure DOWN is seen by service task */
2165 smp_mb__before_atomic();
2166 hclgevf_flush_link_update(hdev);
2170 static int hclgevf_ae_start(struct hnae3_handle *handle)
2172 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
2174 clear_bit(HCLGEVF_STATE_DOWN, &hdev->state);
2175 clear_bit(HCLGEVF_STATE_PF_PUSH_LINK_STATUS, &hdev->state);
2177 hclge_comm_reset_tqp_stats(handle);
2179 hclgevf_request_link_info(hdev);
2181 hclgevf_update_link_mode(hdev);
2186 static void hclgevf_ae_stop(struct hnae3_handle *handle)
2188 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
2190 set_bit(HCLGEVF_STATE_DOWN, &hdev->state);
2192 if (hdev->reset_type != HNAE3_VF_RESET)
2193 hclgevf_reset_tqp(handle);
2195 hclge_comm_reset_tqp_stats(handle);
2196 hclgevf_update_link_status(hdev, 0);
2199 static int hclgevf_set_alive(struct hnae3_handle *handle, bool alive)
2201 #define HCLGEVF_STATE_ALIVE 1
2202 #define HCLGEVF_STATE_NOT_ALIVE 0
2204 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
2205 struct hclge_vf_to_pf_msg send_msg;
2207 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_SET_ALIVE, 0);
2208 send_msg.data[0] = alive ? HCLGEVF_STATE_ALIVE :
2209 HCLGEVF_STATE_NOT_ALIVE;
2210 return hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
2213 static int hclgevf_client_start(struct hnae3_handle *handle)
2215 return hclgevf_set_alive(handle, true);
2218 static void hclgevf_client_stop(struct hnae3_handle *handle)
2220 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
2223 ret = hclgevf_set_alive(handle, false);
2225 dev_warn(&hdev->pdev->dev,
2226 "%s failed %d\n", __func__, ret);
2229 static void hclgevf_state_init(struct hclgevf_dev *hdev)
2231 clear_bit(HCLGEVF_STATE_MBX_SERVICE_SCHED, &hdev->state);
2232 clear_bit(HCLGEVF_STATE_MBX_HANDLING, &hdev->state);
2233 clear_bit(HCLGEVF_STATE_RST_FAIL, &hdev->state);
2235 INIT_DELAYED_WORK(&hdev->service_task, hclgevf_service_task);
2237 mutex_init(&hdev->mbx_resp.mbx_mutex);
2238 sema_init(&hdev->reset_sem, 1);
2240 spin_lock_init(&hdev->mac_table.mac_list_lock);
2241 INIT_LIST_HEAD(&hdev->mac_table.uc_mac_list);
2242 INIT_LIST_HEAD(&hdev->mac_table.mc_mac_list);
2244 /* bring the device down */
2245 set_bit(HCLGEVF_STATE_DOWN, &hdev->state);
2248 static void hclgevf_state_uninit(struct hclgevf_dev *hdev)
2250 set_bit(HCLGEVF_STATE_DOWN, &hdev->state);
2251 set_bit(HCLGEVF_STATE_REMOVING, &hdev->state);
2253 if (hdev->service_task.work.func)
2254 cancel_delayed_work_sync(&hdev->service_task);
2256 mutex_destroy(&hdev->mbx_resp.mbx_mutex);
2259 static int hclgevf_init_msi(struct hclgevf_dev *hdev)
2261 struct pci_dev *pdev = hdev->pdev;
2265 if (hnae3_dev_roce_supported(hdev))
2266 vectors = pci_alloc_irq_vectors(pdev,
2267 hdev->roce_base_msix_offset + 1,
2271 vectors = pci_alloc_irq_vectors(pdev, HNAE3_MIN_VECTOR_NUM,
2273 PCI_IRQ_MSI | PCI_IRQ_MSIX);
2277 "failed(%d) to allocate MSI/MSI-X vectors\n",
2281 if (vectors < hdev->num_msi)
2282 dev_warn(&hdev->pdev->dev,
2283 "requested %u MSI/MSI-X, but allocated %d MSI/MSI-X\n",
2284 hdev->num_msi, vectors);
2286 hdev->num_msi = vectors;
2287 hdev->num_msi_left = vectors;
2289 hdev->vector_status = devm_kcalloc(&pdev->dev, hdev->num_msi,
2290 sizeof(u16), GFP_KERNEL);
2291 if (!hdev->vector_status) {
2292 pci_free_irq_vectors(pdev);
2296 for (i = 0; i < hdev->num_msi; i++)
2297 hdev->vector_status[i] = HCLGEVF_INVALID_VPORT;
2299 hdev->vector_irq = devm_kcalloc(&pdev->dev, hdev->num_msi,
2300 sizeof(int), GFP_KERNEL);
2301 if (!hdev->vector_irq) {
2302 devm_kfree(&pdev->dev, hdev->vector_status);
2303 pci_free_irq_vectors(pdev);
2310 static void hclgevf_uninit_msi(struct hclgevf_dev *hdev)
2312 struct pci_dev *pdev = hdev->pdev;
2314 devm_kfree(&pdev->dev, hdev->vector_status);
2315 devm_kfree(&pdev->dev, hdev->vector_irq);
2316 pci_free_irq_vectors(pdev);
2319 static int hclgevf_misc_irq_init(struct hclgevf_dev *hdev)
2323 hclgevf_get_misc_vector(hdev);
2325 snprintf(hdev->misc_vector.name, HNAE3_INT_NAME_LEN, "%s-misc-%s",
2326 HCLGEVF_NAME, pci_name(hdev->pdev));
2327 ret = request_irq(hdev->misc_vector.vector_irq, hclgevf_misc_irq_handle,
2328 0, hdev->misc_vector.name, hdev);
2330 dev_err(&hdev->pdev->dev, "VF failed to request misc irq(%d)\n",
2331 hdev->misc_vector.vector_irq);
2335 hclgevf_clear_event_cause(hdev, 0);
2337 /* enable misc. vector(vector 0) */
2338 hclgevf_enable_vector(&hdev->misc_vector, true);
2343 static void hclgevf_misc_irq_uninit(struct hclgevf_dev *hdev)
2345 /* disable misc vector(vector 0) */
2346 hclgevf_enable_vector(&hdev->misc_vector, false);
2347 synchronize_irq(hdev->misc_vector.vector_irq);
2348 free_irq(hdev->misc_vector.vector_irq, hdev);
2349 hclgevf_free_vector(hdev, 0);
2352 static void hclgevf_info_show(struct hclgevf_dev *hdev)
2354 struct device *dev = &hdev->pdev->dev;
2356 dev_info(dev, "VF info begin:\n");
2358 dev_info(dev, "Task queue pairs numbers: %u\n", hdev->num_tqps);
2359 dev_info(dev, "Desc num per TX queue: %u\n", hdev->num_tx_desc);
2360 dev_info(dev, "Desc num per RX queue: %u\n", hdev->num_rx_desc);
2361 dev_info(dev, "Numbers of vports: %u\n", hdev->num_alloc_vport);
2362 dev_info(dev, "HW tc map: 0x%x\n", hdev->hw_tc_map);
2363 dev_info(dev, "PF media type of this VF: %u\n",
2364 hdev->hw.mac.media_type);
2366 dev_info(dev, "VF info end.\n");
2369 static int hclgevf_init_nic_client_instance(struct hnae3_ae_dev *ae_dev,
2370 struct hnae3_client *client)
2372 struct hclgevf_dev *hdev = ae_dev->priv;
2373 int rst_cnt = hdev->rst_stats.rst_cnt;
2376 ret = client->ops->init_instance(&hdev->nic);
2380 set_bit(HCLGEVF_STATE_NIC_REGISTERED, &hdev->state);
2381 if (test_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state) ||
2382 rst_cnt != hdev->rst_stats.rst_cnt) {
2383 clear_bit(HCLGEVF_STATE_NIC_REGISTERED, &hdev->state);
2385 client->ops->uninit_instance(&hdev->nic, 0);
2389 hnae3_set_client_init_flag(client, ae_dev, 1);
2391 if (netif_msg_drv(&hdev->nic))
2392 hclgevf_info_show(hdev);
2397 static int hclgevf_init_roce_client_instance(struct hnae3_ae_dev *ae_dev,
2398 struct hnae3_client *client)
2400 struct hclgevf_dev *hdev = ae_dev->priv;
2403 if (!hnae3_dev_roce_supported(hdev) || !hdev->roce_client ||
2407 ret = hclgevf_init_roce_base_info(hdev);
2411 ret = client->ops->init_instance(&hdev->roce);
2415 set_bit(HCLGEVF_STATE_ROCE_REGISTERED, &hdev->state);
2416 hnae3_set_client_init_flag(client, ae_dev, 1);
2421 static int hclgevf_init_client_instance(struct hnae3_client *client,
2422 struct hnae3_ae_dev *ae_dev)
2424 struct hclgevf_dev *hdev = ae_dev->priv;
2427 switch (client->type) {
2428 case HNAE3_CLIENT_KNIC:
2429 hdev->nic_client = client;
2430 hdev->nic.client = client;
2432 ret = hclgevf_init_nic_client_instance(ae_dev, client);
2436 ret = hclgevf_init_roce_client_instance(ae_dev,
2442 case HNAE3_CLIENT_ROCE:
2443 if (hnae3_dev_roce_supported(hdev)) {
2444 hdev->roce_client = client;
2445 hdev->roce.client = client;
2448 ret = hclgevf_init_roce_client_instance(ae_dev, client);
2460 hdev->nic_client = NULL;
2461 hdev->nic.client = NULL;
2464 hdev->roce_client = NULL;
2465 hdev->roce.client = NULL;
2469 static void hclgevf_uninit_client_instance(struct hnae3_client *client,
2470 struct hnae3_ae_dev *ae_dev)
2472 struct hclgevf_dev *hdev = ae_dev->priv;
2474 /* un-init roce, if it exists */
2475 if (hdev->roce_client) {
2476 while (test_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state))
2477 msleep(HCLGEVF_WAIT_RESET_DONE);
2478 clear_bit(HCLGEVF_STATE_ROCE_REGISTERED, &hdev->state);
2480 hdev->roce_client->ops->uninit_instance(&hdev->roce, 0);
2481 hdev->roce_client = NULL;
2482 hdev->roce.client = NULL;
2485 /* un-init nic/unic, if this was not called by roce client */
2486 if (client->ops->uninit_instance && hdev->nic_client &&
2487 client->type != HNAE3_CLIENT_ROCE) {
2488 while (test_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state))
2489 msleep(HCLGEVF_WAIT_RESET_DONE);
2490 clear_bit(HCLGEVF_STATE_NIC_REGISTERED, &hdev->state);
2492 client->ops->uninit_instance(&hdev->nic, 0);
2493 hdev->nic_client = NULL;
2494 hdev->nic.client = NULL;
2498 static int hclgevf_dev_mem_map(struct hclgevf_dev *hdev)
2500 struct pci_dev *pdev = hdev->pdev;
2501 struct hclgevf_hw *hw = &hdev->hw;
2503 /* for device does not have device memory, return directly */
2504 if (!(pci_select_bars(pdev, IORESOURCE_MEM) & BIT(HCLGEVF_MEM_BAR)))
2508 devm_ioremap_wc(&pdev->dev,
2509 pci_resource_start(pdev, HCLGEVF_MEM_BAR),
2510 pci_resource_len(pdev, HCLGEVF_MEM_BAR));
2511 if (!hw->hw.mem_base) {
2512 dev_err(&pdev->dev, "failed to map device memory\n");
2519 static int hclgevf_pci_init(struct hclgevf_dev *hdev)
2521 struct pci_dev *pdev = hdev->pdev;
2522 struct hclgevf_hw *hw;
2525 ret = pci_enable_device(pdev);
2527 dev_err(&pdev->dev, "failed to enable PCI device\n");
2531 ret = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
2533 dev_err(&pdev->dev, "can't set consistent PCI DMA, exiting");
2534 goto err_disable_device;
2537 ret = pci_request_regions(pdev, HCLGEVF_DRIVER_NAME);
2539 dev_err(&pdev->dev, "PCI request regions failed %d\n", ret);
2540 goto err_disable_device;
2543 pci_set_master(pdev);
2545 hw->hw.io_base = pci_iomap(pdev, 2, 0);
2546 if (!hw->hw.io_base) {
2547 dev_err(&pdev->dev, "can't map configuration register space\n");
2549 goto err_release_regions;
2552 ret = hclgevf_dev_mem_map(hdev);
2554 goto err_unmap_io_base;
2559 pci_iounmap(pdev, hdev->hw.hw.io_base);
2560 err_release_regions:
2561 pci_release_regions(pdev);
2563 pci_disable_device(pdev);
2568 static void hclgevf_pci_uninit(struct hclgevf_dev *hdev)
2570 struct pci_dev *pdev = hdev->pdev;
2572 if (hdev->hw.hw.mem_base)
2573 devm_iounmap(&pdev->dev, hdev->hw.hw.mem_base);
2575 pci_iounmap(pdev, hdev->hw.hw.io_base);
2576 pci_release_regions(pdev);
2577 pci_disable_device(pdev);
2580 static int hclgevf_query_vf_resource(struct hclgevf_dev *hdev)
2582 struct hclgevf_query_res_cmd *req;
2583 struct hclge_desc desc;
2586 hclgevf_cmd_setup_basic_desc(&desc, HCLGE_OPC_QUERY_VF_RSRC, true);
2587 ret = hclgevf_cmd_send(&hdev->hw, &desc, 1);
2589 dev_err(&hdev->pdev->dev,
2590 "query vf resource failed, ret = %d.\n", ret);
2594 req = (struct hclgevf_query_res_cmd *)desc.data;
2596 if (hnae3_dev_roce_supported(hdev)) {
2597 hdev->roce_base_msix_offset =
2598 hnae3_get_field(le16_to_cpu(req->msixcap_localid_ba_rocee),
2599 HCLGEVF_MSIX_OFT_ROCEE_M,
2600 HCLGEVF_MSIX_OFT_ROCEE_S);
2601 hdev->num_roce_msix =
2602 hnae3_get_field(le16_to_cpu(req->vf_intr_vector_number),
2603 HCLGEVF_VEC_NUM_M, HCLGEVF_VEC_NUM_S);
2605 /* nic's msix numbers is always equals to the roce's. */
2606 hdev->num_nic_msix = hdev->num_roce_msix;
2608 /* VF should have NIC vectors and Roce vectors, NIC vectors
2609 * are queued before Roce vectors. The offset is fixed to 64.
2611 hdev->num_msi = hdev->num_roce_msix +
2612 hdev->roce_base_msix_offset;
2615 hnae3_get_field(le16_to_cpu(req->vf_intr_vector_number),
2616 HCLGEVF_VEC_NUM_M, HCLGEVF_VEC_NUM_S);
2618 hdev->num_nic_msix = hdev->num_msi;
2621 if (hdev->num_nic_msix < HNAE3_MIN_VECTOR_NUM) {
2622 dev_err(&hdev->pdev->dev,
2623 "Just %u msi resources, not enough for vf(min:2).\n",
2624 hdev->num_nic_msix);
2631 static void hclgevf_set_default_dev_specs(struct hclgevf_dev *hdev)
2633 #define HCLGEVF_MAX_NON_TSO_BD_NUM 8U
2635 struct hnae3_ae_dev *ae_dev = pci_get_drvdata(hdev->pdev);
2637 ae_dev->dev_specs.max_non_tso_bd_num =
2638 HCLGEVF_MAX_NON_TSO_BD_NUM;
2639 ae_dev->dev_specs.rss_ind_tbl_size = HCLGEVF_RSS_IND_TBL_SIZE;
2640 ae_dev->dev_specs.rss_key_size = HCLGE_COMM_RSS_KEY_SIZE;
2641 ae_dev->dev_specs.max_int_gl = HCLGEVF_DEF_MAX_INT_GL;
2642 ae_dev->dev_specs.max_frm_size = HCLGEVF_MAC_MAX_FRAME;
2645 static void hclgevf_parse_dev_specs(struct hclgevf_dev *hdev,
2646 struct hclge_desc *desc)
2648 struct hnae3_ae_dev *ae_dev = pci_get_drvdata(hdev->pdev);
2649 struct hclgevf_dev_specs_0_cmd *req0;
2650 struct hclgevf_dev_specs_1_cmd *req1;
2652 req0 = (struct hclgevf_dev_specs_0_cmd *)desc[0].data;
2653 req1 = (struct hclgevf_dev_specs_1_cmd *)desc[1].data;
2655 ae_dev->dev_specs.max_non_tso_bd_num = req0->max_non_tso_bd_num;
2656 ae_dev->dev_specs.rss_ind_tbl_size =
2657 le16_to_cpu(req0->rss_ind_tbl_size);
2658 ae_dev->dev_specs.int_ql_max = le16_to_cpu(req0->int_ql_max);
2659 ae_dev->dev_specs.rss_key_size = le16_to_cpu(req0->rss_key_size);
2660 ae_dev->dev_specs.max_int_gl = le16_to_cpu(req1->max_int_gl);
2661 ae_dev->dev_specs.max_frm_size = le16_to_cpu(req1->max_frm_size);
2664 static void hclgevf_check_dev_specs(struct hclgevf_dev *hdev)
2666 struct hnae3_dev_specs *dev_specs = &hdev->ae_dev->dev_specs;
2668 if (!dev_specs->max_non_tso_bd_num)
2669 dev_specs->max_non_tso_bd_num = HCLGEVF_MAX_NON_TSO_BD_NUM;
2670 if (!dev_specs->rss_ind_tbl_size)
2671 dev_specs->rss_ind_tbl_size = HCLGEVF_RSS_IND_TBL_SIZE;
2672 if (!dev_specs->rss_key_size)
2673 dev_specs->rss_key_size = HCLGE_COMM_RSS_KEY_SIZE;
2674 if (!dev_specs->max_int_gl)
2675 dev_specs->max_int_gl = HCLGEVF_DEF_MAX_INT_GL;
2676 if (!dev_specs->max_frm_size)
2677 dev_specs->max_frm_size = HCLGEVF_MAC_MAX_FRAME;
2680 static int hclgevf_query_dev_specs(struct hclgevf_dev *hdev)
2682 struct hclge_desc desc[HCLGEVF_QUERY_DEV_SPECS_BD_NUM];
2686 /* set default specifications as devices lower than version V3 do not
2687 * support querying specifications from firmware.
2689 if (hdev->ae_dev->dev_version < HNAE3_DEVICE_VERSION_V3) {
2690 hclgevf_set_default_dev_specs(hdev);
2694 for (i = 0; i < HCLGEVF_QUERY_DEV_SPECS_BD_NUM - 1; i++) {
2695 hclgevf_cmd_setup_basic_desc(&desc[i],
2696 HCLGE_OPC_QUERY_DEV_SPECS, true);
2697 desc[i].flag |= cpu_to_le16(HCLGE_COMM_CMD_FLAG_NEXT);
2699 hclgevf_cmd_setup_basic_desc(&desc[i], HCLGE_OPC_QUERY_DEV_SPECS, true);
2701 ret = hclgevf_cmd_send(&hdev->hw, desc, HCLGEVF_QUERY_DEV_SPECS_BD_NUM);
2705 hclgevf_parse_dev_specs(hdev, desc);
2706 hclgevf_check_dev_specs(hdev);
2711 static int hclgevf_pci_reset(struct hclgevf_dev *hdev)
2713 struct pci_dev *pdev = hdev->pdev;
2716 if ((hdev->reset_type == HNAE3_VF_FULL_RESET ||
2717 hdev->reset_type == HNAE3_FLR_RESET) &&
2718 test_bit(HCLGEVF_STATE_IRQ_INITED, &hdev->state)) {
2719 hclgevf_misc_irq_uninit(hdev);
2720 hclgevf_uninit_msi(hdev);
2721 clear_bit(HCLGEVF_STATE_IRQ_INITED, &hdev->state);
2724 if (!test_bit(HCLGEVF_STATE_IRQ_INITED, &hdev->state)) {
2725 pci_set_master(pdev);
2726 ret = hclgevf_init_msi(hdev);
2729 "failed(%d) to init MSI/MSI-X\n", ret);
2733 ret = hclgevf_misc_irq_init(hdev);
2735 hclgevf_uninit_msi(hdev);
2736 dev_err(&pdev->dev, "failed(%d) to init Misc IRQ(vector0)\n",
2741 set_bit(HCLGEVF_STATE_IRQ_INITED, &hdev->state);
2747 static int hclgevf_clear_vport_list(struct hclgevf_dev *hdev)
2749 struct hclge_vf_to_pf_msg send_msg;
2751 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_HANDLE_VF_TBL,
2752 HCLGE_MBX_VPORT_LIST_CLEAR);
2753 return hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
2756 static void hclgevf_init_rxd_adv_layout(struct hclgevf_dev *hdev)
2758 if (hnae3_ae_dev_rxd_adv_layout_supported(hdev->ae_dev))
2759 hclgevf_write_dev(&hdev->hw, HCLGEVF_RXD_ADV_LAYOUT_EN_REG, 1);
2762 static void hclgevf_uninit_rxd_adv_layout(struct hclgevf_dev *hdev)
2764 if (hnae3_ae_dev_rxd_adv_layout_supported(hdev->ae_dev))
2765 hclgevf_write_dev(&hdev->hw, HCLGEVF_RXD_ADV_LAYOUT_EN_REG, 0);
2768 static int hclgevf_reset_hdev(struct hclgevf_dev *hdev)
2770 struct pci_dev *pdev = hdev->pdev;
2773 ret = hclgevf_pci_reset(hdev);
2775 dev_err(&pdev->dev, "pci reset failed %d\n", ret);
2779 hclgevf_arq_init(hdev);
2780 ret = hclge_comm_cmd_init(hdev->ae_dev, &hdev->hw.hw,
2781 &hdev->fw_version, false,
2782 hdev->reset_pending);
2784 dev_err(&pdev->dev, "cmd failed %d\n", ret);
2788 ret = hclgevf_rss_init_hw(hdev);
2790 dev_err(&hdev->pdev->dev,
2791 "failed(%d) to initialize RSS\n", ret);
2795 ret = hclgevf_config_gro(hdev);
2799 ret = hclgevf_init_vlan_config(hdev);
2801 dev_err(&hdev->pdev->dev,
2802 "failed(%d) to initialize VLAN config\n", ret);
2806 /* get current port based vlan state from PF */
2807 ret = hclgevf_get_port_base_vlan_filter_state(hdev);
2811 set_bit(HCLGEVF_STATE_PROMISC_CHANGED, &hdev->state);
2813 hclgevf_init_rxd_adv_layout(hdev);
2815 dev_info(&hdev->pdev->dev, "Reset done\n");
2820 static int hclgevf_init_hdev(struct hclgevf_dev *hdev)
2822 struct pci_dev *pdev = hdev->pdev;
2825 ret = hclgevf_pci_init(hdev);
2829 ret = hclgevf_devlink_init(hdev);
2831 goto err_devlink_init;
2833 ret = hclge_comm_cmd_queue_init(hdev->pdev, &hdev->hw.hw);
2835 goto err_cmd_queue_init;
2837 hclgevf_arq_init(hdev);
2838 ret = hclge_comm_cmd_init(hdev->ae_dev, &hdev->hw.hw,
2839 &hdev->fw_version, false,
2840 hdev->reset_pending);
2844 /* Get vf resource */
2845 ret = hclgevf_query_vf_resource(hdev);
2849 ret = hclgevf_query_dev_specs(hdev);
2852 "failed to query dev specifications, ret = %d\n", ret);
2856 ret = hclgevf_init_msi(hdev);
2858 dev_err(&pdev->dev, "failed(%d) to init MSI/MSI-X\n", ret);
2862 hclgevf_state_init(hdev);
2863 hdev->reset_level = HNAE3_VF_FUNC_RESET;
2864 hdev->reset_type = HNAE3_NONE_RESET;
2866 ret = hclgevf_misc_irq_init(hdev);
2868 goto err_misc_irq_init;
2870 set_bit(HCLGEVF_STATE_IRQ_INITED, &hdev->state);
2872 ret = hclgevf_configure(hdev);
2874 dev_err(&pdev->dev, "failed(%d) to fetch configuration\n", ret);
2878 ret = hclgevf_alloc_tqps(hdev);
2880 dev_err(&pdev->dev, "failed(%d) to allocate TQPs\n", ret);
2884 ret = hclgevf_set_handle_info(hdev);
2888 ret = hclgevf_config_gro(hdev);
2892 /* Initialize RSS for this VF */
2893 ret = hclge_comm_rss_init_cfg(&hdev->nic, hdev->ae_dev,
2896 dev_err(&pdev->dev, "failed to init rss cfg, ret = %d\n", ret);
2900 ret = hclgevf_rss_init_hw(hdev);
2902 dev_err(&hdev->pdev->dev,
2903 "failed(%d) to initialize RSS\n", ret);
2907 /* ensure vf tbl list as empty before init */
2908 ret = hclgevf_clear_vport_list(hdev);
2911 "failed to clear tbl list configuration, ret = %d.\n",
2916 ret = hclgevf_init_vlan_config(hdev);
2918 dev_err(&hdev->pdev->dev,
2919 "failed(%d) to initialize VLAN config\n", ret);
2923 hclgevf_init_rxd_adv_layout(hdev);
2925 set_bit(HCLGEVF_STATE_SERVICE_INITED, &hdev->state);
2927 hdev->last_reset_time = jiffies;
2928 dev_info(&hdev->pdev->dev, "finished initializing %s driver\n",
2929 HCLGEVF_DRIVER_NAME);
2931 hclgevf_task_schedule(hdev, round_jiffies_relative(HZ));
2936 hclgevf_misc_irq_uninit(hdev);
2938 hclgevf_state_uninit(hdev);
2939 hclgevf_uninit_msi(hdev);
2941 hclge_comm_cmd_uninit(hdev->ae_dev, &hdev->hw.hw);
2943 hclgevf_devlink_uninit(hdev);
2945 hclgevf_pci_uninit(hdev);
2946 clear_bit(HCLGEVF_STATE_IRQ_INITED, &hdev->state);
2950 static void hclgevf_uninit_hdev(struct hclgevf_dev *hdev)
2952 struct hclge_vf_to_pf_msg send_msg;
2954 hclgevf_state_uninit(hdev);
2955 hclgevf_uninit_rxd_adv_layout(hdev);
2957 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_VF_UNINIT, 0);
2958 hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
2960 if (test_bit(HCLGEVF_STATE_IRQ_INITED, &hdev->state)) {
2961 hclgevf_misc_irq_uninit(hdev);
2962 hclgevf_uninit_msi(hdev);
2965 hclge_comm_cmd_uninit(hdev->ae_dev, &hdev->hw.hw);
2966 hclgevf_devlink_uninit(hdev);
2967 hclgevf_pci_uninit(hdev);
2968 hclgevf_uninit_mac_list(hdev);
2971 static int hclgevf_init_ae_dev(struct hnae3_ae_dev *ae_dev)
2973 struct pci_dev *pdev = ae_dev->pdev;
2976 ret = hclgevf_alloc_hdev(ae_dev);
2978 dev_err(&pdev->dev, "hclge device allocation failed\n");
2982 ret = hclgevf_init_hdev(ae_dev->priv);
2984 dev_err(&pdev->dev, "hclge device initialization failed\n");
2991 static void hclgevf_uninit_ae_dev(struct hnae3_ae_dev *ae_dev)
2993 struct hclgevf_dev *hdev = ae_dev->priv;
2995 hclgevf_uninit_hdev(hdev);
2996 ae_dev->priv = NULL;
2999 static u32 hclgevf_get_max_channels(struct hclgevf_dev *hdev)
3001 struct hnae3_handle *nic = &hdev->nic;
3002 struct hnae3_knic_private_info *kinfo = &nic->kinfo;
3004 return min_t(u32, hdev->rss_size_max,
3005 hdev->num_tqps / kinfo->tc_info.num_tc);
3009 * hclgevf_get_channels - Get the current channels enabled and max supported.
3010 * @handle: hardware information for network interface
3011 * @ch: ethtool channels structure
3013 * We don't support separate tx and rx queues as channels. The other count
3014 * represents how many queues are being used for control. max_combined counts
3015 * how many queue pairs we can support. They may not be mapped 1 to 1 with
3016 * q_vectors since we support a lot more queue pairs than q_vectors.
3018 static void hclgevf_get_channels(struct hnae3_handle *handle,
3019 struct ethtool_channels *ch)
3021 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3023 ch->max_combined = hclgevf_get_max_channels(hdev);
3024 ch->other_count = 0;
3026 ch->combined_count = handle->kinfo.rss_size;
3029 static void hclgevf_get_tqps_and_rss_info(struct hnae3_handle *handle,
3030 u16 *alloc_tqps, u16 *max_rss_size)
3032 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3034 *alloc_tqps = hdev->num_tqps;
3035 *max_rss_size = hdev->rss_size_max;
3038 static void hclgevf_update_rss_size(struct hnae3_handle *handle,
3041 struct hnae3_knic_private_info *kinfo = &handle->kinfo;
3042 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3045 kinfo->req_rss_size = new_tqps_num;
3047 max_rss_size = min_t(u16, hdev->rss_size_max,
3048 hdev->num_tqps / kinfo->tc_info.num_tc);
3050 /* Use the user's configuration when it is not larger than
3051 * max_rss_size, otherwise, use the maximum specification value.
3053 if (kinfo->req_rss_size != kinfo->rss_size && kinfo->req_rss_size &&
3054 kinfo->req_rss_size <= max_rss_size)
3055 kinfo->rss_size = kinfo->req_rss_size;
3056 else if (kinfo->rss_size > max_rss_size ||
3057 (!kinfo->req_rss_size && kinfo->rss_size < max_rss_size))
3058 kinfo->rss_size = max_rss_size;
3060 kinfo->num_tqps = kinfo->tc_info.num_tc * kinfo->rss_size;
3063 static int hclgevf_set_channels(struct hnae3_handle *handle, u32 new_tqps_num,
3064 bool rxfh_configured)
3066 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3067 struct hnae3_knic_private_info *kinfo = &handle->kinfo;
3068 u16 tc_offset[HCLGE_COMM_MAX_TC_NUM];
3069 u16 tc_valid[HCLGE_COMM_MAX_TC_NUM];
3070 u16 tc_size[HCLGE_COMM_MAX_TC_NUM];
3071 u16 cur_rss_size = kinfo->rss_size;
3072 u16 cur_tqps = kinfo->num_tqps;
3077 hclgevf_update_rss_size(handle, new_tqps_num);
3079 hclge_comm_get_rss_tc_info(kinfo->rss_size, hdev->hw_tc_map,
3080 tc_offset, tc_valid, tc_size);
3081 ret = hclge_comm_set_rss_tc_mode(&hdev->hw.hw, tc_offset,
3086 /* RSS indirection table has been configured by user */
3087 if (rxfh_configured)
3090 /* Reinitializes the rss indirect table according to the new RSS size */
3091 rss_indir = kcalloc(hdev->ae_dev->dev_specs.rss_ind_tbl_size,
3092 sizeof(u32), GFP_KERNEL);
3096 for (i = 0; i < hdev->ae_dev->dev_specs.rss_ind_tbl_size; i++)
3097 rss_indir[i] = i % kinfo->rss_size;
3099 hdev->rss_cfg.rss_size = kinfo->rss_size;
3101 ret = hclgevf_set_rss(handle, rss_indir, NULL, 0);
3103 dev_err(&hdev->pdev->dev, "set rss indir table fail, ret=%d\n",
3110 dev_info(&hdev->pdev->dev,
3111 "Channels changed, rss_size from %u to %u, tqps from %u to %u",
3112 cur_rss_size, kinfo->rss_size,
3113 cur_tqps, kinfo->rss_size * kinfo->tc_info.num_tc);
3118 static int hclgevf_get_status(struct hnae3_handle *handle)
3120 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3122 return hdev->hw.mac.link;
3125 static void hclgevf_get_ksettings_an_result(struct hnae3_handle *handle,
3126 u8 *auto_neg, u32 *speed,
3127 u8 *duplex, u32 *lane_num)
3129 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3132 *speed = hdev->hw.mac.speed;
3134 *duplex = hdev->hw.mac.duplex;
3136 *auto_neg = AUTONEG_DISABLE;
3139 void hclgevf_update_speed_duplex(struct hclgevf_dev *hdev, u32 speed,
3142 hdev->hw.mac.speed = speed;
3143 hdev->hw.mac.duplex = duplex;
3146 static int hclgevf_gro_en(struct hnae3_handle *handle, bool enable)
3148 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3149 bool gro_en_old = hdev->gro_en;
3152 hdev->gro_en = enable;
3153 ret = hclgevf_config_gro(hdev);
3155 hdev->gro_en = gro_en_old;
3160 static void hclgevf_get_media_type(struct hnae3_handle *handle, u8 *media_type,
3163 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3166 *media_type = hdev->hw.mac.media_type;
3169 *module_type = hdev->hw.mac.module_type;
3172 static bool hclgevf_get_hw_reset_stat(struct hnae3_handle *handle)
3174 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3176 return !!hclgevf_read_dev(&hdev->hw, HCLGEVF_RST_ING);
3179 static bool hclgevf_get_cmdq_stat(struct hnae3_handle *handle)
3181 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3183 return test_bit(HCLGE_COMM_STATE_CMD_DISABLE, &hdev->hw.hw.comm_state);
3186 static bool hclgevf_ae_dev_resetting(struct hnae3_handle *handle)
3188 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3190 return test_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state);
3193 static unsigned long hclgevf_ae_dev_reset_cnt(struct hnae3_handle *handle)
3195 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3197 return hdev->rst_stats.hw_rst_done_cnt;
3200 static void hclgevf_get_link_mode(struct hnae3_handle *handle,
3201 unsigned long *supported,
3202 unsigned long *advertising)
3204 struct hclgevf_dev *hdev = hclgevf_ae_get_hdev(handle);
3206 *supported = hdev->hw.mac.supported;
3207 *advertising = hdev->hw.mac.advertising;
3210 void hclgevf_update_port_base_vlan_info(struct hclgevf_dev *hdev, u16 state,
3211 struct hclge_mbx_port_base_vlan *port_base_vlan)
3213 struct hnae3_handle *nic = &hdev->nic;
3214 struct hclge_vf_to_pf_msg send_msg;
3219 if (test_bit(HCLGEVF_STATE_RST_HANDLING, &hdev->state) ||
3220 test_bit(HCLGEVF_STATE_RST_FAIL, &hdev->state)) {
3221 dev_warn(&hdev->pdev->dev,
3222 "is resetting when updating port based vlan info\n");
3227 ret = hclgevf_notify_client(hdev, HNAE3_DOWN_CLIENT);
3233 /* send msg to PF and wait update port based vlan info */
3234 hclgevf_build_send_msg(&send_msg, HCLGE_MBX_SET_VLAN,
3235 HCLGE_MBX_PORT_BASE_VLAN_CFG);
3236 memcpy(send_msg.data, port_base_vlan, sizeof(*port_base_vlan));
3237 ret = hclgevf_send_mbx_msg(hdev, &send_msg, false, NULL, 0);
3239 if (state == HNAE3_PORT_BASE_VLAN_DISABLE)
3240 nic->port_base_vlan_state = state;
3242 nic->port_base_vlan_state = HNAE3_PORT_BASE_VLAN_ENABLE;
3245 hclgevf_notify_client(hdev, HNAE3_UP_CLIENT);
3249 static const struct hnae3_ae_ops hclgevf_ops = {
3250 .init_ae_dev = hclgevf_init_ae_dev,
3251 .uninit_ae_dev = hclgevf_uninit_ae_dev,
3252 .reset_prepare = hclgevf_reset_prepare_general,
3253 .reset_done = hclgevf_reset_done,
3254 .init_client_instance = hclgevf_init_client_instance,
3255 .uninit_client_instance = hclgevf_uninit_client_instance,
3256 .start = hclgevf_ae_start,
3257 .stop = hclgevf_ae_stop,
3258 .client_start = hclgevf_client_start,
3259 .client_stop = hclgevf_client_stop,
3260 .map_ring_to_vector = hclgevf_map_ring_to_vector,
3261 .unmap_ring_from_vector = hclgevf_unmap_ring_from_vector,
3262 .get_vector = hclgevf_get_vector,
3263 .put_vector = hclgevf_put_vector,
3264 .reset_queue = hclgevf_reset_tqp,
3265 .get_mac_addr = hclgevf_get_mac_addr,
3266 .set_mac_addr = hclgevf_set_mac_addr,
3267 .add_uc_addr = hclgevf_add_uc_addr,
3268 .rm_uc_addr = hclgevf_rm_uc_addr,
3269 .add_mc_addr = hclgevf_add_mc_addr,
3270 .rm_mc_addr = hclgevf_rm_mc_addr,
3271 .get_stats = hclgevf_get_stats,
3272 .update_stats = hclgevf_update_stats,
3273 .get_strings = hclgevf_get_strings,
3274 .get_sset_count = hclgevf_get_sset_count,
3275 .get_rss_key_size = hclge_comm_get_rss_key_size,
3276 .get_rss = hclgevf_get_rss,
3277 .set_rss = hclgevf_set_rss,
3278 .get_rss_tuple = hclgevf_get_rss_tuple,
3279 .set_rss_tuple = hclgevf_set_rss_tuple,
3280 .get_tc_size = hclgevf_get_tc_size,
3281 .get_fw_version = hclgevf_get_fw_version,
3282 .set_vlan_filter = hclgevf_set_vlan_filter,
3283 .enable_vlan_filter = hclgevf_enable_vlan_filter,
3284 .enable_hw_strip_rxvtag = hclgevf_en_hw_strip_rxvtag,
3285 .reset_event = hclgevf_reset_event,
3286 .set_default_reset_request = hclgevf_set_def_reset_request,
3287 .set_channels = hclgevf_set_channels,
3288 .get_channels = hclgevf_get_channels,
3289 .get_tqps_and_rss_info = hclgevf_get_tqps_and_rss_info,
3290 .get_regs_len = hclgevf_get_regs_len,
3291 .get_regs = hclgevf_get_regs,
3292 .get_status = hclgevf_get_status,
3293 .get_ksettings_an_result = hclgevf_get_ksettings_an_result,
3294 .get_media_type = hclgevf_get_media_type,
3295 .get_hw_reset_stat = hclgevf_get_hw_reset_stat,
3296 .ae_dev_resetting = hclgevf_ae_dev_resetting,
3297 .ae_dev_reset_cnt = hclgevf_ae_dev_reset_cnt,
3298 .set_gro_en = hclgevf_gro_en,
3299 .set_mtu = hclgevf_set_mtu,
3300 .get_global_queue_id = hclgevf_get_qid_global,
3301 .set_timer_task = hclgevf_set_timer_task,
3302 .get_link_mode = hclgevf_get_link_mode,
3303 .set_promisc_mode = hclgevf_set_promisc_mode,
3304 .request_update_promisc_mode = hclgevf_request_update_promisc_mode,
3305 .get_cmdq_stat = hclgevf_get_cmdq_stat,
3308 static struct hnae3_ae_algo ae_algovf = {
3309 .ops = &hclgevf_ops,
3310 .pdev_id_table = ae_algovf_pci_tbl,
3313 static int __init hclgevf_init(void)
3315 pr_info("%s is initializing\n", HCLGEVF_NAME);
3317 hclgevf_wq = alloc_workqueue("%s", WQ_UNBOUND, 0, HCLGEVF_NAME);
3319 pr_err("%s: failed to create workqueue\n", HCLGEVF_NAME);
3323 hnae3_register_ae_algo(&ae_algovf);
3328 static void __exit hclgevf_exit(void)
3330 hnae3_unregister_ae_algo(&ae_algovf);
3331 destroy_workqueue(hclgevf_wq);
3333 module_init(hclgevf_init);
3334 module_exit(hclgevf_exit);
3336 MODULE_LICENSE("GPL");
3337 MODULE_AUTHOR("Huawei Tech. Co., Ltd.");
3338 MODULE_DESCRIPTION("HCLGEVF Driver");
3339 MODULE_VERSION(HCLGEVF_MOD_VERSION);