selftests: drivers/dma-buf: Fix implicit declaration warns
[platform/kernel/linux-rpi.git] / drivers / net / ethernet / qlogic / qed / qed_iwarp.c
1 // SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause)
2 /* QLogic qed NIC Driver
3  * Copyright (c) 2015-2017  QLogic Corporation
4  * Copyright (c) 2019-2020 Marvell International Ltd.
5  */
6
7 #include <linux/if_ether.h>
8 #include <linux/if_vlan.h>
9 #include <linux/ip.h>
10 #include <linux/ipv6.h>
11 #include <linux/spinlock.h>
12 #include <linux/tcp.h>
13 #include "qed_cxt.h"
14 #include "qed_hw.h"
15 #include "qed_ll2.h"
16 #include "qed_rdma.h"
17 #include "qed_reg_addr.h"
18 #include "qed_sp.h"
19 #include "qed_ooo.h"
20
21 #define QED_IWARP_ORD_DEFAULT           32
22 #define QED_IWARP_IRD_DEFAULT           32
23 #define QED_IWARP_MAX_FW_MSS            4120
24
25 #define QED_EP_SIG 0xecabcdef
26
27 struct mpa_v2_hdr {
28         __be16 ird;
29         __be16 ord;
30 };
31
32 #define MPA_V2_PEER2PEER_MODEL  0x8000
33 #define MPA_V2_SEND_RTR         0x4000  /* on ird */
34 #define MPA_V2_READ_RTR         0x4000  /* on ord */
35 #define MPA_V2_WRITE_RTR        0x8000
36 #define MPA_V2_IRD_ORD_MASK     0x3FFF
37
38 #define MPA_REV2(_mpa_rev) ((_mpa_rev) == MPA_NEGOTIATION_TYPE_ENHANCED)
39
40 #define QED_IWARP_INVALID_TCP_CID       0xffffffff
41
42 #define QED_IWARP_RCV_WND_SIZE_DEF_BB_2P (200 * 1024)
43 #define QED_IWARP_RCV_WND_SIZE_DEF_BB_4P (100 * 1024)
44 #define QED_IWARP_RCV_WND_SIZE_DEF_AH_2P (150 * 1024)
45 #define QED_IWARP_RCV_WND_SIZE_DEF_AH_4P (90 * 1024)
46
47 #define QED_IWARP_RCV_WND_SIZE_MIN      (0xffff)
48 #define TIMESTAMP_HEADER_SIZE           (12)
49 #define QED_IWARP_MAX_FIN_RT_DEFAULT    (2)
50
51 #define QED_IWARP_TS_EN                 BIT(0)
52 #define QED_IWARP_DA_EN                 BIT(1)
53 #define QED_IWARP_PARAM_CRC_NEEDED      (1)
54 #define QED_IWARP_PARAM_P2P             (1)
55
56 #define QED_IWARP_DEF_MAX_RT_TIME       (0)
57 #define QED_IWARP_DEF_CWND_FACTOR       (4)
58 #define QED_IWARP_DEF_KA_MAX_PROBE_CNT  (5)
59 #define QED_IWARP_DEF_KA_TIMEOUT        (1200000)       /* 20 min */
60 #define QED_IWARP_DEF_KA_INTERVAL       (1000)          /* 1 sec */
61
62 static int qed_iwarp_async_event(struct qed_hwfn *p_hwfn, u8 fw_event_code,
63                                  __le16 echo, union event_ring_data *data,
64                                  u8 fw_return_code);
65
66 /* Override devinfo with iWARP specific values */
67 void qed_iwarp_init_devinfo(struct qed_hwfn *p_hwfn)
68 {
69         struct qed_rdma_device *dev = p_hwfn->p_rdma_info->dev;
70
71         dev->max_inline = IWARP_REQ_MAX_INLINE_DATA_SIZE;
72         dev->max_qp = min_t(u32,
73                             IWARP_MAX_QPS,
74                             p_hwfn->p_rdma_info->num_qps) -
75                       QED_IWARP_PREALLOC_CNT;
76
77         dev->max_cq = dev->max_qp;
78
79         dev->max_qp_resp_rd_atomic_resc = QED_IWARP_IRD_DEFAULT;
80         dev->max_qp_req_rd_atomic_resc = QED_IWARP_ORD_DEFAULT;
81 }
82
83 void qed_iwarp_init_hw(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt)
84 {
85         p_hwfn->rdma_prs_search_reg = PRS_REG_SEARCH_TCP;
86         qed_wr(p_hwfn, p_ptt, p_hwfn->rdma_prs_search_reg, 1);
87         p_hwfn->b_rdma_enabled_in_prs = true;
88 }
89
90 /* We have two cid maps, one for tcp which should be used only from passive
91  * syn processing and replacing a pre-allocated ep in the list. The second
92  * for active tcp and for QPs.
93  */
94 static void qed_iwarp_cid_cleaned(struct qed_hwfn *p_hwfn, u32 cid)
95 {
96         cid -= qed_cxt_get_proto_cid_start(p_hwfn, p_hwfn->p_rdma_info->proto);
97
98         spin_lock_bh(&p_hwfn->p_rdma_info->lock);
99
100         if (cid < QED_IWARP_PREALLOC_CNT)
101                 qed_bmap_release_id(p_hwfn, &p_hwfn->p_rdma_info->tcp_cid_map,
102                                     cid);
103         else
104                 qed_bmap_release_id(p_hwfn, &p_hwfn->p_rdma_info->cid_map, cid);
105
106         spin_unlock_bh(&p_hwfn->p_rdma_info->lock);
107 }
108
109 void
110 qed_iwarp_init_fw_ramrod(struct qed_hwfn *p_hwfn,
111                          struct iwarp_init_func_ramrod_data *p_ramrod)
112 {
113         p_ramrod->iwarp.ll2_ooo_q_index =
114             RESC_START(p_hwfn, QED_LL2_RAM_QUEUE) +
115             p_hwfn->p_rdma_info->iwarp.ll2_ooo_handle;
116
117         p_ramrod->tcp.max_fin_rt = QED_IWARP_MAX_FIN_RT_DEFAULT;
118
119         return;
120 }
121
122 static int qed_iwarp_alloc_cid(struct qed_hwfn *p_hwfn, u32 *cid)
123 {
124         int rc;
125
126         spin_lock_bh(&p_hwfn->p_rdma_info->lock);
127         rc = qed_rdma_bmap_alloc_id(p_hwfn, &p_hwfn->p_rdma_info->cid_map, cid);
128         spin_unlock_bh(&p_hwfn->p_rdma_info->lock);
129         if (rc) {
130                 DP_NOTICE(p_hwfn, "Failed in allocating iwarp cid\n");
131                 return rc;
132         }
133         *cid += qed_cxt_get_proto_cid_start(p_hwfn, p_hwfn->p_rdma_info->proto);
134
135         rc = qed_cxt_dynamic_ilt_alloc(p_hwfn, QED_ELEM_CXT, *cid);
136         if (rc)
137                 qed_iwarp_cid_cleaned(p_hwfn, *cid);
138
139         return rc;
140 }
141
142 static void qed_iwarp_set_tcp_cid(struct qed_hwfn *p_hwfn, u32 cid)
143 {
144         cid -= qed_cxt_get_proto_cid_start(p_hwfn, p_hwfn->p_rdma_info->proto);
145
146         spin_lock_bh(&p_hwfn->p_rdma_info->lock);
147         qed_bmap_set_id(p_hwfn, &p_hwfn->p_rdma_info->tcp_cid_map, cid);
148         spin_unlock_bh(&p_hwfn->p_rdma_info->lock);
149 }
150
151 /* This function allocates a cid for passive tcp (called from syn receive)
152  * the reason it's separate from the regular cid allocation is because it
153  * is assured that these cids already have ilt allocated. They are preallocated
154  * to ensure that we won't need to allocate memory during syn processing
155  */
156 static int qed_iwarp_alloc_tcp_cid(struct qed_hwfn *p_hwfn, u32 *cid)
157 {
158         int rc;
159
160         spin_lock_bh(&p_hwfn->p_rdma_info->lock);
161
162         rc = qed_rdma_bmap_alloc_id(p_hwfn,
163                                     &p_hwfn->p_rdma_info->tcp_cid_map, cid);
164
165         spin_unlock_bh(&p_hwfn->p_rdma_info->lock);
166
167         if (rc) {
168                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
169                            "can't allocate iwarp tcp cid max-count=%d\n",
170                            p_hwfn->p_rdma_info->tcp_cid_map.max_count);
171
172                 *cid = QED_IWARP_INVALID_TCP_CID;
173                 return rc;
174         }
175
176         *cid += qed_cxt_get_proto_cid_start(p_hwfn,
177                                             p_hwfn->p_rdma_info->proto);
178         return 0;
179 }
180
181 int qed_iwarp_create_qp(struct qed_hwfn *p_hwfn,
182                         struct qed_rdma_qp *qp,
183                         struct qed_rdma_create_qp_out_params *out_params)
184 {
185         struct iwarp_create_qp_ramrod_data *p_ramrod;
186         struct qed_sp_init_data init_data;
187         struct qed_spq_entry *p_ent;
188         u16 physical_queue;
189         u32 cid;
190         int rc;
191
192         qp->shared_queue = dma_alloc_coherent(&p_hwfn->cdev->pdev->dev,
193                                               IWARP_SHARED_QUEUE_PAGE_SIZE,
194                                               &qp->shared_queue_phys_addr,
195                                               GFP_KERNEL);
196         if (!qp->shared_queue)
197                 return -ENOMEM;
198
199         out_params->sq_pbl_virt = (u8 *)qp->shared_queue +
200             IWARP_SHARED_QUEUE_PAGE_SQ_PBL_OFFSET;
201         out_params->sq_pbl_phys = qp->shared_queue_phys_addr +
202             IWARP_SHARED_QUEUE_PAGE_SQ_PBL_OFFSET;
203         out_params->rq_pbl_virt = (u8 *)qp->shared_queue +
204             IWARP_SHARED_QUEUE_PAGE_RQ_PBL_OFFSET;
205         out_params->rq_pbl_phys = qp->shared_queue_phys_addr +
206             IWARP_SHARED_QUEUE_PAGE_RQ_PBL_OFFSET;
207
208         rc = qed_iwarp_alloc_cid(p_hwfn, &cid);
209         if (rc)
210                 goto err1;
211
212         qp->icid = (u16)cid;
213
214         memset(&init_data, 0, sizeof(init_data));
215         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
216         init_data.cid = qp->icid;
217         init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
218
219         rc = qed_sp_init_request(p_hwfn, &p_ent,
220                                  IWARP_RAMROD_CMD_ID_CREATE_QP,
221                                  PROTOCOLID_IWARP, &init_data);
222         if (rc)
223                 goto err2;
224
225         p_ramrod = &p_ent->ramrod.iwarp_create_qp;
226
227         SET_FIELD(p_ramrod->flags,
228                   IWARP_CREATE_QP_RAMROD_DATA_FMR_AND_RESERVED_EN,
229                   qp->fmr_and_reserved_lkey);
230
231         SET_FIELD(p_ramrod->flags,
232                   IWARP_CREATE_QP_RAMROD_DATA_SIGNALED_COMP, qp->signal_all);
233
234         SET_FIELD(p_ramrod->flags,
235                   IWARP_CREATE_QP_RAMROD_DATA_RDMA_RD_EN,
236                   qp->incoming_rdma_read_en);
237
238         SET_FIELD(p_ramrod->flags,
239                   IWARP_CREATE_QP_RAMROD_DATA_RDMA_WR_EN,
240                   qp->incoming_rdma_write_en);
241
242         SET_FIELD(p_ramrod->flags,
243                   IWARP_CREATE_QP_RAMROD_DATA_ATOMIC_EN,
244                   qp->incoming_atomic_en);
245
246         SET_FIELD(p_ramrod->flags,
247                   IWARP_CREATE_QP_RAMROD_DATA_SRQ_FLG, qp->use_srq);
248
249         p_ramrod->pd = cpu_to_le16(qp->pd);
250         p_ramrod->sq_num_pages = cpu_to_le16(qp->sq_num_pages);
251         p_ramrod->rq_num_pages = cpu_to_le16(qp->rq_num_pages);
252
253         p_ramrod->srq_id.srq_idx = cpu_to_le16(qp->srq_id);
254         p_ramrod->srq_id.opaque_fid = cpu_to_le16(p_hwfn->hw_info.opaque_fid);
255         p_ramrod->qp_handle_for_cqe.hi = qp->qp_handle.hi;
256         p_ramrod->qp_handle_for_cqe.lo = qp->qp_handle.lo;
257
258         p_ramrod->cq_cid_for_sq =
259             cpu_to_le32((p_hwfn->hw_info.opaque_fid << 16) | qp->sq_cq_id);
260         p_ramrod->cq_cid_for_rq =
261             cpu_to_le32((p_hwfn->hw_info.opaque_fid << 16) | qp->rq_cq_id);
262
263         p_ramrod->dpi = cpu_to_le16(qp->dpi);
264
265         physical_queue = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_OFLD);
266         p_ramrod->physical_q0 = cpu_to_le16(physical_queue);
267         physical_queue = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_ACK);
268         p_ramrod->physical_q1 = cpu_to_le16(physical_queue);
269
270         rc = qed_spq_post(p_hwfn, p_ent, NULL);
271         if (rc)
272                 goto err2;
273
274         return rc;
275
276 err2:
277         qed_iwarp_cid_cleaned(p_hwfn, cid);
278 err1:
279         dma_free_coherent(&p_hwfn->cdev->pdev->dev,
280                           IWARP_SHARED_QUEUE_PAGE_SIZE,
281                           qp->shared_queue, qp->shared_queue_phys_addr);
282
283         return rc;
284 }
285
286 static int qed_iwarp_modify_fw(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp)
287 {
288         struct iwarp_modify_qp_ramrod_data *p_ramrod;
289         struct qed_sp_init_data init_data;
290         struct qed_spq_entry *p_ent;
291         u16 flags, trans_to_state;
292         int rc;
293
294         /* Get SPQ entry */
295         memset(&init_data, 0, sizeof(init_data));
296         init_data.cid = qp->icid;
297         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
298         init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
299
300         rc = qed_sp_init_request(p_hwfn, &p_ent,
301                                  IWARP_RAMROD_CMD_ID_MODIFY_QP,
302                                  p_hwfn->p_rdma_info->proto, &init_data);
303         if (rc)
304                 return rc;
305
306         p_ramrod = &p_ent->ramrod.iwarp_modify_qp;
307
308         flags = le16_to_cpu(p_ramrod->flags);
309         SET_FIELD(flags, IWARP_MODIFY_QP_RAMROD_DATA_STATE_TRANS_EN, 0x1);
310         p_ramrod->flags = cpu_to_le16(flags);
311
312         if (qp->iwarp_state == QED_IWARP_QP_STATE_CLOSING)
313                 trans_to_state = IWARP_MODIFY_QP_STATE_CLOSING;
314         else
315                 trans_to_state = IWARP_MODIFY_QP_STATE_ERROR;
316
317         p_ramrod->transition_to_state = cpu_to_le16(trans_to_state);
318
319         rc = qed_spq_post(p_hwfn, p_ent, NULL);
320
321         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x)rc=%d\n", qp->icid, rc);
322
323         return rc;
324 }
325
326 enum qed_iwarp_qp_state qed_roce2iwarp_state(enum qed_roce_qp_state state)
327 {
328         switch (state) {
329         case QED_ROCE_QP_STATE_RESET:
330         case QED_ROCE_QP_STATE_INIT:
331         case QED_ROCE_QP_STATE_RTR:
332                 return QED_IWARP_QP_STATE_IDLE;
333         case QED_ROCE_QP_STATE_RTS:
334                 return QED_IWARP_QP_STATE_RTS;
335         case QED_ROCE_QP_STATE_SQD:
336                 return QED_IWARP_QP_STATE_CLOSING;
337         case QED_ROCE_QP_STATE_ERR:
338                 return QED_IWARP_QP_STATE_ERROR;
339         case QED_ROCE_QP_STATE_SQE:
340                 return QED_IWARP_QP_STATE_TERMINATE;
341         default:
342                 return QED_IWARP_QP_STATE_ERROR;
343         }
344 }
345
346 static enum qed_roce_qp_state
347 qed_iwarp2roce_state(enum qed_iwarp_qp_state state)
348 {
349         switch (state) {
350         case QED_IWARP_QP_STATE_IDLE:
351                 return QED_ROCE_QP_STATE_INIT;
352         case QED_IWARP_QP_STATE_RTS:
353                 return QED_ROCE_QP_STATE_RTS;
354         case QED_IWARP_QP_STATE_TERMINATE:
355                 return QED_ROCE_QP_STATE_SQE;
356         case QED_IWARP_QP_STATE_CLOSING:
357                 return QED_ROCE_QP_STATE_SQD;
358         case QED_IWARP_QP_STATE_ERROR:
359                 return QED_ROCE_QP_STATE_ERR;
360         default:
361                 return QED_ROCE_QP_STATE_ERR;
362         }
363 }
364
365 static const char * const iwarp_state_names[] = {
366         "IDLE",
367         "RTS",
368         "TERMINATE",
369         "CLOSING",
370         "ERROR",
371 };
372
373 int
374 qed_iwarp_modify_qp(struct qed_hwfn *p_hwfn,
375                     struct qed_rdma_qp *qp,
376                     enum qed_iwarp_qp_state new_state, bool internal)
377 {
378         enum qed_iwarp_qp_state prev_iw_state;
379         bool modify_fw = false;
380         int rc = 0;
381
382         /* modify QP can be called from upper-layer or as a result of async
383          * RST/FIN... therefore need to protect
384          */
385         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.qp_lock);
386         prev_iw_state = qp->iwarp_state;
387
388         if (prev_iw_state == new_state) {
389                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.qp_lock);
390                 return 0;
391         }
392
393         switch (prev_iw_state) {
394         case QED_IWARP_QP_STATE_IDLE:
395                 switch (new_state) {
396                 case QED_IWARP_QP_STATE_RTS:
397                         qp->iwarp_state = QED_IWARP_QP_STATE_RTS;
398                         break;
399                 case QED_IWARP_QP_STATE_ERROR:
400                         qp->iwarp_state = QED_IWARP_QP_STATE_ERROR;
401                         if (!internal)
402                                 modify_fw = true;
403                         break;
404                 default:
405                         break;
406                 }
407                 break;
408         case QED_IWARP_QP_STATE_RTS:
409                 switch (new_state) {
410                 case QED_IWARP_QP_STATE_CLOSING:
411                         if (!internal)
412                                 modify_fw = true;
413
414                         qp->iwarp_state = QED_IWARP_QP_STATE_CLOSING;
415                         break;
416                 case QED_IWARP_QP_STATE_ERROR:
417                         if (!internal)
418                                 modify_fw = true;
419                         qp->iwarp_state = QED_IWARP_QP_STATE_ERROR;
420                         break;
421                 default:
422                         break;
423                 }
424                 break;
425         case QED_IWARP_QP_STATE_ERROR:
426                 switch (new_state) {
427                 case QED_IWARP_QP_STATE_IDLE:
428
429                         qp->iwarp_state = new_state;
430                         break;
431                 case QED_IWARP_QP_STATE_CLOSING:
432                         /* could happen due to race... do nothing.... */
433                         break;
434                 default:
435                         rc = -EINVAL;
436                 }
437                 break;
438         case QED_IWARP_QP_STATE_TERMINATE:
439         case QED_IWARP_QP_STATE_CLOSING:
440                 qp->iwarp_state = new_state;
441                 break;
442         default:
443                 break;
444         }
445
446         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) %s --> %s%s\n",
447                    qp->icid,
448                    iwarp_state_names[prev_iw_state],
449                    iwarp_state_names[qp->iwarp_state],
450                    internal ? "internal" : "");
451
452         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.qp_lock);
453
454         if (modify_fw)
455                 rc = qed_iwarp_modify_fw(p_hwfn, qp);
456
457         return rc;
458 }
459
460 int qed_iwarp_fw_destroy(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp)
461 {
462         struct qed_sp_init_data init_data;
463         struct qed_spq_entry *p_ent;
464         int rc;
465
466         /* Get SPQ entry */
467         memset(&init_data, 0, sizeof(init_data));
468         init_data.cid = qp->icid;
469         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
470         init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
471
472         rc = qed_sp_init_request(p_hwfn, &p_ent,
473                                  IWARP_RAMROD_CMD_ID_DESTROY_QP,
474                                  p_hwfn->p_rdma_info->proto, &init_data);
475         if (rc)
476                 return rc;
477
478         rc = qed_spq_post(p_hwfn, p_ent, NULL);
479
480         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) rc = %d\n", qp->icid, rc);
481
482         return rc;
483 }
484
485 static void qed_iwarp_destroy_ep(struct qed_hwfn *p_hwfn,
486                                  struct qed_iwarp_ep *ep,
487                                  bool remove_from_active_list)
488 {
489         dma_free_coherent(&p_hwfn->cdev->pdev->dev,
490                           sizeof(*ep->ep_buffer_virt),
491                           ep->ep_buffer_virt, ep->ep_buffer_phys);
492
493         if (remove_from_active_list) {
494                 spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
495                 list_del(&ep->list_entry);
496                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
497         }
498
499         if (ep->qp)
500                 ep->qp->ep = NULL;
501
502         kfree(ep);
503 }
504
505 int qed_iwarp_destroy_qp(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp)
506 {
507         struct qed_iwarp_ep *ep = qp->ep;
508         int wait_count = 0;
509         int rc = 0;
510
511         if (qp->iwarp_state != QED_IWARP_QP_STATE_ERROR) {
512                 rc = qed_iwarp_modify_qp(p_hwfn, qp,
513                                          QED_IWARP_QP_STATE_ERROR, false);
514                 if (rc)
515                         return rc;
516         }
517
518         /* Make sure ep is closed before returning and freeing memory. */
519         if (ep) {
520                 while (READ_ONCE(ep->state) != QED_IWARP_EP_CLOSED &&
521                        wait_count++ < 200)
522                         msleep(100);
523
524                 if (ep->state != QED_IWARP_EP_CLOSED)
525                         DP_NOTICE(p_hwfn, "ep state close timeout state=%x\n",
526                                   ep->state);
527
528                 qed_iwarp_destroy_ep(p_hwfn, ep, false);
529         }
530
531         rc = qed_iwarp_fw_destroy(p_hwfn, qp);
532
533         if (qp->shared_queue)
534                 dma_free_coherent(&p_hwfn->cdev->pdev->dev,
535                                   IWARP_SHARED_QUEUE_PAGE_SIZE,
536                                   qp->shared_queue, qp->shared_queue_phys_addr);
537
538         return rc;
539 }
540
541 static int
542 qed_iwarp_create_ep(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep **ep_out)
543 {
544         struct qed_iwarp_ep *ep;
545         int rc;
546
547         ep = kzalloc(sizeof(*ep), GFP_KERNEL);
548         if (!ep)
549                 return -ENOMEM;
550
551         ep->state = QED_IWARP_EP_INIT;
552
553         ep->ep_buffer_virt = dma_alloc_coherent(&p_hwfn->cdev->pdev->dev,
554                                                 sizeof(*ep->ep_buffer_virt),
555                                                 &ep->ep_buffer_phys,
556                                                 GFP_KERNEL);
557         if (!ep->ep_buffer_virt) {
558                 rc = -ENOMEM;
559                 goto err;
560         }
561
562         ep->sig = QED_EP_SIG;
563
564         *ep_out = ep;
565
566         return 0;
567
568 err:
569         kfree(ep);
570         return rc;
571 }
572
573 static void
574 qed_iwarp_print_tcp_ramrod(struct qed_hwfn *p_hwfn,
575                            struct iwarp_tcp_offload_ramrod_data *p_tcp_ramrod)
576 {
577         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "local_mac=%x %x %x, remote_mac=%x %x %x\n",
578                    p_tcp_ramrod->tcp.local_mac_addr_lo,
579                    p_tcp_ramrod->tcp.local_mac_addr_mid,
580                    p_tcp_ramrod->tcp.local_mac_addr_hi,
581                    p_tcp_ramrod->tcp.remote_mac_addr_lo,
582                    p_tcp_ramrod->tcp.remote_mac_addr_mid,
583                    p_tcp_ramrod->tcp.remote_mac_addr_hi);
584
585         if (p_tcp_ramrod->tcp.ip_version == TCP_IPV4) {
586                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
587                            "local_ip=%pI4h:%x, remote_ip=%pI4h:%x, vlan=%x\n",
588                            p_tcp_ramrod->tcp.local_ip,
589                            p_tcp_ramrod->tcp.local_port,
590                            p_tcp_ramrod->tcp.remote_ip,
591                            p_tcp_ramrod->tcp.remote_port,
592                            p_tcp_ramrod->tcp.vlan_id);
593         } else {
594                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
595                            "local_ip=%pI6:%x, remote_ip=%pI6:%x, vlan=%x\n",
596                            p_tcp_ramrod->tcp.local_ip,
597                            p_tcp_ramrod->tcp.local_port,
598                            p_tcp_ramrod->tcp.remote_ip,
599                            p_tcp_ramrod->tcp.remote_port,
600                            p_tcp_ramrod->tcp.vlan_id);
601         }
602
603         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
604                    "flow_label=%x, ttl=%x, tos_or_tc=%x, mss=%x, rcv_wnd_scale=%x, connect_mode=%x, flags=%x\n",
605                    p_tcp_ramrod->tcp.flow_label,
606                    p_tcp_ramrod->tcp.ttl,
607                    p_tcp_ramrod->tcp.tos_or_tc,
608                    p_tcp_ramrod->tcp.mss,
609                    p_tcp_ramrod->tcp.rcv_wnd_scale,
610                    p_tcp_ramrod->tcp.connect_mode,
611                    p_tcp_ramrod->tcp.flags);
612
613         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "syn_ip_payload_length=%x, lo=%x, hi=%x\n",
614                    p_tcp_ramrod->tcp.syn_ip_payload_length,
615                    p_tcp_ramrod->tcp.syn_phy_addr_lo,
616                    p_tcp_ramrod->tcp.syn_phy_addr_hi);
617 }
618
619 static int
620 qed_iwarp_tcp_offload(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
621 {
622         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
623         struct iwarp_tcp_offload_ramrod_data *p_tcp_ramrod;
624         struct tcp_offload_params_opt2 *tcp;
625         struct qed_sp_init_data init_data;
626         struct qed_spq_entry *p_ent;
627         dma_addr_t async_output_phys;
628         dma_addr_t in_pdata_phys;
629         u16 physical_q;
630         u16 flags = 0;
631         u8 tcp_flags;
632         int rc;
633         int i;
634
635         memset(&init_data, 0, sizeof(init_data));
636         init_data.cid = ep->tcp_cid;
637         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
638         if (ep->connect_mode == TCP_CONNECT_PASSIVE)
639                 init_data.comp_mode = QED_SPQ_MODE_CB;
640         else
641                 init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
642
643         rc = qed_sp_init_request(p_hwfn, &p_ent,
644                                  IWARP_RAMROD_CMD_ID_TCP_OFFLOAD,
645                                  PROTOCOLID_IWARP, &init_data);
646         if (rc)
647                 return rc;
648
649         p_tcp_ramrod = &p_ent->ramrod.iwarp_tcp_offload;
650
651         in_pdata_phys = ep->ep_buffer_phys +
652                         offsetof(struct qed_iwarp_ep_memory, in_pdata);
653         DMA_REGPAIR_LE(p_tcp_ramrod->iwarp.incoming_ulp_buffer.addr,
654                        in_pdata_phys);
655
656         p_tcp_ramrod->iwarp.incoming_ulp_buffer.len =
657             cpu_to_le16(sizeof(ep->ep_buffer_virt->in_pdata));
658
659         async_output_phys = ep->ep_buffer_phys +
660                             offsetof(struct qed_iwarp_ep_memory, async_output);
661         DMA_REGPAIR_LE(p_tcp_ramrod->iwarp.async_eqe_output_buf,
662                        async_output_phys);
663
664         p_tcp_ramrod->iwarp.handle_for_async.hi = cpu_to_le32(PTR_HI(ep));
665         p_tcp_ramrod->iwarp.handle_for_async.lo = cpu_to_le32(PTR_LO(ep));
666
667         physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_OFLD);
668         p_tcp_ramrod->iwarp.physical_q0 = cpu_to_le16(physical_q);
669         physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_ACK);
670         p_tcp_ramrod->iwarp.physical_q1 = cpu_to_le16(physical_q);
671         p_tcp_ramrod->iwarp.mpa_mode = iwarp_info->mpa_rev;
672
673         tcp = &p_tcp_ramrod->tcp;
674         qed_set_fw_mac_addr(&tcp->remote_mac_addr_hi,
675                             &tcp->remote_mac_addr_mid,
676                             &tcp->remote_mac_addr_lo, ep->remote_mac_addr);
677         qed_set_fw_mac_addr(&tcp->local_mac_addr_hi, &tcp->local_mac_addr_mid,
678                             &tcp->local_mac_addr_lo, ep->local_mac_addr);
679
680         tcp->vlan_id = cpu_to_le16(ep->cm_info.vlan);
681
682         tcp_flags = p_hwfn->p_rdma_info->iwarp.tcp_flags;
683
684         SET_FIELD(flags, TCP_OFFLOAD_PARAMS_OPT2_TS_EN,
685                   !!(tcp_flags & QED_IWARP_TS_EN));
686
687         SET_FIELD(flags, TCP_OFFLOAD_PARAMS_OPT2_DA_EN,
688                   !!(tcp_flags & QED_IWARP_DA_EN));
689
690         tcp->flags = cpu_to_le16(flags);
691         tcp->ip_version = ep->cm_info.ip_version;
692
693         for (i = 0; i < 4; i++) {
694                 tcp->remote_ip[i] = cpu_to_le32(ep->cm_info.remote_ip[i]);
695                 tcp->local_ip[i] = cpu_to_le32(ep->cm_info.local_ip[i]);
696         }
697
698         tcp->remote_port = cpu_to_le16(ep->cm_info.remote_port);
699         tcp->local_port = cpu_to_le16(ep->cm_info.local_port);
700         tcp->mss = cpu_to_le16(ep->mss);
701         tcp->flow_label = 0;
702         tcp->ttl = 0x40;
703         tcp->tos_or_tc = 0;
704
705         tcp->max_rt_time = QED_IWARP_DEF_MAX_RT_TIME;
706         tcp->cwnd = cpu_to_le32(QED_IWARP_DEF_CWND_FACTOR * ep->mss);
707         tcp->ka_max_probe_cnt = QED_IWARP_DEF_KA_MAX_PROBE_CNT;
708         tcp->ka_timeout = cpu_to_le32(QED_IWARP_DEF_KA_TIMEOUT);
709         tcp->ka_interval = cpu_to_le32(QED_IWARP_DEF_KA_INTERVAL);
710
711         tcp->rcv_wnd_scale = (u8)p_hwfn->p_rdma_info->iwarp.rcv_wnd_scale;
712         tcp->connect_mode = ep->connect_mode;
713
714         if (ep->connect_mode == TCP_CONNECT_PASSIVE) {
715                 tcp->syn_ip_payload_length =
716                         cpu_to_le16(ep->syn_ip_payload_length);
717                 tcp->syn_phy_addr_hi = DMA_HI_LE(ep->syn_phy_addr);
718                 tcp->syn_phy_addr_lo = DMA_LO_LE(ep->syn_phy_addr);
719         }
720
721         qed_iwarp_print_tcp_ramrod(p_hwfn, p_tcp_ramrod);
722
723         rc = qed_spq_post(p_hwfn, p_ent, NULL);
724
725         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
726                    "EP(0x%x) Offload completed rc=%d\n", ep->tcp_cid, rc);
727
728         return rc;
729 }
730
731 static void
732 qed_iwarp_mpa_received(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
733 {
734         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
735         struct qed_iwarp_cm_event_params params;
736         struct mpa_v2_hdr *mpa_v2;
737         union async_output *async_data;
738         u16 mpa_ord, mpa_ird;
739         u8 mpa_hdr_size = 0;
740         u16 ulp_data_len;
741         u8 mpa_rev;
742
743         async_data = &ep->ep_buffer_virt->async_output;
744
745         mpa_rev = async_data->mpa_request.mpa_handshake_mode;
746         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
747                    "private_data_len=%x handshake_mode=%x private_data=(%x)\n",
748                    async_data->mpa_request.ulp_data_len,
749                    mpa_rev, *((u32 *)(ep->ep_buffer_virt->in_pdata)));
750
751         if (mpa_rev == MPA_NEGOTIATION_TYPE_ENHANCED) {
752                 /* Read ord/ird values from private data buffer */
753                 mpa_v2 = (struct mpa_v2_hdr *)ep->ep_buffer_virt->in_pdata;
754                 mpa_hdr_size = sizeof(*mpa_v2);
755
756                 mpa_ord = ntohs(mpa_v2->ord);
757                 mpa_ird = ntohs(mpa_v2->ird);
758
759                 /* Temprary store in cm_info incoming ord/ird requested, later
760                  * replace with negotiated value during accept
761                  */
762                 ep->cm_info.ord = (u8)min_t(u16,
763                                             (mpa_ord & MPA_V2_IRD_ORD_MASK),
764                                             QED_IWARP_ORD_DEFAULT);
765
766                 ep->cm_info.ird = (u8)min_t(u16,
767                                             (mpa_ird & MPA_V2_IRD_ORD_MASK),
768                                             QED_IWARP_IRD_DEFAULT);
769
770                 /* Peer2Peer negotiation */
771                 ep->rtr_type = MPA_RTR_TYPE_NONE;
772                 if (mpa_ird & MPA_V2_PEER2PEER_MODEL) {
773                         if (mpa_ord & MPA_V2_WRITE_RTR)
774                                 ep->rtr_type |= MPA_RTR_TYPE_ZERO_WRITE;
775
776                         if (mpa_ord & MPA_V2_READ_RTR)
777                                 ep->rtr_type |= MPA_RTR_TYPE_ZERO_READ;
778
779                         if (mpa_ird & MPA_V2_SEND_RTR)
780                                 ep->rtr_type |= MPA_RTR_TYPE_ZERO_SEND;
781
782                         ep->rtr_type &= iwarp_info->rtr_type;
783
784                         /* if we're left with no match send our capabilities */
785                         if (ep->rtr_type == MPA_RTR_TYPE_NONE)
786                                 ep->rtr_type = iwarp_info->rtr_type;
787                 }
788
789                 ep->mpa_rev = MPA_NEGOTIATION_TYPE_ENHANCED;
790         } else {
791                 ep->cm_info.ord = QED_IWARP_ORD_DEFAULT;
792                 ep->cm_info.ird = QED_IWARP_IRD_DEFAULT;
793                 ep->mpa_rev = MPA_NEGOTIATION_TYPE_BASIC;
794         }
795
796         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
797                    "MPA_NEGOTIATE (v%d): ORD: 0x%x IRD: 0x%x rtr:0x%x ulp_data_len = %x mpa_hdr_size = %x\n",
798                    mpa_rev, ep->cm_info.ord, ep->cm_info.ird, ep->rtr_type,
799                    async_data->mpa_request.ulp_data_len, mpa_hdr_size);
800
801         /* Strip mpa v2 hdr from private data before sending to upper layer */
802         ep->cm_info.private_data = ep->ep_buffer_virt->in_pdata + mpa_hdr_size;
803
804         ulp_data_len = le16_to_cpu(async_data->mpa_request.ulp_data_len);
805         ep->cm_info.private_data_len = ulp_data_len - mpa_hdr_size;
806
807         params.event = QED_IWARP_EVENT_MPA_REQUEST;
808         params.cm_info = &ep->cm_info;
809         params.ep_context = ep;
810         params.status = 0;
811
812         ep->state = QED_IWARP_EP_MPA_REQ_RCVD;
813         ep->event_cb(ep->cb_context, &params);
814 }
815
816 static int
817 qed_iwarp_mpa_offload(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
818 {
819         struct iwarp_mpa_offload_ramrod_data *p_mpa_ramrod;
820         struct mpa_outgoing_params *common;
821         struct qed_iwarp_info *iwarp_info;
822         struct qed_sp_init_data init_data;
823         dma_addr_t async_output_phys;
824         struct qed_spq_entry *p_ent;
825         dma_addr_t out_pdata_phys;
826         dma_addr_t in_pdata_phys;
827         struct qed_rdma_qp *qp;
828         bool reject;
829         u32 val;
830         int rc;
831
832         if (!ep)
833                 return -EINVAL;
834
835         qp = ep->qp;
836         reject = !qp;
837
838         memset(&init_data, 0, sizeof(init_data));
839         init_data.cid = reject ? ep->tcp_cid : qp->icid;
840         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
841
842         if (ep->connect_mode == TCP_CONNECT_ACTIVE)
843                 init_data.comp_mode = QED_SPQ_MODE_CB;
844         else
845                 init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
846
847         rc = qed_sp_init_request(p_hwfn, &p_ent,
848                                  IWARP_RAMROD_CMD_ID_MPA_OFFLOAD,
849                                  PROTOCOLID_IWARP, &init_data);
850         if (rc)
851                 return rc;
852
853         p_mpa_ramrod = &p_ent->ramrod.iwarp_mpa_offload;
854         common = &p_mpa_ramrod->common;
855
856         out_pdata_phys = ep->ep_buffer_phys +
857                          offsetof(struct qed_iwarp_ep_memory, out_pdata);
858         DMA_REGPAIR_LE(common->outgoing_ulp_buffer.addr, out_pdata_phys);
859
860         val = ep->cm_info.private_data_len;
861         common->outgoing_ulp_buffer.len = cpu_to_le16(val);
862         common->crc_needed = p_hwfn->p_rdma_info->iwarp.crc_needed;
863
864         common->out_rq.ord = cpu_to_le32(ep->cm_info.ord);
865         common->out_rq.ird = cpu_to_le32(ep->cm_info.ird);
866
867         val = p_hwfn->hw_info.opaque_fid << 16 | ep->tcp_cid;
868         p_mpa_ramrod->tcp_cid = cpu_to_le32(val);
869
870         in_pdata_phys = ep->ep_buffer_phys +
871                         offsetof(struct qed_iwarp_ep_memory, in_pdata);
872         p_mpa_ramrod->tcp_connect_side = ep->connect_mode;
873         DMA_REGPAIR_LE(p_mpa_ramrod->incoming_ulp_buffer.addr,
874                        in_pdata_phys);
875         p_mpa_ramrod->incoming_ulp_buffer.len =
876             cpu_to_le16(sizeof(ep->ep_buffer_virt->in_pdata));
877         async_output_phys = ep->ep_buffer_phys +
878                             offsetof(struct qed_iwarp_ep_memory, async_output);
879         DMA_REGPAIR_LE(p_mpa_ramrod->async_eqe_output_buf,
880                        async_output_phys);
881         p_mpa_ramrod->handle_for_async.hi = cpu_to_le32(PTR_HI(ep));
882         p_mpa_ramrod->handle_for_async.lo = cpu_to_le32(PTR_LO(ep));
883
884         if (!reject) {
885                 DMA_REGPAIR_LE(p_mpa_ramrod->shared_queue_addr,
886                                qp->shared_queue_phys_addr);
887                 p_mpa_ramrod->stats_counter_id =
888                     RESC_START(p_hwfn, QED_RDMA_STATS_QUEUE) + qp->stats_queue;
889         } else {
890                 common->reject = 1;
891         }
892
893         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
894         p_mpa_ramrod->rcv_wnd = cpu_to_le16(iwarp_info->rcv_wnd_size);
895         p_mpa_ramrod->mode = ep->mpa_rev;
896         SET_FIELD(p_mpa_ramrod->rtr_pref,
897                   IWARP_MPA_OFFLOAD_RAMROD_DATA_RTR_SUPPORTED, ep->rtr_type);
898
899         ep->state = QED_IWARP_EP_MPA_OFFLOADED;
900         rc = qed_spq_post(p_hwfn, p_ent, NULL);
901         if (!reject)
902                 ep->cid = qp->icid;     /* Now they're migrated. */
903
904         DP_VERBOSE(p_hwfn,
905                    QED_MSG_RDMA,
906                    "QP(0x%x) EP(0x%x) MPA Offload rc = %d IRD=0x%x ORD=0x%x rtr_type=%d mpa_rev=%d reject=%d\n",
907                    reject ? 0xffff : qp->icid,
908                    ep->tcp_cid,
909                    rc,
910                    ep->cm_info.ird,
911                    ep->cm_info.ord, ep->rtr_type, ep->mpa_rev, reject);
912         return rc;
913 }
914
915 static void
916 qed_iwarp_return_ep(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
917 {
918         ep->state = QED_IWARP_EP_INIT;
919         if (ep->qp)
920                 ep->qp->ep = NULL;
921         ep->qp = NULL;
922         memset(&ep->cm_info, 0, sizeof(ep->cm_info));
923
924         if (ep->tcp_cid == QED_IWARP_INVALID_TCP_CID) {
925                 /* We don't care about the return code, it's ok if tcp_cid
926                  * remains invalid...in this case we'll defer allocation
927                  */
928                 qed_iwarp_alloc_tcp_cid(p_hwfn, &ep->tcp_cid);
929         }
930         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
931
932         list_move_tail(&ep->list_entry,
933                        &p_hwfn->p_rdma_info->iwarp.ep_free_list);
934
935         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
936 }
937
938 static void
939 qed_iwarp_parse_private_data(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
940 {
941         struct mpa_v2_hdr *mpa_v2_params;
942         union async_output *async_data;
943         u16 mpa_ird, mpa_ord;
944         u8 mpa_data_size = 0;
945         u16 ulp_data_len;
946
947         if (MPA_REV2(p_hwfn->p_rdma_info->iwarp.mpa_rev)) {
948                 mpa_v2_params =
949                         (struct mpa_v2_hdr *)(ep->ep_buffer_virt->in_pdata);
950                 mpa_data_size = sizeof(*mpa_v2_params);
951                 mpa_ird = ntohs(mpa_v2_params->ird);
952                 mpa_ord = ntohs(mpa_v2_params->ord);
953
954                 ep->cm_info.ird = (u8)(mpa_ord & MPA_V2_IRD_ORD_MASK);
955                 ep->cm_info.ord = (u8)(mpa_ird & MPA_V2_IRD_ORD_MASK);
956         }
957
958         async_data = &ep->ep_buffer_virt->async_output;
959         ep->cm_info.private_data = ep->ep_buffer_virt->in_pdata + mpa_data_size;
960
961         ulp_data_len = le16_to_cpu(async_data->mpa_response.ulp_data_len);
962         ep->cm_info.private_data_len = ulp_data_len - mpa_data_size;
963 }
964
965 static void
966 qed_iwarp_mpa_reply_arrived(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
967 {
968         struct qed_iwarp_cm_event_params params;
969
970         if (ep->connect_mode == TCP_CONNECT_PASSIVE) {
971                 DP_NOTICE(p_hwfn,
972                           "MPA reply event not expected on passive side!\n");
973                 return;
974         }
975
976         params.event = QED_IWARP_EVENT_ACTIVE_MPA_REPLY;
977
978         qed_iwarp_parse_private_data(p_hwfn, ep);
979
980         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
981                    "MPA_NEGOTIATE (v%d): ORD: 0x%x IRD: 0x%x\n",
982                    ep->mpa_rev, ep->cm_info.ord, ep->cm_info.ird);
983
984         params.cm_info = &ep->cm_info;
985         params.ep_context = ep;
986         params.status = 0;
987
988         ep->mpa_reply_processed = true;
989
990         ep->event_cb(ep->cb_context, &params);
991 }
992
993 #define QED_IWARP_CONNECT_MODE_STRING(ep) \
994         ((ep)->connect_mode == TCP_CONNECT_PASSIVE) ? "Passive" : "Active"
995
996 /* Called as a result of the event:
997  * IWARP_EVENT_TYPE_ASYNC_MPA_HANDSHAKE_COMPLETE
998  */
999 static void
1000 qed_iwarp_mpa_complete(struct qed_hwfn *p_hwfn,
1001                        struct qed_iwarp_ep *ep, u8 fw_return_code)
1002 {
1003         struct qed_iwarp_cm_event_params params;
1004
1005         if (ep->connect_mode == TCP_CONNECT_ACTIVE)
1006                 params.event = QED_IWARP_EVENT_ACTIVE_COMPLETE;
1007         else
1008                 params.event = QED_IWARP_EVENT_PASSIVE_COMPLETE;
1009
1010         if (ep->connect_mode == TCP_CONNECT_ACTIVE && !ep->mpa_reply_processed)
1011                 qed_iwarp_parse_private_data(p_hwfn, ep);
1012
1013         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1014                    "MPA_NEGOTIATE (v%d): ORD: 0x%x IRD: 0x%x\n",
1015                    ep->mpa_rev, ep->cm_info.ord, ep->cm_info.ird);
1016
1017         params.cm_info = &ep->cm_info;
1018
1019         params.ep_context = ep;
1020
1021         switch (fw_return_code) {
1022         case RDMA_RETURN_OK:
1023                 ep->qp->max_rd_atomic_req = ep->cm_info.ord;
1024                 ep->qp->max_rd_atomic_resp = ep->cm_info.ird;
1025                 qed_iwarp_modify_qp(p_hwfn, ep->qp, QED_IWARP_QP_STATE_RTS, 1);
1026                 ep->state = QED_IWARP_EP_ESTABLISHED;
1027                 params.status = 0;
1028                 break;
1029         case IWARP_CONN_ERROR_MPA_TIMEOUT:
1030                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA timeout\n",
1031                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1032                 params.status = -EBUSY;
1033                 break;
1034         case IWARP_CONN_ERROR_MPA_ERROR_REJECT:
1035                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA Reject\n",
1036                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1037                 params.status = -ECONNREFUSED;
1038                 break;
1039         case IWARP_CONN_ERROR_MPA_RST:
1040                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA reset(tcp cid: 0x%x)\n",
1041                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid,
1042                           ep->tcp_cid);
1043                 params.status = -ECONNRESET;
1044                 break;
1045         case IWARP_CONN_ERROR_MPA_FIN:
1046                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA received FIN\n",
1047                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1048                 params.status = -ECONNREFUSED;
1049                 break;
1050         case IWARP_CONN_ERROR_MPA_INSUF_IRD:
1051                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA insufficient ird\n",
1052                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1053                 params.status = -ECONNREFUSED;
1054                 break;
1055         case IWARP_CONN_ERROR_MPA_RTR_MISMATCH:
1056                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA RTR MISMATCH\n",
1057                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1058                 params.status = -ECONNREFUSED;
1059                 break;
1060         case IWARP_CONN_ERROR_MPA_INVALID_PACKET:
1061                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA Invalid Packet\n",
1062                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1063                 params.status = -ECONNREFUSED;
1064                 break;
1065         case IWARP_CONN_ERROR_MPA_LOCAL_ERROR:
1066                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA Local Error\n",
1067                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1068                 params.status = -ECONNREFUSED;
1069                 break;
1070         case IWARP_CONN_ERROR_MPA_TERMINATE:
1071                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA TERMINATE\n",
1072                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1073                 params.status = -ECONNREFUSED;
1074                 break;
1075         default:
1076                 params.status = -ECONNRESET;
1077                 break;
1078         }
1079
1080         if (fw_return_code != RDMA_RETURN_OK)
1081                 /* paired with READ_ONCE in destroy_qp */
1082                 smp_store_release(&ep->state, QED_IWARP_EP_CLOSED);
1083
1084         ep->event_cb(ep->cb_context, &params);
1085
1086         /* on passive side, if there is no associated QP (REJECT) we need to
1087          * return the ep to the pool, (in the regular case we add an element
1088          * in accept instead of this one.
1089          * In both cases we need to remove it from the ep_list.
1090          */
1091         if (fw_return_code != RDMA_RETURN_OK) {
1092                 ep->tcp_cid = QED_IWARP_INVALID_TCP_CID;
1093                 if ((ep->connect_mode == TCP_CONNECT_PASSIVE) &&
1094                     (!ep->qp)) {        /* Rejected */
1095                         qed_iwarp_return_ep(p_hwfn, ep);
1096                 } else {
1097                         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1098                         list_del(&ep->list_entry);
1099                         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1100                 }
1101         }
1102 }
1103
1104 static void
1105 qed_iwarp_mpa_v2_set_private(struct qed_hwfn *p_hwfn,
1106                              struct qed_iwarp_ep *ep, u8 *mpa_data_size)
1107 {
1108         struct mpa_v2_hdr *mpa_v2_params;
1109         u16 mpa_ird, mpa_ord;
1110
1111         *mpa_data_size = 0;
1112         if (MPA_REV2(ep->mpa_rev)) {
1113                 mpa_v2_params =
1114                     (struct mpa_v2_hdr *)ep->ep_buffer_virt->out_pdata;
1115                 *mpa_data_size = sizeof(*mpa_v2_params);
1116
1117                 mpa_ird = (u16)ep->cm_info.ird;
1118                 mpa_ord = (u16)ep->cm_info.ord;
1119
1120                 if (ep->rtr_type != MPA_RTR_TYPE_NONE) {
1121                         mpa_ird |= MPA_V2_PEER2PEER_MODEL;
1122
1123                         if (ep->rtr_type & MPA_RTR_TYPE_ZERO_SEND)
1124                                 mpa_ird |= MPA_V2_SEND_RTR;
1125
1126                         if (ep->rtr_type & MPA_RTR_TYPE_ZERO_WRITE)
1127                                 mpa_ord |= MPA_V2_WRITE_RTR;
1128
1129                         if (ep->rtr_type & MPA_RTR_TYPE_ZERO_READ)
1130                                 mpa_ord |= MPA_V2_READ_RTR;
1131                 }
1132
1133                 mpa_v2_params->ird = htons(mpa_ird);
1134                 mpa_v2_params->ord = htons(mpa_ord);
1135
1136                 DP_VERBOSE(p_hwfn,
1137                            QED_MSG_RDMA,
1138                            "MPA_NEGOTIATE Header: [%x ord:%x ird] %x ord:%x ird:%x peer2peer:%x rtr_send:%x rtr_write:%x rtr_read:%x\n",
1139                            mpa_v2_params->ird,
1140                            mpa_v2_params->ord,
1141                            *((u32 *)mpa_v2_params),
1142                            mpa_ord & MPA_V2_IRD_ORD_MASK,
1143                            mpa_ird & MPA_V2_IRD_ORD_MASK,
1144                            !!(mpa_ird & MPA_V2_PEER2PEER_MODEL),
1145                            !!(mpa_ird & MPA_V2_SEND_RTR),
1146                            !!(mpa_ord & MPA_V2_WRITE_RTR),
1147                            !!(mpa_ord & MPA_V2_READ_RTR));
1148         }
1149 }
1150
1151 int qed_iwarp_connect(void *rdma_cxt,
1152                       struct qed_iwarp_connect_in *iparams,
1153                       struct qed_iwarp_connect_out *oparams)
1154 {
1155         struct qed_hwfn *p_hwfn = rdma_cxt;
1156         struct qed_iwarp_info *iwarp_info;
1157         struct qed_iwarp_ep *ep;
1158         u8 mpa_data_size = 0;
1159         u32 cid;
1160         int rc;
1161
1162         if ((iparams->cm_info.ord > QED_IWARP_ORD_DEFAULT) ||
1163             (iparams->cm_info.ird > QED_IWARP_IRD_DEFAULT)) {
1164                 DP_NOTICE(p_hwfn,
1165                           "QP(0x%x) ERROR: Invalid ord(0x%x)/ird(0x%x)\n",
1166                           iparams->qp->icid, iparams->cm_info.ord,
1167                           iparams->cm_info.ird);
1168
1169                 return -EINVAL;
1170         }
1171
1172         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
1173
1174         /* Allocate ep object */
1175         rc = qed_iwarp_alloc_cid(p_hwfn, &cid);
1176         if (rc)
1177                 return rc;
1178
1179         rc = qed_iwarp_create_ep(p_hwfn, &ep);
1180         if (rc)
1181                 goto err;
1182
1183         ep->tcp_cid = cid;
1184
1185         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1186         list_add_tail(&ep->list_entry, &p_hwfn->p_rdma_info->iwarp.ep_list);
1187         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1188
1189         ep->qp = iparams->qp;
1190         ep->qp->ep = ep;
1191         ether_addr_copy(ep->remote_mac_addr, iparams->remote_mac_addr);
1192         ether_addr_copy(ep->local_mac_addr, iparams->local_mac_addr);
1193         memcpy(&ep->cm_info, &iparams->cm_info, sizeof(ep->cm_info));
1194
1195         ep->cm_info.ord = iparams->cm_info.ord;
1196         ep->cm_info.ird = iparams->cm_info.ird;
1197
1198         ep->rtr_type = iwarp_info->rtr_type;
1199         if (!iwarp_info->peer2peer)
1200                 ep->rtr_type = MPA_RTR_TYPE_NONE;
1201
1202         if ((ep->rtr_type & MPA_RTR_TYPE_ZERO_READ) && (ep->cm_info.ord == 0))
1203                 ep->cm_info.ord = 1;
1204
1205         ep->mpa_rev = iwarp_info->mpa_rev;
1206
1207         qed_iwarp_mpa_v2_set_private(p_hwfn, ep, &mpa_data_size);
1208
1209         ep->cm_info.private_data = ep->ep_buffer_virt->out_pdata;
1210         ep->cm_info.private_data_len = iparams->cm_info.private_data_len +
1211                                        mpa_data_size;
1212
1213         memcpy((u8 *)ep->ep_buffer_virt->out_pdata + mpa_data_size,
1214                iparams->cm_info.private_data,
1215                iparams->cm_info.private_data_len);
1216
1217         ep->mss = iparams->mss;
1218         ep->mss = min_t(u16, QED_IWARP_MAX_FW_MSS, ep->mss);
1219
1220         ep->event_cb = iparams->event_cb;
1221         ep->cb_context = iparams->cb_context;
1222         ep->connect_mode = TCP_CONNECT_ACTIVE;
1223
1224         oparams->ep_context = ep;
1225
1226         rc = qed_iwarp_tcp_offload(p_hwfn, ep);
1227
1228         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) EP(0x%x) rc = %d\n",
1229                    iparams->qp->icid, ep->tcp_cid, rc);
1230
1231         if (rc) {
1232                 qed_iwarp_destroy_ep(p_hwfn, ep, true);
1233                 goto err;
1234         }
1235
1236         return rc;
1237 err:
1238         qed_iwarp_cid_cleaned(p_hwfn, cid);
1239
1240         return rc;
1241 }
1242
1243 static struct qed_iwarp_ep *qed_iwarp_get_free_ep(struct qed_hwfn *p_hwfn)
1244 {
1245         struct qed_iwarp_ep *ep = NULL;
1246         int rc;
1247
1248         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1249
1250         if (list_empty(&p_hwfn->p_rdma_info->iwarp.ep_free_list)) {
1251                 DP_ERR(p_hwfn, "Ep list is empty\n");
1252                 goto out;
1253         }
1254
1255         ep = list_first_entry(&p_hwfn->p_rdma_info->iwarp.ep_free_list,
1256                               struct qed_iwarp_ep, list_entry);
1257
1258         /* in some cases we could have failed allocating a tcp cid when added
1259          * from accept / failure... retry now..this is not the common case.
1260          */
1261         if (ep->tcp_cid == QED_IWARP_INVALID_TCP_CID) {
1262                 rc = qed_iwarp_alloc_tcp_cid(p_hwfn, &ep->tcp_cid);
1263
1264                 /* if we fail we could look for another entry with a valid
1265                  * tcp_cid, but since we don't expect to reach this anyway
1266                  * it's not worth the handling
1267                  */
1268                 if (rc) {
1269                         ep->tcp_cid = QED_IWARP_INVALID_TCP_CID;
1270                         ep = NULL;
1271                         goto out;
1272                 }
1273         }
1274
1275         list_del(&ep->list_entry);
1276
1277 out:
1278         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1279         return ep;
1280 }
1281
1282 #define QED_IWARP_MAX_CID_CLEAN_TIME  100
1283 #define QED_IWARP_MAX_NO_PROGRESS_CNT 5
1284
1285 /* This function waits for all the bits of a bmap to be cleared, as long as
1286  * there is progress ( i.e. the number of bits left to be cleared decreases )
1287  * the function continues.
1288  */
1289 static int
1290 qed_iwarp_wait_cid_map_cleared(struct qed_hwfn *p_hwfn, struct qed_bmap *bmap)
1291 {
1292         int prev_weight = 0;
1293         int wait_count = 0;
1294         int weight = 0;
1295
1296         weight = bitmap_weight(bmap->bitmap, bmap->max_count);
1297         prev_weight = weight;
1298
1299         while (weight) {
1300                 msleep(QED_IWARP_MAX_CID_CLEAN_TIME);
1301
1302                 weight = bitmap_weight(bmap->bitmap, bmap->max_count);
1303
1304                 if (prev_weight == weight) {
1305                         wait_count++;
1306                 } else {
1307                         prev_weight = weight;
1308                         wait_count = 0;
1309                 }
1310
1311                 if (wait_count > QED_IWARP_MAX_NO_PROGRESS_CNT) {
1312                         DP_NOTICE(p_hwfn,
1313                                   "%s bitmap wait timed out (%d cids pending)\n",
1314                                   bmap->name, weight);
1315                         return -EBUSY;
1316                 }
1317         }
1318         return 0;
1319 }
1320
1321 static int qed_iwarp_wait_for_all_cids(struct qed_hwfn *p_hwfn)
1322 {
1323         int rc;
1324         int i;
1325
1326         rc = qed_iwarp_wait_cid_map_cleared(p_hwfn,
1327                                             &p_hwfn->p_rdma_info->tcp_cid_map);
1328         if (rc)
1329                 return rc;
1330
1331         /* Now free the tcp cids from the main cid map */
1332         for (i = 0; i < QED_IWARP_PREALLOC_CNT; i++)
1333                 qed_bmap_release_id(p_hwfn, &p_hwfn->p_rdma_info->cid_map, i);
1334
1335         /* Now wait for all cids to be completed */
1336         return qed_iwarp_wait_cid_map_cleared(p_hwfn,
1337                                               &p_hwfn->p_rdma_info->cid_map);
1338 }
1339
1340 static void qed_iwarp_free_prealloc_ep(struct qed_hwfn *p_hwfn)
1341 {
1342         struct qed_iwarp_ep *ep;
1343
1344         while (!list_empty(&p_hwfn->p_rdma_info->iwarp.ep_free_list)) {
1345                 spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1346
1347                 ep = list_first_entry(&p_hwfn->p_rdma_info->iwarp.ep_free_list,
1348                                       struct qed_iwarp_ep, list_entry);
1349
1350                 if (!ep) {
1351                         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1352                         break;
1353                 }
1354                 list_del(&ep->list_entry);
1355
1356                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1357
1358                 if (ep->tcp_cid != QED_IWARP_INVALID_TCP_CID)
1359                         qed_iwarp_cid_cleaned(p_hwfn, ep->tcp_cid);
1360
1361                 qed_iwarp_destroy_ep(p_hwfn, ep, false);
1362         }
1363 }
1364
1365 static int qed_iwarp_prealloc_ep(struct qed_hwfn *p_hwfn, bool init)
1366 {
1367         struct qed_iwarp_ep *ep;
1368         int rc = 0;
1369         int count;
1370         u32 cid;
1371         int i;
1372
1373         count = init ? QED_IWARP_PREALLOC_CNT : 1;
1374         for (i = 0; i < count; i++) {
1375                 rc = qed_iwarp_create_ep(p_hwfn, &ep);
1376                 if (rc)
1377                         return rc;
1378
1379                 /* During initialization we allocate from the main pool,
1380                  * afterwards we allocate only from the tcp_cid.
1381                  */
1382                 if (init) {
1383                         rc = qed_iwarp_alloc_cid(p_hwfn, &cid);
1384                         if (rc)
1385                                 goto err;
1386                         qed_iwarp_set_tcp_cid(p_hwfn, cid);
1387                 } else {
1388                         /* We don't care about the return code, it's ok if
1389                          * tcp_cid remains invalid...in this case we'll
1390                          * defer allocation
1391                          */
1392                         qed_iwarp_alloc_tcp_cid(p_hwfn, &cid);
1393                 }
1394
1395                 ep->tcp_cid = cid;
1396
1397                 spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1398                 list_add_tail(&ep->list_entry,
1399                               &p_hwfn->p_rdma_info->iwarp.ep_free_list);
1400                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1401         }
1402
1403         return rc;
1404
1405 err:
1406         qed_iwarp_destroy_ep(p_hwfn, ep, false);
1407
1408         return rc;
1409 }
1410
1411 int qed_iwarp_alloc(struct qed_hwfn *p_hwfn)
1412 {
1413         int rc;
1414
1415         /* Allocate bitmap for tcp cid. These are used by passive side
1416          * to ensure it can allocate a tcp cid during dpc that was
1417          * pre-acquired and doesn't require dynamic allocation of ilt
1418          */
1419         rc = qed_rdma_bmap_alloc(p_hwfn, &p_hwfn->p_rdma_info->tcp_cid_map,
1420                                  QED_IWARP_PREALLOC_CNT, "TCP_CID");
1421         if (rc) {
1422                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1423                            "Failed to allocate tcp cid, rc = %d\n", rc);
1424                 return rc;
1425         }
1426
1427         INIT_LIST_HEAD(&p_hwfn->p_rdma_info->iwarp.ep_free_list);
1428         spin_lock_init(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1429
1430         rc = qed_iwarp_prealloc_ep(p_hwfn, true);
1431         if (rc)
1432                 return rc;
1433
1434         return qed_ooo_alloc(p_hwfn);
1435 }
1436
1437 void qed_iwarp_resc_free(struct qed_hwfn *p_hwfn)
1438 {
1439         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
1440
1441         qed_ooo_free(p_hwfn);
1442         qed_rdma_bmap_free(p_hwfn, &p_hwfn->p_rdma_info->tcp_cid_map, 1);
1443         kfree(iwarp_info->mpa_bufs);
1444         kfree(iwarp_info->partial_fpdus);
1445         kfree(iwarp_info->mpa_intermediate_buf);
1446 }
1447
1448 int qed_iwarp_accept(void *rdma_cxt, struct qed_iwarp_accept_in *iparams)
1449 {
1450         struct qed_hwfn *p_hwfn = rdma_cxt;
1451         struct qed_iwarp_ep *ep;
1452         u8 mpa_data_size = 0;
1453         int rc;
1454
1455         ep = iparams->ep_context;
1456         if (!ep) {
1457                 DP_ERR(p_hwfn, "Ep Context receive in accept is NULL\n");
1458                 return -EINVAL;
1459         }
1460
1461         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) EP(0x%x)\n",
1462                    iparams->qp->icid, ep->tcp_cid);
1463
1464         if ((iparams->ord > QED_IWARP_ORD_DEFAULT) ||
1465             (iparams->ird > QED_IWARP_IRD_DEFAULT)) {
1466                 DP_VERBOSE(p_hwfn,
1467                            QED_MSG_RDMA,
1468                            "QP(0x%x) EP(0x%x) ERROR: Invalid ord(0x%x)/ird(0x%x)\n",
1469                            iparams->qp->icid,
1470                            ep->tcp_cid, iparams->ord, iparams->ord);
1471                 return -EINVAL;
1472         }
1473
1474         qed_iwarp_prealloc_ep(p_hwfn, false);
1475
1476         ep->cb_context = iparams->cb_context;
1477         ep->qp = iparams->qp;
1478         ep->qp->ep = ep;
1479
1480         if (ep->mpa_rev == MPA_NEGOTIATION_TYPE_ENHANCED) {
1481                 /* Negotiate ord/ird: if upperlayer requested ord larger than
1482                  * ird advertised by remote, we need to decrease our ord
1483                  */
1484                 if (iparams->ord > ep->cm_info.ird)
1485                         iparams->ord = ep->cm_info.ird;
1486
1487                 if ((ep->rtr_type & MPA_RTR_TYPE_ZERO_READ) &&
1488                     (iparams->ird == 0))
1489                         iparams->ird = 1;
1490         }
1491
1492         /* Update cm_info ord/ird to be negotiated values */
1493         ep->cm_info.ord = iparams->ord;
1494         ep->cm_info.ird = iparams->ird;
1495
1496         qed_iwarp_mpa_v2_set_private(p_hwfn, ep, &mpa_data_size);
1497
1498         ep->cm_info.private_data = ep->ep_buffer_virt->out_pdata;
1499         ep->cm_info.private_data_len = iparams->private_data_len +
1500                                        mpa_data_size;
1501
1502         memcpy((u8 *)ep->ep_buffer_virt->out_pdata + mpa_data_size,
1503                iparams->private_data, iparams->private_data_len);
1504
1505         rc = qed_iwarp_mpa_offload(p_hwfn, ep);
1506         if (rc)
1507                 qed_iwarp_modify_qp(p_hwfn,
1508                                     iparams->qp, QED_IWARP_QP_STATE_ERROR, 1);
1509
1510         return rc;
1511 }
1512
1513 int qed_iwarp_reject(void *rdma_cxt, struct qed_iwarp_reject_in *iparams)
1514 {
1515         struct qed_hwfn *p_hwfn = rdma_cxt;
1516         struct qed_iwarp_ep *ep;
1517         u8 mpa_data_size = 0;
1518
1519         ep = iparams->ep_context;
1520         if (!ep) {
1521                 DP_ERR(p_hwfn, "Ep Context receive in reject is NULL\n");
1522                 return -EINVAL;
1523         }
1524
1525         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "EP(0x%x)\n", ep->tcp_cid);
1526
1527         ep->cb_context = iparams->cb_context;
1528         ep->qp = NULL;
1529
1530         qed_iwarp_mpa_v2_set_private(p_hwfn, ep, &mpa_data_size);
1531
1532         ep->cm_info.private_data = ep->ep_buffer_virt->out_pdata;
1533         ep->cm_info.private_data_len = iparams->private_data_len +
1534                                        mpa_data_size;
1535
1536         memcpy((u8 *)ep->ep_buffer_virt->out_pdata + mpa_data_size,
1537                iparams->private_data, iparams->private_data_len);
1538
1539         return qed_iwarp_mpa_offload(p_hwfn, ep);
1540 }
1541
1542 static void
1543 qed_iwarp_print_cm_info(struct qed_hwfn *p_hwfn,
1544                         struct qed_iwarp_cm_info *cm_info)
1545 {
1546         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "ip_version = %d\n",
1547                    cm_info->ip_version);
1548
1549         if (cm_info->ip_version == QED_TCP_IPV4)
1550                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1551                            "remote_ip %pI4h:%x, local_ip %pI4h:%x vlan=%x\n",
1552                            cm_info->remote_ip, cm_info->remote_port,
1553                            cm_info->local_ip, cm_info->local_port,
1554                            cm_info->vlan);
1555         else
1556                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1557                            "remote_ip %pI6:%x, local_ip %pI6:%x vlan=%x\n",
1558                            cm_info->remote_ip, cm_info->remote_port,
1559                            cm_info->local_ip, cm_info->local_port,
1560                            cm_info->vlan);
1561
1562         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1563                    "private_data_len = %x ord = %d, ird = %d\n",
1564                    cm_info->private_data_len, cm_info->ord, cm_info->ird);
1565 }
1566
1567 static int
1568 qed_iwarp_ll2_post_rx(struct qed_hwfn *p_hwfn,
1569                       struct qed_iwarp_ll2_buff *buf, u8 handle)
1570 {
1571         int rc;
1572
1573         rc = qed_ll2_post_rx_buffer(p_hwfn, handle, buf->data_phys_addr,
1574                                     (u16)buf->buff_size, buf, 1);
1575         if (rc) {
1576                 DP_NOTICE(p_hwfn,
1577                           "Failed to repost rx buffer to ll2 rc = %d, handle=%d\n",
1578                           rc, handle);
1579                 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buf->buff_size,
1580                                   buf->data, buf->data_phys_addr);
1581                 kfree(buf);
1582         }
1583
1584         return rc;
1585 }
1586
1587 static bool
1588 qed_iwarp_ep_exists(struct qed_hwfn *p_hwfn, struct qed_iwarp_cm_info *cm_info)
1589 {
1590         struct qed_iwarp_ep *ep = NULL;
1591         bool found = false;
1592
1593         list_for_each_entry(ep,
1594                             &p_hwfn->p_rdma_info->iwarp.ep_list,
1595                             list_entry) {
1596                 if ((ep->cm_info.local_port == cm_info->local_port) &&
1597                     (ep->cm_info.remote_port == cm_info->remote_port) &&
1598                     (ep->cm_info.vlan == cm_info->vlan) &&
1599                     !memcmp(&ep->cm_info.local_ip, cm_info->local_ip,
1600                             sizeof(cm_info->local_ip)) &&
1601                     !memcmp(&ep->cm_info.remote_ip, cm_info->remote_ip,
1602                             sizeof(cm_info->remote_ip))) {
1603                         found = true;
1604                         break;
1605                 }
1606         }
1607
1608         if (found) {
1609                 DP_NOTICE(p_hwfn,
1610                           "SYN received on active connection - dropping\n");
1611                 qed_iwarp_print_cm_info(p_hwfn, cm_info);
1612
1613                 return true;
1614         }
1615
1616         return false;
1617 }
1618
1619 static struct qed_iwarp_listener *
1620 qed_iwarp_get_listener(struct qed_hwfn *p_hwfn,
1621                        struct qed_iwarp_cm_info *cm_info)
1622 {
1623         struct qed_iwarp_listener *listener = NULL;
1624         static const u32 ip_zero[4] = { 0, 0, 0, 0 };
1625         bool found = false;
1626
1627         list_for_each_entry(listener,
1628                             &p_hwfn->p_rdma_info->iwarp.listen_list,
1629                             list_entry) {
1630                 if (listener->port == cm_info->local_port) {
1631                         if (!memcmp(listener->ip_addr,
1632                                     ip_zero, sizeof(ip_zero))) {
1633                                 found = true;
1634                                 break;
1635                         }
1636
1637                         if (!memcmp(listener->ip_addr,
1638                                     cm_info->local_ip,
1639                                     sizeof(cm_info->local_ip)) &&
1640                             (listener->vlan == cm_info->vlan)) {
1641                                 found = true;
1642                                 break;
1643                         }
1644                 }
1645         }
1646
1647         if (found) {
1648                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "listener found = %p\n",
1649                            listener);
1650                 return listener;
1651         }
1652
1653         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "listener not found\n");
1654         return NULL;
1655 }
1656
1657 static int
1658 qed_iwarp_parse_rx_pkt(struct qed_hwfn *p_hwfn,
1659                        struct qed_iwarp_cm_info *cm_info,
1660                        void *buf,
1661                        u8 *remote_mac_addr,
1662                        u8 *local_mac_addr,
1663                        int *payload_len, int *tcp_start_offset)
1664 {
1665         struct vlan_ethhdr *vethh;
1666         bool vlan_valid = false;
1667         struct ipv6hdr *ip6h;
1668         struct ethhdr *ethh;
1669         struct tcphdr *tcph;
1670         struct iphdr *iph;
1671         int eth_hlen;
1672         int ip_hlen;
1673         int eth_type;
1674         int i;
1675
1676         ethh = buf;
1677         eth_type = ntohs(ethh->h_proto);
1678         if (eth_type == ETH_P_8021Q) {
1679                 vlan_valid = true;
1680                 vethh = (struct vlan_ethhdr *)ethh;
1681                 cm_info->vlan = ntohs(vethh->h_vlan_TCI) & VLAN_VID_MASK;
1682                 eth_type = ntohs(vethh->h_vlan_encapsulated_proto);
1683         }
1684
1685         eth_hlen = ETH_HLEN + (vlan_valid ? sizeof(u32) : 0);
1686
1687         if (!ether_addr_equal(ethh->h_dest,
1688                               p_hwfn->p_rdma_info->iwarp.mac_addr)) {
1689                 DP_VERBOSE(p_hwfn,
1690                            QED_MSG_RDMA,
1691                            "Got unexpected mac %pM instead of %pM\n",
1692                            ethh->h_dest, p_hwfn->p_rdma_info->iwarp.mac_addr);
1693                 return -EINVAL;
1694         }
1695
1696         ether_addr_copy(remote_mac_addr, ethh->h_source);
1697         ether_addr_copy(local_mac_addr, ethh->h_dest);
1698
1699         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "eth_type =%d source mac: %pM\n",
1700                    eth_type, ethh->h_source);
1701
1702         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "eth_hlen=%d destination mac: %pM\n",
1703                    eth_hlen, ethh->h_dest);
1704
1705         iph = (struct iphdr *)((u8 *)(ethh) + eth_hlen);
1706
1707         if (eth_type == ETH_P_IP) {
1708                 if (iph->protocol != IPPROTO_TCP) {
1709                         DP_NOTICE(p_hwfn,
1710                                   "Unexpected ip protocol on ll2 %x\n",
1711                                   iph->protocol);
1712                         return -EINVAL;
1713                 }
1714
1715                 cm_info->local_ip[0] = ntohl(iph->daddr);
1716                 cm_info->remote_ip[0] = ntohl(iph->saddr);
1717                 cm_info->ip_version = QED_TCP_IPV4;
1718
1719                 ip_hlen = (iph->ihl) * sizeof(u32);
1720                 *payload_len = ntohs(iph->tot_len) - ip_hlen;
1721         } else if (eth_type == ETH_P_IPV6) {
1722                 ip6h = (struct ipv6hdr *)iph;
1723
1724                 if (ip6h->nexthdr != IPPROTO_TCP) {
1725                         DP_NOTICE(p_hwfn,
1726                                   "Unexpected ip protocol on ll2 %x\n",
1727                                   iph->protocol);
1728                         return -EINVAL;
1729                 }
1730
1731                 for (i = 0; i < 4; i++) {
1732                         cm_info->local_ip[i] =
1733                             ntohl(ip6h->daddr.in6_u.u6_addr32[i]);
1734                         cm_info->remote_ip[i] =
1735                             ntohl(ip6h->saddr.in6_u.u6_addr32[i]);
1736                 }
1737                 cm_info->ip_version = QED_TCP_IPV6;
1738
1739                 ip_hlen = sizeof(*ip6h);
1740                 *payload_len = ntohs(ip6h->payload_len);
1741         } else {
1742                 DP_NOTICE(p_hwfn, "Unexpected ethertype on ll2 %x\n", eth_type);
1743                 return -EINVAL;
1744         }
1745
1746         tcph = (struct tcphdr *)((u8 *)iph + ip_hlen);
1747
1748         if (!tcph->syn) {
1749                 DP_NOTICE(p_hwfn,
1750                           "Only SYN type packet expected on this ll2 conn, iph->ihl=%d source=%d dest=%d\n",
1751                           iph->ihl, tcph->source, tcph->dest);
1752                 return -EINVAL;
1753         }
1754
1755         cm_info->local_port = ntohs(tcph->dest);
1756         cm_info->remote_port = ntohs(tcph->source);
1757
1758         qed_iwarp_print_cm_info(p_hwfn, cm_info);
1759
1760         *tcp_start_offset = eth_hlen + ip_hlen;
1761
1762         return 0;
1763 }
1764
1765 static struct qed_iwarp_fpdu *qed_iwarp_get_curr_fpdu(struct qed_hwfn *p_hwfn,
1766                                                       u16 cid)
1767 {
1768         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
1769         struct qed_iwarp_fpdu *partial_fpdu;
1770         u32 idx;
1771
1772         idx = cid - qed_cxt_get_proto_cid_start(p_hwfn, PROTOCOLID_IWARP);
1773         if (idx >= iwarp_info->max_num_partial_fpdus) {
1774                 DP_ERR(p_hwfn, "Invalid cid %x max_num_partial_fpdus=%x\n", cid,
1775                        iwarp_info->max_num_partial_fpdus);
1776                 return NULL;
1777         }
1778
1779         partial_fpdu = &iwarp_info->partial_fpdus[idx];
1780
1781         return partial_fpdu;
1782 }
1783
1784 enum qed_iwarp_mpa_pkt_type {
1785         QED_IWARP_MPA_PKT_PACKED,
1786         QED_IWARP_MPA_PKT_PARTIAL,
1787         QED_IWARP_MPA_PKT_UNALIGNED
1788 };
1789
1790 #define QED_IWARP_INVALID_FPDU_LENGTH 0xffff
1791 #define QED_IWARP_MPA_FPDU_LENGTH_SIZE (2)
1792 #define QED_IWARP_MPA_CRC32_DIGEST_SIZE (4)
1793
1794 /* Pad to multiple of 4 */
1795 #define QED_IWARP_PDU_DATA_LEN_WITH_PAD(data_len) ALIGN(data_len, 4)
1796 #define QED_IWARP_FPDU_LEN_WITH_PAD(_mpa_len)                              \
1797         (QED_IWARP_PDU_DATA_LEN_WITH_PAD((_mpa_len) +                      \
1798                                          QED_IWARP_MPA_FPDU_LENGTH_SIZE) + \
1799                                          QED_IWARP_MPA_CRC32_DIGEST_SIZE)
1800
1801 /* fpdu can be fragmented over maximum 3 bds: header, partial mpa, unaligned */
1802 #define QED_IWARP_MAX_BDS_PER_FPDU 3
1803
1804 static const char * const pkt_type_str[] = {
1805         "QED_IWARP_MPA_PKT_PACKED",
1806         "QED_IWARP_MPA_PKT_PARTIAL",
1807         "QED_IWARP_MPA_PKT_UNALIGNED"
1808 };
1809
1810 static int
1811 qed_iwarp_recycle_pkt(struct qed_hwfn *p_hwfn,
1812                       struct qed_iwarp_fpdu *fpdu,
1813                       struct qed_iwarp_ll2_buff *buf);
1814
1815 static enum qed_iwarp_mpa_pkt_type
1816 qed_iwarp_mpa_classify(struct qed_hwfn *p_hwfn,
1817                        struct qed_iwarp_fpdu *fpdu,
1818                        u16 tcp_payload_len, u8 *mpa_data)
1819 {
1820         enum qed_iwarp_mpa_pkt_type pkt_type;
1821         u16 mpa_len;
1822
1823         if (fpdu->incomplete_bytes) {
1824                 pkt_type = QED_IWARP_MPA_PKT_UNALIGNED;
1825                 goto out;
1826         }
1827
1828         /* special case of one byte remaining...
1829          * lower byte will be read next packet
1830          */
1831         if (tcp_payload_len == 1) {
1832                 fpdu->fpdu_length = *mpa_data << BITS_PER_BYTE;
1833                 pkt_type = QED_IWARP_MPA_PKT_PARTIAL;
1834                 goto out;
1835         }
1836
1837         mpa_len = ntohs(*(__force __be16 *)mpa_data);
1838         fpdu->fpdu_length = QED_IWARP_FPDU_LEN_WITH_PAD(mpa_len);
1839
1840         if (fpdu->fpdu_length <= tcp_payload_len)
1841                 pkt_type = QED_IWARP_MPA_PKT_PACKED;
1842         else
1843                 pkt_type = QED_IWARP_MPA_PKT_PARTIAL;
1844
1845 out:
1846         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1847                    "MPA_ALIGN: %s: fpdu_length=0x%x tcp_payload_len:0x%x\n",
1848                    pkt_type_str[pkt_type], fpdu->fpdu_length, tcp_payload_len);
1849
1850         return pkt_type;
1851 }
1852
1853 static void
1854 qed_iwarp_init_fpdu(struct qed_iwarp_ll2_buff *buf,
1855                     struct qed_iwarp_fpdu *fpdu,
1856                     struct unaligned_opaque_data *pkt_data,
1857                     u16 tcp_payload_size, u8 placement_offset)
1858 {
1859         u16 first_mpa_offset = le16_to_cpu(pkt_data->first_mpa_offset);
1860
1861         fpdu->mpa_buf = buf;
1862         fpdu->pkt_hdr = buf->data_phys_addr + placement_offset;
1863         fpdu->pkt_hdr_size = pkt_data->tcp_payload_offset;
1864         fpdu->mpa_frag = buf->data_phys_addr + first_mpa_offset;
1865         fpdu->mpa_frag_virt = (u8 *)(buf->data) + first_mpa_offset;
1866
1867         if (tcp_payload_size == 1)
1868                 fpdu->incomplete_bytes = QED_IWARP_INVALID_FPDU_LENGTH;
1869         else if (tcp_payload_size < fpdu->fpdu_length)
1870                 fpdu->incomplete_bytes = fpdu->fpdu_length - tcp_payload_size;
1871         else
1872                 fpdu->incomplete_bytes = 0;     /* complete fpdu */
1873
1874         fpdu->mpa_frag_len = fpdu->fpdu_length - fpdu->incomplete_bytes;
1875 }
1876
1877 static int
1878 qed_iwarp_cp_pkt(struct qed_hwfn *p_hwfn,
1879                  struct qed_iwarp_fpdu *fpdu,
1880                  struct unaligned_opaque_data *pkt_data,
1881                  struct qed_iwarp_ll2_buff *buf, u16 tcp_payload_size)
1882 {
1883         u16 first_mpa_offset = le16_to_cpu(pkt_data->first_mpa_offset);
1884         u8 *tmp_buf = p_hwfn->p_rdma_info->iwarp.mpa_intermediate_buf;
1885         int rc;
1886
1887         /* need to copy the data from the partial packet stored in fpdu
1888          * to the new buf, for this we also need to move the data currently
1889          * placed on the buf. The assumption is that the buffer is big enough
1890          * since fpdu_length <= mss, we use an intermediate buffer since
1891          * we may need to copy the new data to an overlapping location
1892          */
1893         if ((fpdu->mpa_frag_len + tcp_payload_size) > (u16)buf->buff_size) {
1894                 DP_ERR(p_hwfn,
1895                        "MPA ALIGN: Unexpected: buffer is not large enough for split fpdu buff_size = %d mpa_frag_len = %d, tcp_payload_size = %d, incomplete_bytes = %d\n",
1896                        buf->buff_size, fpdu->mpa_frag_len,
1897                        tcp_payload_size, fpdu->incomplete_bytes);
1898                 return -EINVAL;
1899         }
1900
1901         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1902                    "MPA ALIGN Copying fpdu: [%p, %d] [%p, %d]\n",
1903                    fpdu->mpa_frag_virt, fpdu->mpa_frag_len,
1904                    (u8 *)(buf->data) + first_mpa_offset, tcp_payload_size);
1905
1906         memcpy(tmp_buf, fpdu->mpa_frag_virt, fpdu->mpa_frag_len);
1907         memcpy(tmp_buf + fpdu->mpa_frag_len,
1908                (u8 *)(buf->data) + first_mpa_offset, tcp_payload_size);
1909
1910         rc = qed_iwarp_recycle_pkt(p_hwfn, fpdu, fpdu->mpa_buf);
1911         if (rc)
1912                 return rc;
1913
1914         /* If we managed to post the buffer copy the data to the new buffer
1915          * o/w this will occur in the next round...
1916          */
1917         memcpy((u8 *)(buf->data), tmp_buf,
1918                fpdu->mpa_frag_len + tcp_payload_size);
1919
1920         fpdu->mpa_buf = buf;
1921         /* fpdu->pkt_hdr remains as is */
1922         /* fpdu->mpa_frag is overridden with new buf */
1923         fpdu->mpa_frag = buf->data_phys_addr;
1924         fpdu->mpa_frag_virt = buf->data;
1925         fpdu->mpa_frag_len += tcp_payload_size;
1926
1927         fpdu->incomplete_bytes -= tcp_payload_size;
1928
1929         DP_VERBOSE(p_hwfn,
1930                    QED_MSG_RDMA,
1931                    "MPA ALIGN: split fpdu buff_size = %d mpa_frag_len = %d, tcp_payload_size = %d, incomplete_bytes = %d\n",
1932                    buf->buff_size, fpdu->mpa_frag_len, tcp_payload_size,
1933                    fpdu->incomplete_bytes);
1934
1935         return 0;
1936 }
1937
1938 static void
1939 qed_iwarp_update_fpdu_length(struct qed_hwfn *p_hwfn,
1940                              struct qed_iwarp_fpdu *fpdu, u8 *mpa_data)
1941 {
1942         u16 mpa_len;
1943
1944         /* Update incomplete packets if needed */
1945         if (fpdu->incomplete_bytes == QED_IWARP_INVALID_FPDU_LENGTH) {
1946                 /* Missing lower byte is now available */
1947                 mpa_len = fpdu->fpdu_length | *mpa_data;
1948                 fpdu->fpdu_length = QED_IWARP_FPDU_LEN_WITH_PAD(mpa_len);
1949                 /* one byte of hdr */
1950                 fpdu->mpa_frag_len = 1;
1951                 fpdu->incomplete_bytes = fpdu->fpdu_length - 1;
1952                 DP_VERBOSE(p_hwfn,
1953                            QED_MSG_RDMA,
1954                            "MPA_ALIGN: Partial header mpa_len=%x fpdu_length=%x incomplete_bytes=%x\n",
1955                            mpa_len, fpdu->fpdu_length, fpdu->incomplete_bytes);
1956         }
1957 }
1958
1959 #define QED_IWARP_IS_RIGHT_EDGE(_curr_pkt) \
1960         (GET_FIELD((_curr_pkt)->flags,     \
1961                    UNALIGNED_OPAQUE_DATA_PKT_REACHED_WIN_RIGHT_EDGE))
1962
1963 /* This function is used to recycle a buffer using the ll2 drop option. It
1964  * uses the mechanism to ensure that all buffers posted to tx before this one
1965  * were completed. The buffer sent here will be sent as a cookie in the tx
1966  * completion function and can then be reposted to rx chain when done. The flow
1967  * that requires this is the flow where a FPDU splits over more than 3 tcp
1968  * segments. In this case the driver needs to re-post a rx buffer instead of
1969  * the one received, but driver can't simply repost a buffer it copied from
1970  * as there is a case where the buffer was originally a packed FPDU, and is
1971  * partially posted to FW. Driver needs to ensure FW is done with it.
1972  */
1973 static int
1974 qed_iwarp_recycle_pkt(struct qed_hwfn *p_hwfn,
1975                       struct qed_iwarp_fpdu *fpdu,
1976                       struct qed_iwarp_ll2_buff *buf)
1977 {
1978         struct qed_ll2_tx_pkt_info tx_pkt;
1979         u8 ll2_handle;
1980         int rc;
1981
1982         memset(&tx_pkt, 0, sizeof(tx_pkt));
1983         tx_pkt.num_of_bds = 1;
1984         tx_pkt.tx_dest = QED_LL2_TX_DEST_DROP;
1985         tx_pkt.l4_hdr_offset_w = fpdu->pkt_hdr_size >> 2;
1986         tx_pkt.first_frag = fpdu->pkt_hdr;
1987         tx_pkt.first_frag_len = fpdu->pkt_hdr_size;
1988         buf->piggy_buf = NULL;
1989         tx_pkt.cookie = buf;
1990
1991         ll2_handle = p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle;
1992
1993         rc = qed_ll2_prepare_tx_packet(p_hwfn, ll2_handle, &tx_pkt, true);
1994         if (rc)
1995                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1996                            "Can't drop packet rc=%d\n", rc);
1997
1998         DP_VERBOSE(p_hwfn,
1999                    QED_MSG_RDMA,
2000                    "MPA_ALIGN: send drop tx packet [%lx, 0x%x], buf=%p, rc=%d\n",
2001                    (unsigned long int)tx_pkt.first_frag,
2002                    tx_pkt.first_frag_len, buf, rc);
2003
2004         return rc;
2005 }
2006
2007 static int
2008 qed_iwarp_win_right_edge(struct qed_hwfn *p_hwfn, struct qed_iwarp_fpdu *fpdu)
2009 {
2010         struct qed_ll2_tx_pkt_info tx_pkt;
2011         u8 ll2_handle;
2012         int rc;
2013
2014         memset(&tx_pkt, 0, sizeof(tx_pkt));
2015         tx_pkt.num_of_bds = 1;
2016         tx_pkt.tx_dest = QED_LL2_TX_DEST_LB;
2017         tx_pkt.l4_hdr_offset_w = fpdu->pkt_hdr_size >> 2;
2018
2019         tx_pkt.first_frag = fpdu->pkt_hdr;
2020         tx_pkt.first_frag_len = fpdu->pkt_hdr_size;
2021         tx_pkt.enable_ip_cksum = true;
2022         tx_pkt.enable_l4_cksum = true;
2023         tx_pkt.calc_ip_len = true;
2024         /* vlan overload with enum iwarp_ll2_tx_queues */
2025         tx_pkt.vlan = IWARP_LL2_ALIGNED_RIGHT_TRIMMED_TX_QUEUE;
2026
2027         ll2_handle = p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle;
2028
2029         rc = qed_ll2_prepare_tx_packet(p_hwfn, ll2_handle, &tx_pkt, true);
2030         if (rc)
2031                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2032                            "Can't send right edge rc=%d\n", rc);
2033         DP_VERBOSE(p_hwfn,
2034                    QED_MSG_RDMA,
2035                    "MPA_ALIGN: Sent right edge FPDU num_bds=%d [%lx, 0x%x], rc=%d\n",
2036                    tx_pkt.num_of_bds,
2037                    (unsigned long int)tx_pkt.first_frag,
2038                    tx_pkt.first_frag_len, rc);
2039
2040         return rc;
2041 }
2042
2043 static int
2044 qed_iwarp_send_fpdu(struct qed_hwfn *p_hwfn,
2045                     struct qed_iwarp_fpdu *fpdu,
2046                     struct unaligned_opaque_data *curr_pkt,
2047                     struct qed_iwarp_ll2_buff *buf,
2048                     u16 tcp_payload_size, enum qed_iwarp_mpa_pkt_type pkt_type)
2049 {
2050         struct qed_ll2_tx_pkt_info tx_pkt;
2051         u16 first_mpa_offset;
2052         u8 ll2_handle;
2053         int rc;
2054
2055         memset(&tx_pkt, 0, sizeof(tx_pkt));
2056
2057         /* An unaligned packet means it's split over two tcp segments. So the
2058          * complete packet requires 3 bds, one for the header, one for the
2059          * part of the fpdu of the first tcp segment, and the last fragment
2060          * will point to the remainder of the fpdu. A packed pdu, requires only
2061          * two bds, one for the header and one for the data.
2062          */
2063         tx_pkt.num_of_bds = (pkt_type == QED_IWARP_MPA_PKT_UNALIGNED) ? 3 : 2;
2064         tx_pkt.tx_dest = QED_LL2_TX_DEST_LB;
2065         tx_pkt.l4_hdr_offset_w = fpdu->pkt_hdr_size >> 2; /* offset in words */
2066
2067         /* Send the mpa_buf only with the last fpdu (in case of packed) */
2068         if (pkt_type == QED_IWARP_MPA_PKT_UNALIGNED ||
2069             tcp_payload_size <= fpdu->fpdu_length)
2070                 tx_pkt.cookie = fpdu->mpa_buf;
2071
2072         tx_pkt.first_frag = fpdu->pkt_hdr;
2073         tx_pkt.first_frag_len = fpdu->pkt_hdr_size;
2074         tx_pkt.enable_ip_cksum = true;
2075         tx_pkt.enable_l4_cksum = true;
2076         tx_pkt.calc_ip_len = true;
2077         /* vlan overload with enum iwarp_ll2_tx_queues */
2078         tx_pkt.vlan = IWARP_LL2_ALIGNED_TX_QUEUE;
2079
2080         /* special case of unaligned packet and not packed, need to send
2081          * both buffers as cookie to release.
2082          */
2083         if (tcp_payload_size == fpdu->incomplete_bytes)
2084                 fpdu->mpa_buf->piggy_buf = buf;
2085
2086         ll2_handle = p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle;
2087
2088         /* Set first fragment to header */
2089         rc = qed_ll2_prepare_tx_packet(p_hwfn, ll2_handle, &tx_pkt, true);
2090         if (rc)
2091                 goto out;
2092
2093         /* Set second fragment to first part of packet */
2094         rc = qed_ll2_set_fragment_of_tx_packet(p_hwfn, ll2_handle,
2095                                                fpdu->mpa_frag,
2096                                                fpdu->mpa_frag_len);
2097         if (rc)
2098                 goto out;
2099
2100         if (!fpdu->incomplete_bytes)
2101                 goto out;
2102
2103         first_mpa_offset = le16_to_cpu(curr_pkt->first_mpa_offset);
2104
2105         /* Set third fragment to second part of the packet */
2106         rc = qed_ll2_set_fragment_of_tx_packet(p_hwfn,
2107                                                ll2_handle,
2108                                                buf->data_phys_addr +
2109                                                first_mpa_offset,
2110                                                fpdu->incomplete_bytes);
2111 out:
2112         DP_VERBOSE(p_hwfn,
2113                    QED_MSG_RDMA,
2114                    "MPA_ALIGN: Sent FPDU num_bds=%d first_frag_len=%x, mpa_frag_len=0x%x, incomplete_bytes:0x%x rc=%d\n",
2115                    tx_pkt.num_of_bds,
2116                    tx_pkt.first_frag_len,
2117                    fpdu->mpa_frag_len,
2118                    fpdu->incomplete_bytes, rc);
2119
2120         return rc;
2121 }
2122
2123 static void
2124 qed_iwarp_mpa_get_data(struct qed_hwfn *p_hwfn,
2125                        struct unaligned_opaque_data *curr_pkt,
2126                        u32 opaque_data0, u32 opaque_data1)
2127 {
2128         u64 opaque_data;
2129
2130         opaque_data = HILO_64(cpu_to_le32(opaque_data1),
2131                               cpu_to_le32(opaque_data0));
2132         *curr_pkt = *((struct unaligned_opaque_data *)&opaque_data);
2133
2134         le16_add_cpu(&curr_pkt->first_mpa_offset,
2135                      curr_pkt->tcp_payload_offset);
2136 }
2137
2138 /* This function is called when an unaligned or incomplete MPA packet arrives
2139  * driver needs to align the packet, perhaps using previous data and send
2140  * it down to FW once it is aligned.
2141  */
2142 static int
2143 qed_iwarp_process_mpa_pkt(struct qed_hwfn *p_hwfn,
2144                           struct qed_iwarp_ll2_mpa_buf *mpa_buf)
2145 {
2146         struct unaligned_opaque_data *curr_pkt = &mpa_buf->data;
2147         struct qed_iwarp_ll2_buff *buf = mpa_buf->ll2_buf;
2148         enum qed_iwarp_mpa_pkt_type pkt_type;
2149         struct qed_iwarp_fpdu *fpdu;
2150         u16 cid, first_mpa_offset;
2151         int rc = -EINVAL;
2152         u8 *mpa_data;
2153
2154         cid = le32_to_cpu(curr_pkt->cid);
2155
2156         fpdu = qed_iwarp_get_curr_fpdu(p_hwfn, (u16)cid);
2157         if (!fpdu) { /* something corrupt with cid, post rx back */
2158                 DP_ERR(p_hwfn, "Invalid cid, drop and post back to rx cid=%x\n",
2159                        cid);
2160                 goto err;
2161         }
2162
2163         do {
2164                 first_mpa_offset = le16_to_cpu(curr_pkt->first_mpa_offset);
2165                 mpa_data = ((u8 *)(buf->data) + first_mpa_offset);
2166
2167                 pkt_type = qed_iwarp_mpa_classify(p_hwfn, fpdu,
2168                                                   mpa_buf->tcp_payload_len,
2169                                                   mpa_data);
2170
2171                 switch (pkt_type) {
2172                 case QED_IWARP_MPA_PKT_PARTIAL:
2173                         qed_iwarp_init_fpdu(buf, fpdu,
2174                                             curr_pkt,
2175                                             mpa_buf->tcp_payload_len,
2176                                             mpa_buf->placement_offset);
2177
2178                         if (!QED_IWARP_IS_RIGHT_EDGE(curr_pkt)) {
2179                                 mpa_buf->tcp_payload_len = 0;
2180                                 break;
2181                         }
2182
2183                         rc = qed_iwarp_win_right_edge(p_hwfn, fpdu);
2184
2185                         if (rc) {
2186                                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2187                                            "Can't send FPDU:reset rc=%d\n", rc);
2188                                 memset(fpdu, 0, sizeof(*fpdu));
2189                                 break;
2190                         }
2191
2192                         mpa_buf->tcp_payload_len = 0;
2193                         break;
2194                 case QED_IWARP_MPA_PKT_PACKED:
2195                         qed_iwarp_init_fpdu(buf, fpdu,
2196                                             curr_pkt,
2197                                             mpa_buf->tcp_payload_len,
2198                                             mpa_buf->placement_offset);
2199
2200                         rc = qed_iwarp_send_fpdu(p_hwfn, fpdu, curr_pkt, buf,
2201                                                  mpa_buf->tcp_payload_len,
2202                                                  pkt_type);
2203                         if (rc) {
2204                                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2205                                            "Can't send FPDU:reset rc=%d\n", rc);
2206                                 memset(fpdu, 0, sizeof(*fpdu));
2207                                 break;
2208                         }
2209
2210                         mpa_buf->tcp_payload_len -= fpdu->fpdu_length;
2211                         le16_add_cpu(&curr_pkt->first_mpa_offset,
2212                                      fpdu->fpdu_length);
2213                         break;
2214                 case QED_IWARP_MPA_PKT_UNALIGNED:
2215                         qed_iwarp_update_fpdu_length(p_hwfn, fpdu, mpa_data);
2216                         if (mpa_buf->tcp_payload_len < fpdu->incomplete_bytes) {
2217                                 /* special handling of fpdu split over more
2218                                  * than 2 segments
2219                                  */
2220                                 if (QED_IWARP_IS_RIGHT_EDGE(curr_pkt)) {
2221                                         rc = qed_iwarp_win_right_edge(p_hwfn,
2222                                                                       fpdu);
2223                                         /* packet will be re-processed later */
2224                                         if (rc)
2225                                                 return rc;
2226                                 }
2227
2228                                 rc = qed_iwarp_cp_pkt(p_hwfn, fpdu, curr_pkt,
2229                                                       buf,
2230                                                       mpa_buf->tcp_payload_len);
2231                                 if (rc) /* packet will be re-processed later */
2232                                         return rc;
2233
2234                                 mpa_buf->tcp_payload_len = 0;
2235                                 break;
2236                         }
2237
2238                         rc = qed_iwarp_send_fpdu(p_hwfn, fpdu, curr_pkt, buf,
2239                                                  mpa_buf->tcp_payload_len,
2240                                                  pkt_type);
2241                         if (rc) {
2242                                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2243                                            "Can't send FPDU:delay rc=%d\n", rc);
2244                                 /* don't reset fpdu -> we need it for next
2245                                  * classify
2246                                  */
2247                                 break;
2248                         }
2249
2250                         mpa_buf->tcp_payload_len -= fpdu->incomplete_bytes;
2251                         le16_add_cpu(&curr_pkt->first_mpa_offset,
2252                                      fpdu->incomplete_bytes);
2253
2254                         /* The framed PDU was sent - no more incomplete bytes */
2255                         fpdu->incomplete_bytes = 0;
2256                         break;
2257                 }
2258         } while (mpa_buf->tcp_payload_len && !rc);
2259
2260         return rc;
2261
2262 err:
2263         qed_iwarp_ll2_post_rx(p_hwfn,
2264                               buf,
2265                               p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle);
2266         return rc;
2267 }
2268
2269 static void qed_iwarp_process_pending_pkts(struct qed_hwfn *p_hwfn)
2270 {
2271         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2272         struct qed_iwarp_ll2_mpa_buf *mpa_buf = NULL;
2273         int rc;
2274
2275         while (!list_empty(&iwarp_info->mpa_buf_pending_list)) {
2276                 mpa_buf = list_first_entry(&iwarp_info->mpa_buf_pending_list,
2277                                            struct qed_iwarp_ll2_mpa_buf,
2278                                            list_entry);
2279
2280                 rc = qed_iwarp_process_mpa_pkt(p_hwfn, mpa_buf);
2281
2282                 /* busy means break and continue processing later, don't
2283                  * remove the buf from the pending list.
2284                  */
2285                 if (rc == -EBUSY)
2286                         break;
2287
2288                 list_move_tail(&mpa_buf->list_entry,
2289                                &iwarp_info->mpa_buf_list);
2290
2291                 if (rc) {       /* different error, don't continue */
2292                         DP_NOTICE(p_hwfn, "process pkts failed rc=%d\n", rc);
2293                         break;
2294                 }
2295         }
2296 }
2297
2298 static void
2299 qed_iwarp_ll2_comp_mpa_pkt(void *cxt, struct qed_ll2_comp_rx_data *data)
2300 {
2301         struct qed_iwarp_ll2_mpa_buf *mpa_buf;
2302         struct qed_iwarp_info *iwarp_info;
2303         struct qed_hwfn *p_hwfn = cxt;
2304         u16 first_mpa_offset;
2305
2306         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2307         mpa_buf = list_first_entry(&iwarp_info->mpa_buf_list,
2308                                    struct qed_iwarp_ll2_mpa_buf, list_entry);
2309         if (!mpa_buf) {
2310                 DP_ERR(p_hwfn, "No free mpa buf\n");
2311                 goto err;
2312         }
2313
2314         list_del(&mpa_buf->list_entry);
2315         qed_iwarp_mpa_get_data(p_hwfn, &mpa_buf->data,
2316                                data->opaque_data_0, data->opaque_data_1);
2317
2318         first_mpa_offset = le16_to_cpu(mpa_buf->data.first_mpa_offset);
2319
2320         DP_VERBOSE(p_hwfn,
2321                    QED_MSG_RDMA,
2322                    "LL2 MPA CompRx payload_len:0x%x\tfirst_mpa_offset:0x%x\ttcp_payload_offset:0x%x\tflags:0x%x\tcid:0x%x\n",
2323                    data->length.packet_length, first_mpa_offset,
2324                    mpa_buf->data.tcp_payload_offset, mpa_buf->data.flags,
2325                    mpa_buf->data.cid);
2326
2327         mpa_buf->ll2_buf = data->cookie;
2328         mpa_buf->tcp_payload_len = data->length.packet_length -
2329                                    first_mpa_offset;
2330
2331         first_mpa_offset += data->u.placement_offset;
2332         mpa_buf->data.first_mpa_offset = cpu_to_le16(first_mpa_offset);
2333         mpa_buf->placement_offset = data->u.placement_offset;
2334
2335         list_add_tail(&mpa_buf->list_entry, &iwarp_info->mpa_buf_pending_list);
2336
2337         qed_iwarp_process_pending_pkts(p_hwfn);
2338         return;
2339 err:
2340         qed_iwarp_ll2_post_rx(p_hwfn, data->cookie,
2341                               iwarp_info->ll2_mpa_handle);
2342 }
2343
2344 static void
2345 qed_iwarp_ll2_comp_syn_pkt(void *cxt, struct qed_ll2_comp_rx_data *data)
2346 {
2347         struct qed_iwarp_ll2_buff *buf = data->cookie;
2348         struct qed_iwarp_listener *listener;
2349         struct qed_ll2_tx_pkt_info tx_pkt;
2350         struct qed_iwarp_cm_info cm_info;
2351         struct qed_hwfn *p_hwfn = cxt;
2352         u8 remote_mac_addr[ETH_ALEN];
2353         u8 local_mac_addr[ETH_ALEN];
2354         struct qed_iwarp_ep *ep;
2355         int tcp_start_offset;
2356         u8 ll2_syn_handle;
2357         int payload_len;
2358         u32 hdr_size;
2359         int rc;
2360
2361         memset(&cm_info, 0, sizeof(cm_info));
2362         ll2_syn_handle = p_hwfn->p_rdma_info->iwarp.ll2_syn_handle;
2363
2364         /* Check if packet was received with errors... */
2365         if (data->err_flags) {
2366                 DP_NOTICE(p_hwfn, "Error received on SYN packet: 0x%x\n",
2367                           data->err_flags);
2368                 goto err;
2369         }
2370
2371         if (GET_FIELD(data->parse_flags,
2372                       PARSING_AND_ERR_FLAGS_L4CHKSMWASCALCULATED) &&
2373             GET_FIELD(data->parse_flags, PARSING_AND_ERR_FLAGS_L4CHKSMERROR)) {
2374                 DP_NOTICE(p_hwfn, "Syn packet received with checksum error\n");
2375                 goto err;
2376         }
2377
2378         rc = qed_iwarp_parse_rx_pkt(p_hwfn, &cm_info, (u8 *)(buf->data) +
2379                                     data->u.placement_offset, remote_mac_addr,
2380                                     local_mac_addr, &payload_len,
2381                                     &tcp_start_offset);
2382         if (rc)
2383                 goto err;
2384
2385         /* Check if there is a listener for this 4-tuple+vlan */
2386         listener = qed_iwarp_get_listener(p_hwfn, &cm_info);
2387         if (!listener) {
2388                 DP_VERBOSE(p_hwfn,
2389                            QED_MSG_RDMA,
2390                            "SYN received on tuple not listened on parse_flags=%d packet len=%d\n",
2391                            data->parse_flags, data->length.packet_length);
2392
2393                 memset(&tx_pkt, 0, sizeof(tx_pkt));
2394                 tx_pkt.num_of_bds = 1;
2395                 tx_pkt.l4_hdr_offset_w = (data->length.packet_length) >> 2;
2396                 tx_pkt.tx_dest = QED_LL2_TX_DEST_LB;
2397                 tx_pkt.first_frag = buf->data_phys_addr +
2398                                     data->u.placement_offset;
2399                 tx_pkt.first_frag_len = data->length.packet_length;
2400                 tx_pkt.cookie = buf;
2401
2402                 rc = qed_ll2_prepare_tx_packet(p_hwfn, ll2_syn_handle,
2403                                                &tx_pkt, true);
2404
2405                 if (rc) {
2406                         DP_NOTICE(p_hwfn,
2407                                   "Can't post SYN back to chip rc=%d\n", rc);
2408                         goto err;
2409                 }
2410                 return;
2411         }
2412
2413         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "Received syn on listening port\n");
2414         /* There may be an open ep on this connection if this is a syn
2415          * retrasnmit... need to make sure there isn't...
2416          */
2417         if (qed_iwarp_ep_exists(p_hwfn, &cm_info))
2418                 goto err;
2419
2420         ep = qed_iwarp_get_free_ep(p_hwfn);
2421         if (!ep)
2422                 goto err;
2423
2424         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
2425         list_add_tail(&ep->list_entry, &p_hwfn->p_rdma_info->iwarp.ep_list);
2426         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
2427
2428         ether_addr_copy(ep->remote_mac_addr, remote_mac_addr);
2429         ether_addr_copy(ep->local_mac_addr, local_mac_addr);
2430
2431         memcpy(&ep->cm_info, &cm_info, sizeof(ep->cm_info));
2432
2433         hdr_size = ((cm_info.ip_version == QED_TCP_IPV4) ? 40 : 60);
2434         ep->mss = p_hwfn->p_rdma_info->iwarp.max_mtu - hdr_size;
2435         ep->mss = min_t(u16, QED_IWARP_MAX_FW_MSS, ep->mss);
2436
2437         ep->event_cb = listener->event_cb;
2438         ep->cb_context = listener->cb_context;
2439         ep->connect_mode = TCP_CONNECT_PASSIVE;
2440
2441         ep->syn = buf;
2442         ep->syn_ip_payload_length = (u16)payload_len;
2443         ep->syn_phy_addr = buf->data_phys_addr + data->u.placement_offset +
2444                            tcp_start_offset;
2445
2446         rc = qed_iwarp_tcp_offload(p_hwfn, ep);
2447         if (rc) {
2448                 qed_iwarp_return_ep(p_hwfn, ep);
2449                 goto err;
2450         }
2451
2452         return;
2453 err:
2454         qed_iwarp_ll2_post_rx(p_hwfn, buf, ll2_syn_handle);
2455 }
2456
2457 static void qed_iwarp_ll2_rel_rx_pkt(void *cxt, u8 connection_handle,
2458                                      void *cookie, dma_addr_t rx_buf_addr,
2459                                      bool b_last_packet)
2460 {
2461         struct qed_iwarp_ll2_buff *buffer = cookie;
2462         struct qed_hwfn *p_hwfn = cxt;
2463
2464         dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size,
2465                           buffer->data, buffer->data_phys_addr);
2466         kfree(buffer);
2467 }
2468
2469 static void qed_iwarp_ll2_comp_tx_pkt(void *cxt, u8 connection_handle,
2470                                       void *cookie, dma_addr_t first_frag_addr,
2471                                       bool b_last_fragment, bool b_last_packet)
2472 {
2473         struct qed_iwarp_ll2_buff *buffer = cookie;
2474         struct qed_iwarp_ll2_buff *piggy;
2475         struct qed_hwfn *p_hwfn = cxt;
2476
2477         if (!buffer)            /* can happen in packed mpa unaligned... */
2478                 return;
2479
2480         /* this was originally an rx packet, post it back */
2481         piggy = buffer->piggy_buf;
2482         if (piggy) {
2483                 buffer->piggy_buf = NULL;
2484                 qed_iwarp_ll2_post_rx(p_hwfn, piggy, connection_handle);
2485         }
2486
2487         qed_iwarp_ll2_post_rx(p_hwfn, buffer, connection_handle);
2488
2489         if (connection_handle == p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle)
2490                 qed_iwarp_process_pending_pkts(p_hwfn);
2491
2492         return;
2493 }
2494
2495 static void qed_iwarp_ll2_rel_tx_pkt(void *cxt, u8 connection_handle,
2496                                      void *cookie, dma_addr_t first_frag_addr,
2497                                      bool b_last_fragment, bool b_last_packet)
2498 {
2499         struct qed_iwarp_ll2_buff *buffer = cookie;
2500         struct qed_hwfn *p_hwfn = cxt;
2501
2502         if (!buffer)
2503                 return;
2504
2505         if (buffer->piggy_buf) {
2506                 dma_free_coherent(&p_hwfn->cdev->pdev->dev,
2507                                   buffer->piggy_buf->buff_size,
2508                                   buffer->piggy_buf->data,
2509                                   buffer->piggy_buf->data_phys_addr);
2510
2511                 kfree(buffer->piggy_buf);
2512         }
2513
2514         dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size,
2515                           buffer->data, buffer->data_phys_addr);
2516
2517         kfree(buffer);
2518 }
2519
2520 /* The only slowpath for iwarp ll2 is unalign flush. When this completion
2521  * is received, need to reset the FPDU.
2522  */
2523 static void
2524 qed_iwarp_ll2_slowpath(void *cxt,
2525                        u8 connection_handle,
2526                        u32 opaque_data_0, u32 opaque_data_1)
2527 {
2528         struct unaligned_opaque_data unalign_data;
2529         struct qed_hwfn *p_hwfn = cxt;
2530         struct qed_iwarp_fpdu *fpdu;
2531         u32 cid;
2532
2533         qed_iwarp_mpa_get_data(p_hwfn, &unalign_data,
2534                                opaque_data_0, opaque_data_1);
2535
2536         cid = le32_to_cpu(unalign_data.cid);
2537
2538         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "(0x%x) Flush fpdu\n", cid);
2539
2540         fpdu = qed_iwarp_get_curr_fpdu(p_hwfn, (u16)cid);
2541         if (fpdu)
2542                 memset(fpdu, 0, sizeof(*fpdu));
2543 }
2544
2545 static int qed_iwarp_ll2_stop(struct qed_hwfn *p_hwfn)
2546 {
2547         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2548         int rc = 0;
2549
2550         if (iwarp_info->ll2_syn_handle != QED_IWARP_HANDLE_INVAL) {
2551                 rc = qed_ll2_terminate_connection(p_hwfn,
2552                                                   iwarp_info->ll2_syn_handle);
2553                 if (rc)
2554                         DP_INFO(p_hwfn, "Failed to terminate syn connection\n");
2555
2556                 qed_ll2_release_connection(p_hwfn, iwarp_info->ll2_syn_handle);
2557                 iwarp_info->ll2_syn_handle = QED_IWARP_HANDLE_INVAL;
2558         }
2559
2560         if (iwarp_info->ll2_ooo_handle != QED_IWARP_HANDLE_INVAL) {
2561                 rc = qed_ll2_terminate_connection(p_hwfn,
2562                                                   iwarp_info->ll2_ooo_handle);
2563                 if (rc)
2564                         DP_INFO(p_hwfn, "Failed to terminate ooo connection\n");
2565
2566                 qed_ll2_release_connection(p_hwfn, iwarp_info->ll2_ooo_handle);
2567                 iwarp_info->ll2_ooo_handle = QED_IWARP_HANDLE_INVAL;
2568         }
2569
2570         if (iwarp_info->ll2_mpa_handle != QED_IWARP_HANDLE_INVAL) {
2571                 rc = qed_ll2_terminate_connection(p_hwfn,
2572                                                   iwarp_info->ll2_mpa_handle);
2573                 if (rc)
2574                         DP_INFO(p_hwfn, "Failed to terminate mpa connection\n");
2575
2576                 qed_ll2_release_connection(p_hwfn, iwarp_info->ll2_mpa_handle);
2577                 iwarp_info->ll2_mpa_handle = QED_IWARP_HANDLE_INVAL;
2578         }
2579
2580         qed_llh_remove_mac_filter(p_hwfn->cdev, 0,
2581                                   p_hwfn->p_rdma_info->iwarp.mac_addr);
2582
2583         return rc;
2584 }
2585
2586 static int
2587 qed_iwarp_ll2_alloc_buffers(struct qed_hwfn *p_hwfn,
2588                             int num_rx_bufs, int buff_size, u8 ll2_handle)
2589 {
2590         struct qed_iwarp_ll2_buff *buffer;
2591         int rc = 0;
2592         int i;
2593
2594         for (i = 0; i < num_rx_bufs; i++) {
2595                 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL);
2596                 if (!buffer) {
2597                         rc = -ENOMEM;
2598                         break;
2599                 }
2600
2601                 buffer->data = dma_alloc_coherent(&p_hwfn->cdev->pdev->dev,
2602                                                   buff_size,
2603                                                   &buffer->data_phys_addr,
2604                                                   GFP_KERNEL);
2605                 if (!buffer->data) {
2606                         kfree(buffer);
2607                         rc = -ENOMEM;
2608                         break;
2609                 }
2610
2611                 buffer->buff_size = buff_size;
2612                 rc = qed_iwarp_ll2_post_rx(p_hwfn, buffer, ll2_handle);
2613                 if (rc)
2614                         /* buffers will be deallocated by qed_ll2 */
2615                         break;
2616         }
2617         return rc;
2618 }
2619
2620 #define QED_IWARP_MAX_BUF_SIZE(mtu)                                  \
2621         ALIGN((mtu) + ETH_HLEN + 2 * VLAN_HLEN + 2 + ETH_CACHE_LINE_SIZE, \
2622                 ETH_CACHE_LINE_SIZE)
2623
2624 static int
2625 qed_iwarp_ll2_start(struct qed_hwfn *p_hwfn,
2626                     struct qed_rdma_start_in_params *params,
2627                     u32 rcv_wnd_size)
2628 {
2629         struct qed_iwarp_info *iwarp_info;
2630         struct qed_ll2_acquire_data data;
2631         struct qed_ll2_cbs cbs;
2632         u32 buff_size;
2633         u16 n_ooo_bufs;
2634         int rc = 0;
2635         int i;
2636
2637         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2638         iwarp_info->ll2_syn_handle = QED_IWARP_HANDLE_INVAL;
2639         iwarp_info->ll2_ooo_handle = QED_IWARP_HANDLE_INVAL;
2640         iwarp_info->ll2_mpa_handle = QED_IWARP_HANDLE_INVAL;
2641
2642         iwarp_info->max_mtu = params->max_mtu;
2643
2644         ether_addr_copy(p_hwfn->p_rdma_info->iwarp.mac_addr, params->mac_addr);
2645
2646         rc = qed_llh_add_mac_filter(p_hwfn->cdev, 0, params->mac_addr);
2647         if (rc)
2648                 return rc;
2649
2650         /* Start SYN connection */
2651         cbs.rx_comp_cb = qed_iwarp_ll2_comp_syn_pkt;
2652         cbs.rx_release_cb = qed_iwarp_ll2_rel_rx_pkt;
2653         cbs.tx_comp_cb = qed_iwarp_ll2_comp_tx_pkt;
2654         cbs.tx_release_cb = qed_iwarp_ll2_rel_tx_pkt;
2655         cbs.slowpath_cb = NULL;
2656         cbs.cookie = p_hwfn;
2657
2658         memset(&data, 0, sizeof(data));
2659         data.input.conn_type = QED_LL2_TYPE_IWARP;
2660         /* SYN will use ctx based queues */
2661         data.input.rx_conn_type = QED_LL2_RX_TYPE_CTX;
2662         data.input.mtu = params->max_mtu;
2663         data.input.rx_num_desc = QED_IWARP_LL2_SYN_RX_SIZE;
2664         data.input.tx_num_desc = QED_IWARP_LL2_SYN_TX_SIZE;
2665         data.input.tx_max_bds_per_packet = 1;   /* will never be fragmented */
2666         data.input.tx_tc = PKT_LB_TC;
2667         data.input.tx_dest = QED_LL2_TX_DEST_LB;
2668         data.p_connection_handle = &iwarp_info->ll2_syn_handle;
2669         data.cbs = &cbs;
2670
2671         rc = qed_ll2_acquire_connection(p_hwfn, &data);
2672         if (rc) {
2673                 DP_NOTICE(p_hwfn, "Failed to acquire LL2 connection\n");
2674                 qed_llh_remove_mac_filter(p_hwfn->cdev, 0, params->mac_addr);
2675                 return rc;
2676         }
2677
2678         rc = qed_ll2_establish_connection(p_hwfn, iwarp_info->ll2_syn_handle);
2679         if (rc) {
2680                 DP_NOTICE(p_hwfn, "Failed to establish LL2 connection\n");
2681                 goto err;
2682         }
2683
2684         buff_size = QED_IWARP_MAX_BUF_SIZE(params->max_mtu);
2685         rc = qed_iwarp_ll2_alloc_buffers(p_hwfn,
2686                                          QED_IWARP_LL2_SYN_RX_SIZE,
2687                                          buff_size,
2688                                          iwarp_info->ll2_syn_handle);
2689         if (rc)
2690                 goto err;
2691
2692         /* Start OOO connection */
2693         data.input.conn_type = QED_LL2_TYPE_OOO;
2694         /* OOO/unaligned will use legacy ll2 queues (ram based) */
2695         data.input.rx_conn_type = QED_LL2_RX_TYPE_LEGACY;
2696         data.input.mtu = params->max_mtu;
2697
2698         n_ooo_bufs = (QED_IWARP_MAX_OOO * rcv_wnd_size) /
2699                      iwarp_info->max_mtu;
2700         n_ooo_bufs = min_t(u32, n_ooo_bufs, QED_IWARP_LL2_OOO_MAX_RX_SIZE);
2701
2702         data.input.rx_num_desc = n_ooo_bufs;
2703         data.input.rx_num_ooo_buffers = n_ooo_bufs;
2704
2705         data.input.tx_max_bds_per_packet = 1;   /* will never be fragmented */
2706         data.input.tx_num_desc = QED_IWARP_LL2_OOO_DEF_TX_SIZE;
2707         data.p_connection_handle = &iwarp_info->ll2_ooo_handle;
2708
2709         rc = qed_ll2_acquire_connection(p_hwfn, &data);
2710         if (rc)
2711                 goto err;
2712
2713         rc = qed_ll2_establish_connection(p_hwfn, iwarp_info->ll2_ooo_handle);
2714         if (rc)
2715                 goto err;
2716
2717         /* Start Unaligned MPA connection */
2718         cbs.rx_comp_cb = qed_iwarp_ll2_comp_mpa_pkt;
2719         cbs.slowpath_cb = qed_iwarp_ll2_slowpath;
2720
2721         memset(&data, 0, sizeof(data));
2722         data.input.conn_type = QED_LL2_TYPE_IWARP;
2723         data.input.mtu = params->max_mtu;
2724         /* FW requires that once a packet arrives OOO, it must have at
2725          * least 2 rx buffers available on the unaligned connection
2726          * for handling the case that it is a partial fpdu.
2727          */
2728         data.input.rx_num_desc = n_ooo_bufs * 2;
2729         data.input.tx_num_desc = data.input.rx_num_desc;
2730         data.input.tx_max_bds_per_packet = QED_IWARP_MAX_BDS_PER_FPDU;
2731         data.input.tx_tc = PKT_LB_TC;
2732         data.input.tx_dest = QED_LL2_TX_DEST_LB;
2733         data.p_connection_handle = &iwarp_info->ll2_mpa_handle;
2734         data.input.secondary_queue = true;
2735         data.cbs = &cbs;
2736
2737         rc = qed_ll2_acquire_connection(p_hwfn, &data);
2738         if (rc)
2739                 goto err;
2740
2741         rc = qed_ll2_establish_connection(p_hwfn, iwarp_info->ll2_mpa_handle);
2742         if (rc)
2743                 goto err;
2744
2745         rc = qed_iwarp_ll2_alloc_buffers(p_hwfn,
2746                                          data.input.rx_num_desc,
2747                                          buff_size,
2748                                          iwarp_info->ll2_mpa_handle);
2749         if (rc)
2750                 goto err;
2751
2752         iwarp_info->partial_fpdus = kcalloc((u16)p_hwfn->p_rdma_info->num_qps,
2753                                             sizeof(*iwarp_info->partial_fpdus),
2754                                             GFP_KERNEL);
2755         if (!iwarp_info->partial_fpdus) {
2756                 rc = -ENOMEM;
2757                 goto err;
2758         }
2759
2760         iwarp_info->max_num_partial_fpdus = (u16)p_hwfn->p_rdma_info->num_qps;
2761
2762         iwarp_info->mpa_intermediate_buf = kzalloc(buff_size, GFP_KERNEL);
2763         if (!iwarp_info->mpa_intermediate_buf) {
2764                 rc = -ENOMEM;
2765                 goto err;
2766         }
2767
2768         /* The mpa_bufs array serves for pending RX packets received on the
2769          * mpa ll2 that don't have place on the tx ring and require later
2770          * processing. We can't fail on allocation of such a struct therefore
2771          * we allocate enough to take care of all rx packets
2772          */
2773         iwarp_info->mpa_bufs = kcalloc(data.input.rx_num_desc,
2774                                        sizeof(*iwarp_info->mpa_bufs),
2775                                        GFP_KERNEL);
2776         if (!iwarp_info->mpa_bufs) {
2777                 rc = -ENOMEM;
2778                 goto err;
2779         }
2780
2781         INIT_LIST_HEAD(&iwarp_info->mpa_buf_pending_list);
2782         INIT_LIST_HEAD(&iwarp_info->mpa_buf_list);
2783         for (i = 0; i < data.input.rx_num_desc; i++)
2784                 list_add_tail(&iwarp_info->mpa_bufs[i].list_entry,
2785                               &iwarp_info->mpa_buf_list);
2786         return rc;
2787 err:
2788         qed_iwarp_ll2_stop(p_hwfn);
2789
2790         return rc;
2791 }
2792
2793 static struct {
2794         u32 two_ports;
2795         u32 four_ports;
2796 } qed_iwarp_rcv_wnd_size[MAX_CHIP_IDS] = {
2797         {QED_IWARP_RCV_WND_SIZE_DEF_BB_2P, QED_IWARP_RCV_WND_SIZE_DEF_BB_4P},
2798         {QED_IWARP_RCV_WND_SIZE_DEF_AH_2P, QED_IWARP_RCV_WND_SIZE_DEF_AH_4P}
2799 };
2800
2801 int qed_iwarp_setup(struct qed_hwfn *p_hwfn,
2802                     struct qed_rdma_start_in_params *params)
2803 {
2804         struct qed_dev *cdev = p_hwfn->cdev;
2805         struct qed_iwarp_info *iwarp_info;
2806         enum chip_ids chip_id;
2807         u32 rcv_wnd_size;
2808
2809         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2810
2811         iwarp_info->tcp_flags = QED_IWARP_TS_EN;
2812
2813         chip_id = QED_IS_BB(cdev) ? CHIP_BB : CHIP_K2;
2814         rcv_wnd_size = (qed_device_num_ports(cdev) == 4) ?
2815                 qed_iwarp_rcv_wnd_size[chip_id].four_ports :
2816                 qed_iwarp_rcv_wnd_size[chip_id].two_ports;
2817
2818         /* value 0 is used for ilog2(QED_IWARP_RCV_WND_SIZE_MIN) */
2819         iwarp_info->rcv_wnd_scale = ilog2(rcv_wnd_size) -
2820             ilog2(QED_IWARP_RCV_WND_SIZE_MIN);
2821         iwarp_info->rcv_wnd_size = rcv_wnd_size >> iwarp_info->rcv_wnd_scale;
2822         iwarp_info->crc_needed = QED_IWARP_PARAM_CRC_NEEDED;
2823         iwarp_info->mpa_rev = MPA_NEGOTIATION_TYPE_ENHANCED;
2824
2825         iwarp_info->peer2peer = QED_IWARP_PARAM_P2P;
2826
2827         iwarp_info->rtr_type =  MPA_RTR_TYPE_ZERO_SEND |
2828                                 MPA_RTR_TYPE_ZERO_WRITE |
2829                                 MPA_RTR_TYPE_ZERO_READ;
2830
2831         spin_lock_init(&p_hwfn->p_rdma_info->iwarp.qp_lock);
2832         INIT_LIST_HEAD(&p_hwfn->p_rdma_info->iwarp.ep_list);
2833         INIT_LIST_HEAD(&p_hwfn->p_rdma_info->iwarp.listen_list);
2834
2835         qed_spq_register_async_cb(p_hwfn, PROTOCOLID_IWARP,
2836                                   qed_iwarp_async_event);
2837         qed_ooo_setup(p_hwfn);
2838
2839         return qed_iwarp_ll2_start(p_hwfn, params, rcv_wnd_size);
2840 }
2841
2842 int qed_iwarp_stop(struct qed_hwfn *p_hwfn)
2843 {
2844         int rc;
2845
2846         qed_iwarp_free_prealloc_ep(p_hwfn);
2847         rc = qed_iwarp_wait_for_all_cids(p_hwfn);
2848         if (rc)
2849                 return rc;
2850
2851         return qed_iwarp_ll2_stop(p_hwfn);
2852 }
2853
2854 static void qed_iwarp_qp_in_error(struct qed_hwfn *p_hwfn,
2855                                   struct qed_iwarp_ep *ep,
2856                                   u8 fw_return_code)
2857 {
2858         struct qed_iwarp_cm_event_params params;
2859
2860         qed_iwarp_modify_qp(p_hwfn, ep->qp, QED_IWARP_QP_STATE_ERROR, true);
2861
2862         params.event = QED_IWARP_EVENT_CLOSE;
2863         params.ep_context = ep;
2864         params.cm_info = &ep->cm_info;
2865         params.status = (fw_return_code == IWARP_QP_IN_ERROR_GOOD_CLOSE) ?
2866                          0 : -ECONNRESET;
2867
2868         /* paired with READ_ONCE in destroy_qp */
2869         smp_store_release(&ep->state, QED_IWARP_EP_CLOSED);
2870
2871         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
2872         list_del(&ep->list_entry);
2873         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
2874
2875         ep->event_cb(ep->cb_context, &params);
2876 }
2877
2878 static void qed_iwarp_exception_received(struct qed_hwfn *p_hwfn,
2879                                          struct qed_iwarp_ep *ep,
2880                                          int fw_ret_code)
2881 {
2882         struct qed_iwarp_cm_event_params params;
2883         bool event_cb = false;
2884
2885         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "EP(0x%x) fw_ret_code=%d\n",
2886                    ep->cid, fw_ret_code);
2887
2888         switch (fw_ret_code) {
2889         case IWARP_EXCEPTION_DETECTED_LLP_CLOSED:
2890                 params.status = 0;
2891                 params.event = QED_IWARP_EVENT_DISCONNECT;
2892                 event_cb = true;
2893                 break;
2894         case IWARP_EXCEPTION_DETECTED_LLP_RESET:
2895                 params.status = -ECONNRESET;
2896                 params.event = QED_IWARP_EVENT_DISCONNECT;
2897                 event_cb = true;
2898                 break;
2899         case IWARP_EXCEPTION_DETECTED_RQ_EMPTY:
2900                 params.event = QED_IWARP_EVENT_RQ_EMPTY;
2901                 event_cb = true;
2902                 break;
2903         case IWARP_EXCEPTION_DETECTED_IRQ_FULL:
2904                 params.event = QED_IWARP_EVENT_IRQ_FULL;
2905                 event_cb = true;
2906                 break;
2907         case IWARP_EXCEPTION_DETECTED_LLP_TIMEOUT:
2908                 params.event = QED_IWARP_EVENT_LLP_TIMEOUT;
2909                 event_cb = true;
2910                 break;
2911         case IWARP_EXCEPTION_DETECTED_REMOTE_PROTECTION_ERROR:
2912                 params.event = QED_IWARP_EVENT_REMOTE_PROTECTION_ERROR;
2913                 event_cb = true;
2914                 break;
2915         case IWARP_EXCEPTION_DETECTED_CQ_OVERFLOW:
2916                 params.event = QED_IWARP_EVENT_CQ_OVERFLOW;
2917                 event_cb = true;
2918                 break;
2919         case IWARP_EXCEPTION_DETECTED_LOCAL_CATASTROPHIC:
2920                 params.event = QED_IWARP_EVENT_QP_CATASTROPHIC;
2921                 event_cb = true;
2922                 break;
2923         case IWARP_EXCEPTION_DETECTED_LOCAL_ACCESS_ERROR:
2924                 params.event = QED_IWARP_EVENT_LOCAL_ACCESS_ERROR;
2925                 event_cb = true;
2926                 break;
2927         case IWARP_EXCEPTION_DETECTED_REMOTE_OPERATION_ERROR:
2928                 params.event = QED_IWARP_EVENT_REMOTE_OPERATION_ERROR;
2929                 event_cb = true;
2930                 break;
2931         case IWARP_EXCEPTION_DETECTED_TERMINATE_RECEIVED:
2932                 params.event = QED_IWARP_EVENT_TERMINATE_RECEIVED;
2933                 event_cb = true;
2934                 break;
2935         default:
2936                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2937                            "Unhandled exception received...fw_ret_code=%d\n",
2938                            fw_ret_code);
2939                 break;
2940         }
2941
2942         if (event_cb) {
2943                 params.ep_context = ep;
2944                 params.cm_info = &ep->cm_info;
2945                 ep->event_cb(ep->cb_context, &params);
2946         }
2947 }
2948
2949 static void
2950 qed_iwarp_tcp_connect_unsuccessful(struct qed_hwfn *p_hwfn,
2951                                    struct qed_iwarp_ep *ep, u8 fw_return_code)
2952 {
2953         struct qed_iwarp_cm_event_params params;
2954
2955         memset(&params, 0, sizeof(params));
2956         params.event = QED_IWARP_EVENT_ACTIVE_COMPLETE;
2957         params.ep_context = ep;
2958         params.cm_info = &ep->cm_info;
2959         /* paired with READ_ONCE in destroy_qp */
2960         smp_store_release(&ep->state, QED_IWARP_EP_CLOSED);
2961
2962         switch (fw_return_code) {
2963         case IWARP_CONN_ERROR_TCP_CONNECT_INVALID_PACKET:
2964                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2965                            "%s(0x%x) TCP connect got invalid packet\n",
2966                            QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2967                 params.status = -ECONNRESET;
2968                 break;
2969         case IWARP_CONN_ERROR_TCP_CONNECTION_RST:
2970                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2971                            "%s(0x%x) TCP Connection Reset\n",
2972                            QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2973                 params.status = -ECONNRESET;
2974                 break;
2975         case IWARP_CONN_ERROR_TCP_CONNECT_TIMEOUT:
2976                 DP_NOTICE(p_hwfn, "%s(0x%x) TCP timeout\n",
2977                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2978                 params.status = -EBUSY;
2979                 break;
2980         case IWARP_CONN_ERROR_MPA_NOT_SUPPORTED_VER:
2981                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA not supported VER\n",
2982                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2983                 params.status = -ECONNREFUSED;
2984                 break;
2985         case IWARP_CONN_ERROR_MPA_INVALID_PACKET:
2986                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA Invalid Packet\n",
2987                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2988                 params.status = -ECONNRESET;
2989                 break;
2990         default:
2991                 DP_ERR(p_hwfn,
2992                        "%s(0x%x) Unexpected return code tcp connect: %d\n",
2993                        QED_IWARP_CONNECT_MODE_STRING(ep),
2994                        ep->tcp_cid, fw_return_code);
2995                 params.status = -ECONNRESET;
2996                 break;
2997         }
2998
2999         if (ep->connect_mode == TCP_CONNECT_PASSIVE) {
3000                 ep->tcp_cid = QED_IWARP_INVALID_TCP_CID;
3001                 qed_iwarp_return_ep(p_hwfn, ep);
3002         } else {
3003                 ep->event_cb(ep->cb_context, &params);
3004                 spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3005                 list_del(&ep->list_entry);
3006                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3007         }
3008 }
3009
3010 static void
3011 qed_iwarp_connect_complete(struct qed_hwfn *p_hwfn,
3012                            struct qed_iwarp_ep *ep, u8 fw_return_code)
3013 {
3014         u8 ll2_syn_handle = p_hwfn->p_rdma_info->iwarp.ll2_syn_handle;
3015
3016         if (ep->connect_mode == TCP_CONNECT_PASSIVE) {
3017                 /* Done with the SYN packet, post back to ll2 rx */
3018                 qed_iwarp_ll2_post_rx(p_hwfn, ep->syn, ll2_syn_handle);
3019
3020                 ep->syn = NULL;
3021
3022                 /* If connect failed - upper layer doesn't know about it */
3023                 if (fw_return_code == RDMA_RETURN_OK)
3024                         qed_iwarp_mpa_received(p_hwfn, ep);
3025                 else
3026                         qed_iwarp_tcp_connect_unsuccessful(p_hwfn, ep,
3027                                                            fw_return_code);
3028         } else {
3029                 if (fw_return_code == RDMA_RETURN_OK)
3030                         qed_iwarp_mpa_offload(p_hwfn, ep);
3031                 else
3032                         qed_iwarp_tcp_connect_unsuccessful(p_hwfn, ep,
3033                                                            fw_return_code);
3034         }
3035 }
3036
3037 static inline bool
3038 qed_iwarp_check_ep_ok(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
3039 {
3040         if (!ep || (ep->sig != QED_EP_SIG)) {
3041                 DP_ERR(p_hwfn, "ERROR ON ASYNC ep=%p\n", ep);
3042                 return false;
3043         }
3044
3045         return true;
3046 }
3047
3048 static int qed_iwarp_async_event(struct qed_hwfn *p_hwfn, u8 fw_event_code,
3049                                  __le16 echo, union event_ring_data *data,
3050                                  u8 fw_return_code)
3051 {
3052         struct qed_rdma_events events = p_hwfn->p_rdma_info->events;
3053         struct regpair *fw_handle = &data->rdma_data.async_handle;
3054         struct qed_iwarp_ep *ep = NULL;
3055         u16 srq_offset;
3056         u16 srq_id;
3057         u16 cid;
3058
3059         ep = (struct qed_iwarp_ep *)(uintptr_t)HILO_64(fw_handle->hi,
3060                                                        fw_handle->lo);
3061
3062         switch (fw_event_code) {
3063         case IWARP_EVENT_TYPE_ASYNC_CONNECT_COMPLETE:
3064                 /* Async completion after TCP 3-way handshake */
3065                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3066                         return -EINVAL;
3067                 DP_VERBOSE(p_hwfn,
3068                            QED_MSG_RDMA,
3069                            "EP(0x%x) IWARP_EVENT_TYPE_ASYNC_CONNECT_COMPLETE fw_ret_code=%d\n",
3070                            ep->tcp_cid, fw_return_code);
3071                 qed_iwarp_connect_complete(p_hwfn, ep, fw_return_code);
3072                 break;
3073         case IWARP_EVENT_TYPE_ASYNC_EXCEPTION_DETECTED:
3074                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3075                         return -EINVAL;
3076                 DP_VERBOSE(p_hwfn,
3077                            QED_MSG_RDMA,
3078                            "QP(0x%x) IWARP_EVENT_TYPE_ASYNC_EXCEPTION_DETECTED fw_ret_code=%d\n",
3079                            ep->cid, fw_return_code);
3080                 qed_iwarp_exception_received(p_hwfn, ep, fw_return_code);
3081                 break;
3082         case IWARP_EVENT_TYPE_ASYNC_QP_IN_ERROR_STATE:
3083                 /* Async completion for Close Connection ramrod */
3084                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3085                         return -EINVAL;
3086                 DP_VERBOSE(p_hwfn,
3087                            QED_MSG_RDMA,
3088                            "QP(0x%x) IWARP_EVENT_TYPE_ASYNC_QP_IN_ERROR_STATE fw_ret_code=%d\n",
3089                            ep->cid, fw_return_code);
3090                 qed_iwarp_qp_in_error(p_hwfn, ep, fw_return_code);
3091                 break;
3092         case IWARP_EVENT_TYPE_ASYNC_ENHANCED_MPA_REPLY_ARRIVED:
3093                 /* Async event for active side only */
3094                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3095                         return -EINVAL;
3096                 DP_VERBOSE(p_hwfn,
3097                            QED_MSG_RDMA,
3098                            "QP(0x%x) IWARP_EVENT_TYPE_ASYNC_MPA_HANDSHAKE_MPA_REPLY_ARRIVED fw_ret_code=%d\n",
3099                            ep->cid, fw_return_code);
3100                 qed_iwarp_mpa_reply_arrived(p_hwfn, ep);
3101                 break;
3102         case IWARP_EVENT_TYPE_ASYNC_MPA_HANDSHAKE_COMPLETE:
3103                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3104                         return -EINVAL;
3105                 DP_VERBOSE(p_hwfn,
3106                            QED_MSG_RDMA,
3107                            "QP(0x%x) IWARP_EVENT_TYPE_ASYNC_MPA_HANDSHAKE_COMPLETE fw_ret_code=%d\n",
3108                            ep->cid, fw_return_code);
3109                 qed_iwarp_mpa_complete(p_hwfn, ep, fw_return_code);
3110                 break;
3111         case IWARP_EVENT_TYPE_ASYNC_CID_CLEANED:
3112                 cid = (u16)le32_to_cpu(fw_handle->lo);
3113                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
3114                            "(0x%x)IWARP_EVENT_TYPE_ASYNC_CID_CLEANED\n", cid);
3115                 qed_iwarp_cid_cleaned(p_hwfn, cid);
3116
3117                 break;
3118         case IWARP_EVENT_TYPE_ASYNC_SRQ_EMPTY:
3119                 DP_NOTICE(p_hwfn, "IWARP_EVENT_TYPE_ASYNC_SRQ_EMPTY\n");
3120                 srq_offset = p_hwfn->p_rdma_info->srq_id_offset;
3121                 /* FW assigns value that is no greater than u16 */
3122                 srq_id = ((u16)le32_to_cpu(fw_handle->lo)) - srq_offset;
3123                 events.affiliated_event(events.context,
3124                                         QED_IWARP_EVENT_SRQ_EMPTY,
3125                                         &srq_id);
3126                 break;
3127         case IWARP_EVENT_TYPE_ASYNC_SRQ_LIMIT:
3128                 DP_NOTICE(p_hwfn, "IWARP_EVENT_TYPE_ASYNC_SRQ_LIMIT\n");
3129                 srq_offset = p_hwfn->p_rdma_info->srq_id_offset;
3130                 /* FW assigns value that is no greater than u16 */
3131                 srq_id = ((u16)le32_to_cpu(fw_handle->lo)) - srq_offset;
3132                 events.affiliated_event(events.context,
3133                                         QED_IWARP_EVENT_SRQ_LIMIT,
3134                                         &srq_id);
3135                 break;
3136         case IWARP_EVENT_TYPE_ASYNC_CQ_OVERFLOW:
3137                 DP_NOTICE(p_hwfn, "IWARP_EVENT_TYPE_ASYNC_CQ_OVERFLOW\n");
3138
3139                 p_hwfn->p_rdma_info->events.affiliated_event(
3140                         p_hwfn->p_rdma_info->events.context,
3141                         QED_IWARP_EVENT_CQ_OVERFLOW,
3142                         (void *)fw_handle);
3143                 break;
3144         default:
3145                 DP_ERR(p_hwfn, "Received unexpected async iwarp event %d\n",
3146                        fw_event_code);
3147                 return -EINVAL;
3148         }
3149         return 0;
3150 }
3151
3152 int
3153 qed_iwarp_create_listen(void *rdma_cxt,
3154                         struct qed_iwarp_listen_in *iparams,
3155                         struct qed_iwarp_listen_out *oparams)
3156 {
3157         struct qed_hwfn *p_hwfn = rdma_cxt;
3158         struct qed_iwarp_listener *listener;
3159
3160         listener = kzalloc(sizeof(*listener), GFP_KERNEL);
3161         if (!listener)
3162                 return -ENOMEM;
3163
3164         listener->ip_version = iparams->ip_version;
3165         memcpy(listener->ip_addr, iparams->ip_addr, sizeof(listener->ip_addr));
3166         listener->port = iparams->port;
3167         listener->vlan = iparams->vlan;
3168
3169         listener->event_cb = iparams->event_cb;
3170         listener->cb_context = iparams->cb_context;
3171         listener->max_backlog = iparams->max_backlog;
3172         oparams->handle = listener;
3173
3174         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3175         list_add_tail(&listener->list_entry,
3176                       &p_hwfn->p_rdma_info->iwarp.listen_list);
3177         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3178
3179         DP_VERBOSE(p_hwfn,
3180                    QED_MSG_RDMA,
3181                    "callback=%p handle=%p ip=%x:%x:%x:%x port=0x%x vlan=0x%x\n",
3182                    listener->event_cb,
3183                    listener,
3184                    listener->ip_addr[0],
3185                    listener->ip_addr[1],
3186                    listener->ip_addr[2],
3187                    listener->ip_addr[3], listener->port, listener->vlan);
3188
3189         return 0;
3190 }
3191
3192 int qed_iwarp_destroy_listen(void *rdma_cxt, void *handle)
3193 {
3194         struct qed_iwarp_listener *listener = handle;
3195         struct qed_hwfn *p_hwfn = rdma_cxt;
3196
3197         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "handle=%p\n", handle);
3198
3199         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3200         list_del(&listener->list_entry);
3201         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3202
3203         kfree(listener);
3204
3205         return 0;
3206 }
3207
3208 int qed_iwarp_send_rtr(void *rdma_cxt, struct qed_iwarp_send_rtr_in *iparams)
3209 {
3210         struct qed_hwfn *p_hwfn = rdma_cxt;
3211         struct qed_sp_init_data init_data;
3212         struct qed_spq_entry *p_ent;
3213         struct qed_iwarp_ep *ep;
3214         struct qed_rdma_qp *qp;
3215         int rc;
3216
3217         ep = iparams->ep_context;
3218         if (!ep) {
3219                 DP_ERR(p_hwfn, "Ep Context receive in send_rtr is NULL\n");
3220                 return -EINVAL;
3221         }
3222
3223         qp = ep->qp;
3224
3225         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) EP(0x%x)\n",
3226                    qp->icid, ep->tcp_cid);
3227
3228         memset(&init_data, 0, sizeof(init_data));
3229         init_data.cid = qp->icid;
3230         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
3231         init_data.comp_mode = QED_SPQ_MODE_CB;
3232
3233         rc = qed_sp_init_request(p_hwfn, &p_ent,
3234                                  IWARP_RAMROD_CMD_ID_MPA_OFFLOAD_SEND_RTR,
3235                                  PROTOCOLID_IWARP, &init_data);
3236
3237         if (rc)
3238                 return rc;
3239
3240         rc = qed_spq_post(p_hwfn, p_ent, NULL);
3241
3242         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "rc = 0x%x\n", rc);
3243
3244         return rc;
3245 }
3246
3247 void
3248 qed_iwarp_query_qp(struct qed_rdma_qp *qp,
3249                    struct qed_rdma_query_qp_out_params *out_params)
3250 {
3251         out_params->state = qed_iwarp2roce_state(qp->iwarp_state);
3252 }