Merge git://git.kernel.org/pub/scm/linux/kernel/git/kvalo/wireless-drivers.git
[platform/kernel/linux-starfive.git] / drivers / net / ethernet / qlogic / qed / qed_iwarp.c
1 // SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause)
2 /* QLogic qed NIC Driver
3  * Copyright (c) 2015-2017  QLogic Corporation
4  * Copyright (c) 2019-2020 Marvell International Ltd.
5  */
6
7 #include <linux/if_ether.h>
8 #include <linux/if_vlan.h>
9 #include <linux/ip.h>
10 #include <linux/ipv6.h>
11 #include <linux/spinlock.h>
12 #include <linux/tcp.h>
13 #include "qed_cxt.h"
14 #include "qed_hw.h"
15 #include "qed_ll2.h"
16 #include "qed_rdma.h"
17 #include "qed_reg_addr.h"
18 #include "qed_sp.h"
19 #include "qed_ooo.h"
20
21 #define QED_IWARP_ORD_DEFAULT           32
22 #define QED_IWARP_IRD_DEFAULT           32
23 #define QED_IWARP_MAX_FW_MSS            4120
24
25 #define QED_EP_SIG 0xecabcdef
26
27 struct mpa_v2_hdr {
28         __be16 ird;
29         __be16 ord;
30 };
31
32 #define MPA_V2_PEER2PEER_MODEL  0x8000
33 #define MPA_V2_SEND_RTR         0x4000  /* on ird */
34 #define MPA_V2_READ_RTR         0x4000  /* on ord */
35 #define MPA_V2_WRITE_RTR        0x8000
36 #define MPA_V2_IRD_ORD_MASK     0x3FFF
37
38 #define MPA_REV2(_mpa_rev) ((_mpa_rev) == MPA_NEGOTIATION_TYPE_ENHANCED)
39
40 #define QED_IWARP_INVALID_TCP_CID       0xffffffff
41
42 #define QED_IWARP_RCV_WND_SIZE_DEF_BB_2P (200 * 1024)
43 #define QED_IWARP_RCV_WND_SIZE_DEF_BB_4P (100 * 1024)
44 #define QED_IWARP_RCV_WND_SIZE_DEF_AH_2P (150 * 1024)
45 #define QED_IWARP_RCV_WND_SIZE_DEF_AH_4P (90 * 1024)
46
47 #define QED_IWARP_RCV_WND_SIZE_MIN      (0xffff)
48 #define TIMESTAMP_HEADER_SIZE           (12)
49 #define QED_IWARP_MAX_FIN_RT_DEFAULT    (2)
50
51 #define QED_IWARP_TS_EN                 BIT(0)
52 #define QED_IWARP_DA_EN                 BIT(1)
53 #define QED_IWARP_PARAM_CRC_NEEDED      (1)
54 #define QED_IWARP_PARAM_P2P             (1)
55
56 #define QED_IWARP_DEF_MAX_RT_TIME       (0)
57 #define QED_IWARP_DEF_CWND_FACTOR       (4)
58 #define QED_IWARP_DEF_KA_MAX_PROBE_CNT  (5)
59 #define QED_IWARP_DEF_KA_TIMEOUT        (1200000)       /* 20 min */
60 #define QED_IWARP_DEF_KA_INTERVAL       (1000)          /* 1 sec */
61
62 static int qed_iwarp_async_event(struct qed_hwfn *p_hwfn, u8 fw_event_code,
63                                  __le16 echo, union event_ring_data *data,
64                                  u8 fw_return_code);
65
66 /* Override devinfo with iWARP specific values */
67 void qed_iwarp_init_devinfo(struct qed_hwfn *p_hwfn)
68 {
69         struct qed_rdma_device *dev = p_hwfn->p_rdma_info->dev;
70
71         dev->max_inline = IWARP_REQ_MAX_INLINE_DATA_SIZE;
72         dev->max_qp = min_t(u32,
73                             IWARP_MAX_QPS,
74                             p_hwfn->p_rdma_info->num_qps) -
75                       QED_IWARP_PREALLOC_CNT;
76
77         dev->max_cq = dev->max_qp;
78
79         dev->max_qp_resp_rd_atomic_resc = QED_IWARP_IRD_DEFAULT;
80         dev->max_qp_req_rd_atomic_resc = QED_IWARP_ORD_DEFAULT;
81 }
82
83 void qed_iwarp_init_hw(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt)
84 {
85         p_hwfn->rdma_prs_search_reg = PRS_REG_SEARCH_TCP;
86         qed_wr(p_hwfn, p_ptt, p_hwfn->rdma_prs_search_reg, 1);
87         p_hwfn->b_rdma_enabled_in_prs = true;
88 }
89
90 /* We have two cid maps, one for tcp which should be used only from passive
91  * syn processing and replacing a pre-allocated ep in the list. The second
92  * for active tcp and for QPs.
93  */
94 static void qed_iwarp_cid_cleaned(struct qed_hwfn *p_hwfn, u32 cid)
95 {
96         cid -= qed_cxt_get_proto_cid_start(p_hwfn, p_hwfn->p_rdma_info->proto);
97
98         spin_lock_bh(&p_hwfn->p_rdma_info->lock);
99
100         if (cid < QED_IWARP_PREALLOC_CNT)
101                 qed_bmap_release_id(p_hwfn, &p_hwfn->p_rdma_info->tcp_cid_map,
102                                     cid);
103         else
104                 qed_bmap_release_id(p_hwfn, &p_hwfn->p_rdma_info->cid_map, cid);
105
106         spin_unlock_bh(&p_hwfn->p_rdma_info->lock);
107 }
108
109 void
110 qed_iwarp_init_fw_ramrod(struct qed_hwfn *p_hwfn,
111                          struct iwarp_init_func_ramrod_data *p_ramrod)
112 {
113         p_ramrod->iwarp.ll2_ooo_q_index =
114             RESC_START(p_hwfn, QED_LL2_RAM_QUEUE) +
115             p_hwfn->p_rdma_info->iwarp.ll2_ooo_handle;
116
117         p_ramrod->tcp.max_fin_rt = QED_IWARP_MAX_FIN_RT_DEFAULT;
118
119         return;
120 }
121
122 static int qed_iwarp_alloc_cid(struct qed_hwfn *p_hwfn, u32 *cid)
123 {
124         int rc;
125
126         spin_lock_bh(&p_hwfn->p_rdma_info->lock);
127         rc = qed_rdma_bmap_alloc_id(p_hwfn, &p_hwfn->p_rdma_info->cid_map, cid);
128         spin_unlock_bh(&p_hwfn->p_rdma_info->lock);
129         if (rc) {
130                 DP_NOTICE(p_hwfn, "Failed in allocating iwarp cid\n");
131                 return rc;
132         }
133         *cid += qed_cxt_get_proto_cid_start(p_hwfn, p_hwfn->p_rdma_info->proto);
134
135         rc = qed_cxt_dynamic_ilt_alloc(p_hwfn, QED_ELEM_CXT, *cid);
136         if (rc)
137                 qed_iwarp_cid_cleaned(p_hwfn, *cid);
138
139         return rc;
140 }
141
142 static void qed_iwarp_set_tcp_cid(struct qed_hwfn *p_hwfn, u32 cid)
143 {
144         cid -= qed_cxt_get_proto_cid_start(p_hwfn, p_hwfn->p_rdma_info->proto);
145
146         spin_lock_bh(&p_hwfn->p_rdma_info->lock);
147         qed_bmap_set_id(p_hwfn, &p_hwfn->p_rdma_info->tcp_cid_map, cid);
148         spin_unlock_bh(&p_hwfn->p_rdma_info->lock);
149 }
150
151 /* This function allocates a cid for passive tcp (called from syn receive)
152  * the reason it's separate from the regular cid allocation is because it
153  * is assured that these cids already have ilt allocated. They are preallocated
154  * to ensure that we won't need to allocate memory during syn processing
155  */
156 static int qed_iwarp_alloc_tcp_cid(struct qed_hwfn *p_hwfn, u32 *cid)
157 {
158         int rc;
159
160         spin_lock_bh(&p_hwfn->p_rdma_info->lock);
161
162         rc = qed_rdma_bmap_alloc_id(p_hwfn,
163                                     &p_hwfn->p_rdma_info->tcp_cid_map, cid);
164
165         spin_unlock_bh(&p_hwfn->p_rdma_info->lock);
166
167         if (rc) {
168                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
169                            "can't allocate iwarp tcp cid max-count=%d\n",
170                            p_hwfn->p_rdma_info->tcp_cid_map.max_count);
171
172                 *cid = QED_IWARP_INVALID_TCP_CID;
173                 return rc;
174         }
175
176         *cid += qed_cxt_get_proto_cid_start(p_hwfn,
177                                             p_hwfn->p_rdma_info->proto);
178         return 0;
179 }
180
181 int qed_iwarp_create_qp(struct qed_hwfn *p_hwfn,
182                         struct qed_rdma_qp *qp,
183                         struct qed_rdma_create_qp_out_params *out_params)
184 {
185         struct iwarp_create_qp_ramrod_data *p_ramrod;
186         struct qed_sp_init_data init_data;
187         struct qed_spq_entry *p_ent;
188         u16 physical_queue;
189         u32 cid;
190         int rc;
191
192         qp->shared_queue = dma_alloc_coherent(&p_hwfn->cdev->pdev->dev,
193                                               IWARP_SHARED_QUEUE_PAGE_SIZE,
194                                               &qp->shared_queue_phys_addr,
195                                               GFP_KERNEL);
196         if (!qp->shared_queue)
197                 return -ENOMEM;
198
199         out_params->sq_pbl_virt = (u8 *)qp->shared_queue +
200             IWARP_SHARED_QUEUE_PAGE_SQ_PBL_OFFSET;
201         out_params->sq_pbl_phys = qp->shared_queue_phys_addr +
202             IWARP_SHARED_QUEUE_PAGE_SQ_PBL_OFFSET;
203         out_params->rq_pbl_virt = (u8 *)qp->shared_queue +
204             IWARP_SHARED_QUEUE_PAGE_RQ_PBL_OFFSET;
205         out_params->rq_pbl_phys = qp->shared_queue_phys_addr +
206             IWARP_SHARED_QUEUE_PAGE_RQ_PBL_OFFSET;
207
208         rc = qed_iwarp_alloc_cid(p_hwfn, &cid);
209         if (rc)
210                 goto err1;
211
212         qp->icid = (u16)cid;
213
214         memset(&init_data, 0, sizeof(init_data));
215         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
216         init_data.cid = qp->icid;
217         init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
218
219         rc = qed_sp_init_request(p_hwfn, &p_ent,
220                                  IWARP_RAMROD_CMD_ID_CREATE_QP,
221                                  PROTOCOLID_IWARP, &init_data);
222         if (rc)
223                 goto err2;
224
225         p_ramrod = &p_ent->ramrod.iwarp_create_qp;
226
227         SET_FIELD(p_ramrod->flags,
228                   IWARP_CREATE_QP_RAMROD_DATA_FMR_AND_RESERVED_EN,
229                   qp->fmr_and_reserved_lkey);
230
231         SET_FIELD(p_ramrod->flags,
232                   IWARP_CREATE_QP_RAMROD_DATA_SIGNALED_COMP, qp->signal_all);
233
234         SET_FIELD(p_ramrod->flags,
235                   IWARP_CREATE_QP_RAMROD_DATA_RDMA_RD_EN,
236                   qp->incoming_rdma_read_en);
237
238         SET_FIELD(p_ramrod->flags,
239                   IWARP_CREATE_QP_RAMROD_DATA_RDMA_WR_EN,
240                   qp->incoming_rdma_write_en);
241
242         SET_FIELD(p_ramrod->flags,
243                   IWARP_CREATE_QP_RAMROD_DATA_ATOMIC_EN,
244                   qp->incoming_atomic_en);
245
246         SET_FIELD(p_ramrod->flags,
247                   IWARP_CREATE_QP_RAMROD_DATA_SRQ_FLG, qp->use_srq);
248
249         p_ramrod->pd = cpu_to_le16(qp->pd);
250         p_ramrod->sq_num_pages = cpu_to_le16(qp->sq_num_pages);
251         p_ramrod->rq_num_pages = cpu_to_le16(qp->rq_num_pages);
252
253         p_ramrod->srq_id.srq_idx = cpu_to_le16(qp->srq_id);
254         p_ramrod->srq_id.opaque_fid = cpu_to_le16(p_hwfn->hw_info.opaque_fid);
255         p_ramrod->qp_handle_for_cqe.hi = qp->qp_handle.hi;
256         p_ramrod->qp_handle_for_cqe.lo = qp->qp_handle.lo;
257
258         p_ramrod->cq_cid_for_sq =
259             cpu_to_le32((p_hwfn->hw_info.opaque_fid << 16) | qp->sq_cq_id);
260         p_ramrod->cq_cid_for_rq =
261             cpu_to_le32((p_hwfn->hw_info.opaque_fid << 16) | qp->rq_cq_id);
262
263         p_ramrod->dpi = cpu_to_le16(qp->dpi);
264
265         physical_queue = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_OFLD);
266         p_ramrod->physical_q0 = cpu_to_le16(physical_queue);
267         physical_queue = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_ACK);
268         p_ramrod->physical_q1 = cpu_to_le16(physical_queue);
269
270         rc = qed_spq_post(p_hwfn, p_ent, NULL);
271         if (rc)
272                 goto err2;
273
274         return rc;
275
276 err2:
277         qed_iwarp_cid_cleaned(p_hwfn, cid);
278 err1:
279         dma_free_coherent(&p_hwfn->cdev->pdev->dev,
280                           IWARP_SHARED_QUEUE_PAGE_SIZE,
281                           qp->shared_queue, qp->shared_queue_phys_addr);
282
283         return rc;
284 }
285
286 static int qed_iwarp_modify_fw(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp)
287 {
288         struct iwarp_modify_qp_ramrod_data *p_ramrod;
289         struct qed_sp_init_data init_data;
290         struct qed_spq_entry *p_ent;
291         u16 flags, trans_to_state;
292         int rc;
293
294         /* Get SPQ entry */
295         memset(&init_data, 0, sizeof(init_data));
296         init_data.cid = qp->icid;
297         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
298         init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
299
300         rc = qed_sp_init_request(p_hwfn, &p_ent,
301                                  IWARP_RAMROD_CMD_ID_MODIFY_QP,
302                                  p_hwfn->p_rdma_info->proto, &init_data);
303         if (rc)
304                 return rc;
305
306         p_ramrod = &p_ent->ramrod.iwarp_modify_qp;
307
308         flags = le16_to_cpu(p_ramrod->flags);
309         SET_FIELD(flags, IWARP_MODIFY_QP_RAMROD_DATA_STATE_TRANS_EN, 0x1);
310         p_ramrod->flags = cpu_to_le16(flags);
311
312         if (qp->iwarp_state == QED_IWARP_QP_STATE_CLOSING)
313                 trans_to_state = IWARP_MODIFY_QP_STATE_CLOSING;
314         else
315                 trans_to_state = IWARP_MODIFY_QP_STATE_ERROR;
316
317         p_ramrod->transition_to_state = cpu_to_le16(trans_to_state);
318
319         rc = qed_spq_post(p_hwfn, p_ent, NULL);
320
321         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x)rc=%d\n", qp->icid, rc);
322
323         return rc;
324 }
325
326 enum qed_iwarp_qp_state qed_roce2iwarp_state(enum qed_roce_qp_state state)
327 {
328         switch (state) {
329         case QED_ROCE_QP_STATE_RESET:
330         case QED_ROCE_QP_STATE_INIT:
331         case QED_ROCE_QP_STATE_RTR:
332                 return QED_IWARP_QP_STATE_IDLE;
333         case QED_ROCE_QP_STATE_RTS:
334                 return QED_IWARP_QP_STATE_RTS;
335         case QED_ROCE_QP_STATE_SQD:
336                 return QED_IWARP_QP_STATE_CLOSING;
337         case QED_ROCE_QP_STATE_ERR:
338                 return QED_IWARP_QP_STATE_ERROR;
339         case QED_ROCE_QP_STATE_SQE:
340                 return QED_IWARP_QP_STATE_TERMINATE;
341         default:
342                 return QED_IWARP_QP_STATE_ERROR;
343         }
344 }
345
346 static enum qed_roce_qp_state
347 qed_iwarp2roce_state(enum qed_iwarp_qp_state state)
348 {
349         switch (state) {
350         case QED_IWARP_QP_STATE_IDLE:
351                 return QED_ROCE_QP_STATE_INIT;
352         case QED_IWARP_QP_STATE_RTS:
353                 return QED_ROCE_QP_STATE_RTS;
354         case QED_IWARP_QP_STATE_TERMINATE:
355                 return QED_ROCE_QP_STATE_SQE;
356         case QED_IWARP_QP_STATE_CLOSING:
357                 return QED_ROCE_QP_STATE_SQD;
358         case QED_IWARP_QP_STATE_ERROR:
359                 return QED_ROCE_QP_STATE_ERR;
360         default:
361                 return QED_ROCE_QP_STATE_ERR;
362         }
363 }
364
365 static const char * const iwarp_state_names[] = {
366         "IDLE",
367         "RTS",
368         "TERMINATE",
369         "CLOSING",
370         "ERROR",
371 };
372
373 int
374 qed_iwarp_modify_qp(struct qed_hwfn *p_hwfn,
375                     struct qed_rdma_qp *qp,
376                     enum qed_iwarp_qp_state new_state, bool internal)
377 {
378         enum qed_iwarp_qp_state prev_iw_state;
379         bool modify_fw = false;
380         int rc = 0;
381
382         /* modify QP can be called from upper-layer or as a result of async
383          * RST/FIN... therefore need to protect
384          */
385         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.qp_lock);
386         prev_iw_state = qp->iwarp_state;
387
388         if (prev_iw_state == new_state) {
389                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.qp_lock);
390                 return 0;
391         }
392
393         switch (prev_iw_state) {
394         case QED_IWARP_QP_STATE_IDLE:
395                 switch (new_state) {
396                 case QED_IWARP_QP_STATE_RTS:
397                         qp->iwarp_state = QED_IWARP_QP_STATE_RTS;
398                         break;
399                 case QED_IWARP_QP_STATE_ERROR:
400                         qp->iwarp_state = QED_IWARP_QP_STATE_ERROR;
401                         if (!internal)
402                                 modify_fw = true;
403                         break;
404                 default:
405                         break;
406                 }
407                 break;
408         case QED_IWARP_QP_STATE_RTS:
409                 switch (new_state) {
410                 case QED_IWARP_QP_STATE_CLOSING:
411                         if (!internal)
412                                 modify_fw = true;
413
414                         qp->iwarp_state = QED_IWARP_QP_STATE_CLOSING;
415                         break;
416                 case QED_IWARP_QP_STATE_ERROR:
417                         if (!internal)
418                                 modify_fw = true;
419                         qp->iwarp_state = QED_IWARP_QP_STATE_ERROR;
420                         break;
421                 default:
422                         break;
423                 }
424                 break;
425         case QED_IWARP_QP_STATE_ERROR:
426                 switch (new_state) {
427                 case QED_IWARP_QP_STATE_IDLE:
428
429                         qp->iwarp_state = new_state;
430                         break;
431                 case QED_IWARP_QP_STATE_CLOSING:
432                         /* could happen due to race... do nothing.... */
433                         break;
434                 default:
435                         rc = -EINVAL;
436                 }
437                 break;
438         case QED_IWARP_QP_STATE_TERMINATE:
439         case QED_IWARP_QP_STATE_CLOSING:
440                 qp->iwarp_state = new_state;
441                 break;
442         default:
443                 break;
444         }
445
446         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) %s --> %s%s\n",
447                    qp->icid,
448                    iwarp_state_names[prev_iw_state],
449                    iwarp_state_names[qp->iwarp_state],
450                    internal ? "internal" : "");
451
452         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.qp_lock);
453
454         if (modify_fw)
455                 rc = qed_iwarp_modify_fw(p_hwfn, qp);
456
457         return rc;
458 }
459
460 int qed_iwarp_fw_destroy(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp)
461 {
462         struct qed_sp_init_data init_data;
463         struct qed_spq_entry *p_ent;
464         int rc;
465
466         /* Get SPQ entry */
467         memset(&init_data, 0, sizeof(init_data));
468         init_data.cid = qp->icid;
469         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
470         init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
471
472         rc = qed_sp_init_request(p_hwfn, &p_ent,
473                                  IWARP_RAMROD_CMD_ID_DESTROY_QP,
474                                  p_hwfn->p_rdma_info->proto, &init_data);
475         if (rc)
476                 return rc;
477
478         rc = qed_spq_post(p_hwfn, p_ent, NULL);
479
480         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) rc = %d\n", qp->icid, rc);
481
482         return rc;
483 }
484
485 static void qed_iwarp_destroy_ep(struct qed_hwfn *p_hwfn,
486                                  struct qed_iwarp_ep *ep,
487                                  bool remove_from_active_list)
488 {
489         dma_free_coherent(&p_hwfn->cdev->pdev->dev,
490                           sizeof(*ep->ep_buffer_virt),
491                           ep->ep_buffer_virt, ep->ep_buffer_phys);
492
493         if (remove_from_active_list) {
494                 spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
495                 list_del(&ep->list_entry);
496                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
497         }
498
499         if (ep->qp)
500                 ep->qp->ep = NULL;
501
502         kfree(ep);
503 }
504
505 int qed_iwarp_destroy_qp(struct qed_hwfn *p_hwfn, struct qed_rdma_qp *qp)
506 {
507         struct qed_iwarp_ep *ep = qp->ep;
508         int wait_count = 0;
509         int rc = 0;
510
511         if (qp->iwarp_state != QED_IWARP_QP_STATE_ERROR) {
512                 rc = qed_iwarp_modify_qp(p_hwfn, qp,
513                                          QED_IWARP_QP_STATE_ERROR, false);
514                 if (rc)
515                         return rc;
516         }
517
518         /* Make sure ep is closed before returning and freeing memory. */
519         if (ep) {
520                 while (READ_ONCE(ep->state) != QED_IWARP_EP_CLOSED &&
521                        wait_count++ < 200)
522                         msleep(100);
523
524                 if (ep->state != QED_IWARP_EP_CLOSED)
525                         DP_NOTICE(p_hwfn, "ep state close timeout state=%x\n",
526                                   ep->state);
527
528                 qed_iwarp_destroy_ep(p_hwfn, ep, false);
529         }
530
531         rc = qed_iwarp_fw_destroy(p_hwfn, qp);
532
533         if (qp->shared_queue)
534                 dma_free_coherent(&p_hwfn->cdev->pdev->dev,
535                                   IWARP_SHARED_QUEUE_PAGE_SIZE,
536                                   qp->shared_queue, qp->shared_queue_phys_addr);
537
538         return rc;
539 }
540
541 static int
542 qed_iwarp_create_ep(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep **ep_out)
543 {
544         struct qed_iwarp_ep *ep;
545         int rc;
546
547         ep = kzalloc(sizeof(*ep), GFP_KERNEL);
548         if (!ep)
549                 return -ENOMEM;
550
551         ep->state = QED_IWARP_EP_INIT;
552
553         ep->ep_buffer_virt = dma_alloc_coherent(&p_hwfn->cdev->pdev->dev,
554                                                 sizeof(*ep->ep_buffer_virt),
555                                                 &ep->ep_buffer_phys,
556                                                 GFP_KERNEL);
557         if (!ep->ep_buffer_virt) {
558                 rc = -ENOMEM;
559                 goto err;
560         }
561
562         ep->sig = QED_EP_SIG;
563
564         *ep_out = ep;
565
566         return 0;
567
568 err:
569         kfree(ep);
570         return rc;
571 }
572
573 static void
574 qed_iwarp_print_tcp_ramrod(struct qed_hwfn *p_hwfn,
575                            struct iwarp_tcp_offload_ramrod_data *p_tcp_ramrod)
576 {
577         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "local_mac=%x %x %x, remote_mac=%x %x %x\n",
578                    p_tcp_ramrod->tcp.local_mac_addr_lo,
579                    p_tcp_ramrod->tcp.local_mac_addr_mid,
580                    p_tcp_ramrod->tcp.local_mac_addr_hi,
581                    p_tcp_ramrod->tcp.remote_mac_addr_lo,
582                    p_tcp_ramrod->tcp.remote_mac_addr_mid,
583                    p_tcp_ramrod->tcp.remote_mac_addr_hi);
584
585         if (p_tcp_ramrod->tcp.ip_version == TCP_IPV4) {
586                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
587                            "local_ip=%pI4h:%x, remote_ip=%pI4h:%x, vlan=%x\n",
588                            p_tcp_ramrod->tcp.local_ip,
589                            p_tcp_ramrod->tcp.local_port,
590                            p_tcp_ramrod->tcp.remote_ip,
591                            p_tcp_ramrod->tcp.remote_port,
592                            p_tcp_ramrod->tcp.vlan_id);
593         } else {
594                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
595                            "local_ip=%pI6:%x, remote_ip=%pI6:%x, vlan=%x\n",
596                            p_tcp_ramrod->tcp.local_ip,
597                            p_tcp_ramrod->tcp.local_port,
598                            p_tcp_ramrod->tcp.remote_ip,
599                            p_tcp_ramrod->tcp.remote_port,
600                            p_tcp_ramrod->tcp.vlan_id);
601         }
602
603         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
604                    "flow_label=%x, ttl=%x, tos_or_tc=%x, mss=%x, rcv_wnd_scale=%x, connect_mode=%x, flags=%x\n",
605                    p_tcp_ramrod->tcp.flow_label,
606                    p_tcp_ramrod->tcp.ttl,
607                    p_tcp_ramrod->tcp.tos_or_tc,
608                    p_tcp_ramrod->tcp.mss,
609                    p_tcp_ramrod->tcp.rcv_wnd_scale,
610                    p_tcp_ramrod->tcp.connect_mode,
611                    p_tcp_ramrod->tcp.flags);
612
613         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "syn_ip_payload_length=%x, lo=%x, hi=%x\n",
614                    p_tcp_ramrod->tcp.syn_ip_payload_length,
615                    p_tcp_ramrod->tcp.syn_phy_addr_lo,
616                    p_tcp_ramrod->tcp.syn_phy_addr_hi);
617 }
618
619 static int
620 qed_iwarp_tcp_offload(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
621 {
622         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
623         struct iwarp_tcp_offload_ramrod_data *p_tcp_ramrod;
624         struct tcp_offload_params_opt2 *tcp;
625         struct qed_sp_init_data init_data;
626         struct qed_spq_entry *p_ent;
627         dma_addr_t async_output_phys;
628         dma_addr_t in_pdata_phys;
629         u16 physical_q;
630         u16 flags = 0;
631         u8 tcp_flags;
632         int rc;
633         int i;
634
635         memset(&init_data, 0, sizeof(init_data));
636         init_data.cid = ep->tcp_cid;
637         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
638         if (ep->connect_mode == TCP_CONNECT_PASSIVE)
639                 init_data.comp_mode = QED_SPQ_MODE_CB;
640         else
641                 init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
642
643         rc = qed_sp_init_request(p_hwfn, &p_ent,
644                                  IWARP_RAMROD_CMD_ID_TCP_OFFLOAD,
645                                  PROTOCOLID_IWARP, &init_data);
646         if (rc)
647                 return rc;
648
649         p_tcp_ramrod = &p_ent->ramrod.iwarp_tcp_offload;
650
651         in_pdata_phys = ep->ep_buffer_phys +
652                         offsetof(struct qed_iwarp_ep_memory, in_pdata);
653         DMA_REGPAIR_LE(p_tcp_ramrod->iwarp.incoming_ulp_buffer.addr,
654                        in_pdata_phys);
655
656         p_tcp_ramrod->iwarp.incoming_ulp_buffer.len =
657             cpu_to_le16(sizeof(ep->ep_buffer_virt->in_pdata));
658
659         async_output_phys = ep->ep_buffer_phys +
660                             offsetof(struct qed_iwarp_ep_memory, async_output);
661         DMA_REGPAIR_LE(p_tcp_ramrod->iwarp.async_eqe_output_buf,
662                        async_output_phys);
663
664         p_tcp_ramrod->iwarp.handle_for_async.hi = cpu_to_le32(PTR_HI(ep));
665         p_tcp_ramrod->iwarp.handle_for_async.lo = cpu_to_le32(PTR_LO(ep));
666
667         physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_OFLD);
668         p_tcp_ramrod->iwarp.physical_q0 = cpu_to_le16(physical_q);
669         physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_ACK);
670         p_tcp_ramrod->iwarp.physical_q1 = cpu_to_le16(physical_q);
671         p_tcp_ramrod->iwarp.mpa_mode = iwarp_info->mpa_rev;
672
673         tcp = &p_tcp_ramrod->tcp;
674         qed_set_fw_mac_addr(&tcp->remote_mac_addr_hi,
675                             &tcp->remote_mac_addr_mid,
676                             &tcp->remote_mac_addr_lo, ep->remote_mac_addr);
677         qed_set_fw_mac_addr(&tcp->local_mac_addr_hi, &tcp->local_mac_addr_mid,
678                             &tcp->local_mac_addr_lo, ep->local_mac_addr);
679
680         tcp->vlan_id = cpu_to_le16(ep->cm_info.vlan);
681
682         tcp_flags = p_hwfn->p_rdma_info->iwarp.tcp_flags;
683
684         SET_FIELD(flags, TCP_OFFLOAD_PARAMS_OPT2_TS_EN,
685                   !!(tcp_flags & QED_IWARP_TS_EN));
686
687         SET_FIELD(flags, TCP_OFFLOAD_PARAMS_OPT2_DA_EN,
688                   !!(tcp_flags & QED_IWARP_DA_EN));
689
690         tcp->flags = cpu_to_le16(flags);
691         tcp->ip_version = ep->cm_info.ip_version;
692
693         for (i = 0; i < 4; i++) {
694                 tcp->remote_ip[i] = cpu_to_le32(ep->cm_info.remote_ip[i]);
695                 tcp->local_ip[i] = cpu_to_le32(ep->cm_info.local_ip[i]);
696         }
697
698         tcp->remote_port = cpu_to_le16(ep->cm_info.remote_port);
699         tcp->local_port = cpu_to_le16(ep->cm_info.local_port);
700         tcp->mss = cpu_to_le16(ep->mss);
701         tcp->flow_label = 0;
702         tcp->ttl = 0x40;
703         tcp->tos_or_tc = 0;
704
705         tcp->max_rt_time = QED_IWARP_DEF_MAX_RT_TIME;
706         tcp->cwnd = cpu_to_le32(QED_IWARP_DEF_CWND_FACTOR * ep->mss);
707         tcp->ka_max_probe_cnt = QED_IWARP_DEF_KA_MAX_PROBE_CNT;
708         tcp->ka_timeout = cpu_to_le32(QED_IWARP_DEF_KA_TIMEOUT);
709         tcp->ka_interval = cpu_to_le32(QED_IWARP_DEF_KA_INTERVAL);
710
711         tcp->rcv_wnd_scale = (u8)p_hwfn->p_rdma_info->iwarp.rcv_wnd_scale;
712         tcp->connect_mode = ep->connect_mode;
713
714         if (ep->connect_mode == TCP_CONNECT_PASSIVE) {
715                 tcp->syn_ip_payload_length =
716                         cpu_to_le16(ep->syn_ip_payload_length);
717                 tcp->syn_phy_addr_hi = DMA_HI_LE(ep->syn_phy_addr);
718                 tcp->syn_phy_addr_lo = DMA_LO_LE(ep->syn_phy_addr);
719         }
720
721         qed_iwarp_print_tcp_ramrod(p_hwfn, p_tcp_ramrod);
722
723         rc = qed_spq_post(p_hwfn, p_ent, NULL);
724
725         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
726                    "EP(0x%x) Offload completed rc=%d\n", ep->tcp_cid, rc);
727
728         return rc;
729 }
730
731 static void
732 qed_iwarp_mpa_received(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
733 {
734         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
735         struct qed_iwarp_cm_event_params params;
736         struct mpa_v2_hdr *mpa_v2;
737         union async_output *async_data;
738         u16 mpa_ord, mpa_ird;
739         u8 mpa_hdr_size = 0;
740         u16 ulp_data_len;
741         u8 mpa_rev;
742
743         async_data = &ep->ep_buffer_virt->async_output;
744
745         mpa_rev = async_data->mpa_request.mpa_handshake_mode;
746         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
747                    "private_data_len=%x handshake_mode=%x private_data=(%x)\n",
748                    async_data->mpa_request.ulp_data_len,
749                    mpa_rev, *((u32 *)(ep->ep_buffer_virt->in_pdata)));
750
751         if (mpa_rev == MPA_NEGOTIATION_TYPE_ENHANCED) {
752                 /* Read ord/ird values from private data buffer */
753                 mpa_v2 = (struct mpa_v2_hdr *)ep->ep_buffer_virt->in_pdata;
754                 mpa_hdr_size = sizeof(*mpa_v2);
755
756                 mpa_ord = ntohs(mpa_v2->ord);
757                 mpa_ird = ntohs(mpa_v2->ird);
758
759                 /* Temprary store in cm_info incoming ord/ird requested, later
760                  * replace with negotiated value during accept
761                  */
762                 ep->cm_info.ord = (u8)min_t(u16,
763                                             (mpa_ord & MPA_V2_IRD_ORD_MASK),
764                                             QED_IWARP_ORD_DEFAULT);
765
766                 ep->cm_info.ird = (u8)min_t(u16,
767                                             (mpa_ird & MPA_V2_IRD_ORD_MASK),
768                                             QED_IWARP_IRD_DEFAULT);
769
770                 /* Peer2Peer negotiation */
771                 ep->rtr_type = MPA_RTR_TYPE_NONE;
772                 if (mpa_ird & MPA_V2_PEER2PEER_MODEL) {
773                         if (mpa_ord & MPA_V2_WRITE_RTR)
774                                 ep->rtr_type |= MPA_RTR_TYPE_ZERO_WRITE;
775
776                         if (mpa_ord & MPA_V2_READ_RTR)
777                                 ep->rtr_type |= MPA_RTR_TYPE_ZERO_READ;
778
779                         if (mpa_ird & MPA_V2_SEND_RTR)
780                                 ep->rtr_type |= MPA_RTR_TYPE_ZERO_SEND;
781
782                         ep->rtr_type &= iwarp_info->rtr_type;
783
784                         /* if we're left with no match send our capabilities */
785                         if (ep->rtr_type == MPA_RTR_TYPE_NONE)
786                                 ep->rtr_type = iwarp_info->rtr_type;
787                 }
788
789                 ep->mpa_rev = MPA_NEGOTIATION_TYPE_ENHANCED;
790         } else {
791                 ep->cm_info.ord = QED_IWARP_ORD_DEFAULT;
792                 ep->cm_info.ird = QED_IWARP_IRD_DEFAULT;
793                 ep->mpa_rev = MPA_NEGOTIATION_TYPE_BASIC;
794         }
795
796         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
797                    "MPA_NEGOTIATE (v%d): ORD: 0x%x IRD: 0x%x rtr:0x%x ulp_data_len = %x mpa_hdr_size = %x\n",
798                    mpa_rev, ep->cm_info.ord, ep->cm_info.ird, ep->rtr_type,
799                    async_data->mpa_request.ulp_data_len, mpa_hdr_size);
800
801         /* Strip mpa v2 hdr from private data before sending to upper layer */
802         ep->cm_info.private_data = ep->ep_buffer_virt->in_pdata + mpa_hdr_size;
803
804         ulp_data_len = le16_to_cpu(async_data->mpa_request.ulp_data_len);
805         ep->cm_info.private_data_len = ulp_data_len - mpa_hdr_size;
806
807         params.event = QED_IWARP_EVENT_MPA_REQUEST;
808         params.cm_info = &ep->cm_info;
809         params.ep_context = ep;
810         params.status = 0;
811
812         ep->state = QED_IWARP_EP_MPA_REQ_RCVD;
813         ep->event_cb(ep->cb_context, &params);
814 }
815
816 static int
817 qed_iwarp_mpa_offload(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
818 {
819         struct iwarp_mpa_offload_ramrod_data *p_mpa_ramrod;
820         struct mpa_outgoing_params *common;
821         struct qed_iwarp_info *iwarp_info;
822         struct qed_sp_init_data init_data;
823         dma_addr_t async_output_phys;
824         struct qed_spq_entry *p_ent;
825         dma_addr_t out_pdata_phys;
826         dma_addr_t in_pdata_phys;
827         struct qed_rdma_qp *qp;
828         bool reject;
829         u32 val;
830         int rc;
831
832         if (!ep)
833                 return -EINVAL;
834
835         qp = ep->qp;
836         reject = !qp;
837
838         memset(&init_data, 0, sizeof(init_data));
839         init_data.cid = reject ? ep->tcp_cid : qp->icid;
840         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
841
842         if (ep->connect_mode == TCP_CONNECT_ACTIVE)
843                 init_data.comp_mode = QED_SPQ_MODE_CB;
844         else
845                 init_data.comp_mode = QED_SPQ_MODE_EBLOCK;
846
847         rc = qed_sp_init_request(p_hwfn, &p_ent,
848                                  IWARP_RAMROD_CMD_ID_MPA_OFFLOAD,
849                                  PROTOCOLID_IWARP, &init_data);
850         if (rc)
851                 return rc;
852
853         p_mpa_ramrod = &p_ent->ramrod.iwarp_mpa_offload;
854         common = &p_mpa_ramrod->common;
855
856         out_pdata_phys = ep->ep_buffer_phys +
857                          offsetof(struct qed_iwarp_ep_memory, out_pdata);
858         DMA_REGPAIR_LE(common->outgoing_ulp_buffer.addr, out_pdata_phys);
859
860         val = ep->cm_info.private_data_len;
861         common->outgoing_ulp_buffer.len = cpu_to_le16(val);
862         common->crc_needed = p_hwfn->p_rdma_info->iwarp.crc_needed;
863
864         common->out_rq.ord = cpu_to_le32(ep->cm_info.ord);
865         common->out_rq.ird = cpu_to_le32(ep->cm_info.ird);
866
867         val = p_hwfn->hw_info.opaque_fid << 16 | ep->tcp_cid;
868         p_mpa_ramrod->tcp_cid = cpu_to_le32(val);
869
870         in_pdata_phys = ep->ep_buffer_phys +
871                         offsetof(struct qed_iwarp_ep_memory, in_pdata);
872         p_mpa_ramrod->tcp_connect_side = ep->connect_mode;
873         DMA_REGPAIR_LE(p_mpa_ramrod->incoming_ulp_buffer.addr,
874                        in_pdata_phys);
875         p_mpa_ramrod->incoming_ulp_buffer.len =
876             cpu_to_le16(sizeof(ep->ep_buffer_virt->in_pdata));
877         async_output_phys = ep->ep_buffer_phys +
878                             offsetof(struct qed_iwarp_ep_memory, async_output);
879         DMA_REGPAIR_LE(p_mpa_ramrod->async_eqe_output_buf,
880                        async_output_phys);
881         p_mpa_ramrod->handle_for_async.hi = cpu_to_le32(PTR_HI(ep));
882         p_mpa_ramrod->handle_for_async.lo = cpu_to_le32(PTR_LO(ep));
883
884         if (!reject) {
885                 DMA_REGPAIR_LE(p_mpa_ramrod->shared_queue_addr,
886                                qp->shared_queue_phys_addr);
887                 p_mpa_ramrod->stats_counter_id =
888                     RESC_START(p_hwfn, QED_RDMA_STATS_QUEUE) + qp->stats_queue;
889         } else {
890                 common->reject = 1;
891         }
892
893         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
894         p_mpa_ramrod->rcv_wnd = cpu_to_le16(iwarp_info->rcv_wnd_size);
895         p_mpa_ramrod->mode = ep->mpa_rev;
896         SET_FIELD(p_mpa_ramrod->rtr_pref,
897                   IWARP_MPA_OFFLOAD_RAMROD_DATA_RTR_SUPPORTED, ep->rtr_type);
898
899         ep->state = QED_IWARP_EP_MPA_OFFLOADED;
900         rc = qed_spq_post(p_hwfn, p_ent, NULL);
901         if (!reject)
902                 ep->cid = qp->icid;     /* Now they're migrated. */
903
904         DP_VERBOSE(p_hwfn,
905                    QED_MSG_RDMA,
906                    "QP(0x%x) EP(0x%x) MPA Offload rc = %d IRD=0x%x ORD=0x%x rtr_type=%d mpa_rev=%d reject=%d\n",
907                    reject ? 0xffff : qp->icid,
908                    ep->tcp_cid,
909                    rc,
910                    ep->cm_info.ird,
911                    ep->cm_info.ord, ep->rtr_type, ep->mpa_rev, reject);
912         return rc;
913 }
914
915 static void
916 qed_iwarp_return_ep(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
917 {
918         ep->state = QED_IWARP_EP_INIT;
919         if (ep->qp)
920                 ep->qp->ep = NULL;
921         ep->qp = NULL;
922         memset(&ep->cm_info, 0, sizeof(ep->cm_info));
923
924         if (ep->tcp_cid == QED_IWARP_INVALID_TCP_CID) {
925                 /* We don't care about the return code, it's ok if tcp_cid
926                  * remains invalid...in this case we'll defer allocation
927                  */
928                 qed_iwarp_alloc_tcp_cid(p_hwfn, &ep->tcp_cid);
929         }
930         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
931
932         list_move_tail(&ep->list_entry,
933                        &p_hwfn->p_rdma_info->iwarp.ep_free_list);
934
935         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
936 }
937
938 static void
939 qed_iwarp_parse_private_data(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
940 {
941         struct mpa_v2_hdr *mpa_v2_params;
942         union async_output *async_data;
943         u16 mpa_ird, mpa_ord;
944         u8 mpa_data_size = 0;
945         u16 ulp_data_len;
946
947         if (MPA_REV2(p_hwfn->p_rdma_info->iwarp.mpa_rev)) {
948                 mpa_v2_params =
949                         (struct mpa_v2_hdr *)(ep->ep_buffer_virt->in_pdata);
950                 mpa_data_size = sizeof(*mpa_v2_params);
951                 mpa_ird = ntohs(mpa_v2_params->ird);
952                 mpa_ord = ntohs(mpa_v2_params->ord);
953
954                 ep->cm_info.ird = (u8)(mpa_ord & MPA_V2_IRD_ORD_MASK);
955                 ep->cm_info.ord = (u8)(mpa_ird & MPA_V2_IRD_ORD_MASK);
956         }
957
958         async_data = &ep->ep_buffer_virt->async_output;
959         ep->cm_info.private_data = ep->ep_buffer_virt->in_pdata + mpa_data_size;
960
961         ulp_data_len = le16_to_cpu(async_data->mpa_response.ulp_data_len);
962         ep->cm_info.private_data_len = ulp_data_len - mpa_data_size;
963 }
964
965 static void
966 qed_iwarp_mpa_reply_arrived(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
967 {
968         struct qed_iwarp_cm_event_params params;
969
970         if (ep->connect_mode == TCP_CONNECT_PASSIVE) {
971                 DP_NOTICE(p_hwfn,
972                           "MPA reply event not expected on passive side!\n");
973                 return;
974         }
975
976         params.event = QED_IWARP_EVENT_ACTIVE_MPA_REPLY;
977
978         qed_iwarp_parse_private_data(p_hwfn, ep);
979
980         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
981                    "MPA_NEGOTIATE (v%d): ORD: 0x%x IRD: 0x%x\n",
982                    ep->mpa_rev, ep->cm_info.ord, ep->cm_info.ird);
983
984         params.cm_info = &ep->cm_info;
985         params.ep_context = ep;
986         params.status = 0;
987
988         ep->mpa_reply_processed = true;
989
990         ep->event_cb(ep->cb_context, &params);
991 }
992
993 #define QED_IWARP_CONNECT_MODE_STRING(ep) \
994         ((ep)->connect_mode == TCP_CONNECT_PASSIVE) ? "Passive" : "Active"
995
996 /* Called as a result of the event:
997  * IWARP_EVENT_TYPE_ASYNC_MPA_HANDSHAKE_COMPLETE
998  */
999 static void
1000 qed_iwarp_mpa_complete(struct qed_hwfn *p_hwfn,
1001                        struct qed_iwarp_ep *ep, u8 fw_return_code)
1002 {
1003         struct qed_iwarp_cm_event_params params;
1004
1005         if (ep->connect_mode == TCP_CONNECT_ACTIVE)
1006                 params.event = QED_IWARP_EVENT_ACTIVE_COMPLETE;
1007         else
1008                 params.event = QED_IWARP_EVENT_PASSIVE_COMPLETE;
1009
1010         if (ep->connect_mode == TCP_CONNECT_ACTIVE && !ep->mpa_reply_processed)
1011                 qed_iwarp_parse_private_data(p_hwfn, ep);
1012
1013         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1014                    "MPA_NEGOTIATE (v%d): ORD: 0x%x IRD: 0x%x\n",
1015                    ep->mpa_rev, ep->cm_info.ord, ep->cm_info.ird);
1016
1017         params.cm_info = &ep->cm_info;
1018
1019         params.ep_context = ep;
1020
1021         switch (fw_return_code) {
1022         case RDMA_RETURN_OK:
1023                 ep->qp->max_rd_atomic_req = ep->cm_info.ord;
1024                 ep->qp->max_rd_atomic_resp = ep->cm_info.ird;
1025                 qed_iwarp_modify_qp(p_hwfn, ep->qp, QED_IWARP_QP_STATE_RTS, 1);
1026                 ep->state = QED_IWARP_EP_ESTABLISHED;
1027                 params.status = 0;
1028                 break;
1029         case IWARP_CONN_ERROR_MPA_TIMEOUT:
1030                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA timeout\n",
1031                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1032                 params.status = -EBUSY;
1033                 break;
1034         case IWARP_CONN_ERROR_MPA_ERROR_REJECT:
1035                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA Reject\n",
1036                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1037                 params.status = -ECONNREFUSED;
1038                 break;
1039         case IWARP_CONN_ERROR_MPA_RST:
1040                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA reset(tcp cid: 0x%x)\n",
1041                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid,
1042                           ep->tcp_cid);
1043                 params.status = -ECONNRESET;
1044                 break;
1045         case IWARP_CONN_ERROR_MPA_FIN:
1046                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA received FIN\n",
1047                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1048                 params.status = -ECONNREFUSED;
1049                 break;
1050         case IWARP_CONN_ERROR_MPA_INSUF_IRD:
1051                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA insufficient ird\n",
1052                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1053                 params.status = -ECONNREFUSED;
1054                 break;
1055         case IWARP_CONN_ERROR_MPA_RTR_MISMATCH:
1056                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA RTR MISMATCH\n",
1057                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1058                 params.status = -ECONNREFUSED;
1059                 break;
1060         case IWARP_CONN_ERROR_MPA_INVALID_PACKET:
1061                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA Invalid Packet\n",
1062                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1063                 params.status = -ECONNREFUSED;
1064                 break;
1065         case IWARP_CONN_ERROR_MPA_LOCAL_ERROR:
1066                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA Local Error\n",
1067                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1068                 params.status = -ECONNREFUSED;
1069                 break;
1070         case IWARP_CONN_ERROR_MPA_TERMINATE:
1071                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA TERMINATE\n",
1072                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->cid);
1073                 params.status = -ECONNREFUSED;
1074                 break;
1075         default:
1076                 params.status = -ECONNRESET;
1077                 break;
1078         }
1079
1080         if (fw_return_code != RDMA_RETURN_OK)
1081                 /* paired with READ_ONCE in destroy_qp */
1082                 smp_store_release(&ep->state, QED_IWARP_EP_CLOSED);
1083
1084         ep->event_cb(ep->cb_context, &params);
1085
1086         /* on passive side, if there is no associated QP (REJECT) we need to
1087          * return the ep to the pool, (in the regular case we add an element
1088          * in accept instead of this one.
1089          * In both cases we need to remove it from the ep_list.
1090          */
1091         if (fw_return_code != RDMA_RETURN_OK) {
1092                 ep->tcp_cid = QED_IWARP_INVALID_TCP_CID;
1093                 if ((ep->connect_mode == TCP_CONNECT_PASSIVE) &&
1094                     (!ep->qp)) {        /* Rejected */
1095                         qed_iwarp_return_ep(p_hwfn, ep);
1096                 } else {
1097                         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1098                         list_del(&ep->list_entry);
1099                         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1100                 }
1101         }
1102 }
1103
1104 static void
1105 qed_iwarp_mpa_v2_set_private(struct qed_hwfn *p_hwfn,
1106                              struct qed_iwarp_ep *ep, u8 *mpa_data_size)
1107 {
1108         struct mpa_v2_hdr *mpa_v2_params;
1109         u16 mpa_ird, mpa_ord;
1110
1111         *mpa_data_size = 0;
1112         if (MPA_REV2(ep->mpa_rev)) {
1113                 mpa_v2_params =
1114                     (struct mpa_v2_hdr *)ep->ep_buffer_virt->out_pdata;
1115                 *mpa_data_size = sizeof(*mpa_v2_params);
1116
1117                 mpa_ird = (u16)ep->cm_info.ird;
1118                 mpa_ord = (u16)ep->cm_info.ord;
1119
1120                 if (ep->rtr_type != MPA_RTR_TYPE_NONE) {
1121                         mpa_ird |= MPA_V2_PEER2PEER_MODEL;
1122
1123                         if (ep->rtr_type & MPA_RTR_TYPE_ZERO_SEND)
1124                                 mpa_ird |= MPA_V2_SEND_RTR;
1125
1126                         if (ep->rtr_type & MPA_RTR_TYPE_ZERO_WRITE)
1127                                 mpa_ord |= MPA_V2_WRITE_RTR;
1128
1129                         if (ep->rtr_type & MPA_RTR_TYPE_ZERO_READ)
1130                                 mpa_ord |= MPA_V2_READ_RTR;
1131                 }
1132
1133                 mpa_v2_params->ird = htons(mpa_ird);
1134                 mpa_v2_params->ord = htons(mpa_ord);
1135
1136                 DP_VERBOSE(p_hwfn,
1137                            QED_MSG_RDMA,
1138                            "MPA_NEGOTIATE Header: [%x ord:%x ird] %x ord:%x ird:%x peer2peer:%x rtr_send:%x rtr_write:%x rtr_read:%x\n",
1139                            mpa_v2_params->ird,
1140                            mpa_v2_params->ord,
1141                            *((u32 *)mpa_v2_params),
1142                            mpa_ord & MPA_V2_IRD_ORD_MASK,
1143                            mpa_ird & MPA_V2_IRD_ORD_MASK,
1144                            !!(mpa_ird & MPA_V2_PEER2PEER_MODEL),
1145                            !!(mpa_ird & MPA_V2_SEND_RTR),
1146                            !!(mpa_ord & MPA_V2_WRITE_RTR),
1147                            !!(mpa_ord & MPA_V2_READ_RTR));
1148         }
1149 }
1150
1151 int qed_iwarp_connect(void *rdma_cxt,
1152                       struct qed_iwarp_connect_in *iparams,
1153                       struct qed_iwarp_connect_out *oparams)
1154 {
1155         struct qed_hwfn *p_hwfn = rdma_cxt;
1156         struct qed_iwarp_info *iwarp_info;
1157         struct qed_iwarp_ep *ep;
1158         u8 mpa_data_size = 0;
1159         u32 cid;
1160         int rc;
1161
1162         if ((iparams->cm_info.ord > QED_IWARP_ORD_DEFAULT) ||
1163             (iparams->cm_info.ird > QED_IWARP_IRD_DEFAULT)) {
1164                 DP_NOTICE(p_hwfn,
1165                           "QP(0x%x) ERROR: Invalid ord(0x%x)/ird(0x%x)\n",
1166                           iparams->qp->icid, iparams->cm_info.ord,
1167                           iparams->cm_info.ird);
1168
1169                 return -EINVAL;
1170         }
1171
1172         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
1173
1174         /* Allocate ep object */
1175         rc = qed_iwarp_alloc_cid(p_hwfn, &cid);
1176         if (rc)
1177                 return rc;
1178
1179         rc = qed_iwarp_create_ep(p_hwfn, &ep);
1180         if (rc)
1181                 goto err;
1182
1183         ep->tcp_cid = cid;
1184
1185         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1186         list_add_tail(&ep->list_entry, &p_hwfn->p_rdma_info->iwarp.ep_list);
1187         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1188
1189         ep->qp = iparams->qp;
1190         ep->qp->ep = ep;
1191         ether_addr_copy(ep->remote_mac_addr, iparams->remote_mac_addr);
1192         ether_addr_copy(ep->local_mac_addr, iparams->local_mac_addr);
1193         memcpy(&ep->cm_info, &iparams->cm_info, sizeof(ep->cm_info));
1194
1195         ep->cm_info.ord = iparams->cm_info.ord;
1196         ep->cm_info.ird = iparams->cm_info.ird;
1197
1198         ep->rtr_type = iwarp_info->rtr_type;
1199         if (!iwarp_info->peer2peer)
1200                 ep->rtr_type = MPA_RTR_TYPE_NONE;
1201
1202         if ((ep->rtr_type & MPA_RTR_TYPE_ZERO_READ) && (ep->cm_info.ord == 0))
1203                 ep->cm_info.ord = 1;
1204
1205         ep->mpa_rev = iwarp_info->mpa_rev;
1206
1207         qed_iwarp_mpa_v2_set_private(p_hwfn, ep, &mpa_data_size);
1208
1209         ep->cm_info.private_data = ep->ep_buffer_virt->out_pdata;
1210         ep->cm_info.private_data_len = iparams->cm_info.private_data_len +
1211                                        mpa_data_size;
1212
1213         memcpy((u8 *)ep->ep_buffer_virt->out_pdata + mpa_data_size,
1214                iparams->cm_info.private_data,
1215                iparams->cm_info.private_data_len);
1216
1217         ep->mss = iparams->mss;
1218         ep->mss = min_t(u16, QED_IWARP_MAX_FW_MSS, ep->mss);
1219
1220         ep->event_cb = iparams->event_cb;
1221         ep->cb_context = iparams->cb_context;
1222         ep->connect_mode = TCP_CONNECT_ACTIVE;
1223
1224         oparams->ep_context = ep;
1225
1226         rc = qed_iwarp_tcp_offload(p_hwfn, ep);
1227
1228         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) EP(0x%x) rc = %d\n",
1229                    iparams->qp->icid, ep->tcp_cid, rc);
1230
1231         if (rc) {
1232                 qed_iwarp_destroy_ep(p_hwfn, ep, true);
1233                 goto err;
1234         }
1235
1236         return rc;
1237 err:
1238         qed_iwarp_cid_cleaned(p_hwfn, cid);
1239
1240         return rc;
1241 }
1242
1243 static struct qed_iwarp_ep *qed_iwarp_get_free_ep(struct qed_hwfn *p_hwfn)
1244 {
1245         struct qed_iwarp_ep *ep = NULL;
1246         int rc;
1247
1248         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1249
1250         if (list_empty(&p_hwfn->p_rdma_info->iwarp.ep_free_list)) {
1251                 DP_ERR(p_hwfn, "Ep list is empty\n");
1252                 goto out;
1253         }
1254
1255         ep = list_first_entry(&p_hwfn->p_rdma_info->iwarp.ep_free_list,
1256                               struct qed_iwarp_ep, list_entry);
1257
1258         /* in some cases we could have failed allocating a tcp cid when added
1259          * from accept / failure... retry now..this is not the common case.
1260          */
1261         if (ep->tcp_cid == QED_IWARP_INVALID_TCP_CID) {
1262                 rc = qed_iwarp_alloc_tcp_cid(p_hwfn, &ep->tcp_cid);
1263
1264                 /* if we fail we could look for another entry with a valid
1265                  * tcp_cid, but since we don't expect to reach this anyway
1266                  * it's not worth the handling
1267                  */
1268                 if (rc) {
1269                         ep->tcp_cid = QED_IWARP_INVALID_TCP_CID;
1270                         ep = NULL;
1271                         goto out;
1272                 }
1273         }
1274
1275         list_del(&ep->list_entry);
1276
1277 out:
1278         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1279         return ep;
1280 }
1281
1282 #define QED_IWARP_MAX_CID_CLEAN_TIME  100
1283 #define QED_IWARP_MAX_NO_PROGRESS_CNT 5
1284
1285 /* This function waits for all the bits of a bmap to be cleared, as long as
1286  * there is progress ( i.e. the number of bits left to be cleared decreases )
1287  * the function continues.
1288  */
1289 static int
1290 qed_iwarp_wait_cid_map_cleared(struct qed_hwfn *p_hwfn, struct qed_bmap *bmap)
1291 {
1292         int prev_weight = 0;
1293         int wait_count = 0;
1294         int weight = 0;
1295
1296         weight = bitmap_weight(bmap->bitmap, bmap->max_count);
1297         prev_weight = weight;
1298
1299         while (weight) {
1300                 msleep(QED_IWARP_MAX_CID_CLEAN_TIME);
1301
1302                 weight = bitmap_weight(bmap->bitmap, bmap->max_count);
1303
1304                 if (prev_weight == weight) {
1305                         wait_count++;
1306                 } else {
1307                         prev_weight = weight;
1308                         wait_count = 0;
1309                 }
1310
1311                 if (wait_count > QED_IWARP_MAX_NO_PROGRESS_CNT) {
1312                         DP_NOTICE(p_hwfn,
1313                                   "%s bitmap wait timed out (%d cids pending)\n",
1314                                   bmap->name, weight);
1315                         return -EBUSY;
1316                 }
1317         }
1318         return 0;
1319 }
1320
1321 static int qed_iwarp_wait_for_all_cids(struct qed_hwfn *p_hwfn)
1322 {
1323         int rc;
1324         int i;
1325
1326         rc = qed_iwarp_wait_cid_map_cleared(p_hwfn,
1327                                             &p_hwfn->p_rdma_info->tcp_cid_map);
1328         if (rc)
1329                 return rc;
1330
1331         /* Now free the tcp cids from the main cid map */
1332         for (i = 0; i < QED_IWARP_PREALLOC_CNT; i++)
1333                 qed_bmap_release_id(p_hwfn, &p_hwfn->p_rdma_info->cid_map, i);
1334
1335         /* Now wait for all cids to be completed */
1336         return qed_iwarp_wait_cid_map_cleared(p_hwfn,
1337                                               &p_hwfn->p_rdma_info->cid_map);
1338 }
1339
1340 static void qed_iwarp_free_prealloc_ep(struct qed_hwfn *p_hwfn)
1341 {
1342         struct qed_iwarp_ep *ep;
1343
1344         while (!list_empty(&p_hwfn->p_rdma_info->iwarp.ep_free_list)) {
1345                 spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1346
1347                 ep = list_first_entry(&p_hwfn->p_rdma_info->iwarp.ep_free_list,
1348                                       struct qed_iwarp_ep, list_entry);
1349
1350                 if (!ep) {
1351                         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1352                         break;
1353                 }
1354                 list_del(&ep->list_entry);
1355
1356                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1357
1358                 if (ep->tcp_cid != QED_IWARP_INVALID_TCP_CID)
1359                         qed_iwarp_cid_cleaned(p_hwfn, ep->tcp_cid);
1360
1361                 qed_iwarp_destroy_ep(p_hwfn, ep, false);
1362         }
1363 }
1364
1365 static int qed_iwarp_prealloc_ep(struct qed_hwfn *p_hwfn, bool init)
1366 {
1367         struct qed_iwarp_ep *ep;
1368         int rc = 0;
1369         int count;
1370         u32 cid;
1371         int i;
1372
1373         count = init ? QED_IWARP_PREALLOC_CNT : 1;
1374         for (i = 0; i < count; i++) {
1375                 rc = qed_iwarp_create_ep(p_hwfn, &ep);
1376                 if (rc)
1377                         return rc;
1378
1379                 /* During initialization we allocate from the main pool,
1380                  * afterwards we allocate only from the tcp_cid.
1381                  */
1382                 if (init) {
1383                         rc = qed_iwarp_alloc_cid(p_hwfn, &cid);
1384                         if (rc)
1385                                 goto err;
1386                         qed_iwarp_set_tcp_cid(p_hwfn, cid);
1387                 } else {
1388                         /* We don't care about the return code, it's ok if
1389                          * tcp_cid remains invalid...in this case we'll
1390                          * defer allocation
1391                          */
1392                         qed_iwarp_alloc_tcp_cid(p_hwfn, &cid);
1393                 }
1394
1395                 ep->tcp_cid = cid;
1396
1397                 spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1398                 list_add_tail(&ep->list_entry,
1399                               &p_hwfn->p_rdma_info->iwarp.ep_free_list);
1400                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1401         }
1402
1403         return rc;
1404
1405 err:
1406         qed_iwarp_destroy_ep(p_hwfn, ep, false);
1407
1408         return rc;
1409 }
1410
1411 int qed_iwarp_alloc(struct qed_hwfn *p_hwfn)
1412 {
1413         int rc;
1414
1415         /* Allocate bitmap for tcp cid. These are used by passive side
1416          * to ensure it can allocate a tcp cid during dpc that was
1417          * pre-acquired and doesn't require dynamic allocation of ilt
1418          */
1419         rc = qed_rdma_bmap_alloc(p_hwfn, &p_hwfn->p_rdma_info->tcp_cid_map,
1420                                  QED_IWARP_PREALLOC_CNT, "TCP_CID");
1421         if (rc) {
1422                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1423                            "Failed to allocate tcp cid, rc = %d\n", rc);
1424                 return rc;
1425         }
1426
1427         INIT_LIST_HEAD(&p_hwfn->p_rdma_info->iwarp.ep_free_list);
1428         spin_lock_init(&p_hwfn->p_rdma_info->iwarp.iw_lock);
1429
1430         rc = qed_iwarp_prealloc_ep(p_hwfn, true);
1431         if (rc)
1432                 return rc;
1433
1434         return qed_ooo_alloc(p_hwfn);
1435 }
1436
1437 void qed_iwarp_resc_free(struct qed_hwfn *p_hwfn)
1438 {
1439         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
1440
1441         qed_ooo_free(p_hwfn);
1442         qed_rdma_bmap_free(p_hwfn, &p_hwfn->p_rdma_info->tcp_cid_map, 1);
1443         kfree(iwarp_info->mpa_bufs);
1444         kfree(iwarp_info->partial_fpdus);
1445         kfree(iwarp_info->mpa_intermediate_buf);
1446 }
1447
1448 int qed_iwarp_accept(void *rdma_cxt, struct qed_iwarp_accept_in *iparams)
1449 {
1450         struct qed_hwfn *p_hwfn = rdma_cxt;
1451         struct qed_iwarp_ep *ep;
1452         u8 mpa_data_size = 0;
1453         int rc;
1454
1455         ep = iparams->ep_context;
1456         if (!ep) {
1457                 DP_ERR(p_hwfn, "Ep Context receive in accept is NULL\n");
1458                 return -EINVAL;
1459         }
1460
1461         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) EP(0x%x)\n",
1462                    iparams->qp->icid, ep->tcp_cid);
1463
1464         if ((iparams->ord > QED_IWARP_ORD_DEFAULT) ||
1465             (iparams->ird > QED_IWARP_IRD_DEFAULT)) {
1466                 DP_VERBOSE(p_hwfn,
1467                            QED_MSG_RDMA,
1468                            "QP(0x%x) EP(0x%x) ERROR: Invalid ord(0x%x)/ird(0x%x)\n",
1469                            iparams->qp->icid,
1470                            ep->tcp_cid, iparams->ord, iparams->ord);
1471                 return -EINVAL;
1472         }
1473
1474         qed_iwarp_prealloc_ep(p_hwfn, false);
1475
1476         ep->cb_context = iparams->cb_context;
1477         ep->qp = iparams->qp;
1478         ep->qp->ep = ep;
1479
1480         if (ep->mpa_rev == MPA_NEGOTIATION_TYPE_ENHANCED) {
1481                 /* Negotiate ord/ird: if upperlayer requested ord larger than
1482                  * ird advertised by remote, we need to decrease our ord
1483                  */
1484                 if (iparams->ord > ep->cm_info.ird)
1485                         iparams->ord = ep->cm_info.ird;
1486
1487                 if ((ep->rtr_type & MPA_RTR_TYPE_ZERO_READ) &&
1488                     (iparams->ird == 0))
1489                         iparams->ird = 1;
1490         }
1491
1492         /* Update cm_info ord/ird to be negotiated values */
1493         ep->cm_info.ord = iparams->ord;
1494         ep->cm_info.ird = iparams->ird;
1495
1496         qed_iwarp_mpa_v2_set_private(p_hwfn, ep, &mpa_data_size);
1497
1498         ep->cm_info.private_data = ep->ep_buffer_virt->out_pdata;
1499         ep->cm_info.private_data_len = iparams->private_data_len +
1500                                        mpa_data_size;
1501
1502         memcpy((u8 *)ep->ep_buffer_virt->out_pdata + mpa_data_size,
1503                iparams->private_data, iparams->private_data_len);
1504
1505         rc = qed_iwarp_mpa_offload(p_hwfn, ep);
1506         if (rc)
1507                 qed_iwarp_modify_qp(p_hwfn,
1508                                     iparams->qp, QED_IWARP_QP_STATE_ERROR, 1);
1509
1510         return rc;
1511 }
1512
1513 int qed_iwarp_reject(void *rdma_cxt, struct qed_iwarp_reject_in *iparams)
1514 {
1515         struct qed_hwfn *p_hwfn = rdma_cxt;
1516         struct qed_iwarp_ep *ep;
1517         u8 mpa_data_size = 0;
1518
1519         ep = iparams->ep_context;
1520         if (!ep) {
1521                 DP_ERR(p_hwfn, "Ep Context receive in reject is NULL\n");
1522                 return -EINVAL;
1523         }
1524
1525         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "EP(0x%x)\n", ep->tcp_cid);
1526
1527         ep->cb_context = iparams->cb_context;
1528         ep->qp = NULL;
1529
1530         qed_iwarp_mpa_v2_set_private(p_hwfn, ep, &mpa_data_size);
1531
1532         ep->cm_info.private_data = ep->ep_buffer_virt->out_pdata;
1533         ep->cm_info.private_data_len = iparams->private_data_len +
1534                                        mpa_data_size;
1535
1536         memcpy((u8 *)ep->ep_buffer_virt->out_pdata + mpa_data_size,
1537                iparams->private_data, iparams->private_data_len);
1538
1539         return qed_iwarp_mpa_offload(p_hwfn, ep);
1540 }
1541
1542 static void
1543 qed_iwarp_print_cm_info(struct qed_hwfn *p_hwfn,
1544                         struct qed_iwarp_cm_info *cm_info)
1545 {
1546         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "ip_version = %d\n",
1547                    cm_info->ip_version);
1548
1549         if (cm_info->ip_version == QED_TCP_IPV4)
1550                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1551                            "remote_ip %pI4h:%x, local_ip %pI4h:%x vlan=%x\n",
1552                            cm_info->remote_ip, cm_info->remote_port,
1553                            cm_info->local_ip, cm_info->local_port,
1554                            cm_info->vlan);
1555         else
1556                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1557                            "remote_ip %pI6:%x, local_ip %pI6:%x vlan=%x\n",
1558                            cm_info->remote_ip, cm_info->remote_port,
1559                            cm_info->local_ip, cm_info->local_port,
1560                            cm_info->vlan);
1561
1562         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1563                    "private_data_len = %x ord = %d, ird = %d\n",
1564                    cm_info->private_data_len, cm_info->ord, cm_info->ird);
1565 }
1566
1567 static int
1568 qed_iwarp_ll2_post_rx(struct qed_hwfn *p_hwfn,
1569                       struct qed_iwarp_ll2_buff *buf, u8 handle)
1570 {
1571         int rc;
1572
1573         rc = qed_ll2_post_rx_buffer(p_hwfn, handle, buf->data_phys_addr,
1574                                     (u16)buf->buff_size, buf, 1);
1575         if (rc) {
1576                 DP_NOTICE(p_hwfn,
1577                           "Failed to repost rx buffer to ll2 rc = %d, handle=%d\n",
1578                           rc, handle);
1579                 dma_free_coherent(&p_hwfn->cdev->pdev->dev, buf->buff_size,
1580                                   buf->data, buf->data_phys_addr);
1581                 kfree(buf);
1582         }
1583
1584         return rc;
1585 }
1586
1587 static bool
1588 qed_iwarp_ep_exists(struct qed_hwfn *p_hwfn, struct qed_iwarp_cm_info *cm_info)
1589 {
1590         struct qed_iwarp_ep *ep = NULL;
1591         bool found = false;
1592
1593         list_for_each_entry(ep,
1594                             &p_hwfn->p_rdma_info->iwarp.ep_list,
1595                             list_entry) {
1596                 if ((ep->cm_info.local_port == cm_info->local_port) &&
1597                     (ep->cm_info.remote_port == cm_info->remote_port) &&
1598                     (ep->cm_info.vlan == cm_info->vlan) &&
1599                     !memcmp(&ep->cm_info.local_ip, cm_info->local_ip,
1600                             sizeof(cm_info->local_ip)) &&
1601                     !memcmp(&ep->cm_info.remote_ip, cm_info->remote_ip,
1602                             sizeof(cm_info->remote_ip))) {
1603                         found = true;
1604                         break;
1605                 }
1606         }
1607
1608         if (found) {
1609                 DP_NOTICE(p_hwfn,
1610                           "SYN received on active connection - dropping\n");
1611                 qed_iwarp_print_cm_info(p_hwfn, cm_info);
1612
1613                 return true;
1614         }
1615
1616         return false;
1617 }
1618
1619 static struct qed_iwarp_listener *
1620 qed_iwarp_get_listener(struct qed_hwfn *p_hwfn,
1621                        struct qed_iwarp_cm_info *cm_info)
1622 {
1623         struct qed_iwarp_listener *listener = NULL;
1624         static const u32 ip_zero[4] = { 0, 0, 0, 0 };
1625         bool found = false;
1626
1627         qed_iwarp_print_cm_info(p_hwfn, cm_info);
1628
1629         list_for_each_entry(listener,
1630                             &p_hwfn->p_rdma_info->iwarp.listen_list,
1631                             list_entry) {
1632                 if (listener->port == cm_info->local_port) {
1633                         if (!memcmp(listener->ip_addr,
1634                                     ip_zero, sizeof(ip_zero))) {
1635                                 found = true;
1636                                 break;
1637                         }
1638
1639                         if (!memcmp(listener->ip_addr,
1640                                     cm_info->local_ip,
1641                                     sizeof(cm_info->local_ip)) &&
1642                             (listener->vlan == cm_info->vlan)) {
1643                                 found = true;
1644                                 break;
1645                         }
1646                 }
1647         }
1648
1649         if (found) {
1650                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "listener found = %p\n",
1651                            listener);
1652                 return listener;
1653         }
1654
1655         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "listener not found\n");
1656         return NULL;
1657 }
1658
1659 static int
1660 qed_iwarp_parse_rx_pkt(struct qed_hwfn *p_hwfn,
1661                        struct qed_iwarp_cm_info *cm_info,
1662                        void *buf,
1663                        u8 *remote_mac_addr,
1664                        u8 *local_mac_addr,
1665                        int *payload_len, int *tcp_start_offset)
1666 {
1667         struct vlan_ethhdr *vethh;
1668         bool vlan_valid = false;
1669         struct ipv6hdr *ip6h;
1670         struct ethhdr *ethh;
1671         struct tcphdr *tcph;
1672         struct iphdr *iph;
1673         int eth_hlen;
1674         int ip_hlen;
1675         int eth_type;
1676         int i;
1677
1678         ethh = buf;
1679         eth_type = ntohs(ethh->h_proto);
1680         if (eth_type == ETH_P_8021Q) {
1681                 vlan_valid = true;
1682                 vethh = (struct vlan_ethhdr *)ethh;
1683                 cm_info->vlan = ntohs(vethh->h_vlan_TCI) & VLAN_VID_MASK;
1684                 eth_type = ntohs(vethh->h_vlan_encapsulated_proto);
1685         }
1686
1687         eth_hlen = ETH_HLEN + (vlan_valid ? sizeof(u32) : 0);
1688
1689         if (!ether_addr_equal(ethh->h_dest,
1690                               p_hwfn->p_rdma_info->iwarp.mac_addr)) {
1691                 DP_VERBOSE(p_hwfn,
1692                            QED_MSG_RDMA,
1693                            "Got unexpected mac %pM instead of %pM\n",
1694                            ethh->h_dest, p_hwfn->p_rdma_info->iwarp.mac_addr);
1695                 return -EINVAL;
1696         }
1697
1698         ether_addr_copy(remote_mac_addr, ethh->h_source);
1699         ether_addr_copy(local_mac_addr, ethh->h_dest);
1700
1701         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "eth_type =%d source mac: %pM\n",
1702                    eth_type, ethh->h_source);
1703
1704         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "eth_hlen=%d destination mac: %pM\n",
1705                    eth_hlen, ethh->h_dest);
1706
1707         iph = (struct iphdr *)((u8 *)(ethh) + eth_hlen);
1708
1709         if (eth_type == ETH_P_IP) {
1710                 if (iph->protocol != IPPROTO_TCP) {
1711                         DP_NOTICE(p_hwfn,
1712                                   "Unexpected ip protocol on ll2 %x\n",
1713                                   iph->protocol);
1714                         return -EINVAL;
1715                 }
1716
1717                 cm_info->local_ip[0] = ntohl(iph->daddr);
1718                 cm_info->remote_ip[0] = ntohl(iph->saddr);
1719                 cm_info->ip_version = QED_TCP_IPV4;
1720
1721                 ip_hlen = (iph->ihl) * sizeof(u32);
1722                 *payload_len = ntohs(iph->tot_len) - ip_hlen;
1723         } else if (eth_type == ETH_P_IPV6) {
1724                 ip6h = (struct ipv6hdr *)iph;
1725
1726                 if (ip6h->nexthdr != IPPROTO_TCP) {
1727                         DP_NOTICE(p_hwfn,
1728                                   "Unexpected ip protocol on ll2 %x\n",
1729                                   iph->protocol);
1730                         return -EINVAL;
1731                 }
1732
1733                 for (i = 0; i < 4; i++) {
1734                         cm_info->local_ip[i] =
1735                             ntohl(ip6h->daddr.in6_u.u6_addr32[i]);
1736                         cm_info->remote_ip[i] =
1737                             ntohl(ip6h->saddr.in6_u.u6_addr32[i]);
1738                 }
1739                 cm_info->ip_version = QED_TCP_IPV6;
1740
1741                 ip_hlen = sizeof(*ip6h);
1742                 *payload_len = ntohs(ip6h->payload_len);
1743         } else {
1744                 DP_NOTICE(p_hwfn, "Unexpected ethertype on ll2 %x\n", eth_type);
1745                 return -EINVAL;
1746         }
1747
1748         tcph = (struct tcphdr *)((u8 *)iph + ip_hlen);
1749
1750         if (!tcph->syn) {
1751                 DP_NOTICE(p_hwfn,
1752                           "Only SYN type packet expected on this ll2 conn, iph->ihl=%d source=%d dest=%d\n",
1753                           iph->ihl, tcph->source, tcph->dest);
1754                 return -EINVAL;
1755         }
1756
1757         cm_info->local_port = ntohs(tcph->dest);
1758         cm_info->remote_port = ntohs(tcph->source);
1759
1760         qed_iwarp_print_cm_info(p_hwfn, cm_info);
1761
1762         *tcp_start_offset = eth_hlen + ip_hlen;
1763
1764         return 0;
1765 }
1766
1767 static struct qed_iwarp_fpdu *qed_iwarp_get_curr_fpdu(struct qed_hwfn *p_hwfn,
1768                                                       u16 cid)
1769 {
1770         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
1771         struct qed_iwarp_fpdu *partial_fpdu;
1772         u32 idx;
1773
1774         idx = cid - qed_cxt_get_proto_cid_start(p_hwfn, PROTOCOLID_IWARP);
1775         if (idx >= iwarp_info->max_num_partial_fpdus) {
1776                 DP_ERR(p_hwfn, "Invalid cid %x max_num_partial_fpdus=%x\n", cid,
1777                        iwarp_info->max_num_partial_fpdus);
1778                 return NULL;
1779         }
1780
1781         partial_fpdu = &iwarp_info->partial_fpdus[idx];
1782
1783         return partial_fpdu;
1784 }
1785
1786 enum qed_iwarp_mpa_pkt_type {
1787         QED_IWARP_MPA_PKT_PACKED,
1788         QED_IWARP_MPA_PKT_PARTIAL,
1789         QED_IWARP_MPA_PKT_UNALIGNED
1790 };
1791
1792 #define QED_IWARP_INVALID_FPDU_LENGTH 0xffff
1793 #define QED_IWARP_MPA_FPDU_LENGTH_SIZE (2)
1794 #define QED_IWARP_MPA_CRC32_DIGEST_SIZE (4)
1795
1796 /* Pad to multiple of 4 */
1797 #define QED_IWARP_PDU_DATA_LEN_WITH_PAD(data_len) ALIGN(data_len, 4)
1798 #define QED_IWARP_FPDU_LEN_WITH_PAD(_mpa_len)                              \
1799         (QED_IWARP_PDU_DATA_LEN_WITH_PAD((_mpa_len) +                      \
1800                                          QED_IWARP_MPA_FPDU_LENGTH_SIZE) + \
1801                                          QED_IWARP_MPA_CRC32_DIGEST_SIZE)
1802
1803 /* fpdu can be fragmented over maximum 3 bds: header, partial mpa, unaligned */
1804 #define QED_IWARP_MAX_BDS_PER_FPDU 3
1805
1806 static const char * const pkt_type_str[] = {
1807         "QED_IWARP_MPA_PKT_PACKED",
1808         "QED_IWARP_MPA_PKT_PARTIAL",
1809         "QED_IWARP_MPA_PKT_UNALIGNED"
1810 };
1811
1812 static int
1813 qed_iwarp_recycle_pkt(struct qed_hwfn *p_hwfn,
1814                       struct qed_iwarp_fpdu *fpdu,
1815                       struct qed_iwarp_ll2_buff *buf);
1816
1817 static enum qed_iwarp_mpa_pkt_type
1818 qed_iwarp_mpa_classify(struct qed_hwfn *p_hwfn,
1819                        struct qed_iwarp_fpdu *fpdu,
1820                        u16 tcp_payload_len, u8 *mpa_data)
1821 {
1822         enum qed_iwarp_mpa_pkt_type pkt_type;
1823         u16 mpa_len;
1824
1825         if (fpdu->incomplete_bytes) {
1826                 pkt_type = QED_IWARP_MPA_PKT_UNALIGNED;
1827                 goto out;
1828         }
1829
1830         /* special case of one byte remaining...
1831          * lower byte will be read next packet
1832          */
1833         if (tcp_payload_len == 1) {
1834                 fpdu->fpdu_length = *mpa_data << BITS_PER_BYTE;
1835                 pkt_type = QED_IWARP_MPA_PKT_PARTIAL;
1836                 goto out;
1837         }
1838
1839         mpa_len = ntohs(*(__force __be16 *)mpa_data);
1840         fpdu->fpdu_length = QED_IWARP_FPDU_LEN_WITH_PAD(mpa_len);
1841
1842         if (fpdu->fpdu_length <= tcp_payload_len)
1843                 pkt_type = QED_IWARP_MPA_PKT_PACKED;
1844         else
1845                 pkt_type = QED_IWARP_MPA_PKT_PARTIAL;
1846
1847 out:
1848         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1849                    "MPA_ALIGN: %s: fpdu_length=0x%x tcp_payload_len:0x%x\n",
1850                    pkt_type_str[pkt_type], fpdu->fpdu_length, tcp_payload_len);
1851
1852         return pkt_type;
1853 }
1854
1855 static void
1856 qed_iwarp_init_fpdu(struct qed_iwarp_ll2_buff *buf,
1857                     struct qed_iwarp_fpdu *fpdu,
1858                     struct unaligned_opaque_data *pkt_data,
1859                     u16 tcp_payload_size, u8 placement_offset)
1860 {
1861         u16 first_mpa_offset = le16_to_cpu(pkt_data->first_mpa_offset);
1862
1863         fpdu->mpa_buf = buf;
1864         fpdu->pkt_hdr = buf->data_phys_addr + placement_offset;
1865         fpdu->pkt_hdr_size = pkt_data->tcp_payload_offset;
1866         fpdu->mpa_frag = buf->data_phys_addr + first_mpa_offset;
1867         fpdu->mpa_frag_virt = (u8 *)(buf->data) + first_mpa_offset;
1868
1869         if (tcp_payload_size == 1)
1870                 fpdu->incomplete_bytes = QED_IWARP_INVALID_FPDU_LENGTH;
1871         else if (tcp_payload_size < fpdu->fpdu_length)
1872                 fpdu->incomplete_bytes = fpdu->fpdu_length - tcp_payload_size;
1873         else
1874                 fpdu->incomplete_bytes = 0;     /* complete fpdu */
1875
1876         fpdu->mpa_frag_len = fpdu->fpdu_length - fpdu->incomplete_bytes;
1877 }
1878
1879 static int
1880 qed_iwarp_cp_pkt(struct qed_hwfn *p_hwfn,
1881                  struct qed_iwarp_fpdu *fpdu,
1882                  struct unaligned_opaque_data *pkt_data,
1883                  struct qed_iwarp_ll2_buff *buf, u16 tcp_payload_size)
1884 {
1885         u16 first_mpa_offset = le16_to_cpu(pkt_data->first_mpa_offset);
1886         u8 *tmp_buf = p_hwfn->p_rdma_info->iwarp.mpa_intermediate_buf;
1887         int rc;
1888
1889         /* need to copy the data from the partial packet stored in fpdu
1890          * to the new buf, for this we also need to move the data currently
1891          * placed on the buf. The assumption is that the buffer is big enough
1892          * since fpdu_length <= mss, we use an intermediate buffer since
1893          * we may need to copy the new data to an overlapping location
1894          */
1895         if ((fpdu->mpa_frag_len + tcp_payload_size) > (u16)buf->buff_size) {
1896                 DP_ERR(p_hwfn,
1897                        "MPA ALIGN: Unexpected: buffer is not large enough for split fpdu buff_size = %d mpa_frag_len = %d, tcp_payload_size = %d, incomplete_bytes = %d\n",
1898                        buf->buff_size, fpdu->mpa_frag_len,
1899                        tcp_payload_size, fpdu->incomplete_bytes);
1900                 return -EINVAL;
1901         }
1902
1903         DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1904                    "MPA ALIGN Copying fpdu: [%p, %d] [%p, %d]\n",
1905                    fpdu->mpa_frag_virt, fpdu->mpa_frag_len,
1906                    (u8 *)(buf->data) + first_mpa_offset, tcp_payload_size);
1907
1908         memcpy(tmp_buf, fpdu->mpa_frag_virt, fpdu->mpa_frag_len);
1909         memcpy(tmp_buf + fpdu->mpa_frag_len,
1910                (u8 *)(buf->data) + first_mpa_offset, tcp_payload_size);
1911
1912         rc = qed_iwarp_recycle_pkt(p_hwfn, fpdu, fpdu->mpa_buf);
1913         if (rc)
1914                 return rc;
1915
1916         /* If we managed to post the buffer copy the data to the new buffer
1917          * o/w this will occur in the next round...
1918          */
1919         memcpy((u8 *)(buf->data), tmp_buf,
1920                fpdu->mpa_frag_len + tcp_payload_size);
1921
1922         fpdu->mpa_buf = buf;
1923         /* fpdu->pkt_hdr remains as is */
1924         /* fpdu->mpa_frag is overridden with new buf */
1925         fpdu->mpa_frag = buf->data_phys_addr;
1926         fpdu->mpa_frag_virt = buf->data;
1927         fpdu->mpa_frag_len += tcp_payload_size;
1928
1929         fpdu->incomplete_bytes -= tcp_payload_size;
1930
1931         DP_VERBOSE(p_hwfn,
1932                    QED_MSG_RDMA,
1933                    "MPA ALIGN: split fpdu buff_size = %d mpa_frag_len = %d, tcp_payload_size = %d, incomplete_bytes = %d\n",
1934                    buf->buff_size, fpdu->mpa_frag_len, tcp_payload_size,
1935                    fpdu->incomplete_bytes);
1936
1937         return 0;
1938 }
1939
1940 static void
1941 qed_iwarp_update_fpdu_length(struct qed_hwfn *p_hwfn,
1942                              struct qed_iwarp_fpdu *fpdu, u8 *mpa_data)
1943 {
1944         u16 mpa_len;
1945
1946         /* Update incomplete packets if needed */
1947         if (fpdu->incomplete_bytes == QED_IWARP_INVALID_FPDU_LENGTH) {
1948                 /* Missing lower byte is now available */
1949                 mpa_len = fpdu->fpdu_length | *mpa_data;
1950                 fpdu->fpdu_length = QED_IWARP_FPDU_LEN_WITH_PAD(mpa_len);
1951                 /* one byte of hdr */
1952                 fpdu->mpa_frag_len = 1;
1953                 fpdu->incomplete_bytes = fpdu->fpdu_length - 1;
1954                 DP_VERBOSE(p_hwfn,
1955                            QED_MSG_RDMA,
1956                            "MPA_ALIGN: Partial header mpa_len=%x fpdu_length=%x incomplete_bytes=%x\n",
1957                            mpa_len, fpdu->fpdu_length, fpdu->incomplete_bytes);
1958         }
1959 }
1960
1961 #define QED_IWARP_IS_RIGHT_EDGE(_curr_pkt) \
1962         (GET_FIELD((_curr_pkt)->flags,     \
1963                    UNALIGNED_OPAQUE_DATA_PKT_REACHED_WIN_RIGHT_EDGE))
1964
1965 /* This function is used to recycle a buffer using the ll2 drop option. It
1966  * uses the mechanism to ensure that all buffers posted to tx before this one
1967  * were completed. The buffer sent here will be sent as a cookie in the tx
1968  * completion function and can then be reposted to rx chain when done. The flow
1969  * that requires this is the flow where a FPDU splits over more than 3 tcp
1970  * segments. In this case the driver needs to re-post a rx buffer instead of
1971  * the one received, but driver can't simply repost a buffer it copied from
1972  * as there is a case where the buffer was originally a packed FPDU, and is
1973  * partially posted to FW. Driver needs to ensure FW is done with it.
1974  */
1975 static int
1976 qed_iwarp_recycle_pkt(struct qed_hwfn *p_hwfn,
1977                       struct qed_iwarp_fpdu *fpdu,
1978                       struct qed_iwarp_ll2_buff *buf)
1979 {
1980         struct qed_ll2_tx_pkt_info tx_pkt;
1981         u8 ll2_handle;
1982         int rc;
1983
1984         memset(&tx_pkt, 0, sizeof(tx_pkt));
1985         tx_pkt.num_of_bds = 1;
1986         tx_pkt.tx_dest = QED_LL2_TX_DEST_DROP;
1987         tx_pkt.l4_hdr_offset_w = fpdu->pkt_hdr_size >> 2;
1988         tx_pkt.first_frag = fpdu->pkt_hdr;
1989         tx_pkt.first_frag_len = fpdu->pkt_hdr_size;
1990         buf->piggy_buf = NULL;
1991         tx_pkt.cookie = buf;
1992
1993         ll2_handle = p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle;
1994
1995         rc = qed_ll2_prepare_tx_packet(p_hwfn, ll2_handle, &tx_pkt, true);
1996         if (rc)
1997                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
1998                            "Can't drop packet rc=%d\n", rc);
1999
2000         DP_VERBOSE(p_hwfn,
2001                    QED_MSG_RDMA,
2002                    "MPA_ALIGN: send drop tx packet [%lx, 0x%x], buf=%p, rc=%d\n",
2003                    (unsigned long int)tx_pkt.first_frag,
2004                    tx_pkt.first_frag_len, buf, rc);
2005
2006         return rc;
2007 }
2008
2009 static int
2010 qed_iwarp_win_right_edge(struct qed_hwfn *p_hwfn, struct qed_iwarp_fpdu *fpdu)
2011 {
2012         struct qed_ll2_tx_pkt_info tx_pkt;
2013         u8 ll2_handle;
2014         int rc;
2015
2016         memset(&tx_pkt, 0, sizeof(tx_pkt));
2017         tx_pkt.num_of_bds = 1;
2018         tx_pkt.tx_dest = QED_LL2_TX_DEST_LB;
2019         tx_pkt.l4_hdr_offset_w = fpdu->pkt_hdr_size >> 2;
2020
2021         tx_pkt.first_frag = fpdu->pkt_hdr;
2022         tx_pkt.first_frag_len = fpdu->pkt_hdr_size;
2023         tx_pkt.enable_ip_cksum = true;
2024         tx_pkt.enable_l4_cksum = true;
2025         tx_pkt.calc_ip_len = true;
2026         /* vlan overload with enum iwarp_ll2_tx_queues */
2027         tx_pkt.vlan = IWARP_LL2_ALIGNED_RIGHT_TRIMMED_TX_QUEUE;
2028
2029         ll2_handle = p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle;
2030
2031         rc = qed_ll2_prepare_tx_packet(p_hwfn, ll2_handle, &tx_pkt, true);
2032         if (rc)
2033                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2034                            "Can't send right edge rc=%d\n", rc);
2035         DP_VERBOSE(p_hwfn,
2036                    QED_MSG_RDMA,
2037                    "MPA_ALIGN: Sent right edge FPDU num_bds=%d [%lx, 0x%x], rc=%d\n",
2038                    tx_pkt.num_of_bds,
2039                    (unsigned long int)tx_pkt.first_frag,
2040                    tx_pkt.first_frag_len, rc);
2041
2042         return rc;
2043 }
2044
2045 static int
2046 qed_iwarp_send_fpdu(struct qed_hwfn *p_hwfn,
2047                     struct qed_iwarp_fpdu *fpdu,
2048                     struct unaligned_opaque_data *curr_pkt,
2049                     struct qed_iwarp_ll2_buff *buf,
2050                     u16 tcp_payload_size, enum qed_iwarp_mpa_pkt_type pkt_type)
2051 {
2052         struct qed_ll2_tx_pkt_info tx_pkt;
2053         u16 first_mpa_offset;
2054         u8 ll2_handle;
2055         int rc;
2056
2057         memset(&tx_pkt, 0, sizeof(tx_pkt));
2058
2059         /* An unaligned packet means it's split over two tcp segments. So the
2060          * complete packet requires 3 bds, one for the header, one for the
2061          * part of the fpdu of the first tcp segment, and the last fragment
2062          * will point to the remainder of the fpdu. A packed pdu, requires only
2063          * two bds, one for the header and one for the data.
2064          */
2065         tx_pkt.num_of_bds = (pkt_type == QED_IWARP_MPA_PKT_UNALIGNED) ? 3 : 2;
2066         tx_pkt.tx_dest = QED_LL2_TX_DEST_LB;
2067         tx_pkt.l4_hdr_offset_w = fpdu->pkt_hdr_size >> 2; /* offset in words */
2068
2069         /* Send the mpa_buf only with the last fpdu (in case of packed) */
2070         if (pkt_type == QED_IWARP_MPA_PKT_UNALIGNED ||
2071             tcp_payload_size <= fpdu->fpdu_length)
2072                 tx_pkt.cookie = fpdu->mpa_buf;
2073
2074         tx_pkt.first_frag = fpdu->pkt_hdr;
2075         tx_pkt.first_frag_len = fpdu->pkt_hdr_size;
2076         tx_pkt.enable_ip_cksum = true;
2077         tx_pkt.enable_l4_cksum = true;
2078         tx_pkt.calc_ip_len = true;
2079         /* vlan overload with enum iwarp_ll2_tx_queues */
2080         tx_pkt.vlan = IWARP_LL2_ALIGNED_TX_QUEUE;
2081
2082         /* special case of unaligned packet and not packed, need to send
2083          * both buffers as cookie to release.
2084          */
2085         if (tcp_payload_size == fpdu->incomplete_bytes)
2086                 fpdu->mpa_buf->piggy_buf = buf;
2087
2088         ll2_handle = p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle;
2089
2090         /* Set first fragment to header */
2091         rc = qed_ll2_prepare_tx_packet(p_hwfn, ll2_handle, &tx_pkt, true);
2092         if (rc)
2093                 goto out;
2094
2095         /* Set second fragment to first part of packet */
2096         rc = qed_ll2_set_fragment_of_tx_packet(p_hwfn, ll2_handle,
2097                                                fpdu->mpa_frag,
2098                                                fpdu->mpa_frag_len);
2099         if (rc)
2100                 goto out;
2101
2102         if (!fpdu->incomplete_bytes)
2103                 goto out;
2104
2105         first_mpa_offset = le16_to_cpu(curr_pkt->first_mpa_offset);
2106
2107         /* Set third fragment to second part of the packet */
2108         rc = qed_ll2_set_fragment_of_tx_packet(p_hwfn,
2109                                                ll2_handle,
2110                                                buf->data_phys_addr +
2111                                                first_mpa_offset,
2112                                                fpdu->incomplete_bytes);
2113 out:
2114         DP_VERBOSE(p_hwfn,
2115                    QED_MSG_RDMA,
2116                    "MPA_ALIGN: Sent FPDU num_bds=%d first_frag_len=%x, mpa_frag_len=0x%x, incomplete_bytes:0x%x rc=%d\n",
2117                    tx_pkt.num_of_bds,
2118                    tx_pkt.first_frag_len,
2119                    fpdu->mpa_frag_len,
2120                    fpdu->incomplete_bytes, rc);
2121
2122         return rc;
2123 }
2124
2125 static void
2126 qed_iwarp_mpa_get_data(struct qed_hwfn *p_hwfn,
2127                        struct unaligned_opaque_data *curr_pkt,
2128                        u32 opaque_data0, u32 opaque_data1)
2129 {
2130         u64 opaque_data;
2131
2132         opaque_data = HILO_64(cpu_to_le32(opaque_data1),
2133                               cpu_to_le32(opaque_data0));
2134         *curr_pkt = *((struct unaligned_opaque_data *)&opaque_data);
2135
2136         le16_add_cpu(&curr_pkt->first_mpa_offset,
2137                      curr_pkt->tcp_payload_offset);
2138 }
2139
2140 /* This function is called when an unaligned or incomplete MPA packet arrives
2141  * driver needs to align the packet, perhaps using previous data and send
2142  * it down to FW once it is aligned.
2143  */
2144 static int
2145 qed_iwarp_process_mpa_pkt(struct qed_hwfn *p_hwfn,
2146                           struct qed_iwarp_ll2_mpa_buf *mpa_buf)
2147 {
2148         struct unaligned_opaque_data *curr_pkt = &mpa_buf->data;
2149         struct qed_iwarp_ll2_buff *buf = mpa_buf->ll2_buf;
2150         enum qed_iwarp_mpa_pkt_type pkt_type;
2151         struct qed_iwarp_fpdu *fpdu;
2152         u16 cid, first_mpa_offset;
2153         int rc = -EINVAL;
2154         u8 *mpa_data;
2155
2156         cid = le32_to_cpu(curr_pkt->cid);
2157
2158         fpdu = qed_iwarp_get_curr_fpdu(p_hwfn, (u16)cid);
2159         if (!fpdu) { /* something corrupt with cid, post rx back */
2160                 DP_ERR(p_hwfn, "Invalid cid, drop and post back to rx cid=%x\n",
2161                        cid);
2162                 goto err;
2163         }
2164
2165         do {
2166                 first_mpa_offset = le16_to_cpu(curr_pkt->first_mpa_offset);
2167                 mpa_data = ((u8 *)(buf->data) + first_mpa_offset);
2168
2169                 pkt_type = qed_iwarp_mpa_classify(p_hwfn, fpdu,
2170                                                   mpa_buf->tcp_payload_len,
2171                                                   mpa_data);
2172
2173                 switch (pkt_type) {
2174                 case QED_IWARP_MPA_PKT_PARTIAL:
2175                         qed_iwarp_init_fpdu(buf, fpdu,
2176                                             curr_pkt,
2177                                             mpa_buf->tcp_payload_len,
2178                                             mpa_buf->placement_offset);
2179
2180                         if (!QED_IWARP_IS_RIGHT_EDGE(curr_pkt)) {
2181                                 mpa_buf->tcp_payload_len = 0;
2182                                 break;
2183                         }
2184
2185                         rc = qed_iwarp_win_right_edge(p_hwfn, fpdu);
2186
2187                         if (rc) {
2188                                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2189                                            "Can't send FPDU:reset rc=%d\n", rc);
2190                                 memset(fpdu, 0, sizeof(*fpdu));
2191                                 break;
2192                         }
2193
2194                         mpa_buf->tcp_payload_len = 0;
2195                         break;
2196                 case QED_IWARP_MPA_PKT_PACKED:
2197                         qed_iwarp_init_fpdu(buf, fpdu,
2198                                             curr_pkt,
2199                                             mpa_buf->tcp_payload_len,
2200                                             mpa_buf->placement_offset);
2201
2202                         rc = qed_iwarp_send_fpdu(p_hwfn, fpdu, curr_pkt, buf,
2203                                                  mpa_buf->tcp_payload_len,
2204                                                  pkt_type);
2205                         if (rc) {
2206                                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2207                                            "Can't send FPDU:reset rc=%d\n", rc);
2208                                 memset(fpdu, 0, sizeof(*fpdu));
2209                                 break;
2210                         }
2211
2212                         mpa_buf->tcp_payload_len -= fpdu->fpdu_length;
2213                         le16_add_cpu(&curr_pkt->first_mpa_offset,
2214                                      fpdu->fpdu_length);
2215                         break;
2216                 case QED_IWARP_MPA_PKT_UNALIGNED:
2217                         qed_iwarp_update_fpdu_length(p_hwfn, fpdu, mpa_data);
2218                         if (mpa_buf->tcp_payload_len < fpdu->incomplete_bytes) {
2219                                 /* special handling of fpdu split over more
2220                                  * than 2 segments
2221                                  */
2222                                 if (QED_IWARP_IS_RIGHT_EDGE(curr_pkt)) {
2223                                         rc = qed_iwarp_win_right_edge(p_hwfn,
2224                                                                       fpdu);
2225                                         /* packet will be re-processed later */
2226                                         if (rc)
2227                                                 return rc;
2228                                 }
2229
2230                                 rc = qed_iwarp_cp_pkt(p_hwfn, fpdu, curr_pkt,
2231                                                       buf,
2232                                                       mpa_buf->tcp_payload_len);
2233                                 if (rc) /* packet will be re-processed later */
2234                                         return rc;
2235
2236                                 mpa_buf->tcp_payload_len = 0;
2237                                 break;
2238                         }
2239
2240                         rc = qed_iwarp_send_fpdu(p_hwfn, fpdu, curr_pkt, buf,
2241                                                  mpa_buf->tcp_payload_len,
2242                                                  pkt_type);
2243                         if (rc) {
2244                                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2245                                            "Can't send FPDU:delay rc=%d\n", rc);
2246                                 /* don't reset fpdu -> we need it for next
2247                                  * classify
2248                                  */
2249                                 break;
2250                         }
2251
2252                         mpa_buf->tcp_payload_len -= fpdu->incomplete_bytes;
2253                         le16_add_cpu(&curr_pkt->first_mpa_offset,
2254                                      fpdu->incomplete_bytes);
2255
2256                         /* The framed PDU was sent - no more incomplete bytes */
2257                         fpdu->incomplete_bytes = 0;
2258                         break;
2259                 }
2260         } while (mpa_buf->tcp_payload_len && !rc);
2261
2262         return rc;
2263
2264 err:
2265         qed_iwarp_ll2_post_rx(p_hwfn,
2266                               buf,
2267                               p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle);
2268         return rc;
2269 }
2270
2271 static void qed_iwarp_process_pending_pkts(struct qed_hwfn *p_hwfn)
2272 {
2273         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2274         struct qed_iwarp_ll2_mpa_buf *mpa_buf = NULL;
2275         int rc;
2276
2277         while (!list_empty(&iwarp_info->mpa_buf_pending_list)) {
2278                 mpa_buf = list_first_entry(&iwarp_info->mpa_buf_pending_list,
2279                                            struct qed_iwarp_ll2_mpa_buf,
2280                                            list_entry);
2281
2282                 rc = qed_iwarp_process_mpa_pkt(p_hwfn, mpa_buf);
2283
2284                 /* busy means break and continue processing later, don't
2285                  * remove the buf from the pending list.
2286                  */
2287                 if (rc == -EBUSY)
2288                         break;
2289
2290                 list_move_tail(&mpa_buf->list_entry,
2291                                &iwarp_info->mpa_buf_list);
2292
2293                 if (rc) {       /* different error, don't continue */
2294                         DP_NOTICE(p_hwfn, "process pkts failed rc=%d\n", rc);
2295                         break;
2296                 }
2297         }
2298 }
2299
2300 static void
2301 qed_iwarp_ll2_comp_mpa_pkt(void *cxt, struct qed_ll2_comp_rx_data *data)
2302 {
2303         struct qed_iwarp_ll2_mpa_buf *mpa_buf;
2304         struct qed_iwarp_info *iwarp_info;
2305         struct qed_hwfn *p_hwfn = cxt;
2306         u16 first_mpa_offset;
2307
2308         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2309         mpa_buf = list_first_entry(&iwarp_info->mpa_buf_list,
2310                                    struct qed_iwarp_ll2_mpa_buf, list_entry);
2311         if (!mpa_buf) {
2312                 DP_ERR(p_hwfn, "No free mpa buf\n");
2313                 goto err;
2314         }
2315
2316         list_del(&mpa_buf->list_entry);
2317         qed_iwarp_mpa_get_data(p_hwfn, &mpa_buf->data,
2318                                data->opaque_data_0, data->opaque_data_1);
2319
2320         first_mpa_offset = le16_to_cpu(mpa_buf->data.first_mpa_offset);
2321
2322         DP_VERBOSE(p_hwfn,
2323                    QED_MSG_RDMA,
2324                    "LL2 MPA CompRx payload_len:0x%x\tfirst_mpa_offset:0x%x\ttcp_payload_offset:0x%x\tflags:0x%x\tcid:0x%x\n",
2325                    data->length.packet_length, first_mpa_offset,
2326                    mpa_buf->data.tcp_payload_offset, mpa_buf->data.flags,
2327                    mpa_buf->data.cid);
2328
2329         mpa_buf->ll2_buf = data->cookie;
2330         mpa_buf->tcp_payload_len = data->length.packet_length -
2331                                    first_mpa_offset;
2332
2333         first_mpa_offset += data->u.placement_offset;
2334         mpa_buf->data.first_mpa_offset = cpu_to_le16(first_mpa_offset);
2335         mpa_buf->placement_offset = data->u.placement_offset;
2336
2337         list_add_tail(&mpa_buf->list_entry, &iwarp_info->mpa_buf_pending_list);
2338
2339         qed_iwarp_process_pending_pkts(p_hwfn);
2340         return;
2341 err:
2342         qed_iwarp_ll2_post_rx(p_hwfn, data->cookie,
2343                               iwarp_info->ll2_mpa_handle);
2344 }
2345
2346 static void
2347 qed_iwarp_ll2_comp_syn_pkt(void *cxt, struct qed_ll2_comp_rx_data *data)
2348 {
2349         struct qed_iwarp_ll2_buff *buf = data->cookie;
2350         struct qed_iwarp_listener *listener;
2351         struct qed_ll2_tx_pkt_info tx_pkt;
2352         struct qed_iwarp_cm_info cm_info;
2353         struct qed_hwfn *p_hwfn = cxt;
2354         u8 remote_mac_addr[ETH_ALEN];
2355         u8 local_mac_addr[ETH_ALEN];
2356         struct qed_iwarp_ep *ep;
2357         int tcp_start_offset;
2358         u8 ll2_syn_handle;
2359         int payload_len;
2360         u32 hdr_size;
2361         int rc;
2362
2363         memset(&cm_info, 0, sizeof(cm_info));
2364         ll2_syn_handle = p_hwfn->p_rdma_info->iwarp.ll2_syn_handle;
2365
2366         /* Check if packet was received with errors... */
2367         if (data->err_flags) {
2368                 DP_NOTICE(p_hwfn, "Error received on SYN packet: 0x%x\n",
2369                           data->err_flags);
2370                 goto err;
2371         }
2372
2373         if (GET_FIELD(data->parse_flags,
2374                       PARSING_AND_ERR_FLAGS_L4CHKSMWASCALCULATED) &&
2375             GET_FIELD(data->parse_flags, PARSING_AND_ERR_FLAGS_L4CHKSMERROR)) {
2376                 DP_NOTICE(p_hwfn, "Syn packet received with checksum error\n");
2377                 goto err;
2378         }
2379
2380         rc = qed_iwarp_parse_rx_pkt(p_hwfn, &cm_info, (u8 *)(buf->data) +
2381                                     data->u.placement_offset, remote_mac_addr,
2382                                     local_mac_addr, &payload_len,
2383                                     &tcp_start_offset);
2384         if (rc)
2385                 goto err;
2386
2387         /* Check if there is a listener for this 4-tuple+vlan */
2388         listener = qed_iwarp_get_listener(p_hwfn, &cm_info);
2389         if (!listener) {
2390                 DP_VERBOSE(p_hwfn,
2391                            QED_MSG_RDMA,
2392                            "SYN received on tuple not listened on parse_flags=%d packet len=%d\n",
2393                            data->parse_flags, data->length.packet_length);
2394
2395                 memset(&tx_pkt, 0, sizeof(tx_pkt));
2396                 tx_pkt.num_of_bds = 1;
2397                 tx_pkt.l4_hdr_offset_w = (data->length.packet_length) >> 2;
2398                 tx_pkt.tx_dest = QED_LL2_TX_DEST_LB;
2399                 tx_pkt.first_frag = buf->data_phys_addr +
2400                                     data->u.placement_offset;
2401                 tx_pkt.first_frag_len = data->length.packet_length;
2402                 tx_pkt.cookie = buf;
2403
2404                 rc = qed_ll2_prepare_tx_packet(p_hwfn, ll2_syn_handle,
2405                                                &tx_pkt, true);
2406
2407                 if (rc) {
2408                         DP_NOTICE(p_hwfn,
2409                                   "Can't post SYN back to chip rc=%d\n", rc);
2410                         goto err;
2411                 }
2412                 return;
2413         }
2414
2415         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "Received syn on listening port\n");
2416         /* There may be an open ep on this connection if this is a syn
2417          * retrasnmit... need to make sure there isn't...
2418          */
2419         if (qed_iwarp_ep_exists(p_hwfn, &cm_info))
2420                 goto err;
2421
2422         ep = qed_iwarp_get_free_ep(p_hwfn);
2423         if (!ep)
2424                 goto err;
2425
2426         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
2427         list_add_tail(&ep->list_entry, &p_hwfn->p_rdma_info->iwarp.ep_list);
2428         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
2429
2430         ether_addr_copy(ep->remote_mac_addr, remote_mac_addr);
2431         ether_addr_copy(ep->local_mac_addr, local_mac_addr);
2432
2433         memcpy(&ep->cm_info, &cm_info, sizeof(ep->cm_info));
2434
2435         hdr_size = ((cm_info.ip_version == QED_TCP_IPV4) ? 40 : 60);
2436         ep->mss = p_hwfn->p_rdma_info->iwarp.max_mtu - hdr_size;
2437         ep->mss = min_t(u16, QED_IWARP_MAX_FW_MSS, ep->mss);
2438
2439         ep->event_cb = listener->event_cb;
2440         ep->cb_context = listener->cb_context;
2441         ep->connect_mode = TCP_CONNECT_PASSIVE;
2442
2443         ep->syn = buf;
2444         ep->syn_ip_payload_length = (u16)payload_len;
2445         ep->syn_phy_addr = buf->data_phys_addr + data->u.placement_offset +
2446                            tcp_start_offset;
2447
2448         rc = qed_iwarp_tcp_offload(p_hwfn, ep);
2449         if (rc) {
2450                 qed_iwarp_return_ep(p_hwfn, ep);
2451                 goto err;
2452         }
2453
2454         return;
2455 err:
2456         qed_iwarp_ll2_post_rx(p_hwfn, buf, ll2_syn_handle);
2457 }
2458
2459 static void qed_iwarp_ll2_rel_rx_pkt(void *cxt, u8 connection_handle,
2460                                      void *cookie, dma_addr_t rx_buf_addr,
2461                                      bool b_last_packet)
2462 {
2463         struct qed_iwarp_ll2_buff *buffer = cookie;
2464         struct qed_hwfn *p_hwfn = cxt;
2465
2466         dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size,
2467                           buffer->data, buffer->data_phys_addr);
2468         kfree(buffer);
2469 }
2470
2471 static void qed_iwarp_ll2_comp_tx_pkt(void *cxt, u8 connection_handle,
2472                                       void *cookie, dma_addr_t first_frag_addr,
2473                                       bool b_last_fragment, bool b_last_packet)
2474 {
2475         struct qed_iwarp_ll2_buff *buffer = cookie;
2476         struct qed_iwarp_ll2_buff *piggy;
2477         struct qed_hwfn *p_hwfn = cxt;
2478
2479         if (!buffer)            /* can happen in packed mpa unaligned... */
2480                 return;
2481
2482         /* this was originally an rx packet, post it back */
2483         piggy = buffer->piggy_buf;
2484         if (piggy) {
2485                 buffer->piggy_buf = NULL;
2486                 qed_iwarp_ll2_post_rx(p_hwfn, piggy, connection_handle);
2487         }
2488
2489         qed_iwarp_ll2_post_rx(p_hwfn, buffer, connection_handle);
2490
2491         if (connection_handle == p_hwfn->p_rdma_info->iwarp.ll2_mpa_handle)
2492                 qed_iwarp_process_pending_pkts(p_hwfn);
2493
2494         return;
2495 }
2496
2497 static void qed_iwarp_ll2_rel_tx_pkt(void *cxt, u8 connection_handle,
2498                                      void *cookie, dma_addr_t first_frag_addr,
2499                                      bool b_last_fragment, bool b_last_packet)
2500 {
2501         struct qed_iwarp_ll2_buff *buffer = cookie;
2502         struct qed_hwfn *p_hwfn = cxt;
2503
2504         if (!buffer)
2505                 return;
2506
2507         if (buffer->piggy_buf) {
2508                 dma_free_coherent(&p_hwfn->cdev->pdev->dev,
2509                                   buffer->piggy_buf->buff_size,
2510                                   buffer->piggy_buf->data,
2511                                   buffer->piggy_buf->data_phys_addr);
2512
2513                 kfree(buffer->piggy_buf);
2514         }
2515
2516         dma_free_coherent(&p_hwfn->cdev->pdev->dev, buffer->buff_size,
2517                           buffer->data, buffer->data_phys_addr);
2518
2519         kfree(buffer);
2520 }
2521
2522 /* The only slowpath for iwarp ll2 is unalign flush. When this completion
2523  * is received, need to reset the FPDU.
2524  */
2525 static void
2526 qed_iwarp_ll2_slowpath(void *cxt,
2527                        u8 connection_handle,
2528                        u32 opaque_data_0, u32 opaque_data_1)
2529 {
2530         struct unaligned_opaque_data unalign_data;
2531         struct qed_hwfn *p_hwfn = cxt;
2532         struct qed_iwarp_fpdu *fpdu;
2533         u32 cid;
2534
2535         qed_iwarp_mpa_get_data(p_hwfn, &unalign_data,
2536                                opaque_data_0, opaque_data_1);
2537
2538         cid = le32_to_cpu(unalign_data.cid);
2539
2540         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "(0x%x) Flush fpdu\n", cid);
2541
2542         fpdu = qed_iwarp_get_curr_fpdu(p_hwfn, (u16)cid);
2543         if (fpdu)
2544                 memset(fpdu, 0, sizeof(*fpdu));
2545 }
2546
2547 static int qed_iwarp_ll2_stop(struct qed_hwfn *p_hwfn)
2548 {
2549         struct qed_iwarp_info *iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2550         int rc = 0;
2551
2552         if (iwarp_info->ll2_syn_handle != QED_IWARP_HANDLE_INVAL) {
2553                 rc = qed_ll2_terminate_connection(p_hwfn,
2554                                                   iwarp_info->ll2_syn_handle);
2555                 if (rc)
2556                         DP_INFO(p_hwfn, "Failed to terminate syn connection\n");
2557
2558                 qed_ll2_release_connection(p_hwfn, iwarp_info->ll2_syn_handle);
2559                 iwarp_info->ll2_syn_handle = QED_IWARP_HANDLE_INVAL;
2560         }
2561
2562         if (iwarp_info->ll2_ooo_handle != QED_IWARP_HANDLE_INVAL) {
2563                 rc = qed_ll2_terminate_connection(p_hwfn,
2564                                                   iwarp_info->ll2_ooo_handle);
2565                 if (rc)
2566                         DP_INFO(p_hwfn, "Failed to terminate ooo connection\n");
2567
2568                 qed_ll2_release_connection(p_hwfn, iwarp_info->ll2_ooo_handle);
2569                 iwarp_info->ll2_ooo_handle = QED_IWARP_HANDLE_INVAL;
2570         }
2571
2572         if (iwarp_info->ll2_mpa_handle != QED_IWARP_HANDLE_INVAL) {
2573                 rc = qed_ll2_terminate_connection(p_hwfn,
2574                                                   iwarp_info->ll2_mpa_handle);
2575                 if (rc)
2576                         DP_INFO(p_hwfn, "Failed to terminate mpa connection\n");
2577
2578                 qed_ll2_release_connection(p_hwfn, iwarp_info->ll2_mpa_handle);
2579                 iwarp_info->ll2_mpa_handle = QED_IWARP_HANDLE_INVAL;
2580         }
2581
2582         qed_llh_remove_mac_filter(p_hwfn->cdev, 0,
2583                                   p_hwfn->p_rdma_info->iwarp.mac_addr);
2584
2585         return rc;
2586 }
2587
2588 static int
2589 qed_iwarp_ll2_alloc_buffers(struct qed_hwfn *p_hwfn,
2590                             int num_rx_bufs, int buff_size, u8 ll2_handle)
2591 {
2592         struct qed_iwarp_ll2_buff *buffer;
2593         int rc = 0;
2594         int i;
2595
2596         for (i = 0; i < num_rx_bufs; i++) {
2597                 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL);
2598                 if (!buffer) {
2599                         rc = -ENOMEM;
2600                         break;
2601                 }
2602
2603                 buffer->data = dma_alloc_coherent(&p_hwfn->cdev->pdev->dev,
2604                                                   buff_size,
2605                                                   &buffer->data_phys_addr,
2606                                                   GFP_KERNEL);
2607                 if (!buffer->data) {
2608                         kfree(buffer);
2609                         rc = -ENOMEM;
2610                         break;
2611                 }
2612
2613                 buffer->buff_size = buff_size;
2614                 rc = qed_iwarp_ll2_post_rx(p_hwfn, buffer, ll2_handle);
2615                 if (rc)
2616                         /* buffers will be deallocated by qed_ll2 */
2617                         break;
2618         }
2619         return rc;
2620 }
2621
2622 #define QED_IWARP_MAX_BUF_SIZE(mtu)                                  \
2623         ALIGN((mtu) + ETH_HLEN + 2 * VLAN_HLEN + 2 + ETH_CACHE_LINE_SIZE, \
2624                 ETH_CACHE_LINE_SIZE)
2625
2626 static int
2627 qed_iwarp_ll2_start(struct qed_hwfn *p_hwfn,
2628                     struct qed_rdma_start_in_params *params,
2629                     u32 rcv_wnd_size)
2630 {
2631         struct qed_iwarp_info *iwarp_info;
2632         struct qed_ll2_acquire_data data;
2633         struct qed_ll2_cbs cbs;
2634         u32 buff_size;
2635         u16 n_ooo_bufs;
2636         int rc = 0;
2637         int i;
2638
2639         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2640         iwarp_info->ll2_syn_handle = QED_IWARP_HANDLE_INVAL;
2641         iwarp_info->ll2_ooo_handle = QED_IWARP_HANDLE_INVAL;
2642         iwarp_info->ll2_mpa_handle = QED_IWARP_HANDLE_INVAL;
2643
2644         iwarp_info->max_mtu = params->max_mtu;
2645
2646         ether_addr_copy(p_hwfn->p_rdma_info->iwarp.mac_addr, params->mac_addr);
2647
2648         rc = qed_llh_add_mac_filter(p_hwfn->cdev, 0, params->mac_addr);
2649         if (rc)
2650                 return rc;
2651
2652         /* Start SYN connection */
2653         cbs.rx_comp_cb = qed_iwarp_ll2_comp_syn_pkt;
2654         cbs.rx_release_cb = qed_iwarp_ll2_rel_rx_pkt;
2655         cbs.tx_comp_cb = qed_iwarp_ll2_comp_tx_pkt;
2656         cbs.tx_release_cb = qed_iwarp_ll2_rel_tx_pkt;
2657         cbs.slowpath_cb = NULL;
2658         cbs.cookie = p_hwfn;
2659
2660         memset(&data, 0, sizeof(data));
2661         data.input.conn_type = QED_LL2_TYPE_IWARP;
2662         /* SYN will use ctx based queues */
2663         data.input.rx_conn_type = QED_LL2_RX_TYPE_CTX;
2664         data.input.mtu = params->max_mtu;
2665         data.input.rx_num_desc = QED_IWARP_LL2_SYN_RX_SIZE;
2666         data.input.tx_num_desc = QED_IWARP_LL2_SYN_TX_SIZE;
2667         data.input.tx_max_bds_per_packet = 1;   /* will never be fragmented */
2668         data.input.tx_tc = PKT_LB_TC;
2669         data.input.tx_dest = QED_LL2_TX_DEST_LB;
2670         data.p_connection_handle = &iwarp_info->ll2_syn_handle;
2671         data.cbs = &cbs;
2672
2673         rc = qed_ll2_acquire_connection(p_hwfn, &data);
2674         if (rc) {
2675                 DP_NOTICE(p_hwfn, "Failed to acquire LL2 connection\n");
2676                 qed_llh_remove_mac_filter(p_hwfn->cdev, 0, params->mac_addr);
2677                 return rc;
2678         }
2679
2680         rc = qed_ll2_establish_connection(p_hwfn, iwarp_info->ll2_syn_handle);
2681         if (rc) {
2682                 DP_NOTICE(p_hwfn, "Failed to establish LL2 connection\n");
2683                 goto err;
2684         }
2685
2686         buff_size = QED_IWARP_MAX_BUF_SIZE(params->max_mtu);
2687         rc = qed_iwarp_ll2_alloc_buffers(p_hwfn,
2688                                          QED_IWARP_LL2_SYN_RX_SIZE,
2689                                          buff_size,
2690                                          iwarp_info->ll2_syn_handle);
2691         if (rc)
2692                 goto err;
2693
2694         /* Start OOO connection */
2695         data.input.conn_type = QED_LL2_TYPE_OOO;
2696         /* OOO/unaligned will use legacy ll2 queues (ram based) */
2697         data.input.rx_conn_type = QED_LL2_RX_TYPE_LEGACY;
2698         data.input.mtu = params->max_mtu;
2699
2700         n_ooo_bufs = (QED_IWARP_MAX_OOO * rcv_wnd_size) /
2701                      iwarp_info->max_mtu;
2702         n_ooo_bufs = min_t(u32, n_ooo_bufs, QED_IWARP_LL2_OOO_MAX_RX_SIZE);
2703
2704         data.input.rx_num_desc = n_ooo_bufs;
2705         data.input.rx_num_ooo_buffers = n_ooo_bufs;
2706
2707         data.input.tx_max_bds_per_packet = 1;   /* will never be fragmented */
2708         data.input.tx_num_desc = QED_IWARP_LL2_OOO_DEF_TX_SIZE;
2709         data.p_connection_handle = &iwarp_info->ll2_ooo_handle;
2710
2711         rc = qed_ll2_acquire_connection(p_hwfn, &data);
2712         if (rc)
2713                 goto err;
2714
2715         rc = qed_ll2_establish_connection(p_hwfn, iwarp_info->ll2_ooo_handle);
2716         if (rc)
2717                 goto err;
2718
2719         /* Start Unaligned MPA connection */
2720         cbs.rx_comp_cb = qed_iwarp_ll2_comp_mpa_pkt;
2721         cbs.slowpath_cb = qed_iwarp_ll2_slowpath;
2722
2723         memset(&data, 0, sizeof(data));
2724         data.input.conn_type = QED_LL2_TYPE_IWARP;
2725         data.input.mtu = params->max_mtu;
2726         /* FW requires that once a packet arrives OOO, it must have at
2727          * least 2 rx buffers available on the unaligned connection
2728          * for handling the case that it is a partial fpdu.
2729          */
2730         data.input.rx_num_desc = n_ooo_bufs * 2;
2731         data.input.tx_num_desc = data.input.rx_num_desc;
2732         data.input.tx_max_bds_per_packet = QED_IWARP_MAX_BDS_PER_FPDU;
2733         data.input.tx_tc = PKT_LB_TC;
2734         data.input.tx_dest = QED_LL2_TX_DEST_LB;
2735         data.p_connection_handle = &iwarp_info->ll2_mpa_handle;
2736         data.input.secondary_queue = true;
2737         data.cbs = &cbs;
2738
2739         rc = qed_ll2_acquire_connection(p_hwfn, &data);
2740         if (rc)
2741                 goto err;
2742
2743         rc = qed_ll2_establish_connection(p_hwfn, iwarp_info->ll2_mpa_handle);
2744         if (rc)
2745                 goto err;
2746
2747         rc = qed_iwarp_ll2_alloc_buffers(p_hwfn,
2748                                          data.input.rx_num_desc,
2749                                          buff_size,
2750                                          iwarp_info->ll2_mpa_handle);
2751         if (rc)
2752                 goto err;
2753
2754         iwarp_info->partial_fpdus = kcalloc((u16)p_hwfn->p_rdma_info->num_qps,
2755                                             sizeof(*iwarp_info->partial_fpdus),
2756                                             GFP_KERNEL);
2757         if (!iwarp_info->partial_fpdus)
2758                 goto err;
2759
2760         iwarp_info->max_num_partial_fpdus = (u16)p_hwfn->p_rdma_info->num_qps;
2761
2762         iwarp_info->mpa_intermediate_buf = kzalloc(buff_size, GFP_KERNEL);
2763         if (!iwarp_info->mpa_intermediate_buf)
2764                 goto err;
2765
2766         /* The mpa_bufs array serves for pending RX packets received on the
2767          * mpa ll2 that don't have place on the tx ring and require later
2768          * processing. We can't fail on allocation of such a struct therefore
2769          * we allocate enough to take care of all rx packets
2770          */
2771         iwarp_info->mpa_bufs = kcalloc(data.input.rx_num_desc,
2772                                        sizeof(*iwarp_info->mpa_bufs),
2773                                        GFP_KERNEL);
2774         if (!iwarp_info->mpa_bufs)
2775                 goto err;
2776
2777         INIT_LIST_HEAD(&iwarp_info->mpa_buf_pending_list);
2778         INIT_LIST_HEAD(&iwarp_info->mpa_buf_list);
2779         for (i = 0; i < data.input.rx_num_desc; i++)
2780                 list_add_tail(&iwarp_info->mpa_bufs[i].list_entry,
2781                               &iwarp_info->mpa_buf_list);
2782         return rc;
2783 err:
2784         qed_iwarp_ll2_stop(p_hwfn);
2785
2786         return rc;
2787 }
2788
2789 static struct {
2790         u32 two_ports;
2791         u32 four_ports;
2792 } qed_iwarp_rcv_wnd_size[MAX_CHIP_IDS] = {
2793         {QED_IWARP_RCV_WND_SIZE_DEF_BB_2P, QED_IWARP_RCV_WND_SIZE_DEF_BB_4P},
2794         {QED_IWARP_RCV_WND_SIZE_DEF_AH_2P, QED_IWARP_RCV_WND_SIZE_DEF_AH_4P}
2795 };
2796
2797 int qed_iwarp_setup(struct qed_hwfn *p_hwfn,
2798                     struct qed_rdma_start_in_params *params)
2799 {
2800         struct qed_dev *cdev = p_hwfn->cdev;
2801         struct qed_iwarp_info *iwarp_info;
2802         enum chip_ids chip_id;
2803         u32 rcv_wnd_size;
2804
2805         iwarp_info = &p_hwfn->p_rdma_info->iwarp;
2806
2807         iwarp_info->tcp_flags = QED_IWARP_TS_EN;
2808
2809         chip_id = QED_IS_BB(cdev) ? CHIP_BB : CHIP_K2;
2810         rcv_wnd_size = (qed_device_num_ports(cdev) == 4) ?
2811                 qed_iwarp_rcv_wnd_size[chip_id].four_ports :
2812                 qed_iwarp_rcv_wnd_size[chip_id].two_ports;
2813
2814         /* value 0 is used for ilog2(QED_IWARP_RCV_WND_SIZE_MIN) */
2815         iwarp_info->rcv_wnd_scale = ilog2(rcv_wnd_size) -
2816             ilog2(QED_IWARP_RCV_WND_SIZE_MIN);
2817         iwarp_info->rcv_wnd_size = rcv_wnd_size >> iwarp_info->rcv_wnd_scale;
2818         iwarp_info->crc_needed = QED_IWARP_PARAM_CRC_NEEDED;
2819         iwarp_info->mpa_rev = MPA_NEGOTIATION_TYPE_ENHANCED;
2820
2821         iwarp_info->peer2peer = QED_IWARP_PARAM_P2P;
2822
2823         iwarp_info->rtr_type =  MPA_RTR_TYPE_ZERO_SEND |
2824                                 MPA_RTR_TYPE_ZERO_WRITE |
2825                                 MPA_RTR_TYPE_ZERO_READ;
2826
2827         spin_lock_init(&p_hwfn->p_rdma_info->iwarp.qp_lock);
2828         INIT_LIST_HEAD(&p_hwfn->p_rdma_info->iwarp.ep_list);
2829         INIT_LIST_HEAD(&p_hwfn->p_rdma_info->iwarp.listen_list);
2830
2831         qed_spq_register_async_cb(p_hwfn, PROTOCOLID_IWARP,
2832                                   qed_iwarp_async_event);
2833         qed_ooo_setup(p_hwfn);
2834
2835         return qed_iwarp_ll2_start(p_hwfn, params, rcv_wnd_size);
2836 }
2837
2838 int qed_iwarp_stop(struct qed_hwfn *p_hwfn)
2839 {
2840         int rc;
2841
2842         qed_iwarp_free_prealloc_ep(p_hwfn);
2843         rc = qed_iwarp_wait_for_all_cids(p_hwfn);
2844         if (rc)
2845                 return rc;
2846
2847         return qed_iwarp_ll2_stop(p_hwfn);
2848 }
2849
2850 static void qed_iwarp_qp_in_error(struct qed_hwfn *p_hwfn,
2851                                   struct qed_iwarp_ep *ep,
2852                                   u8 fw_return_code)
2853 {
2854         struct qed_iwarp_cm_event_params params;
2855
2856         qed_iwarp_modify_qp(p_hwfn, ep->qp, QED_IWARP_QP_STATE_ERROR, true);
2857
2858         params.event = QED_IWARP_EVENT_CLOSE;
2859         params.ep_context = ep;
2860         params.cm_info = &ep->cm_info;
2861         params.status = (fw_return_code == IWARP_QP_IN_ERROR_GOOD_CLOSE) ?
2862                          0 : -ECONNRESET;
2863
2864         /* paired with READ_ONCE in destroy_qp */
2865         smp_store_release(&ep->state, QED_IWARP_EP_CLOSED);
2866
2867         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
2868         list_del(&ep->list_entry);
2869         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
2870
2871         ep->event_cb(ep->cb_context, &params);
2872 }
2873
2874 static void qed_iwarp_exception_received(struct qed_hwfn *p_hwfn,
2875                                          struct qed_iwarp_ep *ep,
2876                                          int fw_ret_code)
2877 {
2878         struct qed_iwarp_cm_event_params params;
2879         bool event_cb = false;
2880
2881         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "EP(0x%x) fw_ret_code=%d\n",
2882                    ep->cid, fw_ret_code);
2883
2884         switch (fw_ret_code) {
2885         case IWARP_EXCEPTION_DETECTED_LLP_CLOSED:
2886                 params.status = 0;
2887                 params.event = QED_IWARP_EVENT_DISCONNECT;
2888                 event_cb = true;
2889                 break;
2890         case IWARP_EXCEPTION_DETECTED_LLP_RESET:
2891                 params.status = -ECONNRESET;
2892                 params.event = QED_IWARP_EVENT_DISCONNECT;
2893                 event_cb = true;
2894                 break;
2895         case IWARP_EXCEPTION_DETECTED_RQ_EMPTY:
2896                 params.event = QED_IWARP_EVENT_RQ_EMPTY;
2897                 event_cb = true;
2898                 break;
2899         case IWARP_EXCEPTION_DETECTED_IRQ_FULL:
2900                 params.event = QED_IWARP_EVENT_IRQ_FULL;
2901                 event_cb = true;
2902                 break;
2903         case IWARP_EXCEPTION_DETECTED_LLP_TIMEOUT:
2904                 params.event = QED_IWARP_EVENT_LLP_TIMEOUT;
2905                 event_cb = true;
2906                 break;
2907         case IWARP_EXCEPTION_DETECTED_REMOTE_PROTECTION_ERROR:
2908                 params.event = QED_IWARP_EVENT_REMOTE_PROTECTION_ERROR;
2909                 event_cb = true;
2910                 break;
2911         case IWARP_EXCEPTION_DETECTED_CQ_OVERFLOW:
2912                 params.event = QED_IWARP_EVENT_CQ_OVERFLOW;
2913                 event_cb = true;
2914                 break;
2915         case IWARP_EXCEPTION_DETECTED_LOCAL_CATASTROPHIC:
2916                 params.event = QED_IWARP_EVENT_QP_CATASTROPHIC;
2917                 event_cb = true;
2918                 break;
2919         case IWARP_EXCEPTION_DETECTED_LOCAL_ACCESS_ERROR:
2920                 params.event = QED_IWARP_EVENT_LOCAL_ACCESS_ERROR;
2921                 event_cb = true;
2922                 break;
2923         case IWARP_EXCEPTION_DETECTED_REMOTE_OPERATION_ERROR:
2924                 params.event = QED_IWARP_EVENT_REMOTE_OPERATION_ERROR;
2925                 event_cb = true;
2926                 break;
2927         case IWARP_EXCEPTION_DETECTED_TERMINATE_RECEIVED:
2928                 params.event = QED_IWARP_EVENT_TERMINATE_RECEIVED;
2929                 event_cb = true;
2930                 break;
2931         default:
2932                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2933                            "Unhandled exception received...fw_ret_code=%d\n",
2934                            fw_ret_code);
2935                 break;
2936         }
2937
2938         if (event_cb) {
2939                 params.ep_context = ep;
2940                 params.cm_info = &ep->cm_info;
2941                 ep->event_cb(ep->cb_context, &params);
2942         }
2943 }
2944
2945 static void
2946 qed_iwarp_tcp_connect_unsuccessful(struct qed_hwfn *p_hwfn,
2947                                    struct qed_iwarp_ep *ep, u8 fw_return_code)
2948 {
2949         struct qed_iwarp_cm_event_params params;
2950
2951         memset(&params, 0, sizeof(params));
2952         params.event = QED_IWARP_EVENT_ACTIVE_COMPLETE;
2953         params.ep_context = ep;
2954         params.cm_info = &ep->cm_info;
2955         /* paired with READ_ONCE in destroy_qp */
2956         smp_store_release(&ep->state, QED_IWARP_EP_CLOSED);
2957
2958         switch (fw_return_code) {
2959         case IWARP_CONN_ERROR_TCP_CONNECT_INVALID_PACKET:
2960                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2961                            "%s(0x%x) TCP connect got invalid packet\n",
2962                            QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2963                 params.status = -ECONNRESET;
2964                 break;
2965         case IWARP_CONN_ERROR_TCP_CONNECTION_RST:
2966                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
2967                            "%s(0x%x) TCP Connection Reset\n",
2968                            QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2969                 params.status = -ECONNRESET;
2970                 break;
2971         case IWARP_CONN_ERROR_TCP_CONNECT_TIMEOUT:
2972                 DP_NOTICE(p_hwfn, "%s(0x%x) TCP timeout\n",
2973                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2974                 params.status = -EBUSY;
2975                 break;
2976         case IWARP_CONN_ERROR_MPA_NOT_SUPPORTED_VER:
2977                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA not supported VER\n",
2978                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2979                 params.status = -ECONNREFUSED;
2980                 break;
2981         case IWARP_CONN_ERROR_MPA_INVALID_PACKET:
2982                 DP_NOTICE(p_hwfn, "%s(0x%x) MPA Invalid Packet\n",
2983                           QED_IWARP_CONNECT_MODE_STRING(ep), ep->tcp_cid);
2984                 params.status = -ECONNRESET;
2985                 break;
2986         default:
2987                 DP_ERR(p_hwfn,
2988                        "%s(0x%x) Unexpected return code tcp connect: %d\n",
2989                        QED_IWARP_CONNECT_MODE_STRING(ep),
2990                        ep->tcp_cid, fw_return_code);
2991                 params.status = -ECONNRESET;
2992                 break;
2993         }
2994
2995         if (ep->connect_mode == TCP_CONNECT_PASSIVE) {
2996                 ep->tcp_cid = QED_IWARP_INVALID_TCP_CID;
2997                 qed_iwarp_return_ep(p_hwfn, ep);
2998         } else {
2999                 ep->event_cb(ep->cb_context, &params);
3000                 spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3001                 list_del(&ep->list_entry);
3002                 spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3003         }
3004 }
3005
3006 static void
3007 qed_iwarp_connect_complete(struct qed_hwfn *p_hwfn,
3008                            struct qed_iwarp_ep *ep, u8 fw_return_code)
3009 {
3010         u8 ll2_syn_handle = p_hwfn->p_rdma_info->iwarp.ll2_syn_handle;
3011
3012         if (ep->connect_mode == TCP_CONNECT_PASSIVE) {
3013                 /* Done with the SYN packet, post back to ll2 rx */
3014                 qed_iwarp_ll2_post_rx(p_hwfn, ep->syn, ll2_syn_handle);
3015
3016                 ep->syn = NULL;
3017
3018                 /* If connect failed - upper layer doesn't know about it */
3019                 if (fw_return_code == RDMA_RETURN_OK)
3020                         qed_iwarp_mpa_received(p_hwfn, ep);
3021                 else
3022                         qed_iwarp_tcp_connect_unsuccessful(p_hwfn, ep,
3023                                                            fw_return_code);
3024         } else {
3025                 if (fw_return_code == RDMA_RETURN_OK)
3026                         qed_iwarp_mpa_offload(p_hwfn, ep);
3027                 else
3028                         qed_iwarp_tcp_connect_unsuccessful(p_hwfn, ep,
3029                                                            fw_return_code);
3030         }
3031 }
3032
3033 static inline bool
3034 qed_iwarp_check_ep_ok(struct qed_hwfn *p_hwfn, struct qed_iwarp_ep *ep)
3035 {
3036         if (!ep || (ep->sig != QED_EP_SIG)) {
3037                 DP_ERR(p_hwfn, "ERROR ON ASYNC ep=%p\n", ep);
3038                 return false;
3039         }
3040
3041         return true;
3042 }
3043
3044 static int qed_iwarp_async_event(struct qed_hwfn *p_hwfn, u8 fw_event_code,
3045                                  __le16 echo, union event_ring_data *data,
3046                                  u8 fw_return_code)
3047 {
3048         struct qed_rdma_events events = p_hwfn->p_rdma_info->events;
3049         struct regpair *fw_handle = &data->rdma_data.async_handle;
3050         struct qed_iwarp_ep *ep = NULL;
3051         u16 srq_offset;
3052         u16 srq_id;
3053         u16 cid;
3054
3055         ep = (struct qed_iwarp_ep *)(uintptr_t)HILO_64(fw_handle->hi,
3056                                                        fw_handle->lo);
3057
3058         switch (fw_event_code) {
3059         case IWARP_EVENT_TYPE_ASYNC_CONNECT_COMPLETE:
3060                 /* Async completion after TCP 3-way handshake */
3061                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3062                         return -EINVAL;
3063                 DP_VERBOSE(p_hwfn,
3064                            QED_MSG_RDMA,
3065                            "EP(0x%x) IWARP_EVENT_TYPE_ASYNC_CONNECT_COMPLETE fw_ret_code=%d\n",
3066                            ep->tcp_cid, fw_return_code);
3067                 qed_iwarp_connect_complete(p_hwfn, ep, fw_return_code);
3068                 break;
3069         case IWARP_EVENT_TYPE_ASYNC_EXCEPTION_DETECTED:
3070                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3071                         return -EINVAL;
3072                 DP_VERBOSE(p_hwfn,
3073                            QED_MSG_RDMA,
3074                            "QP(0x%x) IWARP_EVENT_TYPE_ASYNC_EXCEPTION_DETECTED fw_ret_code=%d\n",
3075                            ep->cid, fw_return_code);
3076                 qed_iwarp_exception_received(p_hwfn, ep, fw_return_code);
3077                 break;
3078         case IWARP_EVENT_TYPE_ASYNC_QP_IN_ERROR_STATE:
3079                 /* Async completion for Close Connection ramrod */
3080                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3081                         return -EINVAL;
3082                 DP_VERBOSE(p_hwfn,
3083                            QED_MSG_RDMA,
3084                            "QP(0x%x) IWARP_EVENT_TYPE_ASYNC_QP_IN_ERROR_STATE fw_ret_code=%d\n",
3085                            ep->cid, fw_return_code);
3086                 qed_iwarp_qp_in_error(p_hwfn, ep, fw_return_code);
3087                 break;
3088         case IWARP_EVENT_TYPE_ASYNC_ENHANCED_MPA_REPLY_ARRIVED:
3089                 /* Async event for active side only */
3090                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3091                         return -EINVAL;
3092                 DP_VERBOSE(p_hwfn,
3093                            QED_MSG_RDMA,
3094                            "QP(0x%x) IWARP_EVENT_TYPE_ASYNC_MPA_HANDSHAKE_MPA_REPLY_ARRIVED fw_ret_code=%d\n",
3095                            ep->cid, fw_return_code);
3096                 qed_iwarp_mpa_reply_arrived(p_hwfn, ep);
3097                 break;
3098         case IWARP_EVENT_TYPE_ASYNC_MPA_HANDSHAKE_COMPLETE:
3099                 if (!qed_iwarp_check_ep_ok(p_hwfn, ep))
3100                         return -EINVAL;
3101                 DP_VERBOSE(p_hwfn,
3102                            QED_MSG_RDMA,
3103                            "QP(0x%x) IWARP_EVENT_TYPE_ASYNC_MPA_HANDSHAKE_COMPLETE fw_ret_code=%d\n",
3104                            ep->cid, fw_return_code);
3105                 qed_iwarp_mpa_complete(p_hwfn, ep, fw_return_code);
3106                 break;
3107         case IWARP_EVENT_TYPE_ASYNC_CID_CLEANED:
3108                 cid = (u16)le32_to_cpu(fw_handle->lo);
3109                 DP_VERBOSE(p_hwfn, QED_MSG_RDMA,
3110                            "(0x%x)IWARP_EVENT_TYPE_ASYNC_CID_CLEANED\n", cid);
3111                 qed_iwarp_cid_cleaned(p_hwfn, cid);
3112
3113                 break;
3114         case IWARP_EVENT_TYPE_ASYNC_SRQ_EMPTY:
3115                 DP_NOTICE(p_hwfn, "IWARP_EVENT_TYPE_ASYNC_SRQ_EMPTY\n");
3116                 srq_offset = p_hwfn->p_rdma_info->srq_id_offset;
3117                 /* FW assigns value that is no greater than u16 */
3118                 srq_id = ((u16)le32_to_cpu(fw_handle->lo)) - srq_offset;
3119                 events.affiliated_event(events.context,
3120                                         QED_IWARP_EVENT_SRQ_EMPTY,
3121                                         &srq_id);
3122                 break;
3123         case IWARP_EVENT_TYPE_ASYNC_SRQ_LIMIT:
3124                 DP_NOTICE(p_hwfn, "IWARP_EVENT_TYPE_ASYNC_SRQ_LIMIT\n");
3125                 srq_offset = p_hwfn->p_rdma_info->srq_id_offset;
3126                 /* FW assigns value that is no greater than u16 */
3127                 srq_id = ((u16)le32_to_cpu(fw_handle->lo)) - srq_offset;
3128                 events.affiliated_event(events.context,
3129                                         QED_IWARP_EVENT_SRQ_LIMIT,
3130                                         &srq_id);
3131                 break;
3132         case IWARP_EVENT_TYPE_ASYNC_CQ_OVERFLOW:
3133                 DP_NOTICE(p_hwfn, "IWARP_EVENT_TYPE_ASYNC_CQ_OVERFLOW\n");
3134
3135                 p_hwfn->p_rdma_info->events.affiliated_event(
3136                         p_hwfn->p_rdma_info->events.context,
3137                         QED_IWARP_EVENT_CQ_OVERFLOW,
3138                         (void *)fw_handle);
3139                 break;
3140         default:
3141                 DP_ERR(p_hwfn, "Received unexpected async iwarp event %d\n",
3142                        fw_event_code);
3143                 return -EINVAL;
3144         }
3145         return 0;
3146 }
3147
3148 int
3149 qed_iwarp_create_listen(void *rdma_cxt,
3150                         struct qed_iwarp_listen_in *iparams,
3151                         struct qed_iwarp_listen_out *oparams)
3152 {
3153         struct qed_hwfn *p_hwfn = rdma_cxt;
3154         struct qed_iwarp_listener *listener;
3155
3156         listener = kzalloc(sizeof(*listener), GFP_KERNEL);
3157         if (!listener)
3158                 return -ENOMEM;
3159
3160         listener->ip_version = iparams->ip_version;
3161         memcpy(listener->ip_addr, iparams->ip_addr, sizeof(listener->ip_addr));
3162         listener->port = iparams->port;
3163         listener->vlan = iparams->vlan;
3164
3165         listener->event_cb = iparams->event_cb;
3166         listener->cb_context = iparams->cb_context;
3167         listener->max_backlog = iparams->max_backlog;
3168         oparams->handle = listener;
3169
3170         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3171         list_add_tail(&listener->list_entry,
3172                       &p_hwfn->p_rdma_info->iwarp.listen_list);
3173         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3174
3175         DP_VERBOSE(p_hwfn,
3176                    QED_MSG_RDMA,
3177                    "callback=%p handle=%p ip=%x:%x:%x:%x port=0x%x vlan=0x%x\n",
3178                    listener->event_cb,
3179                    listener,
3180                    listener->ip_addr[0],
3181                    listener->ip_addr[1],
3182                    listener->ip_addr[2],
3183                    listener->ip_addr[3], listener->port, listener->vlan);
3184
3185         return 0;
3186 }
3187
3188 int qed_iwarp_destroy_listen(void *rdma_cxt, void *handle)
3189 {
3190         struct qed_iwarp_listener *listener = handle;
3191         struct qed_hwfn *p_hwfn = rdma_cxt;
3192
3193         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "handle=%p\n", handle);
3194
3195         spin_lock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3196         list_del(&listener->list_entry);
3197         spin_unlock_bh(&p_hwfn->p_rdma_info->iwarp.iw_lock);
3198
3199         kfree(listener);
3200
3201         return 0;
3202 }
3203
3204 int qed_iwarp_send_rtr(void *rdma_cxt, struct qed_iwarp_send_rtr_in *iparams)
3205 {
3206         struct qed_hwfn *p_hwfn = rdma_cxt;
3207         struct qed_sp_init_data init_data;
3208         struct qed_spq_entry *p_ent;
3209         struct qed_iwarp_ep *ep;
3210         struct qed_rdma_qp *qp;
3211         int rc;
3212
3213         ep = iparams->ep_context;
3214         if (!ep) {
3215                 DP_ERR(p_hwfn, "Ep Context receive in send_rtr is NULL\n");
3216                 return -EINVAL;
3217         }
3218
3219         qp = ep->qp;
3220
3221         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "QP(0x%x) EP(0x%x)\n",
3222                    qp->icid, ep->tcp_cid);
3223
3224         memset(&init_data, 0, sizeof(init_data));
3225         init_data.cid = qp->icid;
3226         init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
3227         init_data.comp_mode = QED_SPQ_MODE_CB;
3228
3229         rc = qed_sp_init_request(p_hwfn, &p_ent,
3230                                  IWARP_RAMROD_CMD_ID_MPA_OFFLOAD_SEND_RTR,
3231                                  PROTOCOLID_IWARP, &init_data);
3232
3233         if (rc)
3234                 return rc;
3235
3236         rc = qed_spq_post(p_hwfn, p_ent, NULL);
3237
3238         DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "rc = 0x%x\n", rc);
3239
3240         return rc;
3241 }
3242
3243 void
3244 qed_iwarp_query_qp(struct qed_rdma_qp *qp,
3245                    struct qed_rdma_query_qp_out_params *out_params)
3246 {
3247         out_params->state = qed_iwarp2roce_state(qp->iwarp_state);
3248 }