1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
3 * Copyright (c) 2016 Mellanox Technologies Ltd. All rights reserved.
4 * Copyright (c) 2015 System Fabric Works, Inc. All rights reserved.
7 #include <linux/skbuff.h>
12 /* check that QP matches packet opcode type and is in a valid state */
13 static int check_type_state(struct rxe_dev *rxe, struct rxe_pkt_info *pkt,
16 unsigned int pkt_type;
18 if (unlikely(!qp->valid))
21 pkt_type = pkt->opcode & 0xe0;
23 switch (qp_type(qp)) {
25 if (unlikely(pkt_type != IB_OPCODE_RC))
29 if (unlikely(pkt_type != IB_OPCODE_UC))
34 if (unlikely(pkt_type != IB_OPCODE_UD))
41 spin_lock_bh(&qp->state_lock);
42 if (pkt->mask & RXE_REQ_MASK) {
43 if (unlikely(qp_state(qp) < IB_QPS_RTR)) {
44 spin_unlock_bh(&qp->state_lock);
48 if (unlikely(qp_state(qp) < IB_QPS_RTS)) {
49 spin_unlock_bh(&qp->state_lock);
53 spin_unlock_bh(&qp->state_lock);
58 static void set_bad_pkey_cntr(struct rxe_port *port)
60 spin_lock_bh(&port->port_lock);
61 port->attr.bad_pkey_cntr = min((u32)0xffff,
62 port->attr.bad_pkey_cntr + 1);
63 spin_unlock_bh(&port->port_lock);
66 static void set_qkey_viol_cntr(struct rxe_port *port)
68 spin_lock_bh(&port->port_lock);
69 port->attr.qkey_viol_cntr = min((u32)0xffff,
70 port->attr.qkey_viol_cntr + 1);
71 spin_unlock_bh(&port->port_lock);
74 static int check_keys(struct rxe_dev *rxe, struct rxe_pkt_info *pkt,
75 u32 qpn, struct rxe_qp *qp)
77 struct rxe_port *port = &rxe->port;
78 u16 pkey = bth_pkey(pkt);
82 if (!pkey_match(pkey, IB_DEFAULT_PKEY_FULL)) {
83 set_bad_pkey_cntr(port);
87 if (qp_type(qp) == IB_QPT_UD || qp_type(qp) == IB_QPT_GSI) {
88 u32 qkey = (qpn == 1) ? GSI_QKEY : qp->attr.qkey;
90 if (unlikely(deth_qkey(pkt) != qkey)) {
91 set_qkey_viol_cntr(port);
99 static int check_addr(struct rxe_dev *rxe, struct rxe_pkt_info *pkt,
102 struct sk_buff *skb = PKT_TO_SKB(pkt);
104 if (qp_type(qp) != IB_QPT_RC && qp_type(qp) != IB_QPT_UC)
107 if (unlikely(pkt->port_num != qp->attr.port_num))
110 if (skb->protocol == htons(ETH_P_IP)) {
111 struct in_addr *saddr =
112 &qp->pri_av.sgid_addr._sockaddr_in.sin_addr;
113 struct in_addr *daddr =
114 &qp->pri_av.dgid_addr._sockaddr_in.sin_addr;
116 if ((ip_hdr(skb)->daddr != saddr->s_addr) ||
117 (ip_hdr(skb)->saddr != daddr->s_addr))
120 } else if (skb->protocol == htons(ETH_P_IPV6)) {
121 struct in6_addr *saddr =
122 &qp->pri_av.sgid_addr._sockaddr_in6.sin6_addr;
123 struct in6_addr *daddr =
124 &qp->pri_av.dgid_addr._sockaddr_in6.sin6_addr;
126 if (memcmp(&ipv6_hdr(skb)->daddr, saddr, sizeof(*saddr)) ||
127 memcmp(&ipv6_hdr(skb)->saddr, daddr, sizeof(*daddr)))
134 static int hdr_check(struct rxe_pkt_info *pkt)
136 struct rxe_dev *rxe = pkt->rxe;
137 struct rxe_port *port = &rxe->port;
138 struct rxe_qp *qp = NULL;
139 u32 qpn = bth_qpn(pkt);
143 if (unlikely(bth_tver(pkt) != BTH_TVER))
146 if (unlikely(qpn == 0))
149 if (qpn != IB_MULTICAST_QPN) {
150 index = (qpn == 1) ? port->qp_gsi_index : qpn;
152 qp = rxe_pool_get_index(&rxe->qp_pool, index);
156 err = check_type_state(rxe, pkt, qp);
160 err = check_addr(rxe, pkt, qp);
164 err = check_keys(rxe, pkt, qpn, qp);
168 if (unlikely((pkt->mask & RXE_GRH_MASK) == 0))
181 static inline void rxe_rcv_pkt(struct rxe_pkt_info *pkt, struct sk_buff *skb)
183 if (pkt->mask & RXE_REQ_MASK)
184 rxe_resp_queue_pkt(pkt->qp, skb);
186 rxe_comp_queue_pkt(pkt->qp, skb);
189 static void rxe_rcv_mcast_pkt(struct rxe_dev *rxe, struct sk_buff *skb)
191 struct rxe_pkt_info *pkt = SKB_TO_PKT(skb);
198 if (skb->protocol == htons(ETH_P_IP))
199 ipv6_addr_set_v4mapped(ip_hdr(skb)->daddr,
200 (struct in6_addr *)&dgid);
201 else if (skb->protocol == htons(ETH_P_IPV6))
202 memcpy(&dgid, &ipv6_hdr(skb)->daddr, sizeof(dgid));
204 /* lookup mcast group corresponding to mgid, takes a ref */
205 mcg = rxe_lookup_mcg(rxe, &dgid);
207 goto drop; /* mcast group not registered */
209 spin_lock_bh(&rxe->mcg_lock);
211 /* this is unreliable datagram service so we let
212 * failures to deliver a multicast packet to a
213 * single QP happen and just move on and try
214 * the rest of them on the list
216 list_for_each_entry(mca, &mcg->qp_list, qp_list) {
219 /* validate qp for incoming packet */
220 err = check_type_state(rxe, pkt, qp);
224 err = check_keys(rxe, pkt, bth_qpn(pkt), qp);
228 /* for all but the last QP create a new clone of the
229 * skb and pass to the QP. Pass the original skb to
230 * the last QP in the list.
232 if (mca->qp_list.next != &mcg->qp_list) {
233 struct sk_buff *cskb;
234 struct rxe_pkt_info *cpkt;
236 cskb = skb_clone(skb, GFP_ATOMIC);
240 if (WARN_ON(!ib_device_try_get(&rxe->ib_dev))) {
245 cpkt = SKB_TO_PKT(cskb);
248 rxe_rcv_pkt(cpkt, cskb);
252 rxe_rcv_pkt(pkt, skb);
253 skb = NULL; /* mark consumed */
257 spin_unlock_bh(&rxe->mcg_lock);
259 kref_put(&mcg->ref_cnt, rxe_cleanup_mcg);
264 /* This only occurs if one of the checks fails on the last
265 * QP in the list above
270 ib_device_put(&rxe->ib_dev);
274 * rxe_chk_dgid - validate destination IP address
275 * @rxe: rxe device that received packet
276 * @skb: the received packet buffer
278 * Accept any loopback packets
279 * Extract IP address from packet and
280 * Accept if multicast packet
281 * Accept if matches an SGID table entry
283 static int rxe_chk_dgid(struct rxe_dev *rxe, struct sk_buff *skb)
285 struct rxe_pkt_info *pkt = SKB_TO_PKT(skb);
286 const struct ib_gid_attr *gid_attr;
290 if (pkt->mask & RXE_LOOPBACK_MASK)
293 if (skb->protocol == htons(ETH_P_IP)) {
294 ipv6_addr_set_v4mapped(ip_hdr(skb)->daddr,
295 (struct in6_addr *)&dgid);
298 pdgid = (union ib_gid *)&ipv6_hdr(skb)->daddr;
301 if (rdma_is_multicast_addr((struct in6_addr *)pdgid))
304 gid_attr = rdma_find_gid_by_port(&rxe->ib_dev, pdgid,
305 IB_GID_TYPE_ROCE_UDP_ENCAP,
307 if (IS_ERR(gid_attr))
308 return PTR_ERR(gid_attr);
310 rdma_put_gid_attr(gid_attr);
314 /* rxe_rcv is called from the interface driver */
315 void rxe_rcv(struct sk_buff *skb)
318 struct rxe_pkt_info *pkt = SKB_TO_PKT(skb);
319 struct rxe_dev *rxe = pkt->rxe;
321 if (unlikely(skb->len < RXE_BTH_BYTES))
324 if (rxe_chk_dgid(rxe, skb) < 0)
327 pkt->opcode = bth_opcode(pkt);
328 pkt->psn = bth_psn(pkt);
330 pkt->mask |= rxe_opcode[pkt->opcode].mask;
332 if (unlikely(skb->len < header_size(pkt)))
335 err = hdr_check(pkt);
339 err = rxe_icrc_check(skb, pkt);
343 rxe_counter_inc(rxe, RXE_CNT_RCVD_PKTS);
345 if (unlikely(bth_qpn(pkt) == IB_MULTICAST_QPN))
346 rxe_rcv_mcast_pkt(rxe, skb);
348 rxe_rcv_pkt(pkt, skb);
357 ib_device_put(&rxe->ib_dev);