RDMA: Validate grh_required when handling AVs
authorArtemy Kovalyov <artemyko@mellanox.com>
Wed, 4 Jul 2018 12:57:50 +0000 (15:57 +0300)
committerJason Gunthorpe <jgg@mellanox.com>
Tue, 10 Jul 2018 17:13:04 +0000 (11:13 -0600)
Extend the existing grh_required flag to check when AV's are handled that
a GRH is present.

Since we don't want to do query_port during the AV checks for performance
reasons move the flag into the immutable_data.

Signed-off-by: Artemy Kovalyov <artemyko@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
drivers/infiniband/core/sa_query.c
drivers/infiniband/core/verbs.c
drivers/infiniband/hw/mlx5/main.c
include/rdma/ib_verbs.h

index fdfdbb2..7b794a1 100644 (file)
@@ -2276,6 +2276,7 @@ static void update_sm_ah(struct work_struct *work)
        struct ib_sa_sm_ah *new_ah;
        struct ib_port_attr port_attr;
        struct rdma_ah_attr   ah_attr;
+       bool grh_required;
 
        if (ib_query_port(port->agent->device, port->port_num, &port_attr)) {
                pr_warn("Couldn't query port\n");
@@ -2301,6 +2302,9 @@ static void update_sm_ah(struct work_struct *work)
        rdma_ah_set_sl(&ah_attr, port_attr.sm_sl);
        rdma_ah_set_port_num(&ah_attr, port->port_num);
 
+       grh_required = rdma_is_grh_required(port->agent->device,
+                                           port->port_num);
+
        /*
         * The OPA sm_lid of 0xFFFF needs special handling so that it can be
         * differentiated from a permissive LID of 0xFFFF.  We set the
@@ -2308,11 +2312,11 @@ static void update_sm_ah(struct work_struct *work)
         * address handle appropriately
         */
        if (ah_attr.type == RDMA_AH_ATTR_TYPE_OPA &&
-           (port_attr.grh_required ||
+           (grh_required ||
             port_attr.sm_lid == be16_to_cpu(IB_LID_PERMISSIVE)))
                rdma_ah_set_make_grd(&ah_attr, true);
 
-       if (ah_attr.type == RDMA_AH_ATTR_TYPE_IB && port_attr.grh_required) {
+       if (ah_attr.type == RDMA_AH_ATTR_TYPE_IB && grh_required) {
                rdma_ah_set_ah_flags(&ah_attr, IB_AH_GRH);
                rdma_ah_set_subnet_prefix(&ah_attr,
                                          cpu_to_be64(port_attr.subnet_prefix));
index 1bb6b6f..b6ceb6f 100644 (file)
@@ -390,7 +390,8 @@ static int rdma_check_ah_attr(struct ib_device *device,
        if (!rdma_is_port_valid(device, ah_attr->port_num))
                return -EINVAL;
 
-       if (ah_attr->type == RDMA_AH_ATTR_TYPE_ROCE &&
+       if ((rdma_is_grh_required(device, ah_attr->port_num) ||
+            ah_attr->type == RDMA_AH_ATTR_TYPE_ROCE) &&
            !(ah_attr->ah_flags & IB_AH_GRH))
                return -EINVAL;
 
index d1f1bee..b7f94bc 100644 (file)
@@ -1220,7 +1220,6 @@ static int mlx5_query_hca_port(struct ib_device *ibdev, u8 port,
        props->qkey_viol_cntr   = rep->qkey_violation_counter;
        props->subnet_timeout   = rep->subnet_timeout;
        props->init_type_reply  = rep->init_type_reply;
-       props->grh_required     = rep->grh_required;
 
        err = mlx5_query_port_link_width_oper(mdev, &ib_link_width_oper, port);
        if (err)
@@ -4462,7 +4461,8 @@ static void destroy_dev_resources(struct mlx5_ib_resources *devr)
                cancel_work_sync(&devr->ports[port].pkey_change_work);
 }
 
-static u32 get_core_cap_flags(struct ib_device *ibdev)
+static u32 get_core_cap_flags(struct ib_device *ibdev,
+                             struct mlx5_hca_vport_context *rep)
 {
        struct mlx5_ib_dev *dev = to_mdev(ibdev);
        enum rdma_link_layer ll = mlx5_ib_port_link_layer(ibdev, 1);
@@ -4471,11 +4471,14 @@ static u32 get_core_cap_flags(struct ib_device *ibdev)
        bool raw_support = !mlx5_core_mp_enabled(dev->mdev);
        u32 ret = 0;
 
+       if (rep->grh_required)
+               ret |= RDMA_CORE_CAP_IB_GRH_REQUIRED;
+
        if (ll == IB_LINK_LAYER_INFINIBAND)
-               return RDMA_CORE_PORT_IBA_IB;
+               return ret | RDMA_CORE_PORT_IBA_IB;
 
        if (raw_support)
-               ret = RDMA_CORE_PORT_RAW_PACKET;
+               ret |= RDMA_CORE_PORT_RAW_PACKET;
 
        if (!(l3_type_cap & MLX5_ROCE_L3_TYPE_IPV4_CAP))
                return ret;
@@ -4498,17 +4501,23 @@ static int mlx5_port_immutable(struct ib_device *ibdev, u8 port_num,
        struct ib_port_attr attr;
        struct mlx5_ib_dev *dev = to_mdev(ibdev);
        enum rdma_link_layer ll = mlx5_ib_port_link_layer(ibdev, port_num);
+       struct mlx5_hca_vport_context rep = {0};
        int err;
 
-       immutable->core_cap_flags = get_core_cap_flags(ibdev);
-
        err = ib_query_port(ibdev, port_num, &attr);
        if (err)
                return err;
 
+       if (ll == IB_LINK_LAYER_INFINIBAND) {
+               err = mlx5_query_hca_vport_context(dev->mdev, 0, port_num, 0,
+                                                  &rep);
+               if (err)
+                       return err;
+       }
+
        immutable->pkey_tbl_len = attr.pkey_tbl_len;
        immutable->gid_tbl_len = attr.gid_tbl_len;
-       immutable->core_cap_flags = get_core_cap_flags(ibdev);
+       immutable->core_cap_flags = get_core_cap_flags(ibdev, &rep);
        if ((ll == IB_LINK_LAYER_INFINIBAND) || MLX5_CAP_GEN(dev->mdev, roce))
                immutable->max_mad_size = IB_MGMT_MAD_SIZE;
 
index 98e0257..b523298 100644 (file)
@@ -529,6 +529,7 @@ static inline struct rdma_hw_stats *rdma_alloc_hw_stats_struct(
 #define RDMA_CORE_CAP_AF_IB             0x00001000
 #define RDMA_CORE_CAP_ETH_AH            0x00002000
 #define RDMA_CORE_CAP_OPA_AH            0x00004000
+#define RDMA_CORE_CAP_IB_GRH_REQUIRED   0x00008000
 
 /* Protocol                             0xFFF00000 */
 #define RDMA_CORE_CAP_PROT_IB           0x00100000
@@ -538,6 +539,10 @@ static inline struct rdma_hw_stats *rdma_alloc_hw_stats_struct(
 #define RDMA_CORE_CAP_PROT_RAW_PACKET   0x01000000
 #define RDMA_CORE_CAP_PROT_USNIC        0x02000000
 
+#define RDMA_CORE_PORT_IB_GRH_REQUIRED (RDMA_CORE_CAP_IB_GRH_REQUIRED \
+                                       | RDMA_CORE_CAP_PROT_ROCE     \
+                                       | RDMA_CORE_CAP_PROT_ROCE_UDP_ENCAP)
+
 #define RDMA_CORE_PORT_IBA_IB          (RDMA_CORE_CAP_PROT_IB  \
                                        | RDMA_CORE_CAP_IB_MAD \
                                        | RDMA_CORE_CAP_IB_SMI \
@@ -570,7 +575,6 @@ struct ib_port_attr {
        enum ib_mtu             max_mtu;
        enum ib_mtu             active_mtu;
        int                     gid_tbl_len;
-       unsigned int            grh_required:1;
        unsigned int            ip_gids:1;
        /* This is the value from PortInfo CapabilityMask, defined by IBA */
        u32                     port_cap_flags;
@@ -2771,6 +2775,13 @@ static inline int rdma_is_port_valid(const struct ib_device *device,
                port <= rdma_end_port(device));
 }
 
+static inline bool rdma_is_grh_required(const struct ib_device *device,
+                                       u8 port_num)
+{
+       return device->port_immutable[port_num].core_cap_flags &
+               RDMA_CORE_PORT_IB_GRH_REQUIRED;
+}
+
 static inline bool rdma_protocol_ib(const struct ib_device *device, u8 port_num)
 {
        return device->port_immutable[port_num].core_cap_flags & RDMA_CORE_CAP_PROT_IB;