IB/mlx5: Add support to dropless RQ
[platform/kernel/linux-rpi.git] / drivers / infiniband / hw / mlx5 / main.c
index a7f2e60..ad4b12d 100644 (file)
@@ -58,6 +58,7 @@
 #include <linux/mlx5/vport.h>
 #include "mlx5_ib.h"
 #include "cmd.h"
+#include <linux/mlx5/vport.h>
 
 #define DRIVER_NAME "mlx5_ib"
 #define DRIVER_VERSION "5.0-0"
@@ -97,6 +98,20 @@ mlx5_ib_port_link_layer(struct ib_device *device, u8 port_num)
        return mlx5_port_type_cap_to_rdma_ll(port_type_cap);
 }
 
+static int get_port_state(struct ib_device *ibdev,
+                         u8 port_num,
+                         enum ib_port_state *state)
+{
+       struct ib_port_attr attr;
+       int ret;
+
+       memset(&attr, 0, sizeof(attr));
+       ret = mlx5_ib_query_port(ibdev, port_num, &attr);
+       if (!ret)
+               *state = attr.state;
+       return ret;
+}
+
 static int mlx5_netdev_event(struct notifier_block *this,
                             unsigned long event, void *ptr)
 {
@@ -114,6 +129,7 @@ static int mlx5_netdev_event(struct notifier_block *this,
                write_unlock(&ibdev->roce.netdev_lock);
                break;
 
+       case NETDEV_CHANGE:
        case NETDEV_UP:
        case NETDEV_DOWN: {
                struct net_device *lag_ndev = mlx5_lag_get_roce_netdev(ibdev->mdev);
@@ -127,10 +143,23 @@ static int mlx5_netdev_event(struct notifier_block *this,
                if ((upper == ndev || (!upper && ndev == ibdev->roce.netdev))
                    && ibdev->ib_active) {
                        struct ib_event ibev = { };
+                       enum ib_port_state port_state;
+
+                       if (get_port_state(&ibdev->ib_dev, 1, &port_state))
+                               return NOTIFY_DONE;
 
+                       if (ibdev->roce.last_port_state == port_state)
+                               return NOTIFY_DONE;
+
+                       ibdev->roce.last_port_state = port_state;
                        ibev.device = &ibdev->ib_dev;
-                       ibev.event = (event == NETDEV_UP) ?
-                                    IB_EVENT_PORT_ACTIVE : IB_EVENT_PORT_ERR;
+                       if (port_state == IB_PORT_DOWN)
+                               ibev.event = IB_EVENT_PORT_ERR;
+                       else if (port_state == IB_PORT_ACTIVE)
+                               ibev.event = IB_EVENT_PORT_ACTIVE;
+                       else
+                               return NOTIFY_DONE;
+
                        ibev.element.port_num = 1;
                        ib_dispatch_event(&ibev);
                }
@@ -668,6 +697,10 @@ static int mlx5_ib_query_device(struct ib_device *ibdev,
                props->device_cap_flags |= IB_DEVICE_UD_TSO;
        }
 
+       if (MLX5_CAP_GEN(dev->mdev, rq_delay_drop) &&
+           MLX5_CAP_GEN(dev->mdev, general_notification_event))
+               props->raw_packet_caps |= IB_RAW_PACKET_CAP_DELAY_DROP;
+
        if (MLX5_CAP_GEN(dev->mdev, eth_net_offloads) &&
            MLX5_CAP_ETH(dev->mdev, scatter_fcs)) {
                /* Legacy bit to support old userspace libraries */
@@ -1187,6 +1220,45 @@ static int deallocate_uars(struct mlx5_ib_dev *dev, struct mlx5_ib_ucontext *con
        return 0;
 }
 
+static int mlx5_ib_alloc_transport_domain(struct mlx5_ib_dev *dev, u32 *tdn)
+{
+       int err;
+
+       err = mlx5_core_alloc_transport_domain(dev->mdev, tdn);
+       if (err)
+               return err;
+
+       if ((MLX5_CAP_GEN(dev->mdev, port_type) != MLX5_CAP_PORT_TYPE_ETH) ||
+           !MLX5_CAP_GEN(dev->mdev, disable_local_lb))
+               return err;
+
+       mutex_lock(&dev->lb_mutex);
+       dev->user_td++;
+
+       if (dev->user_td == 2)
+               err = mlx5_nic_vport_update_local_lb(dev->mdev, true);
+
+       mutex_unlock(&dev->lb_mutex);
+       return err;
+}
+
+static void mlx5_ib_dealloc_transport_domain(struct mlx5_ib_dev *dev, u32 tdn)
+{
+       mlx5_core_dealloc_transport_domain(dev->mdev, tdn);
+
+       if ((MLX5_CAP_GEN(dev->mdev, port_type) != MLX5_CAP_PORT_TYPE_ETH) ||
+           !MLX5_CAP_GEN(dev->mdev, disable_local_lb))
+               return;
+
+       mutex_lock(&dev->lb_mutex);
+       dev->user_td--;
+
+       if (dev->user_td < 2)
+               mlx5_nic_vport_update_local_lb(dev->mdev, false);
+
+       mutex_unlock(&dev->lb_mutex);
+}
+
 static struct ib_ucontext *mlx5_ib_alloc_ucontext(struct ib_device *ibdev,
                                                  struct ib_udata *udata)
 {
@@ -1295,8 +1367,7 @@ static struct ib_ucontext *mlx5_ib_alloc_ucontext(struct ib_device *ibdev,
        mutex_init(&context->upd_xlt_page_mutex);
 
        if (MLX5_CAP_GEN(dev->mdev, log_max_transport_domain)) {
-               err = mlx5_core_alloc_transport_domain(dev->mdev,
-                                                      &context->tdn);
+               err = mlx5_ib_alloc_transport_domain(dev, &context->tdn);
                if (err)
                        goto out_page;
        }
@@ -1362,7 +1433,7 @@ static struct ib_ucontext *mlx5_ib_alloc_ucontext(struct ib_device *ibdev,
 
 out_td:
        if (MLX5_CAP_GEN(dev->mdev, log_max_transport_domain))
-               mlx5_core_dealloc_transport_domain(dev->mdev, context->tdn);
+               mlx5_ib_dealloc_transport_domain(dev, context->tdn);
 
 out_page:
        free_page(context->upd_xlt_page);
@@ -1390,7 +1461,7 @@ static int mlx5_ib_dealloc_ucontext(struct ib_ucontext *ibcontext)
 
        bfregi = &context->bfregi;
        if (MLX5_CAP_GEN(dev->mdev, log_max_transport_domain))
-               mlx5_core_dealloc_transport_domain(dev->mdev, context->tdn);
+               mlx5_ib_dealloc_transport_domain(dev, context->tdn);
 
        free_page(context->upd_xlt_page);
        deallocate_uars(dev, context);
@@ -2685,6 +2756,24 @@ static void mlx5_ib_handle_internal_error(struct mlx5_ib_dev *ibdev)
        spin_unlock_irqrestore(&ibdev->reset_flow_resource_lock, flags);
 }
 
+static void delay_drop_handler(struct work_struct *work)
+{
+       int err;
+       struct mlx5_ib_delay_drop *delay_drop =
+               container_of(work, struct mlx5_ib_delay_drop,
+                            delay_drop_work);
+
+       mutex_lock(&delay_drop->lock);
+       err = mlx5_core_set_delay_drop(delay_drop->dev->mdev,
+                                      delay_drop->timeout);
+       if (err) {
+               mlx5_ib_warn(delay_drop->dev, "Failed to set delay drop, timeout=%u\n",
+                            delay_drop->timeout);
+               delay_drop->activate = false;
+       }
+       mutex_unlock(&delay_drop->lock);
+}
+
 static void mlx5_ib_event(struct mlx5_core_dev *dev, void *context,
                          enum mlx5_dev_event event, unsigned long param)
 {
@@ -2737,8 +2826,11 @@ static void mlx5_ib_event(struct mlx5_core_dev *dev, void *context,
                ibev.event = IB_EVENT_CLIENT_REREGISTER;
                port = (u8)param;
                break;
+       case MLX5_DEV_EVENT_DELAY_DROP_TIMEOUT:
+               schedule_work(&ibdev->delay_drop.delay_drop_work);
+               goto out;
        default:
-               return;
+               goto out;
        }
 
        ibev.device           = &ibdev->ib_dev;
@@ -2746,7 +2838,7 @@ static void mlx5_ib_event(struct mlx5_core_dev *dev, void *context,
 
        if (port < 1 || port > ibdev->num_ports) {
                mlx5_ib_warn(ibdev, "warning: event on port %d\n", port);
-               return;
+               goto out;
        }
 
        if (ibdev->ib_active)
@@ -2754,6 +2846,9 @@ static void mlx5_ib_event(struct mlx5_core_dev *dev, void *context,
 
        if (fatal)
                ibdev->ib_active = false;
+
+out:
+       return;
 }
 
 static int set_has_smi_cap(struct mlx5_ib_dev *dev)
@@ -3556,6 +3651,26 @@ mlx5_ib_alloc_rdma_netdev(struct ib_device *hca,
        return netdev;
 }
 
+static void cancel_delay_drop(struct mlx5_ib_dev *dev)
+{
+       if (!(dev->ib_dev.attrs.raw_packet_caps & IB_RAW_PACKET_CAP_DELAY_DROP))
+               return;
+
+       cancel_work_sync(&dev->delay_drop.delay_drop_work);
+}
+
+static void init_delay_drop(struct mlx5_ib_dev *dev)
+{
+       if (!(dev->ib_dev.attrs.raw_packet_caps & IB_RAW_PACKET_CAP_DELAY_DROP))
+               return;
+
+       mutex_init(&dev->delay_drop.lock);
+       dev->delay_drop.dev = dev;
+       dev->delay_drop.activate = false;
+       dev->delay_drop.timeout = MLX5_MAX_DELAY_DROP_TIMEOUT_MS * 1000;
+       INIT_WORK(&dev->delay_drop.delay_drop_work, delay_drop_handler);
+}
+
 static void *mlx5_ib_add(struct mlx5_core_dev *mdev)
 {
        struct mlx5_ib_dev *dev;
@@ -3754,6 +3869,7 @@ static void *mlx5_ib_add(struct mlx5_core_dev *mdev)
                err = mlx5_enable_eth(dev);
                if (err)
                        goto err_free_port;
+               dev->roce.last_port_state = IB_PORT_DOWN;
        }
 
        err = create_dev_resources(&dev->devr);
@@ -3770,9 +3886,13 @@ static void *mlx5_ib_add(struct mlx5_core_dev *mdev)
                        goto err_odp;
        }
 
+       err = mlx5_ib_init_cong_debugfs(dev);
+       if (err)
+               goto err_cnt;
+
        dev->mdev->priv.uar = mlx5_get_uars_page(dev->mdev);
        if (!dev->mdev->priv.uar)
-               goto err_cnt;
+               goto err_cong;
 
        err = mlx5_alloc_bfreg(dev->mdev, &dev->bfreg, false, false);
        if (err)
@@ -3790,18 +3910,25 @@ static void *mlx5_ib_add(struct mlx5_core_dev *mdev)
        if (err)
                goto err_dev;
 
+       init_delay_drop(dev);
+
        for (i = 0; i < ARRAY_SIZE(mlx5_class_attributes); i++) {
                err = device_create_file(&dev->ib_dev.dev,
                                         mlx5_class_attributes[i]);
                if (err)
-                       goto err_umrc;
+                       goto err_delay_drop;
        }
 
+       if ((MLX5_CAP_GEN(mdev, port_type) == MLX5_CAP_PORT_TYPE_ETH) &&
+           MLX5_CAP_GEN(mdev, disable_local_lb))
+               mutex_init(&dev->lb_mutex);
+
        dev->ib_active = true;
 
        return dev;
 
-err_umrc:
+err_delay_drop:
+       cancel_delay_drop(dev);
        destroy_umrc_res(dev);
 
 err_dev:
@@ -3817,6 +3944,8 @@ err_uar_page:
        mlx5_put_uars_page(dev->mdev, dev->mdev->priv.uar);
 
 err_cnt:
+       mlx5_ib_cleanup_cong_debugfs(dev);
+err_cong:
        if (MLX5_CAP_GEN(dev->mdev, max_qp_cnt))
                mlx5_ib_dealloc_counters(dev);
 
@@ -3846,11 +3975,13 @@ static void mlx5_ib_remove(struct mlx5_core_dev *mdev, void *context)
        struct mlx5_ib_dev *dev = context;
        enum rdma_link_layer ll = mlx5_ib_port_link_layer(&dev->ib_dev, 1);
 
+       cancel_delay_drop(dev);
        mlx5_remove_netdev_notifier(dev);
        ib_unregister_device(&dev->ib_dev);
        mlx5_free_bfreg(dev->mdev, &dev->fp_bfreg);
        mlx5_free_bfreg(dev->mdev, &dev->bfreg);
        mlx5_put_uars_page(dev->mdev, mdev->priv.uar);
+       mlx5_ib_cleanup_cong_debugfs(dev);
        if (MLX5_CAP_GEN(dev->mdev, max_qp_cnt))
                mlx5_ib_dealloc_counters(dev);
        destroy_umrc_res(dev);