habanalabs: modify export dmabuf API
authorOhad Sharabi <osharabi@habana.ai>
Mon, 14 Nov 2022 10:16:37 +0000 (12:16 +0200)
committerOded Gabbay <ogabbay@kernel.org>
Thu, 26 Jan 2023 08:56:21 +0000 (10:56 +0200)
A previous commit deprecated the option to export from handle, leaving
the code with no support for devices with virtual memory.

This commit modifies the export API in a way that unifies the uAPI to
user address for both cases (i.e. with and without MMU support) and add
the actual support for devices with virtual memory.

Signed-off-by: Ohad Sharabi <osharabi@habana.ai>
Reviewed-by: Oded Gabbay <ogabbay@kernel.org>
Signed-off-by: Oded Gabbay <ogabbay@kernel.org>
drivers/misc/habanalabs/common/habanalabs.h
drivers/misc/habanalabs/common/memory.c
include/uapi/misc/habanalabs.h

index e68928b..ef5a765 100644 (file)
@@ -1744,6 +1744,9 @@ struct hl_cs_counters_atomic {
  * struct hl_dmabuf_priv - a dma-buf private object.
  * @dmabuf: pointer to dma-buf object.
  * @ctx: pointer to the dma-buf owner's context.
+ * @phys_pg_pack: pointer to physical page pack if the dma-buf was exported
+ *                where virtual memory is supported.
+ * @memhash_hnode: pointer to the memhash node. this object holds the export count.
  * @device_address: physical address of the device's memory. Relevant only
  *                  if phys_pg_pack is NULL (dma-buf was exported from address).
  *                  The total size can be taken from the dmabuf object.
@@ -1751,6 +1754,8 @@ struct hl_cs_counters_atomic {
 struct hl_dmabuf_priv {
        struct dma_buf                  *dmabuf;
        struct hl_ctx                   *ctx;
+       struct hl_vm_phys_pg_pack       *phys_pg_pack;
+       struct hl_vm_hash_node          *memhash_hnode;
        uint64_t                        device_address;
 };
 
@@ -2078,12 +2083,16 @@ struct hl_cs_parser {
  *                             hl_userptr).
  * @node: node to hang on the hash table in context object.
  * @vaddr: key virtual address.
+ * @handle: memory handle for device memory allocation.
  * @ptr: value pointer (hl_vm_phys_pg_list or hl_userptr).
+ * @export_cnt: number of exports from within the VA block.
  */
 struct hl_vm_hash_node {
        struct hlist_node       node;
        u64                     vaddr;
+       u64                     handle;
        void                    *ptr;
+       int                     export_cnt;
 };
 
 /**
index ba7b2ee..04a04b5 100644 (file)
@@ -19,7 +19,9 @@ MODULE_IMPORT_NS(DMA_BUF);
 #define HL_MMU_DEBUG   0
 
 /* use small pages for supporting non-pow2 (32M/40M/48M) DRAM phys page sizes */
-#define DRAM_POOL_PAGE_SIZE SZ_8M
+#define DRAM_POOL_PAGE_SIZE    SZ_8M
+
+#define MEM_HANDLE_INVALID     ULONG_MAX
 
 static int allocate_timestamps_buffers(struct hl_fpriv *hpriv,
                        struct hl_mem_in *args, u64 *handle);
@@ -1234,6 +1236,7 @@ static int map_device_va(struct hl_ctx *ctx, struct hl_mem_in *args, u64 *device
 
        hnode->ptr = vm_type;
        hnode->vaddr = ret_vaddr;
+       hnode->handle = is_userptr ? MEM_HANDLE_INVALID : handle;
 
        mutex_lock(&ctx->mem_hash_lock);
        hash_add(ctx->mem_hash, &hnode->node, ret_vaddr);
@@ -1307,6 +1310,12 @@ static int unmap_device_va(struct hl_ctx *ctx, struct hl_mem_in *args,
                return -EINVAL;
        }
 
+       if (hnode->export_cnt) {
+               mutex_unlock(&ctx->mem_hash_lock);
+               dev_err(hdev->dev, "failed to unmap %#llx, memory is exported\n", vaddr);
+               return -EINVAL;
+       }
+
        hash_del(&hnode->node);
        mutex_unlock(&ctx->mem_hash_lock);
 
@@ -1694,19 +1703,29 @@ static struct sg_table *hl_map_dmabuf(struct dma_buf_attachment *attachment,
                                        enum dma_data_direction dir)
 {
        struct dma_buf *dma_buf = attachment->dmabuf;
+       struct hl_vm_phys_pg_pack *phys_pg_pack;
        struct hl_dmabuf_priv *hl_dmabuf;
        struct hl_device *hdev;
        struct sg_table *sgt;
 
        hl_dmabuf = dma_buf->priv;
        hdev = hl_dmabuf->ctx->hdev;
+       phys_pg_pack = hl_dmabuf->phys_pg_pack;
 
        if (!attachment->peer2peer) {
                dev_dbg(hdev->dev, "Failed to map dmabuf because p2p is disabled\n");
                return ERR_PTR(-EPERM);
        }
 
-       sgt = alloc_sgt_from_device_pages(hdev,
+       if (phys_pg_pack)
+               sgt = alloc_sgt_from_device_pages(hdev,
+                                               phys_pg_pack->pages,
+                                               phys_pg_pack->npages,
+                                               phys_pg_pack->page_size,
+                                               attachment->dev,
+                                               dir);
+       else
+               sgt = alloc_sgt_from_device_pages(hdev,
                                                &hl_dmabuf->device_address,
                                                1,
                                                hl_dmabuf->dmabuf->size,
@@ -1747,8 +1766,15 @@ static void hl_unmap_dmabuf(struct dma_buf_attachment *attachment,
 static void hl_release_dmabuf(struct dma_buf *dmabuf)
 {
        struct hl_dmabuf_priv *hl_dmabuf = dmabuf->priv;
+       struct hl_ctx *ctx = hl_dmabuf->ctx;
+
+       if (hl_dmabuf->memhash_hnode) {
+               mutex_lock(&ctx->mem_hash_lock);
+               hl_dmabuf->memhash_hnode->export_cnt--;
+               mutex_unlock(&ctx->mem_hash_lock);
+       }
 
-       hl_ctx_put(hl_dmabuf->ctx);
+       hl_ctx_put(ctx);
        kfree(hl_dmabuf);
 }
 
@@ -1797,11 +1823,8 @@ err_dma_buf_put:
        return rc;
 }
 
-static int validate_export_params(struct hl_device *hdev, u64 device_addr, u64 size)
+static int validate_export_params_common(struct hl_device *hdev, u64 device_addr, u64 size)
 {
-       struct asic_fixed_properties *prop = &hdev->asic_prop;
-       u64 bar_address;
-
        if (!IS_ALIGNED(device_addr, PAGE_SIZE)) {
                dev_dbg(hdev->dev,
                        "exported device memory address 0x%llx should be aligned to 0x%lx\n",
@@ -1816,6 +1839,19 @@ static int validate_export_params(struct hl_device *hdev, u64 device_addr, u64 s
                return -EINVAL;
        }
 
+       return 0;
+}
+
+static int validate_export_params_no_mmu(struct hl_device *hdev, u64 device_addr, u64 size)
+{
+       struct asic_fixed_properties *prop = &hdev->asic_prop;
+       u64 bar_address;
+       int rc;
+
+       rc = validate_export_params_common(hdev, device_addr, size);
+       if (rc)
+               return rc;
+
        if (device_addr < prop->dram_user_base_address ||
                                (device_addr + size) > prop->dram_end_address ||
                                (device_addr + size) < device_addr) {
@@ -1838,12 +1874,115 @@ static int validate_export_params(struct hl_device *hdev, u64 device_addr, u64 s
        return 0;
 }
 
+static int validate_export_params(struct hl_device *hdev, u64 device_addr, u64 size, u64 offset,
+                                       struct hl_vm_phys_pg_pack *phys_pg_pack)
+{
+       struct asic_fixed_properties *prop = &hdev->asic_prop;
+       u64 bar_address;
+       int i, rc;
+
+       rc = validate_export_params_common(hdev, device_addr, size);
+       if (rc)
+               return rc;
+
+       if ((offset + size) > phys_pg_pack->total_size) {
+               dev_dbg(hdev->dev, "offset %#llx and size %#llx exceed total map size %#llx\n",
+                               offset, size, phys_pg_pack->total_size);
+               return -EINVAL;
+       }
+
+       for (i = 0 ; i < phys_pg_pack->npages ; i++) {
+
+               bar_address = hdev->dram_pci_bar_start +
+                                       (phys_pg_pack->pages[i] - prop->dram_base_address);
+
+               if ((bar_address + phys_pg_pack->page_size) >
+                               (hdev->dram_pci_bar_start + prop->dram_pci_bar_size) ||
+                               (bar_address + phys_pg_pack->page_size) < bar_address) {
+                       dev_dbg(hdev->dev,
+                               "DRAM memory range 0x%llx (+0x%x) is outside of PCI BAR boundaries\n",
+                                       phys_pg_pack->pages[i],
+                                       phys_pg_pack->page_size);
+
+                       return -EINVAL;
+               }
+       }
+
+       return 0;
+}
+
+static struct hl_vm_hash_node *memhash_node_export_get(struct hl_ctx *ctx, u64 addr)
+{
+       struct hl_device *hdev = ctx->hdev;
+       struct hl_vm_hash_node *hnode;
+
+       /* get the memory handle */
+       mutex_lock(&ctx->mem_hash_lock);
+       hash_for_each_possible(ctx->mem_hash, hnode, node, (unsigned long)addr)
+               if (addr == hnode->vaddr)
+                       break;
+
+       if (!hnode) {
+               mutex_unlock(&ctx->mem_hash_lock);
+               dev_dbg(hdev->dev, "map address %#llx not found\n", addr);
+               return ERR_PTR(-EINVAL);
+       }
+
+       if (upper_32_bits(hnode->handle)) {
+               mutex_unlock(&ctx->mem_hash_lock);
+               dev_dbg(hdev->dev, "invalid handle %#llx for map address %#llx\n",
+                               hnode->handle, addr);
+               return ERR_PTR(-EINVAL);
+       }
+
+       /*
+        * node found, increase export count so this memory cannot be unmapped
+        * and the hash node cannot be deleted.
+        */
+       hnode->export_cnt++;
+       mutex_unlock(&ctx->mem_hash_lock);
+
+       return hnode;
+}
+
+static void memhash_node_export_put(struct hl_ctx *ctx, struct hl_vm_hash_node *hnode)
+{
+       mutex_lock(&ctx->mem_hash_lock);
+       hnode->export_cnt--;
+       mutex_unlock(&ctx->mem_hash_lock);
+}
+
+static struct hl_vm_phys_pg_pack *get_phys_pg_pack_from_hash_node(struct hl_device *hdev,
+                                                       struct hl_vm_hash_node *hnode)
+{
+       struct hl_vm_phys_pg_pack *phys_pg_pack;
+       struct hl_vm *vm = &hdev->vm;
+
+       spin_lock(&vm->idr_lock);
+       phys_pg_pack = idr_find(&vm->phys_pg_pack_handles, (u32) hnode->handle);
+       if (!phys_pg_pack) {
+               spin_unlock(&vm->idr_lock);
+               dev_dbg(hdev->dev, "no match for handle 0x%x\n", (u32) hnode->handle);
+               return ERR_PTR(-EINVAL);
+       }
+
+       spin_unlock(&vm->idr_lock);
+
+       if (phys_pg_pack->vm_type != VM_TYPE_PHYS_PACK) {
+               dev_dbg(hdev->dev, "handle 0x%llx does not represent DRAM memory\n", hnode->handle);
+               return ERR_PTR(-EINVAL);
+       }
+
+       return phys_pg_pack;
+}
+
 /**
  * export_dmabuf_from_addr() - export a dma-buf object for the given memory
  *                             address and size.
  * @ctx: pointer to the context structure.
- * @device_addr:  device memory physical address.
- * @size: size of device memory.
+ * @addr: device address.
+ * @size: size of device memory to export.
+ * @offset: the offset into the buffer from which to start exporting
  * @flags: DMA-BUF file/FD flags.
  * @dmabuf_fd: pointer to result FD that represents the dma-buf object.
  *
@@ -1853,37 +1992,66 @@ static int validate_export_params(struct hl_device *hdev, u64 device_addr, u64 s
  *
  * Return: 0 on success, non-zero for failure.
  */
-static int export_dmabuf_from_addr(struct hl_ctx *ctx, u64 device_addr,
-                                       u64 size, int flags, int *dmabuf_fd)
+static int export_dmabuf_from_addr(struct hl_ctx *ctx, u64 addr, u64 size, u64 offset,
+                                       int flags, int *dmabuf_fd)
 {
-       struct hl_dmabuf_priv *hl_dmabuf;
-       struct hl_device *hdev = ctx->hdev;
+       struct hl_vm_phys_pg_pack *phys_pg_pack = NULL;
+       struct hl_vm_hash_node *hnode = NULL;
        struct asic_fixed_properties *prop;
+       struct hl_dmabuf_priv *hl_dmabuf;
+       struct hl_device *hdev;
+       u64 export_addr;
        int rc;
 
+       hdev = ctx->hdev;
        prop = &hdev->asic_prop;
 
-       if (prop->dram_supports_virtual_memory) {
-               dev_dbg(hdev->dev, "Export not supported for devices with virtual memory\n");
-               return -EOPNOTSUPP;
+       /* offset must be 0 in devices without virtual memory support */
+       if (!prop->dram_supports_virtual_memory && offset) {
+               dev_dbg(hdev->dev, "offset is not allowed in device without virtual memory\n");
+               return -EINVAL;
        }
 
-       rc = validate_export_params(hdev, device_addr, size);
-       if (rc)
-               return rc;
+       export_addr = addr + offset;
 
        hl_dmabuf = kzalloc(sizeof(*hl_dmabuf), GFP_KERNEL);
        if (!hl_dmabuf)
                return -ENOMEM;
 
-       hl_dmabuf->device_address = device_addr;
+       if (prop->dram_supports_virtual_memory) {
+               hnode = memhash_node_export_get(ctx, addr);
+               if (IS_ERR(hnode)) {
+                       rc = PTR_ERR(hnode);
+                       goto err_free_dmabuf_wrapper;
+               }
+               phys_pg_pack = get_phys_pg_pack_from_hash_node(hdev, hnode);
+               if (IS_ERR(phys_pg_pack)) {
+                       rc = PTR_ERR(phys_pg_pack);
+                       goto dec_memhash_export_cnt;
+               }
+               rc = validate_export_params(hdev, export_addr, size, offset, phys_pg_pack);
+               if (rc)
+                       goto dec_memhash_export_cnt;
+
+               hl_dmabuf->phys_pg_pack = phys_pg_pack;
+               hl_dmabuf->memhash_hnode = hnode;
+       } else {
+               rc = validate_export_params_no_mmu(hdev, export_addr, size);
+               if (rc)
+                       goto err_free_dmabuf_wrapper;
+       }
+
+       hl_dmabuf->device_address = export_addr;
 
        rc = export_dmabuf(ctx, hl_dmabuf, size, flags, dmabuf_fd);
        if (rc)
-               goto err_free_dmabuf_wrapper;
+               goto dec_memhash_export_cnt;
 
        return 0;
 
+dec_memhash_export_cnt:
+       if (prop->dram_supports_virtual_memory)
+               memhash_node_export_put(ctx, hnode);
 err_free_dmabuf_wrapper:
        kfree(hl_dmabuf);
        return rc;
@@ -2160,10 +2328,11 @@ int hl_mem_ioctl(struct hl_fpriv *hpriv, void *data)
 
        case HL_MEM_OP_EXPORT_DMABUF_FD:
                rc = export_dmabuf_from_addr(ctx,
-                                       args->in.export_dmabuf_fd.handle,
-                                       args->in.export_dmabuf_fd.mem_size,
-                                       args->in.flags,
-                                       &dmabuf_fd);
+                               args->in.export_dmabuf_fd.addr,
+                               args->in.export_dmabuf_fd.mem_size,
+                               args->in.export_dmabuf_fd.offset,
+                               args->in.flags,
+                               &dmabuf_fd);
                memset(args, 0, sizeof(*args));
                args->out.fd = dmabuf_fd;
                break;
index 3b995e8..c67d189 100644 (file)
@@ -1851,15 +1851,24 @@ struct hl_mem_in {
                /**
                 * structure for exporting DMABUF object (used with
                 * the HL_MEM_OP_EXPORT_DMABUF_FD op)
-                * @handle: handle returned from HL_MEM_OP_ALLOC.
-                *          in Gaudi, where we don't have MMU for the device memory, the
-                *          driver expects a physical address (instead of a handle) in the
-                *          device memory space.
-                * @mem_size: size of memory allocation. Relevant only for GAUDI
+                * @addr: for Gaudi1, the driver expects a physical address
+                *        inside the device's DRAM. this is because in Gaudi1
+                *        we don't have MMU that covers the device's DRAM.
+                *        for all other ASICs, the driver expects a device
+                *        virtual address that represents the start address of
+                *        a mapped DRAM memory area inside the device.
+                *        the address must be the same as was received from the
+                *        driver during a previous HL_MEM_OP_MAP operation.
+                * @mem_size: size of memory to export.
+                * @offset: for Gaudi1, this value must be 0. For all other ASICs,
+                *          the driver expects an offset inside of the memory area
+                *          describe by addr. the offset represents the start
+                *          address of that the exported dma-buf object describes.
                 */
                struct {
-                       __u64 handle;
+                       __u64 addr;
                        __u64 mem_size;
+                       __u64 offset;
                } export_dmabuf_fd;
        };