Merge tag 'char-misc-5.15-rc1-lkdtm' of git://git.kernel.org/pub/scm/linux/kernel...
[linux-2.6-microblaze.git] / drivers / gpu / drm / amd / amdgpu / amdgpu_vram_mgr.c
index bce105e..7b2b098 100644 (file)
  */
 
 #include <linux/dma-mapping.h>
+#include <drm/ttm/ttm_range_manager.h>
+
 #include "amdgpu.h"
 #include "amdgpu_vm.h"
 #include "amdgpu_res_cursor.h"
 #include "amdgpu_atomfirmware.h"
 #include "atom.h"
 
-static inline struct amdgpu_vram_mgr *to_vram_mgr(struct ttm_resource_manager *man)
+struct amdgpu_vram_reservation {
+       struct list_head node;
+       struct drm_mm_node mm_node;
+};
+
+static inline struct amdgpu_vram_mgr *
+to_vram_mgr(struct ttm_resource_manager *man)
 {
        return container_of(man, struct amdgpu_vram_mgr, manager);
 }
 
-static inline struct amdgpu_device *to_amdgpu_device(struct amdgpu_vram_mgr *mgr)
+static inline struct amdgpu_device *
+to_amdgpu_device(struct amdgpu_vram_mgr *mgr)
 {
        return container_of(mgr, struct amdgpu_device, mman.vram_mgr);
 }
@@ -82,12 +91,14 @@ static ssize_t amdgpu_mem_info_vis_vram_total_show(struct device *dev,
  * amount of currently used VRAM in bytes
  */
 static ssize_t amdgpu_mem_info_vram_used_show(struct device *dev,
-               struct device_attribute *attr, char *buf)
+                                             struct device_attribute *attr,
+                                             char *buf)
 {
        struct drm_device *ddev = dev_get_drvdata(dev);
        struct amdgpu_device *adev = drm_to_adev(ddev);
-       struct ttm_resource_manager *man = ttm_manager_type(&adev->mman.bdev, TTM_PL_VRAM);
+       struct ttm_resource_manager *man;
 
+       man = ttm_manager_type(&adev->mman.bdev, TTM_PL_VRAM);
        return sysfs_emit(buf, "%llu\n", amdgpu_vram_mgr_usage(man));
 }
 
@@ -100,18 +111,28 @@ static ssize_t amdgpu_mem_info_vram_used_show(struct device *dev,
  * amount of currently used visible VRAM in bytes
  */
 static ssize_t amdgpu_mem_info_vis_vram_used_show(struct device *dev,
-               struct device_attribute *attr, char *buf)
+                                                 struct device_attribute *attr,
+                                                 char *buf)
 {
        struct drm_device *ddev = dev_get_drvdata(dev);
        struct amdgpu_device *adev = drm_to_adev(ddev);
-       struct ttm_resource_manager *man = ttm_manager_type(&adev->mman.bdev, TTM_PL_VRAM);
+       struct ttm_resource_manager *man;
 
+       man = ttm_manager_type(&adev->mman.bdev, TTM_PL_VRAM);
        return sysfs_emit(buf, "%llu\n", amdgpu_vram_mgr_vis_usage(man));
 }
 
+/**
+ * DOC: mem_info_vram_vendor
+ *
+ * The amdgpu driver provides a sysfs API for reporting the vendor of the
+ * installed VRAM
+ * The file mem_info_vram_vendor is used for this and returns the name of the
+ * vendor.
+ */
 static ssize_t amdgpu_mem_info_vram_vendor(struct device *dev,
-                                                struct device_attribute *attr,
-                                                char *buf)
+                                          struct device_attribute *attr,
+                                          char *buf)
 {
        struct drm_device *ddev = dev_get_drvdata(dev);
        struct amdgpu_device *adev = drm_to_adev(ddev);
@@ -153,7 +174,7 @@ static DEVICE_ATTR(mem_info_vis_vram_used, S_IRUGO,
 static DEVICE_ATTR(mem_info_vram_vendor, S_IRUGO,
                   amdgpu_mem_info_vram_vendor, NULL);
 
-static const struct attribute *amdgpu_vram_mgr_attributes[] = {
+static struct attribute *amdgpu_vram_mgr_attributes[] = {
        &dev_attr_mem_info_vram_total.attr,
        &dev_attr_mem_info_vis_vram_total.attr,
        &dev_attr_mem_info_vram_used.attr,
@@ -162,77 +183,9 @@ static const struct attribute *amdgpu_vram_mgr_attributes[] = {
        NULL
 };
 
-static const struct ttm_resource_manager_func amdgpu_vram_mgr_func;
-
-/**
- * amdgpu_vram_mgr_init - init VRAM manager and DRM MM
- *
- * @adev: amdgpu_device pointer
- *
- * Allocate and initialize the VRAM manager.
- */
-int amdgpu_vram_mgr_init(struct amdgpu_device *adev)
-{
-       struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr;
-       struct ttm_resource_manager *man = &mgr->manager;
-       int ret;
-
-       ttm_resource_manager_init(man, adev->gmc.real_vram_size >> PAGE_SHIFT);
-
-       man->func = &amdgpu_vram_mgr_func;
-
-       drm_mm_init(&mgr->mm, 0, man->size);
-       spin_lock_init(&mgr->lock);
-       INIT_LIST_HEAD(&mgr->reservations_pending);
-       INIT_LIST_HEAD(&mgr->reserved_pages);
-
-       /* Add the two VRAM-related sysfs files */
-       ret = sysfs_create_files(&adev->dev->kobj, amdgpu_vram_mgr_attributes);
-       if (ret)
-               DRM_ERROR("Failed to register sysfs\n");
-
-       ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_VRAM, &mgr->manager);
-       ttm_resource_manager_set_used(man, true);
-       return 0;
-}
-
-/**
- * amdgpu_vram_mgr_fini - free and destroy VRAM manager
- *
- * @adev: amdgpu_device pointer
- *
- * Destroy and free the VRAM manager, returns -EBUSY if ranges are still
- * allocated inside it.
- */
-void amdgpu_vram_mgr_fini(struct amdgpu_device *adev)
-{
-       struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr;
-       struct ttm_resource_manager *man = &mgr->manager;
-       int ret;
-       struct amdgpu_vram_reservation *rsv, *temp;
-
-       ttm_resource_manager_set_used(man, false);
-
-       ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man);
-       if (ret)
-               return;
-
-       spin_lock(&mgr->lock);
-       list_for_each_entry_safe(rsv, temp, &mgr->reservations_pending, node)
-               kfree(rsv);
-
-       list_for_each_entry_safe(rsv, temp, &mgr->reserved_pages, node) {
-               drm_mm_remove_node(&rsv->mm_node);
-               kfree(rsv);
-       }
-       drm_mm_takedown(&mgr->mm);
-       spin_unlock(&mgr->lock);
-
-       sysfs_remove_files(&adev->dev->kobj, amdgpu_vram_mgr_attributes);
-
-       ttm_resource_manager_cleanup(man);
-       ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_VRAM, NULL);
-}
+const struct attribute_group amdgpu_vram_mgr_attr_group = {
+       .attrs = amdgpu_vram_mgr_attributes
+};
 
 /**
  * amdgpu_vram_mgr_vis_size - Calculate visible node size
@@ -266,23 +219,25 @@ static u64 amdgpu_vram_mgr_vis_size(struct amdgpu_device *adev,
 u64 amdgpu_vram_mgr_bo_visible_size(struct amdgpu_bo *bo)
 {
        struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
-       struct ttm_resource *mem = &bo->tbo.mem;
-       struct drm_mm_node *nodes = mem->mm_node;
-       unsigned pages = mem->num_pages;
+       struct ttm_resource *res = bo->tbo.resource;
+       unsigned pages = res->num_pages;
+       struct drm_mm_node *mm;
        u64 usage;
 
        if (amdgpu_gmc_vram_full_visible(&adev->gmc))
                return amdgpu_bo_size(bo);
 
-       if (mem->start >= adev->gmc.visible_vram_size >> PAGE_SHIFT)
+       if (res->start >= adev->gmc.visible_vram_size >> PAGE_SHIFT)
                return 0;
 
-       for (usage = 0; nodes && pages; pages -= nodes->size, nodes++)
-               usage += amdgpu_vram_mgr_vis_size(adev, nodes);
+       mm = &container_of(res, struct ttm_range_mgr_node, base)->mm_nodes[0];
+       for (usage = 0; pages; pages -= mm->size, mm++)
+               usage += amdgpu_vram_mgr_vis_size(adev, mm);
 
        return usage;
 }
 
+/* Commit the reservation of VRAM pages */
 static void amdgpu_vram_mgr_do_reserve(struct ttm_resource_manager *man)
 {
        struct amdgpu_vram_mgr *mgr = to_vram_mgr(man);
@@ -406,22 +361,22 @@ static void amdgpu_vram_mgr_virt_start(struct ttm_resource *mem,
  * @man: TTM memory type manager
  * @tbo: TTM BO we need this range for
  * @place: placement flags and restrictions
- * @mem: the resulting mem object
+ * @res: the resulting mem object
  *
  * Allocate VRAM for the given BO.
  */
 static int amdgpu_vram_mgr_new(struct ttm_resource_manager *man,
                               struct ttm_buffer_object *tbo,
                               const struct ttm_place *place,
-                              struct ttm_resource *mem)
+                              struct ttm_resource **res)
 {
+       unsigned long lpfn, num_nodes, pages_per_node, pages_left, pages;
        struct amdgpu_vram_mgr *mgr = to_vram_mgr(man);
        struct amdgpu_device *adev = to_amdgpu_device(mgr);
+       uint64_t vis_usage = 0, mem_bytes, max_bytes;
+       struct ttm_range_mgr_node *node;
        struct drm_mm *mm = &mgr->mm;
-       struct drm_mm_node *nodes;
        enum drm_mm_insert_mode mode;
-       unsigned long lpfn, num_nodes, pages_per_node, pages_left;
-       uint64_t vis_usage = 0, mem_bytes, max_bytes;
        unsigned i;
        int r;
 
@@ -434,10 +389,10 @@ static int amdgpu_vram_mgr_new(struct ttm_resource_manager *man,
                max_bytes -= AMDGPU_VM_RESERVED_VRAM;
 
        /* bail out quickly if there's likely not enough VRAM for this BO */
-       mem_bytes = (u64)mem->num_pages << PAGE_SHIFT;
+       mem_bytes = tbo->base.size;
        if (atomic64_add_return(mem_bytes, &mgr->usage) > max_bytes) {
-               atomic64_sub(mem_bytes, &mgr->usage);
-               return -ENOSPC;
+               r = -ENOSPC;
+               goto error_sub;
        }
 
        if (place->flags & TTM_PL_FLAG_CONTIGUOUS) {
@@ -448,78 +403,83 @@ static int amdgpu_vram_mgr_new(struct ttm_resource_manager *man,
                pages_per_node = HPAGE_PMD_NR;
 #else
                /* default to 2MB */
-               pages_per_node = (2UL << (20UL - PAGE_SHIFT));
+               pages_per_node = 2UL << (20UL - PAGE_SHIFT);
 #endif
-               pages_per_node = max((uint32_t)pages_per_node, mem->page_alignment);
-               num_nodes = DIV_ROUND_UP(mem->num_pages, pages_per_node);
+               pages_per_node = max_t(uint32_t, pages_per_node,
+                                      tbo->page_alignment);
+               num_nodes = DIV_ROUND_UP_ULL(PFN_UP(mem_bytes), pages_per_node);
        }
 
-       nodes = kvmalloc_array((uint32_t)num_nodes, sizeof(*nodes),
-                              GFP_KERNEL | __GFP_ZERO);
-       if (!nodes) {
-               atomic64_sub(mem_bytes, &mgr->usage);
-               return -ENOMEM;
+       node = kvmalloc(struct_size(node, mm_nodes, num_nodes),
+                       GFP_KERNEL | __GFP_ZERO);
+       if (!node) {
+               r = -ENOMEM;
+               goto error_sub;
        }
 
+       ttm_resource_init(tbo, place, &node->base);
+
        mode = DRM_MM_INSERT_BEST;
        if (place->flags & TTM_PL_FLAG_TOPDOWN)
                mode = DRM_MM_INSERT_HIGH;
 
-       mem->start = 0;
-       pages_left = mem->num_pages;
+       pages_left = node->base.num_pages;
 
-       spin_lock(&mgr->lock);
-       for (i = 0; pages_left >= pages_per_node; ++i) {
-               unsigned long pages = rounddown_pow_of_two(pages_left);
-
-               /* Limit maximum size to 2GB due to SG table limitations */
-               pages = min(pages, (2UL << (30 - PAGE_SHIFT)));
-
-               r = drm_mm_insert_node_in_range(mm, &nodes[i], pages,
-                                               pages_per_node, 0,
-                                               place->fpfn, lpfn,
-                                               mode);
-               if (unlikely(r))
-                       break;
-
-               vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]);
-               amdgpu_vram_mgr_virt_start(mem, &nodes[i]);
-               pages_left -= pages;
-       }
+       /* Limit maximum size to 2GB due to SG table limitations */
+       pages = min(pages_left, 2UL << (30 - PAGE_SHIFT));
 
-       for (; pages_left; ++i) {
-               unsigned long pages = min(pages_left, pages_per_node);
-               uint32_t alignment = mem->page_alignment;
+       i = 0;
+       spin_lock(&mgr->lock);
+       while (pages_left) {
+               uint32_t alignment = tbo->page_alignment;
 
-               if (pages == pages_per_node)
+               if (pages >= pages_per_node)
                        alignment = pages_per_node;
 
-               r = drm_mm_insert_node_in_range(mm, &nodes[i],
-                                               pages, alignment, 0,
-                                               place->fpfn, lpfn,
-                                               mode);
-               if (unlikely(r))
-                       goto error;
+               r = drm_mm_insert_node_in_range(mm, &node->mm_nodes[i], pages,
+                                               alignment, 0, place->fpfn,
+                                               lpfn, mode);
+               if (unlikely(r)) {
+                       if (pages > pages_per_node) {
+                               if (is_power_of_2(pages))
+                                       pages = pages / 2;
+                               else
+                                       pages = rounddown_pow_of_two(pages);
+                               continue;
+                       }
+                       goto error_free;
+               }
 
-               vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]);
-               amdgpu_vram_mgr_virt_start(mem, &nodes[i]);
+               vis_usage += amdgpu_vram_mgr_vis_size(adev, &node->mm_nodes[i]);
+               amdgpu_vram_mgr_virt_start(&node->base, &node->mm_nodes[i]);
                pages_left -= pages;
+               ++i;
+
+               if (pages > pages_left)
+                       pages = pages_left;
        }
        spin_unlock(&mgr->lock);
 
-       atomic64_add(vis_usage, &mgr->vis_usage);
+       if (i == 1)
+               node->base.placement |= TTM_PL_FLAG_CONTIGUOUS;
 
-       mem->mm_node = nodes;
+       if (adev->gmc.xgmi.connected_to_cpu)
+               node->base.bus.caching = ttm_cached;
+       else
+               node->base.bus.caching = ttm_write_combined;
 
+       atomic64_add(vis_usage, &mgr->vis_usage);
+       *res = &node->base;
        return 0;
 
-error:
+error_free:
        while (i--)
-               drm_mm_remove_node(&nodes[i]);
+               drm_mm_remove_node(&node->mm_nodes[i]);
        spin_unlock(&mgr->lock);
-       atomic64_sub(mem->num_pages << PAGE_SHIFT, &mgr->usage);
+       kvfree(node);
 
-       kvfree(nodes);
+error_sub:
+       atomic64_sub(mem_bytes, &mgr->usage);
        return r;
 }
 
@@ -527,29 +487,27 @@ error:
  * amdgpu_vram_mgr_del - free ranges
  *
  * @man: TTM memory type manager
- * @mem: TTM memory object
+ * @res: TTM memory object
  *
  * Free the allocated VRAM again.
  */
 static void amdgpu_vram_mgr_del(struct ttm_resource_manager *man,
-                               struct ttm_resource *mem)
+                               struct ttm_resource *res)
 {
+       struct ttm_range_mgr_node *node = to_ttm_range_mgr_node(res);
        struct amdgpu_vram_mgr *mgr = to_vram_mgr(man);
        struct amdgpu_device *adev = to_amdgpu_device(mgr);
-       struct drm_mm_node *nodes = mem->mm_node;
        uint64_t usage = 0, vis_usage = 0;
-       unsigned pages = mem->num_pages;
-
-       if (!mem->mm_node)
-               return;
+       unsigned i, pages;
 
        spin_lock(&mgr->lock);
-       while (pages) {
-               pages -= nodes->size;
-               drm_mm_remove_node(nodes);
-               usage += nodes->size << PAGE_SHIFT;
-               vis_usage += amdgpu_vram_mgr_vis_size(adev, nodes);
-               ++nodes;
+       for (i = 0, pages = res->num_pages; pages;
+            pages -= node->mm_nodes[i].size, ++i) {
+               struct drm_mm_node *mm = &node->mm_nodes[i];
+
+               drm_mm_remove_node(mm);
+               usage += mm->size << PAGE_SHIFT;
+               vis_usage += amdgpu_vram_mgr_vis_size(adev, mm);
        }
        amdgpu_vram_mgr_do_reserve(man);
        spin_unlock(&mgr->lock);
@@ -557,15 +515,14 @@ static void amdgpu_vram_mgr_del(struct ttm_resource_manager *man,
        atomic64_sub(usage, &mgr->usage);
        atomic64_sub(vis_usage, &mgr->vis_usage);
 
-       kvfree(mem->mm_node);
-       mem->mm_node = NULL;
+       kvfree(node);
 }
 
 /**
  * amdgpu_vram_mgr_alloc_sgt - allocate and fill a sg table
  *
  * @adev: amdgpu device pointer
- * @mem: TTM memory object
+ * @res: TTM memory object
  * @offset: byte offset from the base of VRAM BO
  * @length: number of bytes to export in sg_table
  * @dev: the other device
@@ -575,7 +532,7 @@ static void amdgpu_vram_mgr_del(struct ttm_resource_manager *man,
  * Allocate and fill a sg table from a VRAM allocation.
  */
 int amdgpu_vram_mgr_alloc_sgt(struct amdgpu_device *adev,
-                             struct ttm_resource *mem,
+                             struct ttm_resource *res,
                              u64 offset, u64 length,
                              struct device *dev,
                              enum dma_data_direction dir,
@@ -591,7 +548,7 @@ int amdgpu_vram_mgr_alloc_sgt(struct amdgpu_device *adev,
                return -ENOMEM;
 
        /* Determine the number of DRM_MM nodes to export */
-       amdgpu_res_first(mem, offset, length, &cursor);
+       amdgpu_res_first(res, offset, length, &cursor);
        while (cursor.remaining) {
                num_entries++;
                amdgpu_res_next(&cursor, cursor.size);
@@ -611,7 +568,7 @@ int amdgpu_vram_mgr_alloc_sgt(struct amdgpu_device *adev,
         * and the number of bytes from it. Access the following
         * DRM_MM node(s) if more buffer needs to exported
         */
-       amdgpu_res_first(mem, offset, length, &cursor);
+       amdgpu_res_first(res, offset, length, &cursor);
        for_each_sgtable_sg((*sgt), sg, i) {
                phys_addr_t phys = cursor.start + adev->gmc.aper_base;
                size_t size = cursor.size;
@@ -727,3 +684,65 @@ static const struct ttm_resource_manager_func amdgpu_vram_mgr_func = {
        .free   = amdgpu_vram_mgr_del,
        .debug  = amdgpu_vram_mgr_debug
 };
+
+/**
+ * amdgpu_vram_mgr_init - init VRAM manager and DRM MM
+ *
+ * @adev: amdgpu_device pointer
+ *
+ * Allocate and initialize the VRAM manager.
+ */
+int amdgpu_vram_mgr_init(struct amdgpu_device *adev)
+{
+       struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr;
+       struct ttm_resource_manager *man = &mgr->manager;
+
+       ttm_resource_manager_init(man, adev->gmc.real_vram_size >> PAGE_SHIFT);
+
+       man->func = &amdgpu_vram_mgr_func;
+
+       drm_mm_init(&mgr->mm, 0, man->size);
+       spin_lock_init(&mgr->lock);
+       INIT_LIST_HEAD(&mgr->reservations_pending);
+       INIT_LIST_HEAD(&mgr->reserved_pages);
+
+       ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_VRAM, &mgr->manager);
+       ttm_resource_manager_set_used(man, true);
+       return 0;
+}
+
+/**
+ * amdgpu_vram_mgr_fini - free and destroy VRAM manager
+ *
+ * @adev: amdgpu_device pointer
+ *
+ * Destroy and free the VRAM manager, returns -EBUSY if ranges are still
+ * allocated inside it.
+ */
+void amdgpu_vram_mgr_fini(struct amdgpu_device *adev)
+{
+       struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr;
+       struct ttm_resource_manager *man = &mgr->manager;
+       int ret;
+       struct amdgpu_vram_reservation *rsv, *temp;
+
+       ttm_resource_manager_set_used(man, false);
+
+       ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man);
+       if (ret)
+               return;
+
+       spin_lock(&mgr->lock);
+       list_for_each_entry_safe(rsv, temp, &mgr->reservations_pending, node)
+               kfree(rsv);
+
+       list_for_each_entry_safe(rsv, temp, &mgr->reserved_pages, node) {
+               drm_mm_remove_node(&rsv->mm_node);
+               kfree(rsv);
+       }
+       drm_mm_takedown(&mgr->mm);
+       spin_unlock(&mgr->lock);
+
+       ttm_resource_manager_cleanup(man);
+       ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_VRAM, NULL);
+}