Merge branch 'stable/for-linus-5.15' of git://git.kernel.org/pub/scm/linux/kernel...
[linux-2.6-microblaze.git] / kernel / dma / direct.c
index 2de33e5..4c6c5e0 100644 (file)
@@ -174,10 +174,16 @@ void *dma_direct_alloc(struct device *dev, size_t size,
        }
 
        if (!IS_ENABLED(CONFIG_ARCH_HAS_DMA_SET_UNCACHED) &&
-           !IS_ENABLED(CONFIG_DMA_DIRECT_REMAP) && !dev_is_dma_coherent(dev) &&
+           !IS_ENABLED(CONFIG_DMA_DIRECT_REMAP) &&
+           !IS_ENABLED(CONFIG_DMA_GLOBAL_POOL) &&
+           !dev_is_dma_coherent(dev) &&
            !is_swiotlb_for_alloc(dev))
                return arch_dma_alloc(dev, size, dma_handle, gfp, attrs);
 
+       if (IS_ENABLED(CONFIG_DMA_GLOBAL_POOL) &&
+           !dev_is_dma_coherent(dev))
+               return dma_alloc_from_global_coherent(dev, size, dma_handle);
+
        /*
         * Remapping or decrypting memory may block. If either is required and
         * we can't block, allocate the memory from the atomic pools.
@@ -278,12 +284,21 @@ void dma_direct_free(struct device *dev, size_t size,
        }
 
        if (!IS_ENABLED(CONFIG_ARCH_HAS_DMA_SET_UNCACHED) &&
-           !IS_ENABLED(CONFIG_DMA_DIRECT_REMAP) && !dev_is_dma_coherent(dev) &&
+           !IS_ENABLED(CONFIG_DMA_DIRECT_REMAP) &&
+           !IS_ENABLED(CONFIG_DMA_GLOBAL_POOL) &&
+           !dev_is_dma_coherent(dev) &&
            !is_swiotlb_for_alloc(dev)) {
                arch_dma_free(dev, size, cpu_addr, dma_addr, attrs);
                return;
        }
 
+       if (IS_ENABLED(CONFIG_DMA_GLOBAL_POOL) &&
+           !dev_is_dma_coherent(dev)) {
+               if (!dma_release_from_global_coherent(page_order, cpu_addr))
+                       WARN_ON_ONCE(1);
+               return;
+       }
+
        /* If cpu_addr is not from an atomic pool, dma_free_from_pool() fails */
        if (IS_ENABLED(CONFIG_DMA_COHERENT_POOL) &&
            dma_free_from_pool(dev, cpu_addr, PAGE_ALIGN(size)))
@@ -436,7 +451,7 @@ int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, int nents,
 
 out_unmap:
        dma_direct_unmap_sg(dev, sgl, i, dir, attrs | DMA_ATTR_SKIP_CPU_SYNC);
-       return 0;
+       return -EIO;
 }
 
 dma_addr_t dma_direct_map_resource(struct device *dev, phys_addr_t paddr,
@@ -487,6 +502,8 @@ int dma_direct_mmap(struct device *dev, struct vm_area_struct *vma,
 
        if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret))
                return ret;
+       if (dma_mmap_from_global_coherent(vma, cpu_addr, size, &ret))
+               return ret;
 
        if (vma->vm_pgoff >= count || user_count > count - vma->vm_pgoff)
                return -ENXIO;