| // SPDX-License-Identifier: GPL-2.0+ |
| /* Copyright (C) 2015-2018 Broadcom */ |
| |
| /** |
| * DOC: V3D GEM BO management support |
| * |
| * Compared to VC4 (V3D 2.x), V3D 3.3 introduces an MMU between the |
| * GPU and the bus, allowing us to use shmem objects for our storage |
| * instead of CMA. |
| * |
| * Physically contiguous objects may still be imported to V3D, but the |
| * driver doesn't allocate physically contiguous objects on its own. |
| * Display engines requiring physically contiguous allocations should |
| * look into Mesa's "renderonly" support (as used by the Mesa pl111 |
| * driver) for an example of how to integrate with V3D. |
| * |
| * Long term, we should support evicting pages from the MMU when under |
| * memory pressure (thus the v3d_bo_get_pages() refcounting), but |
| * that's not a high priority since our systems tend to not have swap. |
| */ |
| |
| #include <linux/dma-buf.h> |
| #include <linux/pfn_t.h> |
| |
| #include "v3d_drv.h" |
| #include "uapi/drm/v3d_drm.h" |
| |
| /* Pins the shmem pages, fills in the .pages and .sgt fields of the BO, and maps |
| * it for DMA. |
| */ |
| static int |
| v3d_bo_get_pages(struct v3d_bo *bo) |
| { |
| struct drm_gem_object *obj = &bo->base; |
| struct drm_device *dev = obj->dev; |
| int npages = obj->size >> PAGE_SHIFT; |
| int ret = 0; |
| |
| mutex_lock(&bo->lock); |
| if (bo->pages_refcount++ != 0) |
| goto unlock; |
| |
| if (!obj->import_attach) { |
| bo->pages = drm_gem_get_pages(obj); |
| if (IS_ERR(bo->pages)) { |
| ret = PTR_ERR(bo->pages); |
| goto unlock; |
| } |
| |
| bo->sgt = drm_prime_pages_to_sg(bo->pages, npages); |
| if (IS_ERR(bo->sgt)) { |
| ret = PTR_ERR(bo->sgt); |
| goto put_pages; |
| } |
| |
| /* Map the pages for use by the GPU. */ |
| dma_map_sg(dev->dev, bo->sgt->sgl, |
| bo->sgt->nents, DMA_BIDIRECTIONAL); |
| } else { |
| bo->pages = kcalloc(npages, sizeof(*bo->pages), GFP_KERNEL); |
| if (!bo->pages) |
| goto put_pages; |
| |
| drm_prime_sg_to_page_addr_arrays(bo->sgt, bo->pages, |
| NULL, npages); |
| |
| /* Note that dma-bufs come in mapped. */ |
| } |
| |
| mutex_unlock(&bo->lock); |
| |
| return 0; |
| |
| put_pages: |
| drm_gem_put_pages(obj, bo->pages, true, true); |
| bo->pages = NULL; |
| unlock: |
| bo->pages_refcount--; |
| mutex_unlock(&bo->lock); |
| return ret; |
| } |
| |
| static void |
| v3d_bo_put_pages(struct v3d_bo *bo) |
| { |
| struct drm_gem_object *obj = &bo->base; |
| |
| mutex_lock(&bo->lock); |
| if (--bo->pages_refcount == 0) { |
| if (!obj->import_attach) { |
| dma_unmap_sg(obj->dev->dev, bo->sgt->sgl, |
| bo->sgt->nents, DMA_BIDIRECTIONAL); |
| sg_free_table(bo->sgt); |
| kfree(bo->sgt); |
| drm_gem_put_pages(obj, bo->pages, true, true); |
| } else { |
| kfree(bo->pages); |
| } |
| } |
| mutex_unlock(&bo->lock); |
| } |
| |
| static struct v3d_bo *v3d_bo_create_struct(struct drm_device *dev, |
| size_t unaligned_size) |
| { |
| struct v3d_dev *v3d = to_v3d_dev(dev); |
| struct drm_gem_object *obj; |
| struct v3d_bo *bo; |
| size_t size = roundup(unaligned_size, PAGE_SIZE); |
| int ret; |
| |
| if (size == 0) |
| return ERR_PTR(-EINVAL); |
| |
| bo = kzalloc(sizeof(*bo), GFP_KERNEL); |
| if (!bo) |
| return ERR_PTR(-ENOMEM); |
| obj = &bo->base; |
| |
| INIT_LIST_HEAD(&bo->vmas); |
| INIT_LIST_HEAD(&bo->unref_head); |
| mutex_init(&bo->lock); |
| |
| ret = drm_gem_object_init(dev, obj, size); |
| if (ret) |
| goto free_bo; |
| |
| spin_lock(&v3d->mm_lock); |
| ret = drm_mm_insert_node_generic(&v3d->mm, &bo->node, |
| obj->size >> PAGE_SHIFT, |
| GMP_GRANULARITY >> PAGE_SHIFT, 0, 0); |
| spin_unlock(&v3d->mm_lock); |
| if (ret) |
| goto free_obj; |
| |
| return bo; |
| |
| free_obj: |
| drm_gem_object_release(obj); |
| free_bo: |
| kfree(bo); |
| return ERR_PTR(ret); |
| } |
| |
| struct v3d_bo *v3d_bo_create(struct drm_device *dev, struct drm_file *file_priv, |
| size_t unaligned_size) |
| { |
| struct v3d_dev *v3d = to_v3d_dev(dev); |
| struct drm_gem_object *obj; |
| struct v3d_bo *bo; |
| int ret; |
| |
| bo = v3d_bo_create_struct(dev, unaligned_size); |
| if (IS_ERR(bo)) |
| return bo; |
| obj = &bo->base; |
| |
| bo->resv = &bo->_resv; |
| reservation_object_init(bo->resv); |
| |
| ret = v3d_bo_get_pages(bo); |
| if (ret) |
| goto free_mm; |
| |
| v3d_mmu_insert_ptes(bo); |
| |
| mutex_lock(&v3d->bo_lock); |
| v3d->bo_stats.num_allocated++; |
| v3d->bo_stats.pages_allocated += obj->size >> PAGE_SHIFT; |
| mutex_unlock(&v3d->bo_lock); |
| |
| return bo; |
| |
| free_mm: |
| spin_lock(&v3d->mm_lock); |
| drm_mm_remove_node(&bo->node); |
| spin_unlock(&v3d->mm_lock); |
| |
| drm_gem_object_release(obj); |
| kfree(bo); |
| return ERR_PTR(ret); |
| } |
| |
| /* Called DRM core on the last userspace/kernel unreference of the |
| * BO. |
| */ |
| void v3d_free_object(struct drm_gem_object *obj) |
| { |
| struct v3d_dev *v3d = to_v3d_dev(obj->dev); |
| struct v3d_bo *bo = to_v3d_bo(obj); |
| |
| mutex_lock(&v3d->bo_lock); |
| v3d->bo_stats.num_allocated--; |
| v3d->bo_stats.pages_allocated -= obj->size >> PAGE_SHIFT; |
| mutex_unlock(&v3d->bo_lock); |
| |
| reservation_object_fini(&bo->_resv); |
| |
| v3d_bo_put_pages(bo); |
| |
| if (obj->import_attach) |
| drm_prime_gem_destroy(obj, bo->sgt); |
| |
| v3d_mmu_remove_ptes(bo); |
| spin_lock(&v3d->mm_lock); |
| drm_mm_remove_node(&bo->node); |
| spin_unlock(&v3d->mm_lock); |
| |
| mutex_destroy(&bo->lock); |
| |
| drm_gem_object_release(obj); |
| kfree(bo); |
| } |
| |
| struct reservation_object *v3d_prime_res_obj(struct drm_gem_object *obj) |
| { |
| struct v3d_bo *bo = to_v3d_bo(obj); |
| |
| return bo->resv; |
| } |
| |
| static void |
| v3d_set_mmap_vma_flags(struct vm_area_struct *vma) |
| { |
| vma->vm_flags &= ~VM_PFNMAP; |
| vma->vm_flags |= VM_MIXEDMAP; |
| vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags)); |
| } |
| |
| vm_fault_t v3d_gem_fault(struct vm_fault *vmf) |
| { |
| struct vm_area_struct *vma = vmf->vma; |
| struct drm_gem_object *obj = vma->vm_private_data; |
| struct v3d_bo *bo = to_v3d_bo(obj); |
| pfn_t pfn; |
| pgoff_t pgoff; |
| |
| /* We don't use vmf->pgoff since that has the fake offset: */ |
| pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT; |
| pfn = __pfn_to_pfn_t(page_to_pfn(bo->pages[pgoff]), PFN_DEV); |
| |
| return vmf_insert_mixed(vma, vmf->address, pfn); |
| } |
| |
| int v3d_mmap(struct file *filp, struct vm_area_struct *vma) |
| { |
| int ret; |
| |
| ret = drm_gem_mmap(filp, vma); |
| if (ret) |
| return ret; |
| |
| v3d_set_mmap_vma_flags(vma); |
| |
| return ret; |
| } |
| |
| int v3d_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) |
| { |
| int ret; |
| |
| ret = drm_gem_mmap_obj(obj, obj->size, vma); |
| if (ret < 0) |
| return ret; |
| |
| v3d_set_mmap_vma_flags(vma); |
| |
| return 0; |
| } |
| |
| struct sg_table * |
| v3d_prime_get_sg_table(struct drm_gem_object *obj) |
| { |
| struct v3d_bo *bo = to_v3d_bo(obj); |
| int npages = obj->size >> PAGE_SHIFT; |
| |
| return drm_prime_pages_to_sg(bo->pages, npages); |
| } |
| |
| struct drm_gem_object * |
| v3d_prime_import_sg_table(struct drm_device *dev, |
| struct dma_buf_attachment *attach, |
| struct sg_table *sgt) |
| { |
| struct drm_gem_object *obj; |
| struct v3d_bo *bo; |
| |
| bo = v3d_bo_create_struct(dev, attach->dmabuf->size); |
| if (IS_ERR(bo)) |
| return ERR_CAST(bo); |
| obj = &bo->base; |
| |
| bo->resv = attach->dmabuf->resv; |
| |
| bo->sgt = sgt; |
| v3d_bo_get_pages(bo); |
| |
| v3d_mmu_insert_ptes(bo); |
| |
| return obj; |
| } |
| |
| int v3d_create_bo_ioctl(struct drm_device *dev, void *data, |
| struct drm_file *file_priv) |
| { |
| struct drm_v3d_create_bo *args = data; |
| struct v3d_bo *bo = NULL; |
| int ret; |
| |
| if (args->flags != 0) { |
| DRM_INFO("unknown create_bo flags: %d\n", args->flags); |
| return -EINVAL; |
| } |
| |
| bo = v3d_bo_create(dev, file_priv, PAGE_ALIGN(args->size)); |
| if (IS_ERR(bo)) |
| return PTR_ERR(bo); |
| |
| args->offset = bo->node.start << PAGE_SHIFT; |
| |
| ret = drm_gem_handle_create(file_priv, &bo->base, &args->handle); |
| drm_gem_object_put_unlocked(&bo->base); |
| |
| return ret; |
| } |
| |
| int v3d_mmap_bo_ioctl(struct drm_device *dev, void *data, |
| struct drm_file *file_priv) |
| { |
| struct drm_v3d_mmap_bo *args = data; |
| struct drm_gem_object *gem_obj; |
| int ret; |
| |
| if (args->flags != 0) { |
| DRM_INFO("unknown mmap_bo flags: %d\n", args->flags); |
| return -EINVAL; |
| } |
| |
| gem_obj = drm_gem_object_lookup(file_priv, args->handle); |
| if (!gem_obj) { |
| DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle); |
| return -ENOENT; |
| } |
| |
| ret = drm_gem_create_mmap_offset(gem_obj); |
| if (ret == 0) |
| args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node); |
| drm_gem_object_put_unlocked(gem_obj); |
| |
| return ret; |
| } |
| |
| int v3d_get_bo_offset_ioctl(struct drm_device *dev, void *data, |
| struct drm_file *file_priv) |
| { |
| struct drm_v3d_get_bo_offset *args = data; |
| struct drm_gem_object *gem_obj; |
| struct v3d_bo *bo; |
| |
| gem_obj = drm_gem_object_lookup(file_priv, args->handle); |
| if (!gem_obj) { |
| DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle); |
| return -ENOENT; |
| } |
| bo = to_v3d_bo(gem_obj); |
| |
| args->offset = bo->node.start << PAGE_SHIFT; |
| |
| drm_gem_object_put_unlocked(gem_obj); |
| return 0; |
| } |