123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373 |
- // SPDX-License-Identifier: GPL-2.0+
- /* Copyright (C) 2015-2018 Broadcom */
- /**
- * DOC: V3D GEM BO management support
- *
- * Compared to VC4 (V3D 2.x), V3D 3.3 introduces an MMU between the
- * GPU and the bus, allowing us to use shmem objects for our storage
- * instead of CMA.
- *
- * Physically contiguous objects may still be imported to V3D, but the
- * driver doesn't allocate physically contiguous objects on its own.
- * Display engines requiring physically contiguous allocations should
- * look into Mesa's "renderonly" support (as used by the Mesa pl111
- * driver) for an example of how to integrate with V3D.
- *
- * Long term, we should support evicting pages from the MMU when under
- * memory pressure (thus the v3d_bo_get_pages() refcounting), but
- * that's not a high priority since our systems tend to not have swap.
- */
- #include <linux/dma-buf.h>
- #include <linux/pfn_t.h>
- #include "v3d_drv.h"
- #include "uapi/drm/v3d_drm.h"
- /* Pins the shmem pages, fills in the .pages and .sgt fields of the BO, and maps
- * it for DMA.
- */
- static int
- v3d_bo_get_pages(struct v3d_bo *bo)
- {
- struct drm_gem_object *obj = &bo->base;
- struct drm_device *dev = obj->dev;
- int npages = obj->size >> PAGE_SHIFT;
- int ret = 0;
- mutex_lock(&bo->lock);
- if (bo->pages_refcount++ != 0)
- goto unlock;
- if (!obj->import_attach) {
- bo->pages = drm_gem_get_pages(obj);
- if (IS_ERR(bo->pages)) {
- ret = PTR_ERR(bo->pages);
- goto unlock;
- }
- bo->sgt = drm_prime_pages_to_sg(bo->pages, npages);
- if (IS_ERR(bo->sgt)) {
- ret = PTR_ERR(bo->sgt);
- goto put_pages;
- }
- /* Map the pages for use by the GPU. */
- dma_map_sg(dev->dev, bo->sgt->sgl,
- bo->sgt->nents, DMA_BIDIRECTIONAL);
- } else {
- bo->pages = kcalloc(npages, sizeof(*bo->pages), GFP_KERNEL);
- if (!bo->pages)
- goto put_pages;
- drm_prime_sg_to_page_addr_arrays(bo->sgt, bo->pages,
- NULL, npages);
- /* Note that dma-bufs come in mapped. */
- }
- mutex_unlock(&bo->lock);
- return 0;
- put_pages:
- drm_gem_put_pages(obj, bo->pages, true, true);
- bo->pages = NULL;
- unlock:
- bo->pages_refcount--;
- mutex_unlock(&bo->lock);
- return ret;
- }
- static void
- v3d_bo_put_pages(struct v3d_bo *bo)
- {
- struct drm_gem_object *obj = &bo->base;
- mutex_lock(&bo->lock);
- if (--bo->pages_refcount == 0) {
- if (!obj->import_attach) {
- dma_unmap_sg(obj->dev->dev, bo->sgt->sgl,
- bo->sgt->nents, DMA_BIDIRECTIONAL);
- sg_free_table(bo->sgt);
- kfree(bo->sgt);
- drm_gem_put_pages(obj, bo->pages, true, true);
- } else {
- kfree(bo->pages);
- }
- }
- mutex_unlock(&bo->lock);
- }
- static struct v3d_bo *v3d_bo_create_struct(struct drm_device *dev,
- size_t unaligned_size)
- {
- struct v3d_dev *v3d = to_v3d_dev(dev);
- struct drm_gem_object *obj;
- struct v3d_bo *bo;
- size_t size = roundup(unaligned_size, PAGE_SIZE);
- int ret;
- if (size == 0)
- return ERR_PTR(-EINVAL);
- bo = kzalloc(sizeof(*bo), GFP_KERNEL);
- if (!bo)
- return ERR_PTR(-ENOMEM);
- obj = &bo->base;
- INIT_LIST_HEAD(&bo->vmas);
- INIT_LIST_HEAD(&bo->unref_head);
- mutex_init(&bo->lock);
- ret = drm_gem_object_init(dev, obj, size);
- if (ret)
- goto free_bo;
- spin_lock(&v3d->mm_lock);
- ret = drm_mm_insert_node_generic(&v3d->mm, &bo->node,
- obj->size >> PAGE_SHIFT,
- GMP_GRANULARITY >> PAGE_SHIFT, 0, 0);
- spin_unlock(&v3d->mm_lock);
- if (ret)
- goto free_obj;
- return bo;
- free_obj:
- drm_gem_object_release(obj);
- free_bo:
- kfree(bo);
- return ERR_PTR(ret);
- }
- struct v3d_bo *v3d_bo_create(struct drm_device *dev, struct drm_file *file_priv,
- size_t unaligned_size)
- {
- struct v3d_dev *v3d = to_v3d_dev(dev);
- struct drm_gem_object *obj;
- struct v3d_bo *bo;
- int ret;
- bo = v3d_bo_create_struct(dev, unaligned_size);
- if (IS_ERR(bo))
- return bo;
- obj = &bo->base;
- bo->resv = &bo->_resv;
- reservation_object_init(bo->resv);
- ret = v3d_bo_get_pages(bo);
- if (ret)
- goto free_mm;
- v3d_mmu_insert_ptes(bo);
- mutex_lock(&v3d->bo_lock);
- v3d->bo_stats.num_allocated++;
- v3d->bo_stats.pages_allocated += obj->size >> PAGE_SHIFT;
- mutex_unlock(&v3d->bo_lock);
- return bo;
- free_mm:
- spin_lock(&v3d->mm_lock);
- drm_mm_remove_node(&bo->node);
- spin_unlock(&v3d->mm_lock);
- drm_gem_object_release(obj);
- kfree(bo);
- return ERR_PTR(ret);
- }
- /* Called DRM core on the last userspace/kernel unreference of the
- * BO.
- */
- void v3d_free_object(struct drm_gem_object *obj)
- {
- struct v3d_dev *v3d = to_v3d_dev(obj->dev);
- struct v3d_bo *bo = to_v3d_bo(obj);
- mutex_lock(&v3d->bo_lock);
- v3d->bo_stats.num_allocated--;
- v3d->bo_stats.pages_allocated -= obj->size >> PAGE_SHIFT;
- mutex_unlock(&v3d->bo_lock);
- reservation_object_fini(&bo->_resv);
- v3d_bo_put_pages(bo);
- if (obj->import_attach)
- drm_prime_gem_destroy(obj, bo->sgt);
- v3d_mmu_remove_ptes(bo);
- spin_lock(&v3d->mm_lock);
- drm_mm_remove_node(&bo->node);
- spin_unlock(&v3d->mm_lock);
- mutex_destroy(&bo->lock);
- drm_gem_object_release(obj);
- kfree(bo);
- }
- struct reservation_object *v3d_prime_res_obj(struct drm_gem_object *obj)
- {
- struct v3d_bo *bo = to_v3d_bo(obj);
- return bo->resv;
- }
- static void
- v3d_set_mmap_vma_flags(struct vm_area_struct *vma)
- {
- vma->vm_flags &= ~VM_PFNMAP;
- vma->vm_flags |= VM_MIXEDMAP;
- vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
- }
- vm_fault_t v3d_gem_fault(struct vm_fault *vmf)
- {
- struct vm_area_struct *vma = vmf->vma;
- struct drm_gem_object *obj = vma->vm_private_data;
- struct v3d_bo *bo = to_v3d_bo(obj);
- pfn_t pfn;
- pgoff_t pgoff;
- /* We don't use vmf->pgoff since that has the fake offset: */
- pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
- pfn = __pfn_to_pfn_t(page_to_pfn(bo->pages[pgoff]), PFN_DEV);
- return vmf_insert_mixed(vma, vmf->address, pfn);
- }
- int v3d_mmap(struct file *filp, struct vm_area_struct *vma)
- {
- int ret;
- ret = drm_gem_mmap(filp, vma);
- if (ret)
- return ret;
- v3d_set_mmap_vma_flags(vma);
- return ret;
- }
- int v3d_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma)
- {
- int ret;
- ret = drm_gem_mmap_obj(obj, obj->size, vma);
- if (ret < 0)
- return ret;
- v3d_set_mmap_vma_flags(vma);
- return 0;
- }
- struct sg_table *
- v3d_prime_get_sg_table(struct drm_gem_object *obj)
- {
- struct v3d_bo *bo = to_v3d_bo(obj);
- int npages = obj->size >> PAGE_SHIFT;
- return drm_prime_pages_to_sg(bo->pages, npages);
- }
- struct drm_gem_object *
- v3d_prime_import_sg_table(struct drm_device *dev,
- struct dma_buf_attachment *attach,
- struct sg_table *sgt)
- {
- struct drm_gem_object *obj;
- struct v3d_bo *bo;
- bo = v3d_bo_create_struct(dev, attach->dmabuf->size);
- if (IS_ERR(bo))
- return ERR_CAST(bo);
- obj = &bo->base;
- bo->resv = attach->dmabuf->resv;
- bo->sgt = sgt;
- obj->import_attach = attach;
- v3d_bo_get_pages(bo);
- v3d_mmu_insert_ptes(bo);
- return obj;
- }
- int v3d_create_bo_ioctl(struct drm_device *dev, void *data,
- struct drm_file *file_priv)
- {
- struct drm_v3d_create_bo *args = data;
- struct v3d_bo *bo = NULL;
- int ret;
- if (args->flags != 0) {
- DRM_INFO("unknown create_bo flags: %d\n", args->flags);
- return -EINVAL;
- }
- bo = v3d_bo_create(dev, file_priv, PAGE_ALIGN(args->size));
- if (IS_ERR(bo))
- return PTR_ERR(bo);
- args->offset = bo->node.start << PAGE_SHIFT;
- ret = drm_gem_handle_create(file_priv, &bo->base, &args->handle);
- drm_gem_object_put_unlocked(&bo->base);
- return ret;
- }
- int v3d_mmap_bo_ioctl(struct drm_device *dev, void *data,
- struct drm_file *file_priv)
- {
- struct drm_v3d_mmap_bo *args = data;
- struct drm_gem_object *gem_obj;
- int ret;
- if (args->flags != 0) {
- DRM_INFO("unknown mmap_bo flags: %d\n", args->flags);
- return -EINVAL;
- }
- gem_obj = drm_gem_object_lookup(file_priv, args->handle);
- if (!gem_obj) {
- DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle);
- return -ENOENT;
- }
- ret = drm_gem_create_mmap_offset(gem_obj);
- if (ret == 0)
- args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node);
- drm_gem_object_put_unlocked(gem_obj);
- return ret;
- }
- int v3d_get_bo_offset_ioctl(struct drm_device *dev, void *data,
- struct drm_file *file_priv)
- {
- struct drm_v3d_get_bo_offset *args = data;
- struct drm_gem_object *gem_obj;
- struct v3d_bo *bo;
- gem_obj = drm_gem_object_lookup(file_priv, args->handle);
- if (!gem_obj) {
- DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle);
- return -ENOENT;
- }
- bo = to_v3d_bo(gem_obj);
- args->offset = bo->node.start << PAGE_SHIFT;
- drm_gem_object_put_unlocked(gem_obj);
- return 0;
- }
|