]> Git Repo - J-linux.git/blobdiff - drivers/gpu/drm/virtio/virtgpu_vq.c
Merge tag '9p-for-4.20' of git://github.com/martinetd/linux
[J-linux.git] / drivers / gpu / drm / virtio / virtgpu_vq.c
index df32811f2c3e21014526bb5a544bbc65f8e3348e..4e2e037aed34a22551225f2cb623f5492484c72a 100644 (file)
@@ -483,28 +483,26 @@ void virtio_gpu_cmd_resource_flush(struct virtio_gpu_device *vgdev,
 }
 
 void virtio_gpu_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev,
-                                       uint32_t resource_id, uint64_t offset,
+                                       struct virtio_gpu_object *bo,
+                                       uint64_t offset,
                                        __le32 width, __le32 height,
                                        __le32 x, __le32 y,
                                        struct virtio_gpu_fence **fence)
 {
        struct virtio_gpu_transfer_to_host_2d *cmd_p;
        struct virtio_gpu_vbuffer *vbuf;
-       struct virtio_gpu_fbdev *vgfbdev = vgdev->vgfbdev;
-       struct virtio_gpu_framebuffer *fb = &vgfbdev->vgfb;
-       struct virtio_gpu_object *obj = gem_to_virtio_gpu_obj(fb->base.obj[0]);
        bool use_dma_api = !virtio_has_iommu_quirk(vgdev->vdev);
 
        if (use_dma_api)
                dma_sync_sg_for_device(vgdev->vdev->dev.parent,
-                                      obj->pages->sgl, obj->pages->nents,
+                                      bo->pages->sgl, bo->pages->nents,
                                       DMA_TO_DEVICE);
 
        cmd_p = virtio_gpu_alloc_cmd(vgdev, &vbuf, sizeof(*cmd_p));
        memset(cmd_p, 0, sizeof(*cmd_p));
 
        cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_TRANSFER_TO_HOST_2D);
-       cmd_p->resource_id = cpu_to_le32(resource_id);
+       cmd_p->resource_id = cpu_to_le32(bo->hw_res_handle);
        cmd_p->offset = cpu_to_le64(offset);
        cmd_p->r.width = width;
        cmd_p->r.height = height;
@@ -791,21 +789,19 @@ virtio_gpu_cmd_resource_create_3d(struct virtio_gpu_device *vgdev,
 }
 
 void virtio_gpu_cmd_transfer_to_host_3d(struct virtio_gpu_device *vgdev,
-                                       uint32_t resource_id, uint32_t ctx_id,
+                                       struct virtio_gpu_object *bo,
+                                       uint32_t ctx_id,
                                        uint64_t offset, uint32_t level,
                                        struct virtio_gpu_box *box,
                                        struct virtio_gpu_fence **fence)
 {
        struct virtio_gpu_transfer_host_3d *cmd_p;
        struct virtio_gpu_vbuffer *vbuf;
-       struct virtio_gpu_fbdev *vgfbdev = vgdev->vgfbdev;
-       struct virtio_gpu_framebuffer *fb = &vgfbdev->vgfb;
-       struct virtio_gpu_object *obj = gem_to_virtio_gpu_obj(fb->base.obj[0]);
        bool use_dma_api = !virtio_has_iommu_quirk(vgdev->vdev);
 
        if (use_dma_api)
                dma_sync_sg_for_device(vgdev->vdev->dev.parent,
-                                      obj->pages->sgl, obj->pages->nents,
+                                      bo->pages->sgl, bo->pages->nents,
                                       DMA_TO_DEVICE);
 
        cmd_p = virtio_gpu_alloc_cmd(vgdev, &vbuf, sizeof(*cmd_p));
@@ -813,7 +809,7 @@ void virtio_gpu_cmd_transfer_to_host_3d(struct virtio_gpu_device *vgdev,
 
        cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_TRANSFER_TO_HOST_3D);
        cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id);
-       cmd_p->resource_id = cpu_to_le32(resource_id);
+       cmd_p->resource_id = cpu_to_le32(bo->hw_res_handle);
        cmd_p->box = *box;
        cmd_p->offset = cpu_to_le64(offset);
        cmd_p->level = cpu_to_le32(level);
This page took 0.032771 seconds and 4 git commands to generate.