Loading drivers/gpu/drm/i915/i915_gem.c +133 −54 Original line number Diff line number Diff line Loading @@ -2713,12 +2713,11 @@ i915_gem_object_set_cpu_read_domain_range(struct drm_gem_object *obj, static int i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, struct drm_file *file_priv, struct drm_i915_gem_exec_object *entry) struct drm_i915_gem_exec_object *entry, struct drm_i915_gem_relocation_entry *relocs) { struct drm_device *dev = obj->dev; drm_i915_private_t *dev_priv = dev->dev_private; struct drm_i915_gem_relocation_entry reloc; struct drm_i915_gem_relocation_entry __user *relocs; struct drm_i915_gem_object *obj_priv = obj->driver_private; int i, ret; void __iomem *reloc_page; Loading @@ -2730,25 +2729,18 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, entry->offset = obj_priv->gtt_offset; relocs = (struct drm_i915_gem_relocation_entry __user *) (uintptr_t) entry->relocs_ptr; /* Apply the relocations, using the GTT aperture to avoid cache * flushing requirements. */ for (i = 0; i < entry->relocation_count; i++) { struct drm_i915_gem_relocation_entry *reloc= &relocs[i]; struct drm_gem_object *target_obj; struct drm_i915_gem_object *target_obj_priv; uint32_t reloc_val, reloc_offset; uint32_t __iomem *reloc_entry; ret = copy_from_user(&reloc, relocs + i, sizeof(reloc)); if (ret != 0) { i915_gem_object_unpin(obj); return ret; } target_obj = drm_gem_object_lookup(obj->dev, file_priv, reloc.target_handle); reloc->target_handle); if (target_obj == NULL) { i915_gem_object_unpin(obj); return -EBADF; Loading @@ -2760,53 +2752,53 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, */ if (target_obj_priv->gtt_space == NULL) { DRM_ERROR("No GTT space found for object %d\n", reloc.target_handle); reloc->target_handle); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return -EINVAL; } if (reloc.offset > obj->size - 4) { if (reloc->offset > obj->size - 4) { DRM_ERROR("Relocation beyond object bounds: " "obj %p target %d offset %d size %d.\n", obj, reloc.target_handle, (int) reloc.offset, (int) obj->size); obj, reloc->target_handle, (int) reloc->offset, (int) obj->size); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return -EINVAL; } if (reloc.offset & 3) { if (reloc->offset & 3) { DRM_ERROR("Relocation not 4-byte aligned: " "obj %p target %d offset %d.\n", obj, reloc.target_handle, (int) reloc.offset); obj, reloc->target_handle, (int) reloc->offset); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return -EINVAL; } if (reloc.write_domain & I915_GEM_DOMAIN_CPU || reloc.read_domains & I915_GEM_DOMAIN_CPU) { if (reloc->write_domain & I915_GEM_DOMAIN_CPU || reloc->read_domains & I915_GEM_DOMAIN_CPU) { DRM_ERROR("reloc with read/write CPU domains: " "obj %p target %d offset %d " "read %08x write %08x", obj, reloc.target_handle, (int) reloc.offset, reloc.read_domains, reloc.write_domain); obj, reloc->target_handle, (int) reloc->offset, reloc->read_domains, reloc->write_domain); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return -EINVAL; } if (reloc.write_domain && target_obj->pending_write_domain && reloc.write_domain != target_obj->pending_write_domain) { if (reloc->write_domain && target_obj->pending_write_domain && reloc->write_domain != target_obj->pending_write_domain) { DRM_ERROR("Write domain conflict: " "obj %p target %d offset %d " "new %08x old %08x\n", obj, reloc.target_handle, (int) reloc.offset, reloc.write_domain, obj, reloc->target_handle, (int) reloc->offset, reloc->write_domain, target_obj->pending_write_domain); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); Loading @@ -2819,22 +2811,22 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, "presumed %08x delta %08x\n", __func__, obj, (int) reloc.offset, (int) reloc.target_handle, (int) reloc.read_domains, (int) reloc.write_domain, (int) reloc->offset, (int) reloc->target_handle, (int) reloc->read_domains, (int) reloc->write_domain, (int) target_obj_priv->gtt_offset, (int) reloc.presumed_offset, reloc.delta); (int) reloc->presumed_offset, reloc->delta); #endif target_obj->pending_read_domains |= reloc.read_domains; target_obj->pending_write_domain |= reloc.write_domain; target_obj->pending_read_domains |= reloc->read_domains; target_obj->pending_write_domain |= reloc->write_domain; /* If the relocation already has the right value in it, no * more work needs to be done. */ if (target_obj_priv->gtt_offset == reloc.presumed_offset) { if (target_obj_priv->gtt_offset == reloc->presumed_offset) { drm_gem_object_unreference(target_obj); continue; } Loading @@ -2849,32 +2841,26 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, /* Map the page containing the relocation we're going to * perform. */ reloc_offset = obj_priv->gtt_offset + reloc.offset; reloc_offset = obj_priv->gtt_offset + reloc->offset; reloc_page = io_mapping_map_atomic_wc(dev_priv->mm.gtt_mapping, (reloc_offset & ~(PAGE_SIZE - 1))); reloc_entry = (uint32_t __iomem *)(reloc_page + (reloc_offset & (PAGE_SIZE - 1))); reloc_val = target_obj_priv->gtt_offset + reloc.delta; reloc_val = target_obj_priv->gtt_offset + reloc->delta; #if WATCH_BUF DRM_INFO("Applied relocation: %p@0x%08x %08x -> %08x\n", obj, (unsigned int) reloc.offset, obj, (unsigned int) reloc->offset, readl(reloc_entry), reloc_val); #endif writel(reloc_val, reloc_entry); io_mapping_unmap_atomic(reloc_page); /* Write the updated presumed offset for this entry back out * to the user. /* The updated presumed offset for this entry will be * copied back out to the user. */ reloc.presumed_offset = target_obj_priv->gtt_offset; ret = copy_to_user(relocs + i, &reloc, sizeof(reloc)); if (ret != 0) { drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return ret; } reloc->presumed_offset = target_obj_priv->gtt_offset; drm_gem_object_unreference(target_obj); } Loading Loading @@ -2971,6 +2957,75 @@ i915_gem_ring_throttle(struct drm_device *dev, struct drm_file *file_priv) return ret; } static int i915_gem_get_relocs_from_user(struct drm_i915_gem_exec_object *exec_list, uint32_t buffer_count, struct drm_i915_gem_relocation_entry **relocs) { uint32_t reloc_count = 0, reloc_index = 0, i; int ret; *relocs = NULL; for (i = 0; i < buffer_count; i++) { if (reloc_count + exec_list[i].relocation_count < reloc_count) return -EINVAL; reloc_count += exec_list[i].relocation_count; } *relocs = drm_calloc(reloc_count, sizeof(**relocs), DRM_MEM_DRIVER); if (*relocs == NULL) return -ENOMEM; for (i = 0; i < buffer_count; i++) { struct drm_i915_gem_relocation_entry __user *user_relocs; user_relocs = (void __user *)(uintptr_t)exec_list[i].relocs_ptr; ret = copy_from_user(&(*relocs)[reloc_index], user_relocs, exec_list[i].relocation_count * sizeof(**relocs)); if (ret != 0) { drm_free(*relocs, reloc_count * sizeof(**relocs), DRM_MEM_DRIVER); *relocs = NULL; return ret; } reloc_index += exec_list[i].relocation_count; } return ret; } static int i915_gem_put_relocs_to_user(struct drm_i915_gem_exec_object *exec_list, uint32_t buffer_count, struct drm_i915_gem_relocation_entry *relocs) { uint32_t reloc_count = 0, i; int ret; for (i = 0; i < buffer_count; i++) { struct drm_i915_gem_relocation_entry __user *user_relocs; user_relocs = (void __user *)(uintptr_t)exec_list[i].relocs_ptr; if (ret == 0) { ret = copy_to_user(user_relocs, &relocs[reloc_count], exec_list[i].relocation_count * sizeof(*relocs)); } reloc_count += exec_list[i].relocation_count; } drm_free(relocs, reloc_count * sizeof(*relocs), DRM_MEM_DRIVER); return ret; } int i915_gem_execbuffer(struct drm_device *dev, void *data, struct drm_file *file_priv) Loading @@ -2983,9 +3038,10 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, struct drm_gem_object *batch_obj; struct drm_i915_gem_object *obj_priv; struct drm_clip_rect *cliprects = NULL; int ret, i, pinned = 0; struct drm_i915_gem_relocation_entry *relocs; int ret, ret2, i, pinned = 0; uint64_t exec_offset; uint32_t seqno, flush_domains; uint32_t seqno, flush_domains, reloc_index; int pin_tries; #if WATCH_EXEC Loading Loading @@ -3036,6 +3092,11 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, } } ret = i915_gem_get_relocs_from_user(exec_list, args->buffer_count, &relocs); if (ret != 0) goto pre_mutex_err; mutex_lock(&dev->struct_mutex); i915_verify_inactive(dev, __FILE__, __LINE__); Loading Loading @@ -3078,15 +3139,19 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, /* Pin and relocate */ for (pin_tries = 0; ; pin_tries++) { ret = 0; reloc_index = 0; for (i = 0; i < args->buffer_count; i++) { object_list[i]->pending_read_domains = 0; object_list[i]->pending_write_domain = 0; ret = i915_gem_object_pin_and_relocate(object_list[i], file_priv, &exec_list[i]); &exec_list[i], &relocs[reloc_index]); if (ret) break; pinned = i + 1; reloc_index += exec_list[i].relocation_count; } /* success */ if (ret == 0) Loading Loading @@ -3236,6 +3301,20 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, args->buffer_count, ret); } /* Copy the updated relocations out regardless of current error * state. Failure to update the relocs would mean that the next * time userland calls execbuf, it would do so with presumed offset * state that didn't match the actual object state. */ ret2 = i915_gem_put_relocs_to_user(exec_list, args->buffer_count, relocs); if (ret2 != 0) { DRM_ERROR("Failed to copy relocations back out: %d\n", ret2); if (ret == 0) ret = ret2; } pre_mutex_err: drm_free(object_list, sizeof(*object_list) * args->buffer_count, DRM_MEM_DRIVER); Loading Loading
drivers/gpu/drm/i915/i915_gem.c +133 −54 Original line number Diff line number Diff line Loading @@ -2713,12 +2713,11 @@ i915_gem_object_set_cpu_read_domain_range(struct drm_gem_object *obj, static int i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, struct drm_file *file_priv, struct drm_i915_gem_exec_object *entry) struct drm_i915_gem_exec_object *entry, struct drm_i915_gem_relocation_entry *relocs) { struct drm_device *dev = obj->dev; drm_i915_private_t *dev_priv = dev->dev_private; struct drm_i915_gem_relocation_entry reloc; struct drm_i915_gem_relocation_entry __user *relocs; struct drm_i915_gem_object *obj_priv = obj->driver_private; int i, ret; void __iomem *reloc_page; Loading @@ -2730,25 +2729,18 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, entry->offset = obj_priv->gtt_offset; relocs = (struct drm_i915_gem_relocation_entry __user *) (uintptr_t) entry->relocs_ptr; /* Apply the relocations, using the GTT aperture to avoid cache * flushing requirements. */ for (i = 0; i < entry->relocation_count; i++) { struct drm_i915_gem_relocation_entry *reloc= &relocs[i]; struct drm_gem_object *target_obj; struct drm_i915_gem_object *target_obj_priv; uint32_t reloc_val, reloc_offset; uint32_t __iomem *reloc_entry; ret = copy_from_user(&reloc, relocs + i, sizeof(reloc)); if (ret != 0) { i915_gem_object_unpin(obj); return ret; } target_obj = drm_gem_object_lookup(obj->dev, file_priv, reloc.target_handle); reloc->target_handle); if (target_obj == NULL) { i915_gem_object_unpin(obj); return -EBADF; Loading @@ -2760,53 +2752,53 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, */ if (target_obj_priv->gtt_space == NULL) { DRM_ERROR("No GTT space found for object %d\n", reloc.target_handle); reloc->target_handle); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return -EINVAL; } if (reloc.offset > obj->size - 4) { if (reloc->offset > obj->size - 4) { DRM_ERROR("Relocation beyond object bounds: " "obj %p target %d offset %d size %d.\n", obj, reloc.target_handle, (int) reloc.offset, (int) obj->size); obj, reloc->target_handle, (int) reloc->offset, (int) obj->size); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return -EINVAL; } if (reloc.offset & 3) { if (reloc->offset & 3) { DRM_ERROR("Relocation not 4-byte aligned: " "obj %p target %d offset %d.\n", obj, reloc.target_handle, (int) reloc.offset); obj, reloc->target_handle, (int) reloc->offset); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return -EINVAL; } if (reloc.write_domain & I915_GEM_DOMAIN_CPU || reloc.read_domains & I915_GEM_DOMAIN_CPU) { if (reloc->write_domain & I915_GEM_DOMAIN_CPU || reloc->read_domains & I915_GEM_DOMAIN_CPU) { DRM_ERROR("reloc with read/write CPU domains: " "obj %p target %d offset %d " "read %08x write %08x", obj, reloc.target_handle, (int) reloc.offset, reloc.read_domains, reloc.write_domain); obj, reloc->target_handle, (int) reloc->offset, reloc->read_domains, reloc->write_domain); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return -EINVAL; } if (reloc.write_domain && target_obj->pending_write_domain && reloc.write_domain != target_obj->pending_write_domain) { if (reloc->write_domain && target_obj->pending_write_domain && reloc->write_domain != target_obj->pending_write_domain) { DRM_ERROR("Write domain conflict: " "obj %p target %d offset %d " "new %08x old %08x\n", obj, reloc.target_handle, (int) reloc.offset, reloc.write_domain, obj, reloc->target_handle, (int) reloc->offset, reloc->write_domain, target_obj->pending_write_domain); drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); Loading @@ -2819,22 +2811,22 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, "presumed %08x delta %08x\n", __func__, obj, (int) reloc.offset, (int) reloc.target_handle, (int) reloc.read_domains, (int) reloc.write_domain, (int) reloc->offset, (int) reloc->target_handle, (int) reloc->read_domains, (int) reloc->write_domain, (int) target_obj_priv->gtt_offset, (int) reloc.presumed_offset, reloc.delta); (int) reloc->presumed_offset, reloc->delta); #endif target_obj->pending_read_domains |= reloc.read_domains; target_obj->pending_write_domain |= reloc.write_domain; target_obj->pending_read_domains |= reloc->read_domains; target_obj->pending_write_domain |= reloc->write_domain; /* If the relocation already has the right value in it, no * more work needs to be done. */ if (target_obj_priv->gtt_offset == reloc.presumed_offset) { if (target_obj_priv->gtt_offset == reloc->presumed_offset) { drm_gem_object_unreference(target_obj); continue; } Loading @@ -2849,32 +2841,26 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, /* Map the page containing the relocation we're going to * perform. */ reloc_offset = obj_priv->gtt_offset + reloc.offset; reloc_offset = obj_priv->gtt_offset + reloc->offset; reloc_page = io_mapping_map_atomic_wc(dev_priv->mm.gtt_mapping, (reloc_offset & ~(PAGE_SIZE - 1))); reloc_entry = (uint32_t __iomem *)(reloc_page + (reloc_offset & (PAGE_SIZE - 1))); reloc_val = target_obj_priv->gtt_offset + reloc.delta; reloc_val = target_obj_priv->gtt_offset + reloc->delta; #if WATCH_BUF DRM_INFO("Applied relocation: %p@0x%08x %08x -> %08x\n", obj, (unsigned int) reloc.offset, obj, (unsigned int) reloc->offset, readl(reloc_entry), reloc_val); #endif writel(reloc_val, reloc_entry); io_mapping_unmap_atomic(reloc_page); /* Write the updated presumed offset for this entry back out * to the user. /* The updated presumed offset for this entry will be * copied back out to the user. */ reloc.presumed_offset = target_obj_priv->gtt_offset; ret = copy_to_user(relocs + i, &reloc, sizeof(reloc)); if (ret != 0) { drm_gem_object_unreference(target_obj); i915_gem_object_unpin(obj); return ret; } reloc->presumed_offset = target_obj_priv->gtt_offset; drm_gem_object_unreference(target_obj); } Loading Loading @@ -2971,6 +2957,75 @@ i915_gem_ring_throttle(struct drm_device *dev, struct drm_file *file_priv) return ret; } static int i915_gem_get_relocs_from_user(struct drm_i915_gem_exec_object *exec_list, uint32_t buffer_count, struct drm_i915_gem_relocation_entry **relocs) { uint32_t reloc_count = 0, reloc_index = 0, i; int ret; *relocs = NULL; for (i = 0; i < buffer_count; i++) { if (reloc_count + exec_list[i].relocation_count < reloc_count) return -EINVAL; reloc_count += exec_list[i].relocation_count; } *relocs = drm_calloc(reloc_count, sizeof(**relocs), DRM_MEM_DRIVER); if (*relocs == NULL) return -ENOMEM; for (i = 0; i < buffer_count; i++) { struct drm_i915_gem_relocation_entry __user *user_relocs; user_relocs = (void __user *)(uintptr_t)exec_list[i].relocs_ptr; ret = copy_from_user(&(*relocs)[reloc_index], user_relocs, exec_list[i].relocation_count * sizeof(**relocs)); if (ret != 0) { drm_free(*relocs, reloc_count * sizeof(**relocs), DRM_MEM_DRIVER); *relocs = NULL; return ret; } reloc_index += exec_list[i].relocation_count; } return ret; } static int i915_gem_put_relocs_to_user(struct drm_i915_gem_exec_object *exec_list, uint32_t buffer_count, struct drm_i915_gem_relocation_entry *relocs) { uint32_t reloc_count = 0, i; int ret; for (i = 0; i < buffer_count; i++) { struct drm_i915_gem_relocation_entry __user *user_relocs; user_relocs = (void __user *)(uintptr_t)exec_list[i].relocs_ptr; if (ret == 0) { ret = copy_to_user(user_relocs, &relocs[reloc_count], exec_list[i].relocation_count * sizeof(*relocs)); } reloc_count += exec_list[i].relocation_count; } drm_free(relocs, reloc_count * sizeof(*relocs), DRM_MEM_DRIVER); return ret; } int i915_gem_execbuffer(struct drm_device *dev, void *data, struct drm_file *file_priv) Loading @@ -2983,9 +3038,10 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, struct drm_gem_object *batch_obj; struct drm_i915_gem_object *obj_priv; struct drm_clip_rect *cliprects = NULL; int ret, i, pinned = 0; struct drm_i915_gem_relocation_entry *relocs; int ret, ret2, i, pinned = 0; uint64_t exec_offset; uint32_t seqno, flush_domains; uint32_t seqno, flush_domains, reloc_index; int pin_tries; #if WATCH_EXEC Loading Loading @@ -3036,6 +3092,11 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, } } ret = i915_gem_get_relocs_from_user(exec_list, args->buffer_count, &relocs); if (ret != 0) goto pre_mutex_err; mutex_lock(&dev->struct_mutex); i915_verify_inactive(dev, __FILE__, __LINE__); Loading Loading @@ -3078,15 +3139,19 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, /* Pin and relocate */ for (pin_tries = 0; ; pin_tries++) { ret = 0; reloc_index = 0; for (i = 0; i < args->buffer_count; i++) { object_list[i]->pending_read_domains = 0; object_list[i]->pending_write_domain = 0; ret = i915_gem_object_pin_and_relocate(object_list[i], file_priv, &exec_list[i]); &exec_list[i], &relocs[reloc_index]); if (ret) break; pinned = i + 1; reloc_index += exec_list[i].relocation_count; } /* success */ if (ret == 0) Loading Loading @@ -3236,6 +3301,20 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, args->buffer_count, ret); } /* Copy the updated relocations out regardless of current error * state. Failure to update the relocs would mean that the next * time userland calls execbuf, it would do so with presumed offset * state that didn't match the actual object state. */ ret2 = i915_gem_put_relocs_to_user(exec_list, args->buffer_count, relocs); if (ret2 != 0) { DRM_ERROR("Failed to copy relocations back out: %d\n", ret2); if (ret == 0) ret = ret2; } pre_mutex_err: drm_free(object_list, sizeof(*object_list) * args->buffer_count, DRM_MEM_DRIVER); Loading