Commit a9f58c45 authored by Thomas Hellstrom's avatar Thomas Hellstrom Committed by Deepak Rawat

drm/vmwgfx: Be more restrictive when dirtying resources

Currently we flag resources as dirty (GPU contents not yet read back to
the backing MOB) whenever they have been part of a command stream.
Obviously many resources can't be dirty and others can only be dirty when
written to by the GPU. That is when they are either bound to the context as
render-targets, depth-stencil, copy / clear destinations and
stream-output targets, or similarly when there are corresponding views into
them.
So mark resources dirty only in these special cases. Context- and cotable
resources are always marked dirty when referenced.
This is important for upcoming emulated coherent memory, since we can avoid
issuing automatic readbacks to non-dirty resources when the CPU tries to
access part of the backing MOB.

Testing: Unigine Heaven with max GPU memory set to 256MB resulting in
heavy resource thrashing.
---
v2: Addressed review comments by Deepak Rawat.
v3: Added some documentation
Signed-off-by: default avatarThomas Hellstrom <thellstrom@vmware.com>
Reviewed-by: default avatarDeepak Rawat <drawat@vmware.com>
parent 14d2bd53
...@@ -1269,6 +1269,32 @@ void vmw_binding_state_reset(struct vmw_ctx_binding_state *cbs) ...@@ -1269,6 +1269,32 @@ void vmw_binding_state_reset(struct vmw_ctx_binding_state *cbs)
vmw_binding_drop(entry); vmw_binding_drop(entry);
} }
/**
* vmw_binding_dirtying - Return whether a binding type is dirtying its resource
* @binding_type: The binding type
*
* Each time a resource is put on the validation list as the result of a
* context binding referencing it, we need to determine whether that resource
* will be dirtied (written to by the GPU) as a result of the corresponding
* GPU operation. Currently rendertarget-, depth-stencil-, and
* stream-output-target bindings are capable of dirtying its resource.
*
* Return: Whether the binding type dirties the resource its binding points to.
*/
u32 vmw_binding_dirtying(enum vmw_ctx_binding_type binding_type)
{
static u32 is_binding_dirtying[vmw_ctx_binding_max] = {
[vmw_ctx_binding_rt] = VMW_RES_DIRTY_SET,
[vmw_ctx_binding_dx_rt] = VMW_RES_DIRTY_SET,
[vmw_ctx_binding_ds] = VMW_RES_DIRTY_SET,
[vmw_ctx_binding_so] = VMW_RES_DIRTY_SET,
};
/* Review this function as new bindings are added. */
BUILD_BUG_ON(vmw_ctx_binding_max != 11);
return is_binding_dirtying[binding_type];
}
/* /*
* This function is unused at run-time, and only used to hold various build * This function is unused at run-time, and only used to hold various build
* asserts important for code optimization assumptions. * asserts important for code optimization assumptions.
......
...@@ -205,5 +205,7 @@ extern void vmw_binding_state_free(struct vmw_ctx_binding_state *cbs); ...@@ -205,5 +205,7 @@ extern void vmw_binding_state_free(struct vmw_ctx_binding_state *cbs);
extern struct list_head * extern struct list_head *
vmw_binding_state_list(struct vmw_ctx_binding_state *cbs); vmw_binding_state_list(struct vmw_ctx_binding_state *cbs);
extern void vmw_binding_state_reset(struct vmw_ctx_binding_state *cbs); extern void vmw_binding_state_reset(struct vmw_ctx_binding_state *cbs);
extern u32 vmw_binding_dirtying(enum vmw_ctx_binding_type binding_type);
#endif #endif
...@@ -699,6 +699,8 @@ extern int vmw_user_stream_lookup(struct vmw_private *dev_priv, ...@@ -699,6 +699,8 @@ extern int vmw_user_stream_lookup(struct vmw_private *dev_priv,
uint32_t *inout_id, uint32_t *inout_id,
struct vmw_resource **out); struct vmw_resource **out);
extern void vmw_resource_unreserve(struct vmw_resource *res, extern void vmw_resource_unreserve(struct vmw_resource *res,
bool dirty_set,
bool dirty,
bool switch_backup, bool switch_backup,
struct vmw_buffer_object *new_backup, struct vmw_buffer_object *new_backup,
unsigned long new_backup_offset); unsigned long new_backup_offset);
......
This diff is collapsed.
...@@ -1202,7 +1202,7 @@ static int vmw_create_bo_proxy(struct drm_device *dev, ...@@ -1202,7 +1202,7 @@ static int vmw_create_bo_proxy(struct drm_device *dev,
vmw_bo_unreference(&res->backup); vmw_bo_unreference(&res->backup);
res->backup = vmw_bo_reference(bo_mob); res->backup = vmw_bo_reference(bo_mob);
res->backup_offset = 0; res->backup_offset = 0;
vmw_resource_unreserve(res, false, NULL, 0); vmw_resource_unreserve(res, false, false, false, NULL, 0);
mutex_unlock(&res->dev_priv->cmdbuf_mutex); mutex_unlock(&res->dev_priv->cmdbuf_mutex);
return 0; return 0;
...@@ -2827,7 +2827,8 @@ int vmw_du_helper_plane_update(struct vmw_du_update_plane *update) ...@@ -2827,7 +2827,8 @@ int vmw_du_helper_plane_update(struct vmw_du_update_plane *update)
container_of(update->vfb, typeof(*vfbs), base); container_of(update->vfb, typeof(*vfbs), base);
ret = vmw_validation_add_resource(&val_ctx, &vfbs->surface->res, ret = vmw_validation_add_resource(&val_ctx, &vfbs->surface->res,
0, NULL, NULL); 0, VMW_RES_DIRTY_NONE, NULL,
NULL);
} }
if (ret) if (ret)
......
...@@ -365,14 +365,6 @@ static int vmw_resource_do_validate(struct vmw_resource *res, ...@@ -365,14 +365,6 @@ static int vmw_resource_do_validate(struct vmw_resource *res,
list_add_tail(&res->mob_head, &res->backup->res_list); list_add_tail(&res->mob_head, &res->backup->res_list);
} }
/*
* Only do this on write operations, and move to
* vmw_resource_unreserve if it can be called after
* backup buffers have been unreserved. Otherwise
* sort out locking.
*/
res->res_dirty = true;
return 0; return 0;
out_bind_failed: out_bind_failed:
...@@ -386,6 +378,8 @@ static int vmw_resource_do_validate(struct vmw_resource *res, ...@@ -386,6 +378,8 @@ static int vmw_resource_do_validate(struct vmw_resource *res,
* command submission. * command submission.
* *
* @res: Pointer to the struct vmw_resource to unreserve. * @res: Pointer to the struct vmw_resource to unreserve.
* @dirty_set: Change dirty status of the resource.
* @dirty: When changing dirty status indicates the new status.
* @switch_backup: Backup buffer has been switched. * @switch_backup: Backup buffer has been switched.
* @new_backup: Pointer to new backup buffer if command submission * @new_backup: Pointer to new backup buffer if command submission
* switched. May be NULL. * switched. May be NULL.
...@@ -395,6 +389,8 @@ static int vmw_resource_do_validate(struct vmw_resource *res, ...@@ -395,6 +389,8 @@ static int vmw_resource_do_validate(struct vmw_resource *res,
* resource lru list, so that it can be evicted if necessary. * resource lru list, so that it can be evicted if necessary.
*/ */
void vmw_resource_unreserve(struct vmw_resource *res, void vmw_resource_unreserve(struct vmw_resource *res,
bool dirty_set,
bool dirty,
bool switch_backup, bool switch_backup,
struct vmw_buffer_object *new_backup, struct vmw_buffer_object *new_backup,
unsigned long new_backup_offset) unsigned long new_backup_offset)
...@@ -422,6 +418,9 @@ void vmw_resource_unreserve(struct vmw_resource *res, ...@@ -422,6 +418,9 @@ void vmw_resource_unreserve(struct vmw_resource *res,
if (switch_backup) if (switch_backup)
res->backup_offset = new_backup_offset; res->backup_offset = new_backup_offset;
if (dirty_set)
res->res_dirty = dirty;
if (!res->func->may_evict || res->id == -1 || res->pin_count) if (!res->func->may_evict || res->id == -1 || res->pin_count)
return; return;
...@@ -696,7 +695,7 @@ void vmw_resource_unbind_list(struct vmw_buffer_object *vbo) ...@@ -696,7 +695,7 @@ void vmw_resource_unbind_list(struct vmw_buffer_object *vbo)
if (!res->func->unbind) if (!res->func->unbind)
continue; continue;
(void) res->func->unbind(res, true, &val_buf); (void) res->func->unbind(res, res->res_dirty, &val_buf);
res->backup_dirty = true; res->backup_dirty = true;
res->res_dirty = false; res->res_dirty = false;
list_del_init(&res->mob_head); list_del_init(&res->mob_head);
...@@ -932,7 +931,7 @@ int vmw_resource_pin(struct vmw_resource *res, bool interruptible) ...@@ -932,7 +931,7 @@ int vmw_resource_pin(struct vmw_resource *res, bool interruptible)
res->pin_count++; res->pin_count++;
out_no_validate: out_no_validate:
vmw_resource_unreserve(res, false, NULL, 0UL); vmw_resource_unreserve(res, false, false, false, NULL, 0UL);
out_no_reserve: out_no_reserve:
mutex_unlock(&dev_priv->cmdbuf_mutex); mutex_unlock(&dev_priv->cmdbuf_mutex);
ttm_write_unlock(&dev_priv->reservation_sem); ttm_write_unlock(&dev_priv->reservation_sem);
...@@ -968,7 +967,7 @@ void vmw_resource_unpin(struct vmw_resource *res) ...@@ -968,7 +967,7 @@ void vmw_resource_unpin(struct vmw_resource *res)
ttm_bo_unreserve(&vbo->base); ttm_bo_unreserve(&vbo->base);
} }
vmw_resource_unreserve(res, false, NULL, 0UL); vmw_resource_unreserve(res, false, false, false, NULL, 0UL);
mutex_unlock(&dev_priv->cmdbuf_mutex); mutex_unlock(&dev_priv->cmdbuf_mutex);
ttm_read_unlock(&dev_priv->reservation_sem); ttm_read_unlock(&dev_priv->reservation_sem);
......
...@@ -1148,7 +1148,8 @@ int vmw_kms_sou_do_surface_dirty(struct vmw_private *dev_priv, ...@@ -1148,7 +1148,8 @@ int vmw_kms_sou_do_surface_dirty(struct vmw_private *dev_priv,
if (!srf) if (!srf)
srf = &vfbs->surface->res; srf = &vfbs->surface->res;
ret = vmw_validation_add_resource(&val_ctx, srf, 0, NULL, NULL); ret = vmw_validation_add_resource(&val_ctx, srf, 0, VMW_RES_DIRTY_NONE,
NULL, NULL);
if (ret) if (ret)
return ret; return ret;
......
...@@ -497,6 +497,30 @@ struct vmw_resource *vmw_view_lookup(struct vmw_cmdbuf_res_manager *man, ...@@ -497,6 +497,30 @@ struct vmw_resource *vmw_view_lookup(struct vmw_cmdbuf_res_manager *man,
vmw_view_key(user_key, view_type)); vmw_view_key(user_key, view_type));
} }
/**
* vmw_view_dirtying - Return whether a view type is dirtying its resource
* @res: Pointer to the view
*
* Each time a resource is put on the validation list as the result of a
* view pointing to it, we need to determine whether that resource will
* be dirtied (written to by the GPU) as a result of the corresponding
* GPU operation. Currently only rendertarget- and depth-stencil views are
* capable of dirtying its resource.
*
* Return: Whether the view type of @res dirties the resource it points to.
*/
u32 vmw_view_dirtying(struct vmw_resource *res)
{
static u32 view_is_dirtying[vmw_view_max] = {
[vmw_view_rt] = VMW_RES_DIRTY_SET,
[vmw_view_ds] = VMW_RES_DIRTY_SET,
};
/* Update this function as we add more view types */
BUILD_BUG_ON(vmw_view_max != 3);
return view_is_dirtying[vmw_view(res)->view_type];
}
const u32 vmw_view_destroy_cmds[] = { const u32 vmw_view_destroy_cmds[] = {
[vmw_view_sr] = SVGA_3D_CMD_DX_DESTROY_SHADERRESOURCE_VIEW, [vmw_view_sr] = SVGA_3D_CMD_DX_DESTROY_SHADERRESOURCE_VIEW,
[vmw_view_rt] = SVGA_3D_CMD_DX_DESTROY_RENDERTARGET_VIEW, [vmw_view_rt] = SVGA_3D_CMD_DX_DESTROY_RENDERTARGET_VIEW,
......
...@@ -157,4 +157,5 @@ extern struct vmw_resource *vmw_view_srf(struct vmw_resource *res); ...@@ -157,4 +157,5 @@ extern struct vmw_resource *vmw_view_srf(struct vmw_resource *res);
extern struct vmw_resource *vmw_view_lookup(struct vmw_cmdbuf_res_manager *man, extern struct vmw_resource *vmw_view_lookup(struct vmw_cmdbuf_res_manager *man,
enum vmw_view_type view_type, enum vmw_view_type view_type,
u32 user_key); u32 user_key);
extern u32 vmw_view_dirtying(struct vmw_resource *res);
#endif #endif
...@@ -111,7 +111,7 @@ struct vmw_stdu_update_gb_image { ...@@ -111,7 +111,7 @@ struct vmw_stdu_update_gb_image {
*/ */
struct vmw_screen_target_display_unit { struct vmw_screen_target_display_unit {
struct vmw_display_unit base; struct vmw_display_unit base;
const struct vmw_surface *display_srf; struct vmw_surface *display_srf;
enum stdu_content_type content_fb_type; enum stdu_content_type content_fb_type;
s32 display_width, display_height; s32 display_width, display_height;
...@@ -533,6 +533,7 @@ static void vmw_stdu_bo_fifo_commit(struct vmw_kms_dirty *dirty) ...@@ -533,6 +533,7 @@ static void vmw_stdu_bo_fifo_commit(struct vmw_kms_dirty *dirty)
vmw_fifo_commit(dirty->dev_priv, sizeof(*cmd) + blit_size); vmw_fifo_commit(dirty->dev_priv, sizeof(*cmd) + blit_size);
stdu->display_srf->res.res_dirty = true;
ddirty->left = ddirty->top = S32_MAX; ddirty->left = ddirty->top = S32_MAX;
ddirty->right = ddirty->bottom = S32_MIN; ddirty->right = ddirty->bottom = S32_MIN;
} }
...@@ -629,9 +630,8 @@ static void vmw_stdu_bo_cpu_commit(struct vmw_kms_dirty *dirty) ...@@ -629,9 +630,8 @@ static void vmw_stdu_bo_cpu_commit(struct vmw_kms_dirty *dirty)
region.x2 = diff.rect.x2; region.x2 = diff.rect.x2;
region.y1 = diff.rect.y1; region.y1 = diff.rect.y1;
region.y2 = diff.rect.y2; region.y2 = diff.rect.y2;
ret = vmw_kms_update_proxy( ret = vmw_kms_update_proxy(&stdu->display_srf->res, &region,
(struct vmw_resource *) &stdu->display_srf->res, 1, 1);
(const struct drm_clip_rect *) &region, 1, 1);
if (ret) if (ret)
goto out_cleanup; goto out_cleanup;
...@@ -820,6 +820,7 @@ static void vmw_kms_stdu_surface_fifo_commit(struct vmw_kms_dirty *dirty) ...@@ -820,6 +820,7 @@ static void vmw_kms_stdu_surface_fifo_commit(struct vmw_kms_dirty *dirty)
cmd->body.dest.sid = stdu->display_srf->res.id; cmd->body.dest.sid = stdu->display_srf->res.id;
update = (struct vmw_stdu_update *) &blit[dirty->num_hits]; update = (struct vmw_stdu_update *) &blit[dirty->num_hits];
commit_size = sizeof(*cmd) + blit_size + sizeof(*update); commit_size = sizeof(*cmd) + blit_size + sizeof(*update);
stdu->display_srf->res.res_dirty = true;
} else { } else {
update = dirty->cmd; update = dirty->cmd;
commit_size = sizeof(*update); commit_size = sizeof(*update);
...@@ -876,7 +877,8 @@ int vmw_kms_stdu_surface_dirty(struct vmw_private *dev_priv, ...@@ -876,7 +877,8 @@ int vmw_kms_stdu_surface_dirty(struct vmw_private *dev_priv,
if (!srf) if (!srf)
srf = &vfbs->surface->res; srf = &vfbs->surface->res;
ret = vmw_validation_add_resource(&val_ctx, srf, 0, NULL, NULL); ret = vmw_validation_add_resource(&val_ctx, srf, 0, VMW_RES_DIRTY_NONE,
NULL, NULL);
if (ret) if (ret)
return ret; return ret;
......
...@@ -76,6 +76,8 @@ struct vmw_validation_res_node { ...@@ -76,6 +76,8 @@ struct vmw_validation_res_node {
u32 switching_backup : 1; u32 switching_backup : 1;
u32 first_usage : 1; u32 first_usage : 1;
u32 reserved : 1; u32 reserved : 1;
u32 dirty : 1;
u32 dirty_set : 1;
unsigned long private[0]; unsigned long private[0];
}; };
...@@ -299,6 +301,7 @@ int vmw_validation_add_bo(struct vmw_validation_context *ctx, ...@@ -299,6 +301,7 @@ int vmw_validation_add_bo(struct vmw_validation_context *ctx,
* @ctx: The validation context. * @ctx: The validation context.
* @res: The resource. * @res: The resource.
* @priv_size: Size of private, additional metadata. * @priv_size: Size of private, additional metadata.
* @dirty: Whether to change dirty status.
* @p_node: Output pointer of additional metadata address. * @p_node: Output pointer of additional metadata address.
* @first_usage: Whether this was the first time this resource was seen. * @first_usage: Whether this was the first time this resource was seen.
* *
...@@ -307,6 +310,7 @@ int vmw_validation_add_bo(struct vmw_validation_context *ctx, ...@@ -307,6 +310,7 @@ int vmw_validation_add_bo(struct vmw_validation_context *ctx,
int vmw_validation_add_resource(struct vmw_validation_context *ctx, int vmw_validation_add_resource(struct vmw_validation_context *ctx,
struct vmw_resource *res, struct vmw_resource *res,
size_t priv_size, size_t priv_size,
u32 dirty,
void **p_node, void **p_node,
bool *first_usage) bool *first_usage)
{ {
...@@ -358,6 +362,11 @@ int vmw_validation_add_resource(struct vmw_validation_context *ctx, ...@@ -358,6 +362,11 @@ int vmw_validation_add_resource(struct vmw_validation_context *ctx,
} }
out_fill: out_fill:
if (dirty) {
node->dirty_set = 1;
/* Overwriting previous information here is intentional! */
node->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0;
}
if (first_usage) if (first_usage)
*first_usage = node->first_usage; *first_usage = node->first_usage;
if (p_node) if (p_node)
...@@ -366,6 +375,29 @@ int vmw_validation_add_resource(struct vmw_validation_context *ctx, ...@@ -366,6 +375,29 @@ int vmw_validation_add_resource(struct vmw_validation_context *ctx,
return 0; return 0;
} }
/**
* vmw_validation_res_set_dirty - Register a resource dirty set or clear during
* validation.
* @ctx: The validation context.
* @val_private: The additional meta-data pointer returned when the
* resource was registered with the validation context. Used to identify
* the resource.
* @dirty: Dirty information VMW_RES_DIRTY_XX
*/
void vmw_validation_res_set_dirty(struct vmw_validation_context *ctx,
void *val_private, u32 dirty)
{
struct vmw_validation_res_node *val;
if (!dirty)
return;
val = container_of(val_private, typeof(*val), private);
val->dirty_set = 1;
/* Overwriting previous information here is intentional! */
val->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0;
}
/** /**
* vmw_validation_res_switch_backup - Register a backup MOB switch during * vmw_validation_res_switch_backup - Register a backup MOB switch during
* validation. * validation.
...@@ -450,15 +482,23 @@ void vmw_validation_res_unreserve(struct vmw_validation_context *ctx, ...@@ -450,15 +482,23 @@ void vmw_validation_res_unreserve(struct vmw_validation_context *ctx,
struct vmw_validation_res_node *val; struct vmw_validation_res_node *val;
list_splice_init(&ctx->resource_ctx_list, &ctx->resource_list); list_splice_init(&ctx->resource_ctx_list, &ctx->resource_list);
if (backoff)
list_for_each_entry(val, &ctx->resource_list, head) { list_for_each_entry(val, &ctx->resource_list, head) {
if (val->reserved) if (val->reserved)
vmw_resource_unreserve(val->res, vmw_resource_unreserve(val->res,
!backoff && false, false, false,
val->switching_backup, NULL, 0);
val->new_backup, }
val->new_backup_offset); else
} list_for_each_entry(val, &ctx->resource_list, head) {
if (val->reserved)
vmw_resource_unreserve(val->res,
val->dirty_set,
val->dirty,
val->switching_backup,
val->new_backup,
val->new_backup_offset);
}
} }
/** /**
......
...@@ -33,6 +33,10 @@ ...@@ -33,6 +33,10 @@
#include <linux/ww_mutex.h> #include <linux/ww_mutex.h>
#include <drm/ttm/ttm_execbuf_util.h> #include <drm/ttm/ttm_execbuf_util.h>
#define VMW_RES_DIRTY_NONE 0
#define VMW_RES_DIRTY_SET BIT(0)
#define VMW_RES_DIRTY_CLEAR BIT(1)
/** /**
* struct vmw_validation_mem - Custom interface to provide memory reservations * struct vmw_validation_mem - Custom interface to provide memory reservations
* for the validation code. * for the validation code.
...@@ -237,6 +241,7 @@ void vmw_validation_unref_lists(struct vmw_validation_context *ctx); ...@@ -237,6 +241,7 @@ void vmw_validation_unref_lists(struct vmw_validation_context *ctx);
int vmw_validation_add_resource(struct vmw_validation_context *ctx, int vmw_validation_add_resource(struct vmw_validation_context *ctx,
struct vmw_resource *res, struct vmw_resource *res,
size_t priv_size, size_t priv_size,
u32 dirty,
void **p_node, void **p_node,
bool *first_usage); bool *first_usage);
void vmw_validation_drop_ht(struct vmw_validation_context *ctx); void vmw_validation_drop_ht(struct vmw_validation_context *ctx);
...@@ -261,4 +266,6 @@ void *vmw_validation_mem_alloc(struct vmw_validation_context *ctx, ...@@ -261,4 +266,6 @@ void *vmw_validation_mem_alloc(struct vmw_validation_context *ctx,
int vmw_validation_preload_bo(struct vmw_validation_context *ctx); int vmw_validation_preload_bo(struct vmw_validation_context *ctx);
int vmw_validation_preload_res(struct vmw_validation_context *ctx, int vmw_validation_preload_res(struct vmw_validation_context *ctx,
unsigned int size); unsigned int size);
void vmw_validation_res_set_dirty(struct vmw_validation_context *ctx,
void *val_private, u32 dirty);
#endif #endif
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment