zink: move 'batch_id' and 'is_compute' members to fence
prep for future work, no functional changes Reviewed-by: Adam Jackson <ajax@redhat.com> Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/9753>
This commit is contained in:
committed by
Marge Bot
parent
93e19b8b40
commit
a315e8ce3e
@@ -28,8 +28,8 @@ zink_batch_state_clear_resources(struct zink_screen *screen, struct zink_batch_s
|
||||
/* unref all used resources */
|
||||
set_foreach(bs->resources, entry) {
|
||||
struct zink_resource_object *obj = (struct zink_resource_object *)entry->key;
|
||||
batch_usage_unset(&obj->reads, !!bs->is_compute, bs->batch_id);
|
||||
batch_usage_unset(&obj->writes, !!bs->is_compute, bs->batch_id);
|
||||
batch_usage_unset(&obj->reads, !!bs->fence.is_compute, bs->fence.batch_id);
|
||||
batch_usage_unset(&obj->writes, !!bs->fence.is_compute, bs->fence.batch_id);
|
||||
zink_resource_object_reference(screen, &obj, NULL);
|
||||
_mesa_set_remove(bs->resources, entry);
|
||||
}
|
||||
@@ -50,13 +50,13 @@ zink_reset_batch_state(struct zink_context *ctx, struct zink_batch_state *bs)
|
||||
|
||||
set_foreach(bs->surfaces, entry) {
|
||||
struct zink_surface *surf = (struct zink_surface *)entry->key;
|
||||
batch_usage_unset(&surf->batch_uses, !!bs->is_compute, bs->batch_id);
|
||||
batch_usage_unset(&surf->batch_uses, !!bs->fence.is_compute, bs->fence.batch_id);
|
||||
zink_surface_reference(screen, &surf, NULL);
|
||||
_mesa_set_remove(bs->surfaces, entry);
|
||||
}
|
||||
set_foreach(bs->bufferviews, entry) {
|
||||
struct zink_buffer_view *buffer_view = (struct zink_buffer_view *)entry->key;
|
||||
batch_usage_unset(&buffer_view->batch_uses, !!bs->is_compute, bs->batch_id);
|
||||
batch_usage_unset(&buffer_view->batch_uses, !!bs->fence.is_compute, bs->fence.batch_id);
|
||||
zink_buffer_view_reference(screen, &buffer_view, NULL);
|
||||
_mesa_set_remove(bs->bufferviews, entry);
|
||||
}
|
||||
@@ -69,7 +69,7 @@ zink_reset_batch_state(struct zink_context *ctx, struct zink_batch_state *bs)
|
||||
|
||||
set_foreach(bs->desc_sets, entry) {
|
||||
struct zink_descriptor_set *zds = (void*)entry->key;
|
||||
batch_usage_unset(&zds->batch_uses, !!bs->is_compute, bs->batch_id);
|
||||
batch_usage_unset(&zds->batch_uses, !!bs->fence.is_compute, bs->fence.batch_id);
|
||||
/* reset descriptor pools when no bs is using this program to avoid
|
||||
* having some inactive program hogging a billion descriptors
|
||||
*/
|
||||
@@ -79,7 +79,7 @@ zink_reset_batch_state(struct zink_context *ctx, struct zink_batch_state *bs)
|
||||
}
|
||||
|
||||
set_foreach(bs->programs, entry) {
|
||||
if (bs->is_compute) {
|
||||
if (bs->fence.is_compute) {
|
||||
struct zink_compute_program *comp = (struct zink_compute_program*)entry->key;
|
||||
bool in_use = comp == ctx->curr_compute;
|
||||
if (zink_compute_program_reference(screen, &comp, NULL) && in_use)
|
||||
@@ -102,7 +102,7 @@ zink_reset_batch_state(struct zink_context *ctx, struct zink_batch_state *bs)
|
||||
bs->flush_res = NULL;
|
||||
|
||||
bs->descs_used = 0;
|
||||
ctx->resource_size[bs->is_compute] -= bs->resource_size;
|
||||
ctx->resource_size[bs->fence.is_compute] -= bs->resource_size;
|
||||
bs->resource_size = 0;
|
||||
}
|
||||
|
||||
@@ -179,7 +179,7 @@ create_batch_state(struct zink_context *ctx, enum zink_queue queue)
|
||||
/* this destroys the batch state on failure */
|
||||
return NULL;
|
||||
|
||||
bs->is_compute = queue == ZINK_QUEUE_COMPUTE;
|
||||
bs->fence.is_compute = queue == ZINK_QUEUE_COMPUTE;
|
||||
|
||||
return bs;
|
||||
fail:
|
||||
@@ -251,10 +251,10 @@ zink_start_batch(struct zink_context *ctx, struct zink_batch *batch)
|
||||
if (vkBeginCommandBuffer(batch->state->cmdbuf, &cbbi) != VK_SUCCESS)
|
||||
debug_printf("vkBeginCommandBuffer failed\n");
|
||||
|
||||
batch->state->batch_id = ctx->curr_batch;
|
||||
batch->state->fence.batch_id = ctx->curr_batch;
|
||||
if (ctx->last_fence[batch->queue]) {
|
||||
struct zink_batch_state *last_state = zink_batch_state(ctx->last_fence[batch->queue]);
|
||||
batch->last_batch_id = last_state->batch_id;
|
||||
batch->last_batch_id = last_state->fence.batch_id;
|
||||
}
|
||||
if (!ctx->queries_disabled)
|
||||
zink_resume_queries(ctx, batch);
|
||||
@@ -316,7 +316,7 @@ zink_end_batch(struct zink_context *ctx, struct zink_batch *batch)
|
||||
}
|
||||
|
||||
ctx->last_fence[batch->queue] = &batch->state->fence;
|
||||
_mesa_hash_table_insert_pre_hashed(&ctx->batch_states[batch->queue], batch->state->batch_id, (void*)(uintptr_t)batch->state->batch_id, batch->state);
|
||||
_mesa_hash_table_insert_pre_hashed(&ctx->batch_states[batch->queue], batch->state->fence.batch_id, (void*)(uintptr_t)batch->state->fence.batch_id, batch->state);
|
||||
ctx->resource_size[batch->queue] += batch->state->resource_size;
|
||||
}
|
||||
|
||||
@@ -346,8 +346,8 @@ zink_batch_reference_resource_rw(struct zink_batch *batch, struct zink_resource
|
||||
}
|
||||
|
||||
/* if the resource already has usage of any sort set for this batch, we can skip hashing */
|
||||
if (!zink_batch_usage_matches(&res->obj->reads, batch->queue, batch->state->batch_id) &&
|
||||
!zink_batch_usage_matches(&res->obj->writes, batch->queue, batch->state->batch_id)) {
|
||||
if (!zink_batch_usage_matches(&res->obj->reads, batch->queue, batch->state->fence.batch_id) &&
|
||||
!zink_batch_usage_matches(&res->obj->writes, batch->queue, batch->state->fence.batch_id)) {
|
||||
bool found = false;
|
||||
_mesa_set_search_and_add(batch->state->resources, res->obj, &found);
|
||||
if (!found) {
|
||||
@@ -366,12 +366,12 @@ zink_batch_reference_resource_rw(struct zink_batch *batch, struct zink_resource
|
||||
}
|
||||
if (write) {
|
||||
if (stencil)
|
||||
zink_batch_usage_set(&stencil->obj->writes, batch->queue, batch->state->batch_id);
|
||||
zink_batch_usage_set(&res->obj->writes, batch->queue, batch->state->batch_id);
|
||||
zink_batch_usage_set(&stencil->obj->writes, batch->queue, batch->state->fence.batch_id);
|
||||
zink_batch_usage_set(&res->obj->writes, batch->queue, batch->state->fence.batch_id);
|
||||
} else {
|
||||
if (stencil)
|
||||
zink_batch_usage_set(&stencil->obj->reads, batch->queue, batch->state->batch_id);
|
||||
zink_batch_usage_set(&res->obj->reads, batch->queue, batch->state->batch_id);
|
||||
zink_batch_usage_set(&stencil->obj->reads, batch->queue, batch->state->fence.batch_id);
|
||||
zink_batch_usage_set(&res->obj->reads, batch->queue, batch->state->fence.batch_id);
|
||||
}
|
||||
/* multiple array entries are fine */
|
||||
if (res->obj->persistent_maps)
|
||||
@@ -385,11 +385,11 @@ static bool
|
||||
ptr_add_usage(struct zink_batch *batch, struct set *s, void *ptr, struct zink_batch_usage *u)
|
||||
{
|
||||
bool found = false;
|
||||
if (zink_batch_usage_matches(u, batch->queue, batch->state->batch_id))
|
||||
if (zink_batch_usage_matches(u, batch->queue, batch->state->fence.batch_id))
|
||||
return false;
|
||||
_mesa_set_search_and_add(s, ptr, &found);
|
||||
assert(!found);
|
||||
zink_batch_usage_set(u, batch->queue, batch->state->batch_id);
|
||||
zink_batch_usage_set(u, batch->queue, batch->state->fence.batch_id);
|
||||
return true;
|
||||
}
|
||||
|
||||
|
||||
@@ -78,9 +78,6 @@ struct zink_batch_state {
|
||||
struct set *active_queries; /* zink_query objects which were active at some point in this batch */
|
||||
|
||||
VkDeviceSize resource_size;
|
||||
|
||||
uint32_t batch_id;
|
||||
bool is_compute;
|
||||
};
|
||||
|
||||
struct zink_batch {
|
||||
|
||||
@@ -1823,20 +1823,20 @@ zink_wait_on_batch(struct zink_context *ctx, enum zink_queue queue, uint32_t bat
|
||||
{
|
||||
struct zink_batch_state *bs = zink_batch_queue(ctx, queue)->state;
|
||||
assert(bs);
|
||||
if (!batch_id || bs->batch_id == batch_id)
|
||||
if (!batch_id || bs->fence.batch_id == batch_id)
|
||||
/* not submitted yet */
|
||||
flush_batch(ctx, queue);
|
||||
|
||||
struct zink_fence *fence;
|
||||
|
||||
assert(batch_id || ctx->last_fence[queue]);
|
||||
if (ctx->last_fence[queue] && (!batch_id || batch_id == zink_batch_state(ctx->last_fence[queue])->batch_id))
|
||||
if (ctx->last_fence[queue] && (!batch_id || batch_id == zink_batch_state(ctx->last_fence[queue])->fence.batch_id))
|
||||
fence = ctx->last_fence[queue];
|
||||
else {
|
||||
struct hash_entry *he = _mesa_hash_table_search_pre_hashed(&ctx->batch_states[queue], batch_id, (void*)(uintptr_t)batch_id);
|
||||
if (!he) {
|
||||
util_dynarray_foreach(&ctx->free_batch_states[queue], struct zink_batch_state*, bs) {
|
||||
if ((*bs)->batch_id == batch_id)
|
||||
if ((*bs)->fence.batch_id == batch_id)
|
||||
return;
|
||||
}
|
||||
unreachable("should've found batch state");
|
||||
|
||||
@@ -649,7 +649,7 @@ update_sampler_descriptors(struct zink_context *ctx, struct zink_descriptor_set
|
||||
zink_batch_reference_sampler_view(batch, sampler_view);
|
||||
if (sampler)
|
||||
/* this only tracks the most recent usage for now */
|
||||
zink_batch_usage_set(&sampler->batch_uses, batch->queue, batch->state->batch_id);
|
||||
zink_batch_usage_set(&sampler->batch_uses, batch->queue, batch->state->fence.batch_id);
|
||||
}
|
||||
assert(num_wds < num_descriptors);
|
||||
|
||||
|
||||
@@ -39,7 +39,9 @@ struct zink_fence {
|
||||
struct pipe_reference reference;
|
||||
VkFence fence;
|
||||
struct pipe_context *deferred_ctx;
|
||||
uint32_t batch_id;
|
||||
bool submitted;
|
||||
bool is_compute;
|
||||
};
|
||||
|
||||
static inline struct zink_fence *
|
||||
|
||||
@@ -406,7 +406,7 @@ force_cpu_read(struct zink_context *ctx, struct pipe_query *pquery, bool wait, e
|
||||
unsigned result_size = result_type <= PIPE_QUERY_TYPE_U32 ? sizeof(uint32_t) : sizeof(uint64_t);
|
||||
struct zink_query *query = (struct zink_query*)pquery;
|
||||
union pipe_query_result result;
|
||||
if (zink_batch_usage_matches(&query->batch_id, ZINK_QUEUE_GFX, zink_batch_g(ctx)->state->batch_id))
|
||||
if (zink_batch_usage_matches(&query->batch_id, ZINK_QUEUE_GFX, zink_batch_g(ctx)->state->fence.batch_id))
|
||||
pctx->flush(pctx, NULL, PIPE_FLUSH_HINT_FINISH);
|
||||
else if (is_cs_query(query))
|
||||
zink_flush_compute(ctx);
|
||||
@@ -535,7 +535,7 @@ begin_query(struct zink_context *ctx, struct zink_batch *batch, struct zink_quer
|
||||
if (needs_stats_list(q))
|
||||
list_addtail(&q->stats_list, &ctx->primitives_generated_queries);
|
||||
p_atomic_inc(&q->fences);
|
||||
zink_batch_usage_set(&q->batch_id, batch->queue, batch->state->batch_id);
|
||||
zink_batch_usage_set(&q->batch_id, batch->queue, batch->state->fence.batch_id);
|
||||
_mesa_set_add(batch->state->active_queries, q);
|
||||
}
|
||||
|
||||
@@ -565,7 +565,7 @@ end_query(struct zink_context *ctx, struct zink_batch *batch, struct zink_query
|
||||
if (is_time_query(q)) {
|
||||
vkCmdWriteTimestamp(batch->state->cmdbuf, VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT,
|
||||
q->query_pool, q->curr_query);
|
||||
zink_batch_usage_set(&q->batch_id, batch->queue, batch->state->batch_id);
|
||||
zink_batch_usage_set(&q->batch_id, batch->queue, batch->state->fence.batch_id);
|
||||
} else if (q->type == PIPE_QUERY_PRIMITIVES_EMITTED ||
|
||||
q->type == PIPE_QUERY_PRIMITIVES_GENERATED ||
|
||||
q->type == PIPE_QUERY_SO_OVERFLOW_PREDICATE)
|
||||
|
||||
Reference in New Issue
Block a user