summaryrefslogtreecommitdiff
path: root/benchmarks
diff options
context:
space:
mode:
authorMarcin Bernatowicz <marcin.bernatowicz@linux.intel.com>2023-10-06 16:06:50 +0000
committerKamil Konieczny <kamil.konieczny@linux.intel.com>2023-10-10 14:17:41 +0200
commit2d7db49a2bbaa5fbba8ef9250286f83b75dd80e0 (patch)
tree6e9c92b1a679c73f62aec714ccd9bcc909f45729 /benchmarks
parentdb38359d49c734a8329f9dcacd9265af2ae288ea (diff)
benchmarks/gem_wsim: for_each_w_step macro
for_each_w_step macro to easy traverse workload steps. v2: - added comment on igt_unique in macro Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Signed-off-by: Marcin Bernatowicz <marcin.bernatowicz@linux.intel.com>
Diffstat (limited to 'benchmarks')
-rw-r--r--benchmarks/gem_wsim.c81
1 files changed, 43 insertions, 38 deletions
diff --git a/benchmarks/gem_wsim.c b/benchmarks/gem_wsim.c
index fc2960c4b..95c1e71f7 100644
--- a/benchmarks/gem_wsim.c
+++ b/benchmarks/gem_wsim.c
@@ -238,6 +238,11 @@ struct workload {
#define for_each_ctx(__ctx, __wrk) \
__for_each_ctx(__ctx, __wrk, igt_unique(__ctx_idx))
+/* igt_unique(idx) is same on both lines as macro when expanded comes out on one line */
+#define for_each_w_step(__w_step, __wrk) \
+ for (typeof(__wrk->nr_steps) igt_unique(idx) = ({__w_step = __wrk->steps; 0; }); \
+ igt_unique(idx) < __wrk->nr_steps; igt_unique(idx)++, __w_step++)
+
static unsigned int master_prng;
static int verbose = 1;
@@ -1183,14 +1188,14 @@ add_step:
/*
* Check no duplicate working set ids.
*/
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
struct w_step *w2;
if (w->type != WORKINGSET)
continue;
- for (j = 0, w2 = wrk->steps; j < wrk->nr_steps; w2++, j++) {
- if (j == i)
+ for_each_w_step(w2, wrk) {
+ if (w->idx == w2->idx)
continue;
if (w2->type != WORKINGSET)
continue;
@@ -1203,7 +1208,7 @@ add_step:
/*
* Allocate shared working sets.
*/
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
if (w->type == WORKINGSET && w->working_set.shared) {
unsigned long total =
allocate_working_set(wrk, &w->working_set);
@@ -1215,7 +1220,7 @@ add_step:
}
wrk->max_working_set_id = -1;
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
if (w->type == WORKINGSET &&
w->working_set.shared &&
w->working_set.id > wrk->max_working_set_id)
@@ -1226,7 +1231,7 @@ add_step:
sizeof(*wrk->working_sets));
igt_assert(wrk->working_sets);
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
if (w->type == WORKINGSET && w->working_set.shared)
wrk->working_sets[w->working_set.id] = &w->working_set;
}
@@ -1238,6 +1243,7 @@ static struct workload *
clone_workload(struct workload *_wrk)
{
struct workload *wrk;
+ struct w_step *w;
int i;
wrk = malloc(sizeof(*wrk));
@@ -1265,8 +1271,8 @@ clone_workload(struct workload *_wrk)
}
/* Check if we need a sw sync timeline. */
- for (i = 0; i < wrk->nr_steps; i++) {
- if (wrk->steps[i].type == SW_FENCE) {
+ for_each_w_step(w, wrk) {
+ if (w->type == SW_FENCE) {
wrk->sync_timeline = sw_sync_timeline_create();
igt_assert(wrk->sync_timeline >= 0);
break;
@@ -1722,13 +1728,13 @@ static void measure_active_set(struct workload *wrk)
{
unsigned long total = 0, batch_sizes = 0;
struct dep_entry *dep, *deps = NULL;
- unsigned int nr = 0, i;
+ unsigned int nr = 0;
struct w_step *w;
if (verbose < 3)
return;
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
if (w->type != BATCH)
continue;
@@ -1781,12 +1787,11 @@ static void allocate_contexts(unsigned int id, struct workload *wrk)
{
int max_ctx = -1;
struct w_step *w;
- int i;
/*
* Pre-scan workload steps to allocate context list storage.
*/
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
int ctx = w->context + 1;
int delta;
@@ -1812,13 +1817,13 @@ static int prepare_contexts(unsigned int id, struct workload *wrk)
uint32_t share_vm = 0;
struct w_step *w;
struct ctx *ctx, *ctx2;
- unsigned int i, j;
+ unsigned int j;
/*
* Transfer over engine map configuration from the workload step.
*/
__for_each_ctx(ctx, wrk, ctx_idx) {
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
if (w->context != ctx_idx)
continue;
@@ -1985,12 +1990,11 @@ static void prepare_working_sets(unsigned int id, struct workload *wrk)
struct working_set **sets;
unsigned long total = 0;
struct w_step *w;
- int i;
/*
* Allocate working sets.
*/
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
if (w->type == WORKINGSET && !w->working_set.shared)
total += allocate_working_set(wrk, &w->working_set);
}
@@ -2002,7 +2006,7 @@ static void prepare_working_sets(unsigned int id, struct workload *wrk)
* Map of working set ids.
*/
wrk->max_working_set_id = -1;
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
if (w->type == WORKINGSET &&
w->working_set.id > wrk->max_working_set_id)
wrk->max_working_set_id = w->working_set.id;
@@ -2013,7 +2017,7 @@ static void prepare_working_sets(unsigned int id, struct workload *wrk)
sizeof(*wrk->working_sets));
igt_assert(wrk->working_sets);
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
struct working_set *set;
if (w->type != WORKINGSET)
@@ -2039,7 +2043,6 @@ static void prepare_working_sets(unsigned int id, struct workload *wrk)
static int prepare_workload(unsigned int id, struct workload *wrk)
{
struct w_step *w;
- int i, j;
int ret = 0;
wrk->id = id;
@@ -2054,23 +2057,22 @@ static int prepare_workload(unsigned int id, struct workload *wrk)
return ret;
/* Record default preemption. */
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk)
if (w->type == BATCH)
w->preempt_us = 100;
- }
/*
* Scan for contexts with modified preemption config and record their
* preemption period for the following steps belonging to the same
* context.
*/
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
struct w_step *w2;
if (w->type != PREEMPTION)
continue;
- for (j = i + 1; j < wrk->nr_steps; j++) {
+ for (int j = w->idx + 1; j < wrk->nr_steps; j++) {
w2 = &wrk->steps[j];
if (w2->context != w->context)
@@ -2087,7 +2089,7 @@ static int prepare_workload(unsigned int id, struct workload *wrk)
/*
* Scan for SSEU control steps.
*/
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
if (w->type == SSEU) {
get_device_sseu();
break;
@@ -2099,7 +2101,7 @@ static int prepare_workload(unsigned int id, struct workload *wrk)
/*
* Allocate batch buffers.
*/
- for (i = 0, w = wrk->steps; i < wrk->nr_steps; i++, w++) {
+ for_each_w_step(w, wrk) {
if (w->type != BATCH)
continue;
@@ -2223,7 +2225,6 @@ static void *run_workload(void *data)
int qd_throttle = -1;
int count, missed = 0;
unsigned long time_tot = 0, time_min = ULONG_MAX, time_max = 0;
- int i;
clock_gettime(CLOCK_MONOTONIC, &t_start);
@@ -2233,11 +2234,13 @@ static void *run_workload(void *data)
clock_gettime(CLOCK_MONOTONIC, &repeat_start);
- for (i = 0, w = wrk->steps; wrk->run && (i < wrk->nr_steps);
- i++, w++) {
+ for_each_w_step(w, wrk) {
enum intel_engine_id engine = w->engine;
int do_sleep = 0;
+ if (!wrk->run)
+ break;
+
if (w->type == DELAY) {
do_sleep = w->delay;
} else if (w->type == PERIOD) {
@@ -2256,13 +2259,13 @@ static void *run_workload(void *data)
missed++;
if (verbose > 2)
printf("%u: Dropped period @ %u/%u (%dus late)!\n",
- wrk->id, count, i, do_sleep);
+ wrk->id, count, w->idx, do_sleep);
continue;
}
} else if (w->type == SYNC) {
- unsigned int s_idx = i + w->target;
+ unsigned int s_idx = w->idx + w->target;
- igt_assert(s_idx >= 0 && s_idx < i);
+ igt_assert(s_idx >= 0 && s_idx < w->idx);
igt_assert(wrk->steps[s_idx].type == BATCH);
w_step_sync(&wrk->steps[s_idx]);
continue;
@@ -2283,7 +2286,7 @@ static void *run_workload(void *data)
int tgt = w->idx + w->target;
int inc;
- igt_assert(tgt >= 0 && tgt < i);
+ igt_assert(tgt >= 0 && tgt < w->idx);
igt_assert(wrk->steps[tgt].type == SW_FENCE);
cur_seqno += wrk->steps[tgt].idx;
inc = cur_seqno - wrk->sync_seqno;
@@ -2303,9 +2306,9 @@ static void *run_workload(void *data)
}
continue;
} else if (w->type == TERMINATE) {
- unsigned int t_idx = i + w->target;
+ unsigned int t_idx = w->idx + w->target;
- igt_assert(t_idx >= 0 && t_idx < i);
+ igt_assert(t_idx >= 0 && t_idx < w->idx);
igt_assert(wrk->steps[t_idx].type == BATCH);
igt_assert(wrk->steps[t_idx].duration.unbound);
@@ -2339,7 +2342,7 @@ static void *run_workload(void *data)
sync_deps(wrk, w);
if (throttle > 0)
- w_sync_to(wrk, w, i - throttle);
+ w_sync_to(wrk, w, w->idx - throttle);
do_eb(wrk, w, engine);
@@ -2382,8 +2385,10 @@ static void *run_workload(void *data)
}
/* Cleanup all fences instantiated in this iteration. */
- for (i = 0, w = wrk->steps; wrk->run && (i < wrk->nr_steps);
- i++, w++) {
+ for_each_w_step(w, wrk) {
+ if (!wrk->run)
+ break;
+
if (w->emit_fence > 0) {
close(w->emit_fence);
w->emit_fence = -1;
@@ -2391,7 +2396,7 @@ static void *run_workload(void *data)
}
}
- for (i = 0; i < NUM_ENGINES; i++) {
+ for (int i = 0; i < NUM_ENGINES; i++) {
if (!wrk->nrequest[i])
continue;