summaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
authorAlyssa Rosenzweig <alyssa.rosenzweig@collabora.com>2019-07-16 09:45:11 -0700
committerAlyssa Rosenzweig <alyssa.rosenzweig@collabora.com>2019-07-22 08:20:33 -0700
commitaa03159120f31a45c795bb1d15113d69c485c2bb (patch)
treed1ae44e3524052670e91bdc53610c6aae7efffa8 /src
parent1cabb8a706c86574a509f885065ac1e182bc04bc (diff)
pan/midgard: Call scheduler/RA in a loop
This will allow us to insert instructions as a result of register allocation, permitting spilling to be implemented. As a side effect, with the assert commented out this would fix a bunch of glamor crashes (due to RA failures) so MATE becomes useable. Ideally we'll have scheduling or RA actually sorted out before the branch point but if not this gives us a one-line out to get X working... Signed-off-by: Alyssa Rosenzweig <alyssa.rosenzweig@collabora.com>
Diffstat (limited to 'src')
-rw-r--r--src/panfrost/midgard/compiler.h2
-rw-r--r--src/panfrost/midgard/midgard_ra.c11
-rw-r--r--src/panfrost/midgard/midgard_schedule.c27
3 files changed, 27 insertions, 13 deletions
diff --git a/src/panfrost/midgard/compiler.h b/src/panfrost/midgard/compiler.h
index 59176f3c0ee..bd7c11efab3 100644
--- a/src/panfrost/midgard/compiler.h
+++ b/src/panfrost/midgard/compiler.h
@@ -433,7 +433,7 @@ void schedule_program(compiler_context *ctx);
struct ra_graph;
-struct ra_graph* allocate_registers(compiler_context *ctx);
+struct ra_graph* allocate_registers(compiler_context *ctx, bool *spilled);
void install_registers(compiler_context *ctx, struct ra_graph *g);
bool mir_is_live_after(compiler_context *ctx, midgard_block *block, midgard_instruction *start, int src);
bool mir_has_multiple_writes(compiler_context *ctx, int src);
diff --git a/src/panfrost/midgard/midgard_ra.c b/src/panfrost/midgard/midgard_ra.c
index c089a7be3e8..1505022f451 100644
--- a/src/panfrost/midgard/midgard_ra.c
+++ b/src/panfrost/midgard/midgard_ra.c
@@ -186,7 +186,7 @@ index_to_reg(compiler_context *ctx, struct ra_graph *g, int reg)
* by install_registers */
struct ra_graph *
-allocate_registers(compiler_context *ctx)
+allocate_registers(compiler_context *ctx, bool *spilled)
{
/* The number of vec4 work registers available depends on when the
* uniforms start, so compute that first */
@@ -372,14 +372,15 @@ allocate_registers(compiler_context *ctx)
}
}
- if (!ra_allocate(g)) {
- unreachable("Error allocating registers\n");
- }
-
/* Cleanup */
free(live_start);
free(live_end);
+ if (!ra_allocate(g)) {
+ *spilled = true;
+ return NULL;
+ }
+
return g;
}
diff --git a/src/panfrost/midgard/midgard_schedule.c b/src/panfrost/midgard/midgard_schedule.c
index 7a3841e4d44..ccc641c7b86 100644
--- a/src/panfrost/midgard/midgard_schedule.c
+++ b/src/panfrost/midgard/midgard_schedule.c
@@ -527,15 +527,28 @@ schedule_block(compiler_context *ctx, midgard_block *block)
void
schedule_program(compiler_context *ctx)
{
- /* We run RA prior to scheduling */
+ struct ra_graph *g = NULL;
+ bool spilled = false;
+ int iter_count = 10; /* max iterations */
- mir_foreach_block(ctx, block) {
- schedule_block(ctx, block);
- }
+ do {
+ /* We would like to run RA after scheduling, but spilling can
+ * complicate this */
+
+ mir_foreach_block(ctx, block) {
+ schedule_block(ctx, block);
+ }
- /* Pipeline registers creation is a prepass before RA */
- mir_create_pipeline_registers(ctx);
+ /* Pipeline registers creation is a prepass before RA */
+ mir_create_pipeline_registers(ctx);
+
+ g = allocate_registers(ctx, &spilled);
+ } while(spilled && ((iter_count--) > 0));
+
+ if (iter_count <= 0) {
+ fprintf(stderr, "panfrost: Gave up allocating registers, rendering will be incomplete\n");
+ assert(0);
+ }
- struct ra_graph *g = allocate_registers(ctx);
install_registers(ctx, g);
}