On Sat, Sep 3, 2016 at 12:50 AM, Pohjolainen, Topi < topi.pohjolai...@gmail.com> wrote:
> On Wed, Aug 31, 2016 at 02:22:40PM -0700, Jason Ekstrand wrote: > > --- > > src/intel/vulkan/anv_pipeline.c | 142 > --------------------------------- > > src/intel/vulkan/anv_private.h | 3 - > > src/intel/vulkan/genX_pipeline_util.h | 146 > ++++++++++++++++++++++++++++++++-- > > 3 files changed, 138 insertions(+), 153 deletions(-) > > > > diff --git a/src/intel/vulkan/anv_pipeline.c b/src/intel/vulkan/anv_ > pipeline.c > > index 35acc4a..750f2fa 100644 > > --- a/src/intel/vulkan/anv_pipeline.c > > +++ b/src/intel/vulkan/anv_pipeline.c > > @@ -803,147 +803,6 @@ anv_pipeline_compile_cs(struct anv_pipeline > *pipeline, > > return VK_SUCCESS; > > } > > > > -void > > -anv_compute_urb_partition(struct anv_pipeline *pipeline) > > There was no separate declaration for this function that would need to be > removed? > Good catch! Fixed locally. > > -{ > > - const struct gen_device_info *devinfo = &pipeline->device->info; > > - > > - bool vs_present = pipeline->active_stages & > VK_SHADER_STAGE_VERTEX_BIT; > > - unsigned vs_size = vs_present ? > > - get_vs_prog_data(pipeline)->base.urb_entry_size : 1; > > - unsigned vs_entry_size_bytes = vs_size * 64; > > - bool gs_present = pipeline->active_stages & > VK_SHADER_STAGE_GEOMETRY_BIT; > > - unsigned gs_size = gs_present ? > > - get_gs_prog_data(pipeline)->base.urb_entry_size : 1; > > - unsigned gs_entry_size_bytes = gs_size * 64; > > - > > - /* From p35 of the Ivy Bridge PRM (section 1.7.1: 3DSTATE_URB_GS): > > - * > > - * VS Number of URB Entries must be divisible by 8 if the VS URB > Entry > > - * Allocation Size is less than 9 512-bit URB entries. > > - * > > - * Similar text exists for GS. > > - */ > > - unsigned vs_granularity = (vs_size < 9) ? 8 : 1; > > - unsigned gs_granularity = (gs_size < 9) ? 8 : 1; > > - > > - /* URB allocations must be done in 8k chunks. */ > > - unsigned chunk_size_bytes = 8192; > > - > > - /* Determine the size of the URB in chunks. */ > > - unsigned urb_chunks = pipeline->urb.total_size * 1024 / > chunk_size_bytes; > > - > > - /* Reserve space for push constants */ > > - unsigned push_constant_kb; > > - if (pipeline->device->info.gen >= 8) > > - push_constant_kb = 32; > > - else if (pipeline->device->info.is_haswell) > > - push_constant_kb = pipeline->device->info.gt == 3 ? 32 : 16; > > - else > > - push_constant_kb = 16; > > - > > - unsigned push_constant_bytes = push_constant_kb * 1024; > > - unsigned push_constant_chunks = > > - push_constant_bytes / chunk_size_bytes; > > - > > - /* Initially, assign each stage the minimum amount of URB space it > needs, > > - * and make a note of how much additional space it "wants" (the > amount of > > - * additional space it could actually make use of). > > - */ > > - > > - /* VS has a lower limit on the number of URB entries */ > > - unsigned vs_chunks = > > - ALIGN(devinfo->urb.min_vs_entries * vs_entry_size_bytes, > > - chunk_size_bytes) / chunk_size_bytes; > > - unsigned vs_wants = > > - ALIGN(devinfo->urb.max_vs_entries * vs_entry_size_bytes, > > - chunk_size_bytes) / chunk_size_bytes - vs_chunks; > > - > > - unsigned gs_chunks = 0; > > - unsigned gs_wants = 0; > > - if (gs_present) { > > - /* There are two constraints on the minimum amount of URB space > we can > > - * allocate: > > - * > > - * (1) We need room for at least 2 URB entries, since we always > operate > > - * the GS in DUAL_OBJECT mode. > > - * > > - * (2) We can't allocate less than nr_gs_entries_granularity. > > - */ > > - gs_chunks = ALIGN(MAX2(gs_granularity, 2) * gs_entry_size_bytes, > > - chunk_size_bytes) / chunk_size_bytes; > > - gs_wants = > > - ALIGN(devinfo->urb.max_gs_entries * gs_entry_size_bytes, > > - chunk_size_bytes) / chunk_size_bytes - gs_chunks; > > - } > > - > > - /* There should always be enough URB space to satisfy the minimum > > - * requirements of each stage. > > - */ > > - unsigned total_needs = push_constant_chunks + vs_chunks + gs_chunks; > > - assert(total_needs <= urb_chunks); > > - > > - /* Mete out remaining space (if any) in proportion to "wants". */ > > - unsigned total_wants = vs_wants + gs_wants; > > - unsigned remaining_space = urb_chunks - total_needs; > > - if (remaining_space > total_wants) > > - remaining_space = total_wants; > > - if (remaining_space > 0) { > > - unsigned vs_additional = (unsigned) > > - round(vs_wants * (((double) remaining_space) / total_wants)); > > - vs_chunks += vs_additional; > > - remaining_space -= vs_additional; > > - gs_chunks += remaining_space; > > - } > > - > > - /* Sanity check that we haven't over-allocated. */ > > - assert(push_constant_chunks + vs_chunks + gs_chunks <= urb_chunks); > > - > > - /* Finally, compute the number of entries that can fit in the space > > - * allocated to each stage. > > - */ > > - unsigned nr_vs_entries = vs_chunks * chunk_size_bytes / > vs_entry_size_bytes; > > - unsigned nr_gs_entries = gs_chunks * chunk_size_bytes / > gs_entry_size_bytes; > > - > > - /* Since we rounded up when computing *_wants, this may be slightly > more > > - * than the maximum allowed amount, so correct for that. > > - */ > > - nr_vs_entries = MIN2(nr_vs_entries, devinfo->urb.max_vs_entries); > > - nr_gs_entries = MIN2(nr_gs_entries, devinfo->urb.max_gs_entries); > > - > > - /* Ensure that we program a multiple of the granularity. */ > > - nr_vs_entries = ROUND_DOWN_TO(nr_vs_entries, vs_granularity); > > - nr_gs_entries = ROUND_DOWN_TO(nr_gs_entries, gs_granularity); > > - > > - /* Finally, sanity check to make sure we have at least the minimum > number > > - * of entries needed for each stage. > > - */ > > - assert(nr_vs_entries >= devinfo->urb.min_vs_entries); > > - if (gs_present) > > - assert(nr_gs_entries >= 2); > > - > > - /* Lay out the URB in the following order: > > - * - push constants > > - * - VS > > - * - GS > > - */ > > - pipeline->urb.start[MESA_SHADER_VERTEX] = push_constant_chunks; > > - pipeline->urb.size[MESA_SHADER_VERTEX] = vs_size; > > - pipeline->urb.entries[MESA_SHADER_VERTEX] = nr_vs_entries; > > - > > - pipeline->urb.start[MESA_SHADER_GEOMETRY] = push_constant_chunks + > vs_chunks; > > - pipeline->urb.size[MESA_SHADER_GEOMETRY] = gs_size; > > - pipeline->urb.entries[MESA_SHADER_GEOMETRY] = nr_gs_entries; > > - > > - pipeline->urb.start[MESA_SHADER_TESS_CTRL] = push_constant_chunks; > > - pipeline->urb.size[MESA_SHADER_TESS_CTRL] = 1; > > - pipeline->urb.entries[MESA_SHADER_TESS_CTRL] = 0; > > - > > - pipeline->urb.start[MESA_SHADER_TESS_EVAL] = push_constant_chunks; > > - pipeline->urb.size[MESA_SHADER_TESS_EVAL] = 1; > > - pipeline->urb.entries[MESA_SHADER_TESS_EVAL] = 0; > > -} > > - > > /** > > * Copy pipeline state not marked as dynamic. > > * Dynamic state is pipeline state which hasn't been provided at > pipeline > > @@ -1242,7 +1101,6 @@ anv_pipeline_init(struct anv_pipeline *pipeline, > > } > > > > anv_pipeline_setup_l3_config(pipeline, false); > > - anv_compute_urb_partition(pipeline); > > > > const VkPipelineVertexInputStateCreateInfo *vi_info = > > pCreateInfo->pVertexInputState; > > diff --git a/src/intel/vulkan/anv_private.h b/src/intel/vulkan/anv_ > private.h > > index f32806c..805b3aa 100644 > > --- a/src/intel/vulkan/anv_private.h > > +++ b/src/intel/vulkan/anv_private.h > > @@ -1521,9 +1521,6 @@ struct anv_pipeline { > > struct anv_shader_bin * > shaders[MESA_SHADER_STAGES]; > > > > struct { > > - uint32_t > start[MESA_SHADER_GEOMETRY + 1]; > > - uint32_t > size[MESA_SHADER_GEOMETRY + 1]; > > - uint32_t > entries[MESA_SHADER_GEOMETRY + 1]; > > const struct gen_l3_config * l3_config; > > uint32_t total_size; > > } urb; > > diff --git a/src/intel/vulkan/genX_pipeline_util.h > b/src/intel/vulkan/genX_pipeline_util.h > > index 6518fae..b23a421 100644 > > --- a/src/intel/vulkan/genX_pipeline_util.h > > +++ b/src/intel/vulkan/genX_pipeline_util.h > > @@ -190,9 +190,123 @@ emit_vertex_input(struct anv_pipeline *pipeline, > > static inline void > > emit_urb_setup(struct anv_pipeline *pipeline) > > { > > -#if GEN_GEN == 7 && !GEN_IS_HASWELL > > struct anv_device *device = pipeline->device; > > > > + bool vs_present = pipeline->active_stages & > VK_SHADER_STAGE_VERTEX_BIT; > > + unsigned vs_size = vs_present ? > > + get_vs_prog_data(pipeline)->base.urb_entry_size : 1; > > + unsigned vs_entry_size_bytes = vs_size * 64; > > + bool gs_present = pipeline->active_stages & > VK_SHADER_STAGE_GEOMETRY_BIT; > > + unsigned gs_size = gs_present ? > > + get_gs_prog_data(pipeline)->base.urb_entry_size : 1; > > + unsigned gs_entry_size_bytes = gs_size * 64; > > + > > + /* From p35 of the Ivy Bridge PRM (section 1.7.1: 3DSTATE_URB_GS): > > + * > > + * VS Number of URB Entries must be divisible by 8 if the VS URB > Entry > > + * Allocation Size is less than 9 512-bit URB entries. > > + * > > + * Similar text exists for GS. > > + */ > > + unsigned vs_granularity = (vs_size < 9) ? 8 : 1; > > + unsigned gs_granularity = (gs_size < 9) ? 8 : 1; > > + > > + /* URB allocations must be done in 8k chunks. */ > > + unsigned chunk_size_bytes = 8192; > > + > > + /* Determine the size of the URB in chunks. */ > > + unsigned urb_chunks = pipeline->urb.total_size * 1024 / > chunk_size_bytes; > > + > > + /* Reserve space for push constants */ > > + unsigned push_constant_kb; > > + if (pipeline->device->info.gen >= 8) > > + push_constant_kb = 32; > > + else if (pipeline->device->info.is_haswell) > > + push_constant_kb = pipeline->device->info.gt == 3 ? 32 : 16; > > + else > > + push_constant_kb = 16; > > + > > + unsigned push_constant_bytes = push_constant_kb * 1024; > > + unsigned push_constant_chunks = > > + push_constant_bytes / chunk_size_bytes; > > + > > + /* Initially, assign each stage the minimum amount of URB space it > needs, > > + * and make a note of how much additional space it "wants" (the > amount of > > + * additional space it could actually make use of). > > + */ > > + > > + /* VS has a lower limit on the number of URB entries */ > > + unsigned vs_chunks = > > + ALIGN(device->info.urb.min_vs_entries * vs_entry_size_bytes, > > + chunk_size_bytes) / chunk_size_bytes; > > + unsigned vs_wants = > > + ALIGN(device->info.urb.max_vs_entries * vs_entry_size_bytes, > > + chunk_size_bytes) / chunk_size_bytes - vs_chunks; > > + > > + unsigned gs_chunks = 0; > > + unsigned gs_wants = 0; > > + if (gs_present) { > > + /* There are two constraints on the minimum amount of URB space > we can > > + * allocate: > > + * > > + * (1) We need room for at least 2 URB entries, since we always > operate > > + * the GS in DUAL_OBJECT mode. > > + * > > + * (2) We can't allocate less than nr_gs_entries_granularity. > > + */ > > + gs_chunks = ALIGN(MAX2(gs_granularity, 2) * gs_entry_size_bytes, > > + chunk_size_bytes) / chunk_size_bytes; > > + gs_wants = > > + ALIGN(device->info.urb.max_gs_entries * gs_entry_size_bytes, > > + chunk_size_bytes) / chunk_size_bytes - gs_chunks; > > + } > > + > > + /* There should always be enough URB space to satisfy the minimum > > + * requirements of each stage. > > + */ > > + unsigned total_needs = push_constant_chunks + vs_chunks + gs_chunks; > > + assert(total_needs <= urb_chunks); > > + > > + /* Mete out remaining space (if any) in proportion to "wants". */ > > + unsigned total_wants = vs_wants + gs_wants; > > + unsigned remaining_space = urb_chunks - total_needs; > > + if (remaining_space > total_wants) > > + remaining_space = total_wants; > > + if (remaining_space > 0) { > > + unsigned vs_additional = (unsigned) > > + round(vs_wants * (((double) remaining_space) / total_wants)); > > + vs_chunks += vs_additional; > > + remaining_space -= vs_additional; > > + gs_chunks += remaining_space; > > + } > > + > > + /* Sanity check that we haven't over-allocated. */ > > + assert(push_constant_chunks + vs_chunks + gs_chunks <= urb_chunks); > > + > > + /* Finally, compute the number of entries that can fit in the space > > + * allocated to each stage. > > + */ > > + unsigned nr_vs_entries = vs_chunks * chunk_size_bytes / > vs_entry_size_bytes; > > + unsigned nr_gs_entries = gs_chunks * chunk_size_bytes / > gs_entry_size_bytes; > > + > > + /* Since we rounded up when computing *_wants, this may be slightly > more > > + * than the maximum allowed amount, so correct for that. > > + */ > > + nr_vs_entries = MIN2(nr_vs_entries, device->info.urb.max_vs_ > entries); > > + nr_gs_entries = MIN2(nr_gs_entries, device->info.urb.max_gs_ > entries); > > + > > + /* Ensure that we program a multiple of the granularity. */ > > + nr_vs_entries = ROUND_DOWN_TO(nr_vs_entries, vs_granularity); > > + nr_gs_entries = ROUND_DOWN_TO(nr_gs_entries, gs_granularity); > > + > > + /* Finally, sanity check to make sure we have at least the minimum > number > > + * of entries needed for each stage. > > + */ > > + assert(nr_vs_entries >= device->info.urb.min_vs_entries); > > + if (gs_present) > > + assert(nr_gs_entries >= 2); > > + > > +#if GEN_GEN == 7 && !GEN_IS_HASWELL > > /* From the IVB PRM Vol. 2, Part 1, Section 3.2.1: > > * > > * "A PIPE_CONTROL with Post-Sync Operation set to 1h and a depth > stall > > @@ -208,13 +322,29 @@ emit_urb_setup(struct anv_pipeline *pipeline) > > } > > #endif > > > > - for (int i = MESA_SHADER_VERTEX; i <= MESA_SHADER_GEOMETRY; i++) { > > - anv_batch_emit(&pipeline->batch, GENX(3DSTATE_URB_VS), urb) { > > - urb._3DCommandSubOpcode = 48 + i; > > - urb.VSURBStartingAddress = pipeline->urb.start[i]; > > - urb.VSURBEntryAllocationSize = pipeline->urb.size[i] - 1; > > - urb.VSNumberofURBEntries = pipeline->urb.entries[i]; > > - } > > + /* Lay out the URB in the following order: > > + * - push constants > > + * - VS > > + * - GS > > + */ > > + anv_batch_emit(&pipeline->batch, GENX(3DSTATE_URB_VS), urb) { > > + urb.VSURBStartingAddress = push_constant_chunks; > > + urb.VSURBEntryAllocationSize = vs_size - 1; > > + urb.VSNumberofURBEntries = nr_vs_entries; > > + } > > + > > + anv_batch_emit(&pipeline->batch, GENX(3DSTATE_URB_HS), urb) { > > + urb.HSURBStartingAddress = push_constant_chunks; > > + } > > + > > + anv_batch_emit(&pipeline->batch, GENX(3DSTATE_URB_DS), urb) { > > + urb.DSURBStartingAddress = push_constant_chunks; > > + } > > + > > + anv_batch_emit(&pipeline->batch, GENX(3DSTATE_URB_GS), urb) { > > + urb.GSURBStartingAddress = push_constant_chunks + vs_chunks; > > + urb.GSURBEntryAllocationSize = gs_size - 1; > > + urb.GSNumberofURBEntries = nr_gs_entries; > > } > > } > > > > -- > > 2.5.0.400.gff86faf > > > > _______________________________________________ > > mesa-dev mailing list > > mesa-dev@lists.freedesktop.org > > https://lists.freedesktop.org/mailman/listinfo/mesa-dev >
_______________________________________________ mesa-dev mailing list mesa-dev@lists.freedesktop.org https://lists.freedesktop.org/mailman/listinfo/mesa-dev