On Fri, Sep 22, 2017 at 04:11:33PM +0100, Lionel Landwerlin wrote:
> Verifies that the kernel programs slices correctly based by reading
> the value of PWR_CLK_STATE register.
> 
> v2: Add subslice tests (Lionel)
>     Use MI_SET_PREDICATE for further verification when available (Lionel)
> 
> Signed-off-by: Lionel Landwerlin <lionel.g.landwer...@intel.com>
> ---
>  tests/Makefile.sources |   1 +
>  tests/ctx_rpcs.c       | 453 
> +++++++++++++++++++++++++++++++++++++++++++++++++
>  tests/meson.build      |   1 +
>  3 files changed, 455 insertions(+)
>  create mode 100644 tests/ctx_rpcs.c
> 
> diff --git a/tests/Makefile.sources b/tests/Makefile.sources
> index 0adc28a0..18d18f08 100644
> --- a/tests/Makefile.sources
> +++ b/tests/Makefile.sources
> @@ -26,6 +26,7 @@ TESTS_progs = \
>       core_getversion \
>       core_prop_blob \
>       core_setmaster_vs_auth \
> +     ctx_rpcs \
>       debugfs_test \
>       drm_import_export \
>       drm_mm \
> diff --git a/tests/ctx_rpcs.c b/tests/ctx_rpcs.c
> new file mode 100644
> index 00000000..a127084e
> --- /dev/null
> +++ b/tests/ctx_rpcs.c
> @@ -0,0 +1,453 @@
> +/*
> + * Copyright © 2017 Intel Corporation
> + *
> + * Permission is hereby granted, free of charge, to any person obtaining a
> + * copy of this software and associated documentation files (the "Software"),
> + * to deal in the Software without restriction, including without limitation
> + * the rights to use, copy, modify, merge, publish, distribute, sublicense,
> + * and/or sell copies of the Software, and to permit persons to whom the
> + * Software is furnished to do so, subject to the following conditions:
> + *
> + * The above copyright notice and this permission notice (including the next
> + * paragraph) shall be included in all copies or substantial portions of the
> + * Software.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
> + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
> + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
> + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
> + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER 
> DEALINGS
> + * IN THE SOFTWARE.
> + *
> + * Authors:
> + *    Lionel Landwerlin <lionel.g.landwer...@intel.com>
> + *
> + */
> +
> +#define _GNU_SOURCE
> +#include "igt.h"
> +#include <stdio.h>
> +#include <stdlib.h>
> +#include <string.h>
> +#include <unistd.h>
> +#include <fcntl.h>
> +#include <signal.h>
> +#include <errno.h>
> +#include <time.h>
> +#include <sys/wait.h>
> +
> +#include "intel_bufmgr.h"
> +
> +#define MI_STORE_REGISTER_MEM (0x24 << 23)
> +
> +#define MI_SET_PREDICATE      (0x1 << 23)
> +#define  MI_SET_PREDICATE_NOOP_NEVER (0)
> +#define  MI_SET_PREDICATE_1_SLICES   (5)
> +#define  MI_SET_PREDICATE_2_SLICES   (6)
> +#define  MI_SET_PREDICATE_3_SLICES   (7)
> +
> +#define GEN8_R_PWR_CLK_STATE         0x20C8
> +#define   GEN8_RPCS_ENABLE           (1 << 31)
> +#define   GEN8_RPCS_S_CNT_ENABLE     (1 << 18)
> +#define   GEN8_RPCS_S_CNT_SHIFT              15
> +#define   GEN8_RPCS_S_CNT_MASK               (0x7 << GEN8_RPCS_S_CNT_SHIFT)
> +#define   GEN8_RPCS_SS_CNT_ENABLE    (1 << 11)
> +#define   GEN8_RPCS_SS_CNT_SHIFT     8
> +#define   GEN8_RPCS_SS_CNT_MASK              (0x7 << GEN8_RPCS_SS_CNT_SHIFT)
> +#define   GEN8_RPCS_EU_MAX_SHIFT     4
> +#define   GEN8_RPCS_EU_MAX_MASK              (0xf << GEN8_RPCS_EU_MAX_SHIFT)
> +#define   GEN8_RPCS_EU_MIN_SHIFT     0
> +#define   GEN8_RPCS_EU_MIN_MASK              (0xf << GEN8_RPCS_EU_MIN_SHIFT)
> +
> +#ifndef I915_PARAM_SLICE_MASK
> +#define I915_PARAM_SLICE_MASK                (46)
> +#define I915_PARAM_SUBSLICE_MASK     (47)
> +#endif /* I915_PARAM_SLICE_MASK */
> +
> +#ifndef I915_CONTEXT_PARAM_SSEU
> +#define I915_CONTEXT_PARAM_SSEU              0x6
> +
> +struct drm_i915_gem_context_param_sseu {
> +     /*
> +      * Engine to be configured or queried. Same value you would use with
> +      * drm_i915_gem_execbuffer2.
> +      */
> +     __u64 flags;
> +
> +     union {
> +             struct {
> +                     __u8 slice_mask;
> +                     __u8 subslice_mask;
> +                     __u8 min_eu_per_subslice;
> +                     __u8 max_eu_per_subslice;
> +             } packed;
> +             __u64 value;
> +     };
> +};
> +#endif /* I915_CONTEXT_PARAM_SSEU */
> +
> +static int drm_fd;
> +static int devid;
> +static uint64_t device_slice_mask = 0;
> +static uint64_t device_subslice_mask = 0;
> +static uint32_t device_slice_count = 0;
> +static uint32_t device_subslice_count = 0;
> +
> +static uint64_t mask_minus(uint64_t mask)
> +{
> +     int i;
> +
> +     for (i = 0; i < (sizeof(mask) * 8 - 1); i++) {
> +             if ((1UL << i) & mask) {
> +                     return mask & ~(1UL << i);
> +             }
> +     }
> +
> +     igt_assert(!"reached");
> +     return 0;
> +}
> +
> +static uint32_t
> +read_rpcs_reg(drm_intel_bufmgr *bufmgr,
> +           drm_intel_context *context,
> +           uint64_t engine,
> +           uint32_t expected_slices)
> +{
> +     struct intel_batchbuffer *batch;
> +     drm_intel_bo *bo;
> +     uint32_t rpcs;
> +     int ret;
> +
> +     batch = intel_batchbuffer_alloc(bufmgr, devid);
> +     igt_assert(batch);
> +
> +     intel_batchbuffer_set_context(batch, context);
> +
> +     bo = drm_intel_bo_alloc(bufmgr, "target bo", 4096, 4096);
> +     igt_assert(bo);
> +
> +     /* Clear destination buffer. */
> +     ret = drm_intel_bo_map(bo, false /* write enable */);
> +     igt_assert_eq(ret, 0);
> +     memset(bo->virtual, 0, bo->size);
> +     drm_intel_bo_unmap(bo);
> +
> +     /*
> +      * Prior to Gen10 we can use the predicate to further verify
> +      * that the hardware has been programmed correctly.
> +      */
> +     if (expected_slices != 0 && intel_gen(devid) < 10) {
> +             BEGIN_BATCH(5, 1);
> +             OUT_BATCH(MI_SET_PREDICATE | (1 - 1) |
> +                       (MI_SET_PREDICATE_1_SLICES + expected_slices - 1));
> +     } else {
> +             BEGIN_BATCH(3, 1);
> +     }
> +
> +     OUT_BATCH(MI_STORE_REGISTER_MEM | (4 - 2));
> +     OUT_BATCH(GEN8_R_PWR_CLK_STATE);
> +     OUT_RELOC(bo, I915_GEM_DOMAIN_INSTRUCTION, I915_GEM_DOMAIN_INSTRUCTION, 
> 0);
> +
> +     if (expected_slices != 0 && intel_gen(devid) < 10)
> +             OUT_BATCH(MI_SET_PREDICATE | (1 - 1) | 
> MI_SET_PREDICATE_NOOP_NEVER);
> +
> +     ADVANCE_BATCH();
> +
> +     intel_batchbuffer_flush_on_ring(batch, engine);
> +
> +     drm_intel_bo_wait_rendering(bo);
> +
> +     ret = drm_intel_bo_map(bo, false /* write enable */);
> +     igt_assert_eq(ret, 0);
> +
> +     rpcs = *((uint32_t *) bo->virtual);
> +
> +     drm_intel_bo_unmap(bo);
> +     drm_intel_bo_unreference(bo);
> +
> +     intel_batchbuffer_free(batch);
> +
> +     return rpcs;
> +}
> +
> +
> +static uint32_t
> +read_slice_count(drm_intel_bufmgr *bufmgr,
> +              drm_intel_context *context,
> +              uint32_t expected_slice_count)
> +{
> +     return (read_rpcs_reg(bufmgr, context, I915_EXEC_RENDER,
> +                           expected_slice_count) & GEN8_RPCS_S_CNT_MASK)
> +             >> GEN8_RPCS_S_CNT_SHIFT;
> +}
> +
> +static uint32_t
> +read_subslice_count(drm_intel_bufmgr *bufmgr,
> +                 drm_intel_context *context)
> +{
> +     return (read_rpcs_reg(bufmgr, context, I915_EXEC_RENDER, 0) & 
> GEN8_RPCS_SS_CNT_MASK)
> +             >> GEN8_RPCS_SS_CNT_SHIFT;
> +}
> +
> +static void
> +context_get_sseu_masks(drm_intel_context *context, uint64_t engine,
> +                    uint32_t *slice_mask, uint32_t *subslice_mask)
> +{
> +     struct drm_i915_gem_context_param arg;
> +     struct drm_i915_gem_context_param_sseu sseu;
> +     uint32_t context_id;
> +     int ret;
> +
> +     memset(&sseu, 0, sizeof(sseu));
> +     sseu.flags = engine;
> +
> +     ret = drm_intel_gem_context_get_id(context, &context_id);
> +     igt_assert_eq(ret, 0);
> +
> +     memset(&arg, 0, sizeof(arg));
> +     arg.ctx_id = context_id;
> +     arg.param = I915_CONTEXT_PARAM_SSEU;
> +     arg.value = (uintptr_t) &sseu;
> +
> +     do_ioctl(drm_fd, DRM_IOCTL_I915_GEM_CONTEXT_GETPARAM, &arg);
> +
> +     if (slice_mask)
> +             *slice_mask = sseu.packed.slice_mask;
> +     if (subslice_mask)
> +             *subslice_mask = sseu.packed.subslice_mask;
> +}
> +
> +static void
> +context_set_slice_mask(drm_intel_context *context, uint64_t engine,
> +                    uint32_t slice_mask)
> +{
> +     struct drm_i915_gem_context_param arg;
> +     struct drm_i915_gem_context_param_sseu sseu;
> +     uint32_t context_id;
> +     int ret;
> +
> +     memset(&sseu, 0, sizeof(sseu));
> +     sseu.flags = engine;
> +
> +     ret = drm_intel_gem_context_get_id(context, &context_id);
> +     igt_assert_eq(ret, 0);
> +
> +     memset(&arg, 0, sizeof(arg));
> +     arg.ctx_id = context_id;
> +     arg.param = I915_CONTEXT_PARAM_SSEU;
> +     arg.value = (uintptr_t) &sseu;
> +
> +     do_ioctl(drm_fd, DRM_IOCTL_I915_GEM_CONTEXT_GETPARAM, &arg);
> +
> +     sseu.packed.slice_mask = slice_mask;
> +
> +     do_ioctl(drm_fd, DRM_IOCTL_I915_GEM_CONTEXT_SETPARAM, &arg);
> +}
> +
> +static void
> +context_set_subslice_mask(drm_intel_context *context, uint64_t engine,
> +                       uint32_t subslice_mask)
> +{
> +     struct drm_i915_gem_context_param arg;
> +     struct drm_i915_gem_context_param_sseu sseu;
> +     uint32_t context_id;
> +     int ret;
> +
> +     memset(&sseu, 0, sizeof(sseu));
> +     sseu.flags = engine;
> +
> +     ret = drm_intel_gem_context_get_id(context, &context_id);
> +     igt_assert_eq(ret, 0);
> +
> +     memset(&arg, 0, sizeof(arg));
> +     arg.ctx_id = context_id;
> +     arg.param = I915_CONTEXT_PARAM_SSEU;
> +     arg.value = (uintptr_t) &sseu;
> +
> +     do_ioctl(drm_fd, DRM_IOCTL_I915_GEM_CONTEXT_GETPARAM, &arg);
> +
> +     sseu.packed.subslice_mask = subslice_mask;
> +
> +     do_ioctl(drm_fd, DRM_IOCTL_I915_GEM_CONTEXT_SETPARAM, &arg);
> +}
> +
> +static void rpcs_slice_program_gt(uint64_t engine)
> +{
> +     drm_intel_bufmgr *bufmgr;
> +     drm_intel_context *context1, *context2;
> +     uint32_t slice_mask;
> +
> +     bufmgr = drm_intel_bufmgr_gem_init(drm_fd, 4096);
> +     igt_assert(bufmgr);
> +
> +     context1 = drm_intel_gem_context_create(bufmgr);
> +     igt_assert(context1);
> +
> +     context2 = drm_intel_gem_context_create(bufmgr);
> +     igt_assert(context2);
> +
> +     context_set_slice_mask(context1, engine, mask_minus(device_slice_mask));
> +     context_set_slice_mask(context2, engine, device_slice_mask);
> +
> +     context_get_sseu_masks(context1, engine, &slice_mask, NULL);
> +     igt_assert_eq(mask_minus(device_slice_mask), slice_mask);
> +     context_get_sseu_masks(context2, engine, &slice_mask, NULL);
> +     igt_assert_eq(device_slice_mask, slice_mask);
> +
> +     /*
> +      * Test false positives with predicates (only available on
> +      * before Gen10).
> +      */
> +     if (intel_gen(devid) < 10) {
> +             igt_assert_eq(0, read_slice_count(bufmgr, context1, 
> device_slice_count));
> +     }
> +
> +     igt_assert_eq(device_slice_count - 1,
> +                   read_slice_count(bufmgr, context1, device_slice_count - 
> 1));
> +     igt_assert_eq(device_slice_count,
> +                   read_slice_count(bufmgr, context2, device_slice_count));
> +
> +     context_set_slice_mask(context1, engine, device_slice_mask);
> +     context_set_slice_mask(context2, engine, mask_minus(device_slice_mask));
> +
> +     context_get_sseu_masks(context1, engine, &slice_mask, NULL);
> +     igt_assert_eq(device_slice_mask, slice_mask);
> +     context_get_sseu_masks(context2, engine, &slice_mask, NULL);
> +     igt_assert_eq(mask_minus(device_slice_mask), slice_mask);
> +
> +     igt_assert_eq(device_slice_count,
> +                   read_slice_count(bufmgr, context1, device_slice_count));
> +     igt_assert_eq(device_slice_count - 1,
> +                   read_slice_count(bufmgr, context2, device_slice_count - 
> 1));
> +
> +     if (device_slice_count >= 3) {
> +             context_set_slice_mask(context1, engine, device_slice_mask);
> +             context_set_slice_mask(context2, engine, 
> mask_minus(mask_minus(device_slice_mask)));
> +
> +             context_get_sseu_masks(context1, engine, &slice_mask, NULL);
> +             igt_assert_eq(device_slice_mask, slice_mask);
> +             context_get_sseu_masks(context2, engine, &slice_mask, NULL);
> +             igt_assert_eq(mask_minus(mask_minus(device_slice_mask)), 
> slice_mask);
> +
> +             igt_assert_eq(device_slice_count,
> +                           read_slice_count(bufmgr, context1, 
> device_slice_count));
> +             igt_assert_eq(device_slice_count - 2,
> +                           read_slice_count(bufmgr, context2, 
> device_slice_count - 2));
> +     }
> +
> +     drm_intel_gem_context_destroy(context1);
> +     drm_intel_gem_context_destroy(context2);
> +
> +     drm_intel_bufmgr_destroy(bufmgr);
> +}
> +
> +static void rpcs_subslice_program_gt(uint64_t engine)
> +{
> +     drm_intel_bufmgr *bufmgr;
> +     drm_intel_context *context1, *context2;
> +     uint32_t subslice_mask;
> +
> +     bufmgr = drm_intel_bufmgr_gem_init(drm_fd, 4096);
> +     igt_assert(bufmgr);
> +
> +     context1 = drm_intel_gem_context_create(bufmgr);
> +     igt_assert(context1);
> +
> +     context2 = drm_intel_gem_context_create(bufmgr);
> +     igt_assert(context2);
> +
> +     context_set_subslice_mask(context1, engine, 
> mask_minus(device_subslice_mask));
> +     context_set_subslice_mask(context2, engine, device_subslice_mask);
> +
> +     context_get_sseu_masks(context1, engine, NULL, &subslice_mask);
> +     igt_assert_eq(mask_minus(device_subslice_mask), subslice_mask);
> +     context_get_sseu_masks(context2, engine, NULL, &subslice_mask);
> +     igt_assert_eq(device_subslice_mask, subslice_mask);
> +
> +     igt_assert_eq(device_subslice_count - 1, read_subslice_count(bufmgr, 
> context1));
> +     igt_assert_eq(device_subslice_count, read_subslice_count(bufmgr, 
> context2));
> +
> +     context_set_subslice_mask(context1, engine, device_subslice_mask);
> +     context_set_subslice_mask(context2, engine, 
> mask_minus(device_subslice_mask));
> +
> +     context_get_sseu_masks(context1, engine, NULL, &subslice_mask);
> +     igt_assert_eq(device_subslice_mask, subslice_mask);
> +     context_get_sseu_masks(context2, engine, NULL, &subslice_mask);
> +     igt_assert_eq(mask_minus(device_subslice_mask), subslice_mask);
> +
> +     igt_assert_eq(device_subslice_count, read_subslice_count(bufmgr, 
> context1));
> +     igt_assert_eq(device_subslice_count - 1, read_subslice_count(bufmgr, 
> context2));
> +
> +     if (device_subslice_count >= 3) {
> +             context_set_subslice_mask(context1, engine, 
> device_subslice_mask);
> +             context_set_subslice_mask(context2, engine,
> +                                       
> mask_minus(mask_minus(device_subslice_mask)));
> +
> +             context_get_sseu_masks(context1, engine, &subslice_mask, NULL);
> +             igt_assert_eq(device_subslice_mask, subslice_mask);
> +             context_get_sseu_masks(context2, engine, &subslice_mask, NULL);
> +             igt_assert_eq(mask_minus(mask_minus(device_subslice_mask)), 
> subslice_mask);
> +
> +             igt_assert_eq(device_subslice_count, 
> read_subslice_count(bufmgr, context2));
> +             igt_assert_eq(device_subslice_count - 2, 
> read_subslice_count(bufmgr, context1));
> +     }
> +
> +     drm_intel_gem_context_destroy(context1);
> +     drm_intel_gem_context_destroy(context2);
> +
> +     drm_intel_bufmgr_destroy(bufmgr);
> +}
> +
> +igt_main
> +{
> +     uint64_t engines[] = {
> +             I915_EXEC_RENDER,
> +             I915_EXEC_BSD,
> +             I915_EXEC_VEBOX,
> +     };
> +     int i;
> +     drm_i915_getparam_t gp;
> +
> +     igt_fixture {
> +             /* Use drm_open_driver to verify device existence */
> +             drm_fd = drm_open_driver(DRIVER_INTEL);
> +             igt_require_gem(drm_fd);
> +
> +             devid = intel_get_drm_devid(drm_fd);
> +
> +             /* We can only program slice count from Gen8. */
> +             igt_skip_on(intel_gen(devid) < 8);
> +     }
> +
> +     gp.param = I915_PARAM_SLICE_MASK;
> +     gp.value = (int *) &device_slice_mask;
> +     do_ioctl(drm_fd, DRM_IOCTL_I915_GETPARAM, &gp);
> +     device_slice_count = __builtin_popcount(device_slice_mask);

As you might see from the CI results, you can't do ioctls without
being in a subtest or a fixture, or you break --list-subtests.


-- 
Petri Latvala


> +
> +     gp.param = I915_PARAM_SUBSLICE_MASK;
> +     gp.value = (int *) &device_subslice_mask;
> +     do_ioctl(drm_fd, DRM_IOCTL_I915_GETPARAM, &gp);
> +     device_subslice_count = __builtin_popcount(device_subslice_mask);
> +
> +     igt_subtest("rpcs-slice-program-rcs") {
> +             igt_require(device_slice_count >= 2);
> +
> +             for (i = 0; i < ARRAY_SIZE(engines); i++)
> +                     rpcs_slice_program_gt(engines[i]);
> +     }
> +
> +     igt_subtest("rpcs-subslice-program-rcs") {
> +             igt_require(device_subslice_count >= 2);
> +             /* Only available on some Atom platforms and Gen10+. */
> +             igt_require(IS_BROXTON(devid) || IS_GEMINILAKE(devid) ||
> +                         intel_gen(devid) >= 10);
> +
> +             for (i = 0; i < ARRAY_SIZE(engines); i++)
> +                     rpcs_subslice_program_gt(engines[i]);
> +     }
> +
> +     igt_fixture {
> +             close(drm_fd);
> +     }
> +}
> diff --git a/tests/meson.build b/tests/meson.build
> index 1c619179..18e656a0 100644
> --- a/tests/meson.build
> +++ b/tests/meson.build
> @@ -6,6 +6,7 @@ test_progs = [
>       'core_getversion',
>       'core_prop_blob',
>       'core_setmaster_vs_auth',
> +     'ctx_rpcs',
>       'debugfs_test',
>       'drm_import_export',
>       'drm_mm',
> -- 
> 2.14.1
> 
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx@lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/intel-gfx
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

Reply via email to