On 08/13/2015 04:01 AM, Daniel Vetter wrote: > On Wed, Aug 12, 2015 at 08:29:16PM -0300, Tiago Vignatti wrote: >> This patch adds test_correct_cpu_write, which maps the texture buffer >> through a >> prime fd and then writes directly to it using the CPU. It stresses the driver >> to guarantee cache synchronization among the different domains. >> >> This test also adds test_forked_cpu_write, which creates the GEM bo in one >> process and pass the prime handle of the it to another process, which in turn >> uses the handle only to map and write. Grossly speaking this test simulates >> Chrome OS architecture, where the Web content ("unpriviledged process") maps >> and CPU-draws a buffer, which was previously allocated in the GPU process >> ("priviledged process"). >> >> This requires kernel modifications (Daniel Thompson's "drm: prime: Honour >> O_RDWR during prime-handle-to-fd"). >> >> Signed-off-by: Tiago Vignatti <tiago.vignatti at intel.com> > > Squash with previous patch?
why? if the whole point is to decrease the amount of patches, then I prefer to squash 2/7 with the 1/7 (although they're from different authors and would be nice to keep separately the changes from each). This patch here introduces this writing to mmap'ed dma-buf fd, a concept that is still in debate, requiring a kernel counter-part so that's why I preferred to keep it away. >> --- >> lib/ioctl_wrappers.c | 5 +++- >> tests/prime_mmap.c | 65 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++ >> 2 files changed, 69 insertions(+), 1 deletion(-) >> >> diff --git a/lib/ioctl_wrappers.c b/lib/ioctl_wrappers.c >> index 53bd635..941fa66 100644 >> --- a/lib/ioctl_wrappers.c >> +++ b/lib/ioctl_wrappers.c >> @@ -1125,6 +1125,9 @@ void gem_require_ring(int fd, int ring_id) >> >> /* prime */ >> >> +#ifndef DRM_RDWR >> +#define DRM_RDWR O_RDWR >> +#endif >> /** >> * prime_handle_to_fd: >> * @fd: open i915 drm file descriptor >> @@ -1142,7 +1145,7 @@ int prime_handle_to_fd(int fd, uint32_t handle) >> >> memset(&args, 0, sizeof(args)); >> args.handle = handle; >> - args.flags = DRM_CLOEXEC; >> + args.flags = DRM_CLOEXEC | DRM_RDWR; > > This needs to be optional otherwise all the existing prime tests start > falling over on older kernels. Probably need a > prime_handle_to_fd_with_mmap, which doesn an igt_skip if it fails. true. Thank you. > -Daniel > >> args.fd = -1; >> >> do_ioctl(fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &args); >> diff --git a/tests/prime_mmap.c b/tests/prime_mmap.c >> index dc59e8f..ad91371 100644 >> --- a/tests/prime_mmap.c >> +++ b/tests/prime_mmap.c >> @@ -22,6 +22,7 @@ >> * >> * Authors: >> * Rob Bradford <rob at linux.intel.com> >> + * Tiago Vignatti <tiago.vignatti at intel.com> >> * >> */ >> >> @@ -66,6 +67,12 @@ fill_bo(uint32_t handle, size_t size) >> } >> >> static void >> +fill_bo_cpu(char *ptr) >> +{ >> + memcpy(ptr, pattern, sizeof(pattern)); >> +} >> + >> +static void >> test_correct(void) >> { >> int dma_buf_fd; >> @@ -180,6 +187,62 @@ test_forked(void) >> gem_close(fd, handle); >> } >> >> +/* test CPU write. This has a rather big implication for the driver which >> must >> + * guarantee cache synchronization when writing the bo using CPU. */ >> +static void >> +test_correct_cpu_write(void) >> +{ >> + int dma_buf_fd; >> + char *ptr; >> + uint32_t handle; >> + >> + handle = gem_create(fd, BO_SIZE); >> + >> + dma_buf_fd = prime_handle_to_fd(fd, handle); >> + igt_assert(errno == 0); >> + >> + /* Check correctness of map using write protection (PROT_WRITE) */ >> + ptr = mmap(NULL, BO_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED, >> dma_buf_fd, 0); >> + igt_assert(ptr != MAP_FAILED); >> + >> + /* Fill bo using CPU */ >> + fill_bo_cpu(ptr); >> + >> + /* Check pattern correctness */ >> + igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); >> + >> + munmap(ptr, BO_SIZE); >> + close(dma_buf_fd); >> + gem_close(fd, handle); >> +} >> + >> +/* map from another process and then write using CPU */ >> +static void >> +test_forked_cpu_write(void) >> +{ >> + int dma_buf_fd; >> + char *ptr; >> + uint32_t handle; >> + >> + handle = gem_create(fd, BO_SIZE); >> + >> + dma_buf_fd = prime_handle_to_fd(fd, handle); >> + igt_assert(errno == 0); >> + >> + igt_fork(childno, 1) { >> + ptr = mmap(NULL, BO_SIZE, PROT_READ | PROT_WRITE , MAP_SHARED, >> dma_buf_fd, 0); >> + igt_assert(ptr != MAP_FAILED); >> + fill_bo_cpu(ptr); >> + >> + igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); >> + munmap(ptr, BO_SIZE); >> + close(dma_buf_fd); >> + } >> + close(dma_buf_fd); >> + igt_waitchildren(); >> + gem_close(fd, handle); >> +} >> + >> static void >> test_refcounting(void) >> { >> @@ -346,6 +409,8 @@ igt_main >> { "test_map_unmap", test_map_unmap }, >> { "test_reprime", test_reprime }, >> { "test_forked", test_forked }, >> + { "test_correct_cpu_write", test_correct_cpu_write }, >> + { "test_forked_cpu_write", test_forked_cpu_write }, >> { "test_refcounting", test_refcounting }, >> { "test_dup", test_dup }, >> { "test_errors", test_errors }, >> -- >> 2.1.0 >> >