/* * Copyright © 2014 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. * * Authors: * Rob Bradford * Tiago Vignatti * */ /* * Testcase: Check whether mmap()ing dma-buf works */ #include #include #include #include #include #include #include #include #include #include #include "drm.h" #include "i915_drm.h" #include "drmtest.h" #include "igt_debugfs.h" #include "ioctl_wrappers.h" #include "i915/gem_mman.h" #define BO_SIZE (16*1024) static int fd; char pattern[] = {0xff, 0x00, 0x00, 0x00, 0x00, 0xff, 0x00, 0x00, 0x00, 0x00, 0xff, 0x00, 0x00, 0x00, 0x00, 0xff}; static void fill_bo(uint32_t handle, size_t size) { off_t i; for (i = 0; i < size; i+=sizeof(pattern)) { gem_write(fd, handle, i, pattern, sizeof(pattern)); } } static void fill_bo_cpu(char *ptr) { memcpy(ptr, pattern, sizeof(pattern)); } static void test_correct(void) { int dma_buf_fd; char *ptr1, *ptr2; uint32_t handle; handle = gem_create(fd, BO_SIZE); fill_bo(handle, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); igt_assert(errno == 0); /* Check correctness vs GEM_MMAP_GTT */ ptr1 = gem_mmap__gtt(fd, handle, BO_SIZE, PROT_READ); ptr2 = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr1 != MAP_FAILED); igt_assert(ptr2 != MAP_FAILED); igt_assert(memcmp(ptr1, ptr2, BO_SIZE) == 0); /* Check pattern correctness */ igt_assert(memcmp(ptr2, pattern, sizeof(pattern)) == 0); munmap(ptr1, BO_SIZE); munmap(ptr2, BO_SIZE); close(dma_buf_fd); gem_close(fd, handle); } static void test_map_unmap(void) { int dma_buf_fd; char *ptr; uint32_t handle; handle = gem_create(fd, BO_SIZE); fill_bo(handle, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); igt_assert(errno == 0); ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr != MAP_FAILED); igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); /* Unmap and remap */ munmap(ptr, BO_SIZE); ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr != MAP_FAILED); igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); munmap(ptr, BO_SIZE); close(dma_buf_fd); gem_close(fd, handle); } /* prime and then unprime and then prime again the same handle */ static void test_reprime(void) { int dma_buf_fd; char *ptr; uint32_t handle; handle = gem_create(fd, BO_SIZE); fill_bo(handle, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); igt_assert(errno == 0); ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr != MAP_FAILED); igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); close (dma_buf_fd); igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); munmap(ptr, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr != MAP_FAILED); igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); munmap(ptr, BO_SIZE); close(dma_buf_fd); gem_close(fd, handle); } /* map from another process */ static void test_forked(void) { int dma_buf_fd; char *ptr; uint32_t handle; handle = gem_create(fd, BO_SIZE); fill_bo(handle, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); igt_assert(errno == 0); igt_fork(childno, 1) { ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr != MAP_FAILED); igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); munmap(ptr, BO_SIZE); close(dma_buf_fd); } close(dma_buf_fd); igt_waitchildren(); gem_close(fd, handle); } /* test simple CPU write */ static void test_correct_cpu_write(void) { int dma_buf_fd; char *ptr; uint32_t handle; handle = gem_create(fd, BO_SIZE); dma_buf_fd = prime_handle_to_fd_for_mmap(fd, handle); /* Skip if DRM_RDWR is not supported */ igt_skip_on(errno == EINVAL); /* Check correctness of map using write protection (PROT_WRITE) */ ptr = mmap(NULL, BO_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr != MAP_FAILED); /* Fill bo using CPU */ fill_bo_cpu(ptr); /* Check pattern correctness */ igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); munmap(ptr, BO_SIZE); close(dma_buf_fd); gem_close(fd, handle); } /* map from another process and then write using CPU */ static void test_forked_cpu_write(void) { int dma_buf_fd; char *ptr; uint32_t handle; handle = gem_create(fd, BO_SIZE); dma_buf_fd = prime_handle_to_fd_for_mmap(fd, handle); /* Skip if DRM_RDWR is not supported */ igt_skip_on(errno == EINVAL); igt_fork(childno, 1) { ptr = mmap(NULL, BO_SIZE, PROT_READ | PROT_WRITE , MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr != MAP_FAILED); fill_bo_cpu(ptr); igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); munmap(ptr, BO_SIZE); close(dma_buf_fd); } close(dma_buf_fd); igt_waitchildren(); gem_close(fd, handle); } static void test_refcounting(void) { int dma_buf_fd; char *ptr; uint32_t handle; handle = gem_create(fd, BO_SIZE); fill_bo(handle, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); igt_assert(errno == 0); /* Close gem object before mapping */ gem_close(fd, handle); ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr != MAP_FAILED); igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); munmap(ptr, BO_SIZE); close (dma_buf_fd); } /* dup before mmap */ static void test_dup(void) { int dma_buf_fd; char *ptr; uint32_t handle; handle = gem_create(fd, BO_SIZE); fill_bo(handle, BO_SIZE); dma_buf_fd = dup(prime_handle_to_fd(fd, handle)); igt_assert(errno == 0); ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr != MAP_FAILED); igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0); munmap(ptr, BO_SIZE); gem_close(fd, handle); close (dma_buf_fd); } /* Used for error case testing to avoid wrapper */ static int prime_handle_to_fd_no_assert(uint32_t handle, int flags, int *fd_out) { struct drm_prime_handle args; int ret; args.handle = handle; args.flags = flags; args.fd = -1; ret = drmIoctl(fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &args); if (ret) ret = errno; *fd_out = args.fd; return ret; } static bool has_userptr(void) { uint32_t handle = 0; void *ptr; igt_assert(posix_memalign(&ptr, 4096, 4096) == 0); if ( __gem_userptr(fd, ptr, 4096, 0, 0, &handle) == 0) gem_close(fd, handle); free(ptr); return handle; } /* test for mmap(dma_buf_export(userptr)) */ static void test_userptr(void) { int ret, dma_buf_fd; void *ptr; uint32_t handle; igt_require(has_userptr()); /* create userptr bo */ ret = posix_memalign(&ptr, 4096, BO_SIZE); igt_assert_eq(ret, 0); /* we are not allowed to export unsynchronized userptr. Just create a normal * one */ gem_userptr(fd, (uint32_t *)ptr, BO_SIZE, 0, 0, &handle); /* export userptr */ ret = prime_handle_to_fd_no_assert(handle, DRM_CLOEXEC, &dma_buf_fd); if (ret) { igt_assert(ret == EINVAL || ret == ENODEV); goto free_userptr; } else { igt_assert_eq(ret, 0); igt_assert_lte(0, dma_buf_fd); } /* a userptr doesn't have the obj->base.filp, but can be exported via * dma-buf, so make sure it fails here */ ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr == MAP_FAILED && errno == ENODEV); free_userptr: gem_close(fd, handle); close(dma_buf_fd); } static void test_errors(void) { int i, dma_buf_fd; char *ptr; uint32_t handle; int invalid_flags[] = {DRM_CLOEXEC - 1, DRM_CLOEXEC + 1, DRM_RDWR - 1, DRM_RDWR + 1}; /* Test for invalid flags */ handle = gem_create(fd, BO_SIZE); for (i = 0; i < sizeof(invalid_flags) / sizeof(invalid_flags[0]); i++) { prime_handle_to_fd_no_assert(handle, invalid_flags[i], &dma_buf_fd); igt_assert_eq(errno, EINVAL); errno = 0; } /* Close gem object before priming */ handle = gem_create(fd, BO_SIZE); fill_bo(handle, BO_SIZE); gem_close(fd, handle); prime_handle_to_fd_no_assert(handle, DRM_CLOEXEC, &dma_buf_fd); igt_assert(dma_buf_fd == -1 && errno == ENOENT); errno = 0; /* close fd before mapping */ handle = gem_create(fd, BO_SIZE); fill_bo(handle, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); igt_assert(errno == 0); close(dma_buf_fd); ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr == MAP_FAILED && errno == EBADF); errno = 0; gem_close(fd, handle); /* Map too big */ handle = gem_create(fd, BO_SIZE); fill_bo(handle, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); igt_assert(errno == 0); ptr = mmap(NULL, BO_SIZE * 2, PROT_READ, MAP_SHARED, dma_buf_fd, 0); igt_assert(ptr == MAP_FAILED && errno == EINVAL); errno = 0; close(dma_buf_fd); gem_close(fd, handle); /* Overlapping the end of the buffer */ handle = gem_create(fd, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); igt_assert(errno == 0); ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, BO_SIZE / 2); igt_assert(ptr == MAP_FAILED && errno == EINVAL); errno = 0; close(dma_buf_fd); gem_close(fd, handle); } /* Test for invalid flags on sync ioctl */ static void test_invalid_sync_flags(void) { int i, dma_buf_fd; uint32_t handle; struct local_dma_buf_sync sync; int invalid_flags[] = {-1, 0x00, LOCAL_DMA_BUF_SYNC_RW + 1, LOCAL_DMA_BUF_SYNC_VALID_FLAGS_MASK + 1}; handle = gem_create(fd, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); for (i = 0; i < sizeof(invalid_flags) / sizeof(invalid_flags[0]); i++) { memset(&sync, 0, sizeof(sync)); sync.flags = invalid_flags[i]; drmIoctl(dma_buf_fd, LOCAL_DMA_BUF_IOCTL_SYNC, &sync); igt_assert_eq(errno, EINVAL); errno = 0; } } static void test_aperture_limit(void) { int dma_buf_fd1, dma_buf_fd2; char *ptr1, *ptr2; uint32_t handle1, handle2; /* Two buffers the sum of which > mappable aperture */ uint64_t size1 = (gem_mappable_aperture_size() * 7) / 8; uint64_t size2 = (gem_mappable_aperture_size() * 3) / 8; handle1 = gem_create(fd, size1); fill_bo(handle1, BO_SIZE); dma_buf_fd1 = prime_handle_to_fd(fd, handle1); igt_assert(errno == 0); ptr1 = mmap(NULL, size1, PROT_READ, MAP_SHARED, dma_buf_fd1, 0); igt_assert(ptr1 != MAP_FAILED); igt_assert(memcmp(ptr1, pattern, sizeof(pattern)) == 0); handle2 = gem_create(fd, size1); fill_bo(handle2, BO_SIZE); dma_buf_fd2 = prime_handle_to_fd(fd, handle2); igt_assert(errno == 0); ptr2 = mmap(NULL, size2, PROT_READ, MAP_SHARED, dma_buf_fd2, 0); igt_assert(ptr2 != MAP_FAILED); igt_assert(memcmp(ptr2, pattern, sizeof(pattern)) == 0); igt_assert(memcmp(ptr1, ptr2, BO_SIZE) == 0); munmap(ptr1, size1); munmap(ptr2, size2); close(dma_buf_fd1); close(dma_buf_fd2); gem_close(fd, handle1); gem_close(fd, handle2); } static int check_for_dma_buf_mmap(void) { int dma_buf_fd; char *ptr; uint32_t handle; int ret = 1; handle = gem_create(fd, BO_SIZE); dma_buf_fd = prime_handle_to_fd(fd, handle); ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0); if (ptr != MAP_FAILED) ret = 0; munmap(ptr, BO_SIZE); gem_close(fd, handle); close(dma_buf_fd); return ret; } igt_main { struct { const char *name; void (*fn)(void); } tests[] = { { "test_correct", test_correct }, { "test_map_unmap", test_map_unmap }, { "test_reprime", test_reprime }, { "test_forked", test_forked }, { "test_correct_cpu_write", test_correct_cpu_write }, { "test_forked_cpu_write", test_forked_cpu_write }, { "test_refcounting", test_refcounting }, { "test_dup", test_dup }, { "test_userptr", test_userptr }, { "test_errors", test_errors }, { "test_invalid_sync_flags", test_invalid_sync_flags }, { "test_aperture_limit", test_aperture_limit }, }; int i; igt_fixture { fd = drm_open_driver(DRIVER_INTEL); igt_skip_on((check_for_dma_buf_mmap() != 0)); errno = 0; } for (i = 0; i < ARRAY_SIZE(tests); i++) { igt_subtest(tests[i].name) tests[i].fn(); } igt_fixture close(fd); }