/* * Copyright (C) 2020 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include "dmabuf_heap_test.h" #include #include #include #include #include #include #include #include class DmaBufHeapConcurrentAccessTest : public ::testing::Test { public: virtual void SetUp() { allocator = new BufferAllocator(); } void DoAlloc(bool cpu_access_needed) { static const size_t kAllocSizeInBytes = 4096; int map_fd = allocator->AllocSystem(cpu_access_needed, kAllocSizeInBytes); ASSERT_GE(map_fd, 0); void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0); ASSERT_TRUE(ptr != MAP_FAILED); int ret = allocator->CpuSyncStart(map_fd, kSyncReadWrite); ASSERT_EQ(0, ret); ret = allocator->CpuSyncEnd(map_fd, kSyncReadWrite); ASSERT_EQ(0, ret); ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes)); ASSERT_EQ(0, close(map_fd)); } void DoConcurrentAlloc() { DoAlloc(true /* cpu_access_needed */); DoAlloc(false /* cpu_access_needed */); } void DoConcurrentAllocWithMapName() { allocator->MapNameToIonHeap(kDmabufSystemHeapName, "" /* no mapping for non-legacy */, 0 /* no mapping for non-legacy ion */, ~0 /* legacy ion heap mask */, ION_FLAG_CACHED); DoAlloc(true /* cpu_access_needed */); allocator->MapNameToIonHeap( kDmabufSystemUncachedHeapName, "" /* no mapping for non-legacy */, 0 /* no mapping for non-legacy ion */, ~0 /* legacy ion heap mask */); DoAlloc(false /* cpu_access_needed */); } virtual void TearDown() { delete allocator; } BufferAllocator* allocator = nullptr; }; static constexpr size_t NUM_CONCURRENT_THREADS = 100; TEST_F(DmaBufHeapConcurrentAccessTest, ConcurrentAllocTest) { using android::vintf::KernelVersion; KernelVersion min_kernel_version = KernelVersion(5, 10, 0); KernelVersion kernel_version = android::vintf::VintfObject::GetInstance() ->getRuntimeInfo(android::vintf::RuntimeInfo::FetchFlag::CPU_VERSION) ->kernelVersion(); if (kernel_version < min_kernel_version) { GTEST_SKIP(); } std::vector threads(NUM_CONCURRENT_THREADS); for (int i = 0; i < NUM_CONCURRENT_THREADS; i++) { threads[i] = std::thread(&DmaBufHeapConcurrentAccessTest::DoConcurrentAlloc, this); } for (auto& thread : threads) { thread.join(); } } TEST_F(DmaBufHeapConcurrentAccessTest, ConcurrentAllocWithMapNameTest) { std::vector threads(NUM_CONCURRENT_THREADS); for (int i = 0; i < NUM_CONCURRENT_THREADS; i++) { threads[i] = std::thread(&DmaBufHeapConcurrentAccessTest::DoConcurrentAllocWithMapName, this); } for (auto& thread : threads) { thread.join(); } } DmaBufHeapTest::DmaBufHeapTest() : allocator(new BufferAllocator()) { /* * Legacy ion devices may have hardcoded heap IDs that do not * match the ion UAPI header. Map heap name 'system'/'system-uncached' to a heap mask * of all 1s so that these devices will allocate from the first * available heap when asked to allocate from the system or system-uncached * heap. */ if (BufferAllocator::CheckIonSupport()) { allocator->MapNameToIonHeap(kDmabufSystemHeapName, "" /* no mapping for non-legacy */, 0 /* no mapping for non-legacy ion */, ~0 /* legacy ion heap mask */); allocator->MapNameToIonHeap( kDmabufSystemUncachedHeapName, "" /* no mapping for non-legacy */, 0 /* no mapping for non-legacy ion */, ~0 /* legacy ion heap mask */); } } TEST_F(DmaBufHeapTest, Allocate) { static const size_t allocationSizes[] = {4 * 1024, 64 * 1024, 1024 * 1024, 2 * 1024 * 1024}; for (bool cpu_access_needed : {false, true}) { for (size_t size : allocationSizes) { SCOPED_TRACE(::testing::Message() << "cpu_access_needed: " << cpu_access_needed << " size: " << size); int fd = allocator->AllocSystem(cpu_access_needed, size); ASSERT_GE(fd, 0); ASSERT_EQ(close(fd), 0); // free the buffer } } } TEST_F(DmaBufHeapTest, AllocateCachedNeedsSync) { static const size_t allocationSizes[] = {4 * 1024, 64 * 1024, 1024 * 1024, 2 * 1024 * 1024}; for (size_t size : allocationSizes) { SCOPED_TRACE(::testing::Message() << "heap: " << kDmabufSystemHeapName << " size: " << size); int fd = allocator->Alloc(kDmabufSystemHeapName, size, ION_FLAG_CACHED_NEEDS_SYNC /* ion heap flags will be ignored if using dmabuf heaps */); ASSERT_GE(fd, 0); ASSERT_EQ(close(fd), 0); // free the buffer } } TEST_F(DmaBufHeapTest, RepeatedAllocate) { static const size_t allocationSizes[] = {4 * 1024, 64 * 1024, 1024 * 1024, 2 * 1024 * 1024}; for (bool cpu_access_needed : {false, true}) { for (size_t size : allocationSizes) { SCOPED_TRACE(::testing::Message() << "cpu_access_needed: " << cpu_access_needed << " size: " << size); for (unsigned int i = 0; i < 1024; i++) { SCOPED_TRACE(::testing::Message() << "iteration " << i); int fd = allocator->AllocSystem(cpu_access_needed, size); ASSERT_GE(fd, 0); ASSERT_EQ(close(fd), 0); // free the buffer } } } } /* * Make sure all heaps always return zeroed pages */ TEST_F(DmaBufHeapTest, Zeroed) { static const size_t kAllocSizeInBytes = 4096; static const size_t kNumFds = 16; auto zeroes_ptr = std::make_unique(kAllocSizeInBytes); int fds[kNumFds]; int ret = 0, map_fd = -1; for (unsigned int i = 0; i < kNumFds; i++) { map_fd = allocator->Alloc(kDmabufSystemHeapName, kAllocSizeInBytes); ASSERT_GE(map_fd, 0); void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_WRITE, MAP_SHARED, map_fd, 0); ASSERT_TRUE(ptr != MAP_FAILED); ret = allocator->CpuSyncStart(map_fd, kSyncWrite); ASSERT_EQ(0, ret); memset(ptr, 0xaa, kAllocSizeInBytes); ret = allocator->CpuSyncEnd(map_fd, kSyncWrite); ASSERT_EQ(0, ret); ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes)); fds[i] = map_fd; } for (unsigned int i = 0; i < kNumFds; i++) { ASSERT_EQ(0, close(fds[i])); } map_fd = allocator->Alloc(kDmabufSystemHeapName, kAllocSizeInBytes); ASSERT_GE(map_fd, 0); void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_READ, MAP_SHARED, map_fd, 0); ASSERT_TRUE(ptr != MAP_FAILED); ret = allocator->CpuSyncStart(map_fd); ASSERT_EQ(0, ret); ASSERT_EQ(0, memcmp(ptr, zeroes_ptr.get(), kAllocSizeInBytes)); ret = allocator->CpuSyncEnd(map_fd); ASSERT_EQ(0, ret); ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes)); ASSERT_EQ(0, close(map_fd)); } TEST_F(DmaBufHeapTest, TestCpuSync) { static const size_t kAllocSizeInBytes = 4096; auto vec_sync_type = {kSyncRead, kSyncWrite, kSyncReadWrite}; for (auto sync_type : vec_sync_type) { int map_fd = allocator->Alloc(kDmabufSystemHeapName, kAllocSizeInBytes); ASSERT_GE(map_fd, 0); void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0); ASSERT_TRUE(ptr != MAP_FAILED); int ret = allocator->CpuSyncStart(map_fd, sync_type); ASSERT_EQ(0, ret); ret = allocator->CpuSyncEnd(map_fd, sync_type); ASSERT_EQ(0, ret); ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes)); ASSERT_EQ(0, close(map_fd)); } } int CustomCpuSyncStart(int /* ion_fd */, int /* dma_buf fd */, void* /* custom_data pointer */) { LOG(INFO) << "In custom cpu sync start callback"; return 0; } int CustomCpuSyncEnd(int /* ion_fd */, int /* dma_buf fd */, void* /* custom_data pointer */) { LOG(INFO) << "In custom cpu sync end callback"; return 0; } TEST_F(DmaBufHeapTest, TestCustomLegacyIonSyncCallback) { static const size_t allocationSizes[] = {4 * 1024, 64 * 1024, 1024 * 1024, 2 * 1024 * 1024}; for (size_t size : allocationSizes) { SCOPED_TRACE(::testing::Message() << "heap: " << kDmabufSystemHeapName << " size: " << size); int map_fd = allocator->Alloc(kDmabufSystemHeapName, size); ASSERT_GE(map_fd, 0); void* ptr = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0); ASSERT_TRUE(ptr != MAP_FAILED); int ret = allocator->CpuSyncStart(map_fd, kSyncWrite, CustomCpuSyncStart); ASSERT_EQ(0, ret); memset(ptr, 0xaa, size); ret = allocator->CpuSyncEnd(map_fd, kSyncWrite, CustomCpuSyncEnd); ASSERT_EQ(0, ret); ASSERT_EQ(0, munmap(ptr, size)); ASSERT_EQ(0, close(map_fd)); } } TEST_F(DmaBufHeapTest, TestDeviceCapabilityCheck) { auto heap_list = allocator->GetDmabufHeapList(); ASSERT_TRUE(!heap_list.empty() || BufferAllocator::CheckIonSupport()); } TEST_F(DmaBufHeapTest, TestDmabufSystemHeapCompliance) { using android::vintf::KernelVersion; if (android::base::GetIntProperty("ro.product.first_api_level", 0) < __ANDROID_API_S__) { GTEST_SKIP(); } KernelVersion min_kernel_version = KernelVersion(5, 10, 0); KernelVersion kernel_version = android::vintf::VintfObject::GetInstance() ->getRuntimeInfo(android::vintf::RuntimeInfo::FetchFlag::CPU_VERSION) ->kernelVersion(); if (kernel_version < min_kernel_version) { GTEST_SKIP(); } auto heap_list = allocator->GetDmabufHeapList(); ASSERT_TRUE(heap_list.find("system") != heap_list.end()); for (bool cpu_access_needed : {false, true}) { static const size_t kAllocSizeInBytes = 4096; /* * Test that system heap can be allocated from. */ SCOPED_TRACE(::testing::Message() << "cpu_access_needed: " << cpu_access_needed); int map_fd = allocator->AllocSystem(cpu_access_needed, kAllocSizeInBytes); ASSERT_GE(map_fd, 0); /* * Test that system heap can be mmapped by the CPU. */ void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0); ASSERT_TRUE(ptr != MAP_FAILED); /* * Test that the allocated memory is zeroed. */ auto zeroes_ptr = std::make_unique(kAllocSizeInBytes); int ret = allocator->CpuSyncStart(map_fd); ASSERT_EQ(0, ret); ASSERT_EQ(0, memcmp(ptr, zeroes_ptr.get(), kAllocSizeInBytes)); ret = allocator->CpuSyncEnd(map_fd); ASSERT_EQ(0, ret); ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes)); ASSERT_EQ(0, close(map_fd)); } }