You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
335 lines
12 KiB
335 lines
12 KiB
/*
|
|
* Copyright (C) 2020 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include <BufferAllocator/BufferAllocator.h>
|
|
#include "dmabuf_heap_test.h"
|
|
|
|
#include <linux/ion.h>
|
|
#include <sys/mman.h>
|
|
|
|
#include <gtest/gtest.h>
|
|
|
|
#include <android-base/logging.h>
|
|
#include <android-base/properties.h>
|
|
#include <android-base/unique_fd.h>
|
|
#include <vintf/VintfObject.h>
|
|
|
|
#include <thread>
|
|
|
|
class DmaBufHeapConcurrentAccessTest : public ::testing::Test {
|
|
public:
|
|
virtual void SetUp() { allocator = new BufferAllocator(); }
|
|
|
|
void DoAlloc(bool cpu_access_needed) {
|
|
static const size_t kAllocSizeInBytes = 4096;
|
|
int map_fd = allocator->AllocSystem(cpu_access_needed, kAllocSizeInBytes);
|
|
ASSERT_GE(map_fd, 0);
|
|
|
|
void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0);
|
|
ASSERT_TRUE(ptr != MAP_FAILED);
|
|
|
|
int ret = allocator->CpuSyncStart(map_fd, kSyncReadWrite);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ret = allocator->CpuSyncEnd(map_fd, kSyncReadWrite);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes));
|
|
ASSERT_EQ(0, close(map_fd));
|
|
}
|
|
|
|
void DoConcurrentAlloc() {
|
|
DoAlloc(true /* cpu_access_needed */);
|
|
DoAlloc(false /* cpu_access_needed */);
|
|
}
|
|
|
|
void DoConcurrentAllocWithMapName() {
|
|
allocator->MapNameToIonHeap(kDmabufSystemHeapName, "" /* no mapping for non-legacy */,
|
|
0 /* no mapping for non-legacy ion */,
|
|
~0 /* legacy ion heap mask */, ION_FLAG_CACHED);
|
|
DoAlloc(true /* cpu_access_needed */);
|
|
allocator->MapNameToIonHeap(
|
|
kDmabufSystemUncachedHeapName, "" /* no mapping for non-legacy */,
|
|
0 /* no mapping for non-legacy ion */, ~0 /* legacy ion heap mask */);
|
|
DoAlloc(false /* cpu_access_needed */);
|
|
}
|
|
|
|
virtual void TearDown() { delete allocator; }
|
|
|
|
BufferAllocator* allocator = nullptr;
|
|
};
|
|
|
|
static constexpr size_t NUM_CONCURRENT_THREADS = 100;
|
|
|
|
TEST_F(DmaBufHeapConcurrentAccessTest, ConcurrentAllocTest) {
|
|
using android::vintf::KernelVersion;
|
|
|
|
KernelVersion min_kernel_version = KernelVersion(5, 10, 0);
|
|
KernelVersion kernel_version =
|
|
android::vintf::VintfObject::GetInstance()
|
|
->getRuntimeInfo(android::vintf::RuntimeInfo::FetchFlag::CPU_VERSION)
|
|
->kernelVersion();
|
|
if (kernel_version < min_kernel_version) {
|
|
GTEST_SKIP();
|
|
}
|
|
|
|
std::vector<std::thread> threads(NUM_CONCURRENT_THREADS);
|
|
for (int i = 0; i < NUM_CONCURRENT_THREADS; i++) {
|
|
threads[i] = std::thread(&DmaBufHeapConcurrentAccessTest::DoConcurrentAlloc, this);
|
|
}
|
|
|
|
for (auto& thread : threads) {
|
|
thread.join();
|
|
}
|
|
}
|
|
|
|
TEST_F(DmaBufHeapConcurrentAccessTest, ConcurrentAllocWithMapNameTest) {
|
|
std::vector<std::thread> threads(NUM_CONCURRENT_THREADS);
|
|
for (int i = 0; i < NUM_CONCURRENT_THREADS; i++) {
|
|
threads[i] =
|
|
std::thread(&DmaBufHeapConcurrentAccessTest::DoConcurrentAllocWithMapName, this);
|
|
}
|
|
|
|
for (auto& thread : threads) {
|
|
thread.join();
|
|
}
|
|
}
|
|
|
|
DmaBufHeapTest::DmaBufHeapTest() : allocator(new BufferAllocator()) {
|
|
/*
|
|
* Legacy ion devices may have hardcoded heap IDs that do not
|
|
* match the ion UAPI header. Map heap name 'system'/'system-uncached' to a heap mask
|
|
* of all 1s so that these devices will allocate from the first
|
|
* available heap when asked to allocate from the system or system-uncached
|
|
* heap.
|
|
*/
|
|
if (BufferAllocator::CheckIonSupport()) {
|
|
allocator->MapNameToIonHeap(kDmabufSystemHeapName, "" /* no mapping for non-legacy */,
|
|
0 /* no mapping for non-legacy ion */,
|
|
~0 /* legacy ion heap mask */);
|
|
allocator->MapNameToIonHeap(
|
|
kDmabufSystemUncachedHeapName, "" /* no mapping for non-legacy */,
|
|
0 /* no mapping for non-legacy ion */, ~0 /* legacy ion heap mask */);
|
|
}
|
|
}
|
|
|
|
TEST_F(DmaBufHeapTest, Allocate) {
|
|
static const size_t allocationSizes[] = {4 * 1024, 64 * 1024, 1024 * 1024, 2 * 1024 * 1024};
|
|
for (bool cpu_access_needed : {false, true}) {
|
|
for (size_t size : allocationSizes) {
|
|
SCOPED_TRACE(::testing::Message()
|
|
<< "cpu_access_needed: " << cpu_access_needed << " size: " << size);
|
|
int fd = allocator->AllocSystem(cpu_access_needed, size);
|
|
ASSERT_GE(fd, 0);
|
|
ASSERT_EQ(close(fd), 0); // free the buffer
|
|
}
|
|
}
|
|
}
|
|
|
|
TEST_F(DmaBufHeapTest, AllocateCachedNeedsSync) {
|
|
static const size_t allocationSizes[] = {4 * 1024, 64 * 1024, 1024 * 1024, 2 * 1024 * 1024};
|
|
for (size_t size : allocationSizes) {
|
|
SCOPED_TRACE(::testing::Message()
|
|
<< "heap: " << kDmabufSystemHeapName << " size: " << size);
|
|
int fd = allocator->Alloc(kDmabufSystemHeapName, size, ION_FLAG_CACHED_NEEDS_SYNC
|
|
/* ion heap flags will be ignored if using dmabuf heaps */);
|
|
ASSERT_GE(fd, 0);
|
|
ASSERT_EQ(close(fd), 0); // free the buffer
|
|
}
|
|
}
|
|
|
|
TEST_F(DmaBufHeapTest, RepeatedAllocate) {
|
|
static const size_t allocationSizes[] = {4 * 1024, 64 * 1024, 1024 * 1024, 2 * 1024 * 1024};
|
|
for (bool cpu_access_needed : {false, true}) {
|
|
for (size_t size : allocationSizes) {
|
|
SCOPED_TRACE(::testing::Message()
|
|
<< "cpu_access_needed: " << cpu_access_needed << " size: " << size);
|
|
for (unsigned int i = 0; i < 1024; i++) {
|
|
SCOPED_TRACE(::testing::Message() << "iteration " << i);
|
|
int fd = allocator->AllocSystem(cpu_access_needed, size);
|
|
ASSERT_GE(fd, 0);
|
|
ASSERT_EQ(close(fd), 0); // free the buffer
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Make sure all heaps always return zeroed pages
|
|
*/
|
|
TEST_F(DmaBufHeapTest, Zeroed) {
|
|
static const size_t kAllocSizeInBytes = 4096;
|
|
static const size_t kNumFds = 16;
|
|
|
|
auto zeroes_ptr = std::make_unique<char[]>(kAllocSizeInBytes);
|
|
int fds[kNumFds];
|
|
int ret = 0, map_fd = -1;
|
|
for (unsigned int i = 0; i < kNumFds; i++) {
|
|
map_fd = allocator->Alloc(kDmabufSystemHeapName, kAllocSizeInBytes);
|
|
ASSERT_GE(map_fd, 0);
|
|
|
|
void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_WRITE, MAP_SHARED, map_fd, 0);
|
|
ASSERT_TRUE(ptr != MAP_FAILED);
|
|
|
|
ret = allocator->CpuSyncStart(map_fd, kSyncWrite);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
memset(ptr, 0xaa, kAllocSizeInBytes);
|
|
|
|
ret = allocator->CpuSyncEnd(map_fd, kSyncWrite);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes));
|
|
fds[i] = map_fd;
|
|
}
|
|
|
|
for (unsigned int i = 0; i < kNumFds; i++) {
|
|
ASSERT_EQ(0, close(fds[i]));
|
|
}
|
|
|
|
map_fd = allocator->Alloc(kDmabufSystemHeapName, kAllocSizeInBytes);
|
|
ASSERT_GE(map_fd, 0);
|
|
|
|
void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_READ, MAP_SHARED, map_fd, 0);
|
|
ASSERT_TRUE(ptr != MAP_FAILED);
|
|
|
|
ret = allocator->CpuSyncStart(map_fd);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ASSERT_EQ(0, memcmp(ptr, zeroes_ptr.get(), kAllocSizeInBytes));
|
|
|
|
ret = allocator->CpuSyncEnd(map_fd);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes));
|
|
ASSERT_EQ(0, close(map_fd));
|
|
}
|
|
|
|
TEST_F(DmaBufHeapTest, TestCpuSync) {
|
|
static const size_t kAllocSizeInBytes = 4096;
|
|
auto vec_sync_type = {kSyncRead, kSyncWrite, kSyncReadWrite};
|
|
for (auto sync_type : vec_sync_type) {
|
|
int map_fd = allocator->Alloc(kDmabufSystemHeapName, kAllocSizeInBytes);
|
|
ASSERT_GE(map_fd, 0);
|
|
|
|
void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0);
|
|
ASSERT_TRUE(ptr != MAP_FAILED);
|
|
|
|
int ret = allocator->CpuSyncStart(map_fd, sync_type);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ret = allocator->CpuSyncEnd(map_fd, sync_type);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes));
|
|
ASSERT_EQ(0, close(map_fd));
|
|
}
|
|
}
|
|
|
|
int CustomCpuSyncStart(int /* ion_fd */, int /* dma_buf fd */,
|
|
void* /* custom_data pointer */) {
|
|
LOG(INFO) << "In custom cpu sync start callback";
|
|
return 0;
|
|
}
|
|
|
|
int CustomCpuSyncEnd(int /* ion_fd */, int /* dma_buf fd */,
|
|
void* /* custom_data pointer */) {
|
|
LOG(INFO) << "In custom cpu sync end callback";
|
|
return 0;
|
|
}
|
|
|
|
TEST_F(DmaBufHeapTest, TestCustomLegacyIonSyncCallback) {
|
|
static const size_t allocationSizes[] = {4 * 1024, 64 * 1024, 1024 * 1024, 2 * 1024 * 1024};
|
|
for (size_t size : allocationSizes) {
|
|
SCOPED_TRACE(::testing::Message()
|
|
<< "heap: " << kDmabufSystemHeapName << " size: " << size);
|
|
|
|
int map_fd = allocator->Alloc(kDmabufSystemHeapName, size);
|
|
ASSERT_GE(map_fd, 0);
|
|
|
|
void* ptr = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0);
|
|
ASSERT_TRUE(ptr != MAP_FAILED);
|
|
|
|
int ret = allocator->CpuSyncStart(map_fd, kSyncWrite, CustomCpuSyncStart);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
memset(ptr, 0xaa, size);
|
|
|
|
ret = allocator->CpuSyncEnd(map_fd, kSyncWrite, CustomCpuSyncEnd);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ASSERT_EQ(0, munmap(ptr, size));
|
|
ASSERT_EQ(0, close(map_fd));
|
|
}
|
|
}
|
|
|
|
TEST_F(DmaBufHeapTest, TestDeviceCapabilityCheck) {
|
|
auto heap_list = allocator->GetDmabufHeapList();
|
|
|
|
ASSERT_TRUE(!heap_list.empty() || BufferAllocator::CheckIonSupport());
|
|
}
|
|
|
|
TEST_F(DmaBufHeapTest, TestDmabufSystemHeapCompliance) {
|
|
using android::vintf::KernelVersion;
|
|
|
|
if (android::base::GetIntProperty("ro.product.first_api_level", 0) < __ANDROID_API_S__) {
|
|
GTEST_SKIP();
|
|
}
|
|
|
|
KernelVersion min_kernel_version = KernelVersion(5, 10, 0);
|
|
KernelVersion kernel_version =
|
|
android::vintf::VintfObject::GetInstance()
|
|
->getRuntimeInfo(android::vintf::RuntimeInfo::FetchFlag::CPU_VERSION)
|
|
->kernelVersion();
|
|
if (kernel_version < min_kernel_version) {
|
|
GTEST_SKIP();
|
|
}
|
|
|
|
auto heap_list = allocator->GetDmabufHeapList();
|
|
ASSERT_TRUE(heap_list.find("system") != heap_list.end());
|
|
|
|
for (bool cpu_access_needed : {false, true}) {
|
|
static const size_t kAllocSizeInBytes = 4096;
|
|
/*
|
|
* Test that system heap can be allocated from.
|
|
*/
|
|
SCOPED_TRACE(::testing::Message() << "cpu_access_needed: " << cpu_access_needed);
|
|
int map_fd = allocator->AllocSystem(cpu_access_needed, kAllocSizeInBytes);
|
|
ASSERT_GE(map_fd, 0);
|
|
|
|
/*
|
|
* Test that system heap can be mmapped by the CPU.
|
|
*/
|
|
void* ptr = mmap(NULL, kAllocSizeInBytes, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0);
|
|
ASSERT_TRUE(ptr != MAP_FAILED);
|
|
|
|
/*
|
|
* Test that the allocated memory is zeroed.
|
|
*/
|
|
auto zeroes_ptr = std::make_unique<char[]>(kAllocSizeInBytes);
|
|
int ret = allocator->CpuSyncStart(map_fd);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ASSERT_EQ(0, memcmp(ptr, zeroes_ptr.get(), kAllocSizeInBytes));
|
|
|
|
ret = allocator->CpuSyncEnd(map_fd);
|
|
ASSERT_EQ(0, ret);
|
|
|
|
ASSERT_EQ(0, munmap(ptr, kAllocSizeInBytes));
|
|
ASSERT_EQ(0, close(map_fd));
|
|
}
|
|
}
|