Skip to content

Commit

Permalink
EXSWHTEC-70 - Reimplement tests for hipMemPrefetchAsync (#17)
Browse files Browse the repository at this point in the history
- Basic positive tests
- Negative parameter tests
  • Loading branch information
music-dino authored Jan 17, 2023
1 parent f5eb0da commit 47d9bf2
Showing 1 changed file with 131 additions and 89 deletions.
220 changes: 131 additions & 89 deletions catch/unit/memory/hipMemPrefetchAsync.cc
Original file line number Diff line number Diff line change
@@ -1,13 +1,15 @@
/*
Copyright (c) 2021 Advanced Micro Devices, Inc. All rights reserved.
Copyright (c) 2022 Advanced Micro Devices, Inc. All rights reserved.
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
Expand All @@ -17,108 +19,148 @@ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.
*/

#include <vector>

#include <hip_test_common.hh>
// Kernel function
__global__ void MemPrftchAsyncKernel(int* C_d, const int* A_d, size_t N) {
#include <hip/hip_runtime_api.h>
#include <utils.hh>
#include <resource_guards.hh>

std::vector<int> GetDevicesWithPrefetchSupport() {
const auto device_count = HipTest::getDeviceCount();
std::vector<int> supported_devices;
supported_devices.reserve(device_count + 1);
for (int i = 0; i < device_count; ++i) {
if (DeviceAttributesSupport(i, hipDeviceAttributeManagedMemory,
hipDeviceAttributeConcurrentManagedAccess)) {
supported_devices.push_back(i);
}
}
return supported_devices;
}

__global__ void MemPrefetchAsyncKernel(int* C_d, const int* A_d, size_t N) {
size_t offset = (blockIdx.x * blockDim.x + threadIdx.x);
size_t stride = blockDim.x * gridDim.x;
for (size_t i = offset; i < N; i += stride) {
C_d[i] = A_d[i] * A_d[i];
}
}

TEST_CASE("Unit_hipMemPrefetchAsync_Basic") {
const auto supported_devices = GetDevicesWithPrefetchSupport();
if (supported_devices.empty()) {
HipTest::HIP_SKIP_TEST("Test need at least one device with managed memory support");
}

static int HmmAttrPrint() {
int managed = 0;
INFO("The following are the attribute values related to HMM for"
" device 0:\n");
HIP_CHECK(hipDeviceGetAttribute(&managed,
hipDeviceAttributeDirectManagedMemAccessFromHost, 0));
INFO("hipDeviceAttributeDirectManagedMemAccessFromHost: " << managed);
HIP_CHECK(hipDeviceGetAttribute(&managed,
hipDeviceAttributeConcurrentManagedAccess, 0));
INFO("hipDeviceAttributeConcurrentManagedAccess: " << managed);
HIP_CHECK(hipDeviceGetAttribute(&managed,
hipDeviceAttributePageableMemoryAccess, 0));
INFO("hipDeviceAttributePageableMemoryAccess: " << managed);
HIP_CHECK(hipDeviceGetAttribute(&managed,
hipDeviceAttributePageableMemoryAccessUsesHostPageTables, 0));
INFO("hipDeviceAttributePageableMemoryAccessUsesHostPageTables:"
<< managed);

HIP_CHECK(hipDeviceGetAttribute(&managed, hipDeviceAttributeManagedMemory,
0));
INFO("hipDeviceAttributeManagedMemory: " << managed);
return managed;
LinearAllocGuard<int> alloc1(LinearAllocs::hipMallocManaged, kPageSize);
const auto count = kPageSize / sizeof(*alloc1.ptr());
constexpr auto fill_value = 42;
std::fill_n(alloc1.ptr(), count, fill_value);

for (const auto device : supported_devices) {
HIP_CHECK(hipSetDevice(device));
LinearAllocGuard<int> alloc2(LinearAllocs::hipMallocManaged, kPageSize);
StreamGuard sg(Streams::created);
HIP_CHECK(hipMemPrefetchAsync(alloc1.ptr(), kPageSize, device, sg.stream()));
MemPrefetchAsyncKernel<<<count / 1024 + 1, 1024, 0, sg.stream()>>>(alloc2.ptr(), alloc1.ptr(),
count);
HIP_CHECK(hipGetLastError());
HIP_CHECK(hipStreamSynchronize(sg.stream()));
ArrayFindIfNot(alloc1.ptr(), fill_value, count);
ArrayFindIfNot(alloc2.ptr(), fill_value * fill_value, count);
}

HIP_CHECK(hipMemPrefetchAsync(alloc1.ptr(), kPageSize, hipCpuDeviceId));
HIP_CHECK(hipStreamSynchronize(nullptr));
ArrayFindIfNot(alloc1.ptr(), fill_value, count);
}

/*
Test Description: This test prefetches the memory to each of the available
devices and launch kernel followed by result verification
At the end the memory is prefetched to Host and kernel is launched followed
by result verification.
*/
TEST_CASE("Unit_hipMemPrefetchAsync_Sync_Behavior") {
const auto supported_devices = GetDevicesWithPrefetchSupport();
if (supported_devices.empty()) {
HipTest::HIP_SKIP_TEST("Test need at least one device with managed memory support");
}
const auto device = supported_devices.front();
const auto stream_type = GENERATE(Streams::nullstream, Streams::perThread, Streams::created);

TEST_CASE("Unit_hipMemPrefetchAsync") {
int MangdMem = HmmAttrPrint();
if (MangdMem == 1) {
bool IfTestPassed = true;
int A_CONST = 123, MEM_SIZE = (8192 * sizeof(int));
int *devPtr1 = NULL, *devPtr2 = NULL, NumDevs = 0, flag = 0;
hipStream_t strm;
HIP_CHECK(hipMallocManaged(&devPtr1, MEM_SIZE));
HIP_CHECK(hipMallocManaged(&devPtr2, MEM_SIZE));
HIP_CHECK(hipGetDeviceCount(&NumDevs));
// Initializing the memory
for (uint32_t k = 0; k < (MEM_SIZE/sizeof(int)); ++k) {
devPtr1[k] = A_CONST;
devPtr2[k] = 0;
}
StreamGuard sg(stream_type);
LinearAllocGuard<void> alloc(LinearAllocs::hipMallocManaged, kPageSize);
LaunchDelayKernel(std::chrono::milliseconds{100}, sg.stream());
HIP_CHECK(hipMemPrefetchAsync(alloc.ptr(), kPageSize, device, sg.stream()));
HIP_CHECK_ERROR(hipStreamQuery(sg.stream()), hipErrorNotReady);
HIP_CHECK(hipStreamSynchronize(sg.stream()));
}

TEST_CASE("Unit_hipMemPrefetchAsync_Rounding_Behavior") {
auto supported_devices = GetDevicesWithPrefetchSupport();
if (supported_devices.empty()) {
HipTest::HIP_SKIP_TEST("Test need at least one device with managed memory support");
}
const auto device = supported_devices.front();
LinearAllocGuard<uint8_t> alloc(LinearAllocs::hipMallocManaged, 3 * kPageSize);
REQUIRE_FALSE(reinterpret_cast<intptr_t>(alloc.ptr()) % kPageSize);
const auto [offset, width] =
GENERATE_COPY(std::make_pair(kPageSize / 4, kPageSize / 2), // Withing page
std::make_pair(kPageSize / 2, kPageSize), // Across page border
std::make_pair(kPageSize / 2, kPageSize * 2)); // Across two page borders
HIP_CHECK(hipMemPrefetchAsync(alloc.ptr() + offset, width, device));
HIP_CHECK(hipStreamSynchronize(nullptr));
constexpr auto RoundDown = [](const intptr_t a, const intptr_t n) { return a - a % n; };
constexpr auto RoundUp = [RoundDown](const intptr_t a, const intptr_t n) {
return RoundDown(a + n - 1, n);
};
const auto base = alloc.ptr();
const auto rounded_up = RoundUp(offset + width, kPageSize);
unsigned int attribute = 0;
HIP_CHECK(hipMemRangeGetAttribute(&attribute, sizeof(attribute),
hipMemRangeAttributeLastPrefetchLocation,
reinterpret_cast<void*>(base), rounded_up));
REQUIRE(device == attribute);
HIP_CHECK(hipMemRangeGetAttribute(&attribute, sizeof(attribute),
hipMemRangeAttributeLastPrefetchLocation, alloc.ptr(),
3 * kPageSize));
REQUIRE((rounded_up == 3 * kPageSize ? device : hipInvalidDeviceId) == attribute);
}

for (int i = 0; i < NumDevs; ++i) {
HIP_CHECK(hipSetDevice(i));
HIP_CHECK(hipStreamCreate(&strm));
HIP_CHECK(hipMemPrefetchAsync(devPtr1, MEM_SIZE, i, strm));
HIP_CHECK(hipStreamSynchronize(strm));
MemPrftchAsyncKernel<<<32, (MEM_SIZE/sizeof(int)/32)>>>(devPtr2, devPtr1,
MEM_SIZE/sizeof(int));
for (uint32_t m = 0; m < (MEM_SIZE/sizeof(int)); ++m) {
if (devPtr1[m] != (A_CONST * A_CONST)) {
flag = 1;
}
}
HIP_CHECK(hipStreamDestroy(strm));
if (!flag) {
INFO("Test failed for device: " << i);
IfTestPassed = false;
flag = 0;
}
}
// The memory will be prefetched from last gpu in the system to the host
// memory and kernel is launched followed by result verification.
HIP_CHECK(hipStreamCreate(&strm));
HIP_CHECK(hipMemPrefetchAsync(devPtr1, MEM_SIZE, hipCpuDeviceId, strm));
HIP_CHECK(hipStreamSynchronize(strm));
MemPrftchAsyncKernel<<<32, (MEM_SIZE/sizeof(int)/32)>>>(devPtr2, devPtr1,
MEM_SIZE/sizeof(int));
for (uint32_t m = 0; m < (MEM_SIZE/sizeof(int)); ++m) {
if (devPtr1[m] != (A_CONST * A_CONST)) {
flag = 1;
}
}
HIP_CHECK(hipStreamDestroy(strm));
if (!flag) {
INFO("Failed to prefetch the memory to System space.\n");
IfTestPassed = false;
flag = 0;
}
TEST_CASE("Unit_hipMemPrefetchAsync_Negative_Parameters") {
auto supported_devices = GetDevicesWithPrefetchSupport();
if (supported_devices.empty()) {
HipTest::HIP_SKIP_TEST("Test need at least one device with managed memory support");
}
supported_devices.push_back(hipCpuDeviceId);
const auto device = GENERATE_COPY(from_range(supported_devices));

LinearAllocGuard<void> alloc(LinearAllocs::hipMallocManaged, kPageSize);
SECTION("dev_ptr == nullptr") {
HIP_CHECK_ERROR(hipMemPrefetchAsync(nullptr, kPageSize, device), hipErrorInvalidValue);
}

#if HT_NVIDIA
SECTION("dev_ptr points to non-managed memory") {
LinearAllocGuard<void> alloc(LinearAllocs::hipMalloc, kPageSize);
HIP_CHECK_ERROR(hipMemPrefetchAsync(alloc.ptr(), kPageSize, device), hipErrorInvalidValue);
}
#endif

SECTION("count == 0") {
HIP_CHECK_ERROR(hipMemPrefetchAsync(alloc.ptr(), 0, device), hipErrorInvalidValue);
}

SECTION("count larger than allocation size") {
HIP_CHECK_ERROR(hipMemPrefetchAsync(alloc.ptr(), kPageSize + 1, device), hipErrorInvalidValue);
}

SECTION("Invalid device") {
HIP_CHECK_ERROR(hipMemPrefetchAsync(alloc.ptr(), kPageSize, hipInvalidDeviceId),
hipErrorInvalidDevice);
}

HIP_CHECK(hipFree(devPtr1));
HIP_CHECK(hipFree(devPtr2));
REQUIRE(IfTestPassed);
} else {
SUCCEED("GPU 0 doesn't support hipDeviceAttributeManagedMemory "
"attribute. Hence skipping the testing with Pass result.\n");
SECTION("Invalid stream") {
hipStream_t stream;
HIP_CHECK(hipStreamCreate(&stream));
HIP_CHECK(hipStreamDestroy(stream));
HIP_CHECK_ERROR(hipMemPrefetchAsync(alloc.ptr(), kPageSize, device, stream),
hipErrorContextIsDestroyed);
}
}

0 comments on commit 47d9bf2

Please sign in to comment.