onnxruntime/winml/test/image/imageTestHelper.cpp
Justin Chu 416dc2e84d
Fix clang-format comment indents on Windows for winml/ (#17144)
On Windows, clang-format has a bug when AlignTrailingComments.Kind is
set to `Leave`
(https://clang.llvm.org/docs/ClangFormatStyleOptions.html#aligntrailingcomments),
where it will keep adding indentation to comments after each formatting
runs.

This PR changes to always align comments so we do not hit the bug.

As a consequence of the options change we need to reformat some of the
files. Note that this option is aligned with the rest of the repository.
2023-08-14 23:50:14 -04:00

274 lines
11 KiB
C++

#include "testPch.h"
#include "imageTestHelper.h"
#include "robuffer.h"
#include "winrt/Windows.Storage.h"
#include "winrt/Windows.Storage.Streams.h"
#include <d3dx12.h>
#include <MemoryBuffer.h>
#include <wil\Resource.h>
#define FENCE_SIGNAL_VALUE 1
using namespace winrt;
using namespace winml;
using namespace wfc;
using namespace wm;
using namespace wgi;
namespace ImageTestHelper {
BitmapPixelFormat GetPixelFormat(const std::wstring& inputPixelFormat) {
// Return corresponding BitmapPixelFormat according to input string
if (L"Bgra8" == inputPixelFormat || L"Bgr8" == inputPixelFormat) {
return BitmapPixelFormat::Bgra8;
} else if (L"Rgba8" == inputPixelFormat || L"Rgb8" == inputPixelFormat) {
return BitmapPixelFormat::Rgba8;
} else if (L"Gray8" == inputPixelFormat) {
return BitmapPixelFormat::Gray8;
} else {
throw std::invalid_argument("Unsupported pixelFormat");
}
}
TensorFloat LoadInputImageFromCPU(SoftwareBitmap softwareBitmap, const std::wstring& modelPixelFormat) {
softwareBitmap = SoftwareBitmap::Convert(softwareBitmap, BitmapPixelFormat::Bgra8);
BYTE* pData = nullptr;
UINT32 size = 0;
wgi::BitmapBuffer spBitmapBuffer(softwareBitmap.LockBuffer(wgi::BitmapBufferAccessMode::Read));
wf::IMemoryBufferReference reference = spBitmapBuffer.CreateReference();
auto spByteAccess = reference.as<::Windows::Foundation::IMemoryBufferByteAccess>();
spByteAccess->GetBuffer(&pData, &size);
uint32_t height = softwareBitmap.PixelHeight();
uint32_t width = softwareBitmap.PixelWidth();
// TODO: Need modification for Gray8
std::vector<int64_t> shape = {1, 3, height, width};
float* pCPUTensor;
uint32_t uCapacity;
TensorFloat tf = TensorFloat::Create(shape);
com_ptr<ITensorNative> itn = tf.as<ITensorNative>();
itn->GetBuffer(reinterpret_cast<BYTE**>(&pCPUTensor), &uCapacity);
if (BitmapPixelFormat::Bgra8 == GetPixelFormat(modelPixelFormat)) {
// loop condition is i < size - 2 to avoid potential for extending past the memory buffer
for (UINT32 i = 0; i < size - 2; i += 4) {
UINT32 pixelInd = i / 4;
pCPUTensor[pixelInd] = (float)pData[i];
pCPUTensor[(height * width) + pixelInd] = (float)pData[i + 1];
pCPUTensor[(height * width * 2) + pixelInd] = (float)pData[i + 2];
}
} else if (BitmapPixelFormat::Rgba8 == GetPixelFormat(modelPixelFormat)) {
for (UINT32 i = 0; i < size - 2; i += 4) {
UINT32 pixelInd = i / 4;
pCPUTensor[pixelInd] = (float)pData[i + 2];
pCPUTensor[(height * width) + pixelInd] = (float)pData[i + 1];
pCPUTensor[(height * width * 2) + pixelInd] = (float)pData[i];
}
}
// else if()
// TODO: for Gray8
else {
std::cerr << "Unsupported pixelFormat";
}
return tf;
}
TensorFloat LoadInputImageFromGPU(SoftwareBitmap softwareBitmap, const std::wstring& modelPixelFormat) {
softwareBitmap = SoftwareBitmap::Convert(softwareBitmap, BitmapPixelFormat::Bgra8);
BYTE* pData = nullptr;
UINT32 size = 0;
BitmapBuffer spBitmapBuffer(softwareBitmap.LockBuffer(wgi::BitmapBufferAccessMode::Read));
wf::IMemoryBufferReference reference = spBitmapBuffer.CreateReference();
com_ptr<::Windows::Foundation::IMemoryBufferByteAccess> spByteAccess =
reference.as<::Windows::Foundation::IMemoryBufferByteAccess>();
spByteAccess->GetBuffer(&pData, &size);
std::vector<int64_t> shape = {1, 3, softwareBitmap.PixelHeight(), softwareBitmap.PixelWidth()};
float* pCPUTensor;
uint32_t uCapacity;
// CPU tensor initialization
TensorFloat tf = TensorFloat::Create(shape);
com_ptr<ITensorNative> itn = tf.as<ITensorNative>();
itn->GetBuffer(reinterpret_cast<BYTE**>(&pCPUTensor), &uCapacity);
uint32_t height = softwareBitmap.PixelHeight();
uint32_t width = softwareBitmap.PixelWidth();
if (BitmapPixelFormat::Bgra8 == GetPixelFormat(modelPixelFormat)) {
// loop condition is i < size - 2 to avoid potential for extending past the memory buffer
for (UINT32 i = 0; i < size - 2; i += 4) {
UINT32 pixelInd = i / 4;
pCPUTensor[pixelInd] = (float)pData[i];
pCPUTensor[(height * width) + pixelInd] = (float)pData[i + 1];
pCPUTensor[(height * width * 2) + pixelInd] = (float)pData[i + 2];
}
} else if (BitmapPixelFormat::Rgba8 == GetPixelFormat(modelPixelFormat)) {
for (UINT32 i = 0; i < size - 2; i += 4) {
UINT32 pixelInd = i / 4;
pCPUTensor[pixelInd] = (float)pData[i + 2];
pCPUTensor[(height * width) + pixelInd] = (float)pData[i + 1];
pCPUTensor[(height * width * 2) + pixelInd] = (float)pData[i];
}
}
// else if()
// TODO: for Gray8
else {
std::cerr << "unsupported pixelFormat";
}
// create the d3d device.
com_ptr<ID3D12Device> pD3D12Device = nullptr;
WINML_EXPECT_NO_THROW(D3D12CreateDevice(
nullptr, D3D_FEATURE_LEVEL::D3D_FEATURE_LEVEL_11_0, __uuidof(ID3D12Device), reinterpret_cast<void**>(&pD3D12Device)
));
// create the command queue.
com_ptr<ID3D12CommandQueue> dxQueue = nullptr;
D3D12_COMMAND_QUEUE_DESC commandQueueDesc = {};
commandQueueDesc.Type = D3D12_COMMAND_LIST_TYPE_DIRECT;
pD3D12Device->CreateCommandQueue(&commandQueueDesc, __uuidof(ID3D12CommandQueue), reinterpret_cast<void**>(&dxQueue));
com_ptr<ILearningModelDeviceFactoryNative> devicefactory =
get_activation_factory<LearningModelDevice, ILearningModelDeviceFactoryNative>();
com_ptr<ITensorStaticsNative> tensorfactory = get_activation_factory<TensorFloat, ITensorStaticsNative>();
com_ptr<::IUnknown> spUnk;
devicefactory->CreateFromD3D12CommandQueue(dxQueue.get(), spUnk.put());
// Create ID3D12GraphicsCommandList and Allocator
D3D12_COMMAND_LIST_TYPE queuetype = dxQueue->GetDesc().Type;
com_ptr<ID3D12CommandAllocator> alloctor;
com_ptr<ID3D12GraphicsCommandList> cmdList;
pD3D12Device->CreateCommandAllocator(queuetype, winrt::guid_of<ID3D12CommandAllocator>(), alloctor.put_void());
pD3D12Device->CreateCommandList(
0, queuetype, alloctor.get(), nullptr, winrt::guid_of<ID3D12CommandList>(), cmdList.put_void()
);
// Create Committed Resource
// 3 is number of channels we use. R G B without alpha.
UINT64 bufferbytesize = 3 * sizeof(float) * softwareBitmap.PixelWidth() * softwareBitmap.PixelHeight();
D3D12_HEAP_PROPERTIES heapProperties = {
D3D12_HEAP_TYPE_DEFAULT, D3D12_CPU_PAGE_PROPERTY_UNKNOWN, D3D12_MEMORY_POOL_UNKNOWN, 0, 0};
D3D12_RESOURCE_DESC resourceDesc = {
D3D12_RESOURCE_DIMENSION_BUFFER,
0,
bufferbytesize,
1,
1,
1,
DXGI_FORMAT_UNKNOWN,
{1, 0},
D3D12_TEXTURE_LAYOUT_ROW_MAJOR,
D3D12_RESOURCE_FLAG_ALLOW_UNORDERED_ACCESS
};
com_ptr<ID3D12Resource> pGPUResource = nullptr;
com_ptr<ID3D12Resource> imageUploadHeap;
pD3D12Device->CreateCommittedResource(
&heapProperties,
D3D12_HEAP_FLAG_NONE,
&resourceDesc,
D3D12_RESOURCE_STATE_COMMON,
nullptr,
__uuidof(ID3D12Resource),
pGPUResource.put_void()
);
// Create the GPU upload buffer.
auto heap_properties = CD3DX12_HEAP_PROPERTIES(D3D12_HEAP_TYPE_UPLOAD);
auto buffer_desc = CD3DX12_RESOURCE_DESC::Buffer(bufferbytesize);
WINML_EXPECT_NO_THROW(pD3D12Device->CreateCommittedResource(
&heap_properties,
D3D12_HEAP_FLAG_NONE,
&buffer_desc,
D3D12_RESOURCE_STATE_GENERIC_READ,
nullptr,
__uuidof(ID3D12Resource),
imageUploadHeap.put_void()
));
// Copy from Cpu to GPU
D3D12_SUBRESOURCE_DATA CPUData = {};
CPUData.pData = reinterpret_cast<BYTE*>(pCPUTensor);
CPUData.RowPitch = static_cast<LONG_PTR>(bufferbytesize);
CPUData.SlicePitch = static_cast<LONG_PTR>(bufferbytesize);
UpdateSubresources(cmdList.get(), pGPUResource.get(), imageUploadHeap.get(), 0, 0, 1, &CPUData);
// Close the command list and execute it to begin the initial GPU setup.
WINML_EXPECT_NO_THROW(cmdList->Close());
ID3D12CommandList* ppCommandLists[] = {cmdList.get()};
dxQueue->ExecuteCommandLists(_countof(ppCommandLists), ppCommandLists);
//Create Event
HANDLE directEvent = CreateEvent(nullptr, FALSE, FALSE, nullptr);
wil::unique_event hDirectEvent(directEvent);
//Create Fence
::Microsoft::WRL::ComPtr<ID3D12Fence> spDirectFence = nullptr;
WINML_EXPECT_HRESULT_SUCCEEDED(
pD3D12Device->CreateFence(0, D3D12_FENCE_FLAG_NONE, IID_PPV_ARGS(spDirectFence.ReleaseAndGetAddressOf()))
);
//Adds fence to queue
WINML_EXPECT_HRESULT_SUCCEEDED(dxQueue->Signal(spDirectFence.Get(), FENCE_SIGNAL_VALUE));
WINML_EXPECT_HRESULT_SUCCEEDED(spDirectFence->SetEventOnCompletion(FENCE_SIGNAL_VALUE, hDirectEvent.get()));
//Wait for signal
DWORD retVal = WaitForSingleObject(hDirectEvent.get(), INFINITE);
if (retVal != WAIT_OBJECT_0) {
WINML_EXPECT_HRESULT_SUCCEEDED(E_UNEXPECTED);
}
// GPU tensorize
com_ptr<::IUnknown> spUnkTensor;
TensorFloat input1imagetensor(nullptr);
int64_t shapes[4] = {1, 3, softwareBitmap.PixelWidth(), softwareBitmap.PixelHeight()};
tensorfactory->CreateFromD3D12Resource(pGPUResource.get(), shapes, 4, spUnkTensor.put());
spUnkTensor.try_as(input1imagetensor);
return input1imagetensor;
}
bool VerifyHelper(VideoFrame actual, VideoFrame expected) {
// Verify two input ImageFeatureValues are identified.
auto softwareBitmapActual = actual.SoftwareBitmap();
auto softwareBitmapExpected = expected.SoftwareBitmap();
WINML_EXPECT_TRUE(softwareBitmapActual.PixelHeight() == softwareBitmapExpected.PixelHeight());
WINML_EXPECT_TRUE(softwareBitmapActual.PixelWidth() == softwareBitmapExpected.PixelWidth());
WINML_EXPECT_TRUE(softwareBitmapActual.BitmapPixelFormat() == softwareBitmapExpected.BitmapPixelFormat());
uint32_t size = 4 * softwareBitmapActual.PixelHeight() * softwareBitmapActual.PixelWidth();
ws::Streams::Buffer actualOutputBuffer(size);
ws::Streams::Buffer expectedOutputBuffer(size);
softwareBitmapActual.CopyToBuffer(actualOutputBuffer);
softwareBitmapExpected.CopyToBuffer(expectedOutputBuffer);
byte* actualBytes;
actualOutputBuffer.try_as<::Windows::Storage::Streams::IBufferByteAccess>()->Buffer(&actualBytes);
byte* expectedBytes;
expectedOutputBuffer.try_as<::Windows::Storage::Streams::IBufferByteAccess>()->Buffer(&expectedBytes);
byte* pActualByte = actualBytes;
byte* pExpectedByte = expectedBytes;
// hard code, might need to be modified later.
const float cMaxErrorRate = 0.4f;
int8_t epsilon = 20;
// Even given two same ImageFeatureValues, the comparison cannot exactly match.
// So we use error rate.
UINT errors = 0;
for (uint32_t i = 0; i < size; i++, pActualByte++, pExpectedByte++) {
// Only the check the first three channels, which are (B, G, R)
if ((i + 1) % 4 == 0)
continue;
auto diff = (*pActualByte - *pExpectedByte);
if (diff > epsilon) {
errors++;
}
}
std::cerr << "total errors is " << errors << "/" << size << ", errors rate is " << (float)errors / size << std::endl;
return (float)errors / size < cMaxErrorRate;
}
} // namespace ImageTestHelper