#include "testPch.h" #include "imageTestHelper.h" #include "robuffer.h" #include "winrt/Windows.Storage.h" #include "winrt/Windows.Storage.Streams.h" #include #include #include #define FENCE_SIGNAL_VALUE 1 using namespace winrt; using namespace winml; using namespace wfc; using namespace wm; using namespace wgi; namespace ImageTestHelper { BitmapPixelFormat GetPixelFormat(const std::wstring& inputPixelFormat) { // Return corresponding BitmapPixelFormat according to input string if (L"Bgra8" == inputPixelFormat || L"Bgr8" == inputPixelFormat) { return BitmapPixelFormat::Bgra8; } else if (L"Rgba8" == inputPixelFormat || L"Rgb8" == inputPixelFormat) { return BitmapPixelFormat::Rgba8; } else if (L"Gray8" == inputPixelFormat) { return BitmapPixelFormat::Gray8; } else { throw std::invalid_argument("Unsupported pixelFormat"); } } TensorFloat LoadInputImageFromCPU(SoftwareBitmap softwareBitmap, const std::wstring& modelPixelFormat) { softwareBitmap = SoftwareBitmap::Convert(softwareBitmap, BitmapPixelFormat::Bgra8); BYTE* pData = nullptr; UINT32 size = 0; wgi::BitmapBuffer spBitmapBuffer(softwareBitmap.LockBuffer(wgi::BitmapBufferAccessMode::Read)); wf::IMemoryBufferReference reference = spBitmapBuffer.CreateReference(); auto spByteAccess = reference.as<::Windows::Foundation::IMemoryBufferByteAccess>(); spByteAccess->GetBuffer(&pData, &size); uint32_t height = softwareBitmap.PixelHeight(); uint32_t width = softwareBitmap.PixelWidth(); // TODO: Need modification for Gray8 std::vector shape = {1, 3, height, width}; float* pCPUTensor; uint32_t uCapacity; TensorFloat tf = TensorFloat::Create(shape); com_ptr itn = tf.as(); itn->GetBuffer(reinterpret_cast(&pCPUTensor), &uCapacity); if (BitmapPixelFormat::Bgra8 == GetPixelFormat(modelPixelFormat)) { // loop condition is i < size - 2 to avoid potential for extending past the memory buffer for (UINT32 i = 0; i < size - 2; i += 4) { UINT32 pixelInd = i / 4; pCPUTensor[pixelInd] = (float)pData[i]; pCPUTensor[(height * width) + pixelInd] = (float)pData[i + 1]; pCPUTensor[(height * width * 2) + pixelInd] = (float)pData[i + 2]; } } else if (BitmapPixelFormat::Rgba8 == GetPixelFormat(modelPixelFormat)) { for (UINT32 i = 0; i < size - 2; i += 4) { UINT32 pixelInd = i / 4; pCPUTensor[pixelInd] = (float)pData[i + 2]; pCPUTensor[(height * width) + pixelInd] = (float)pData[i + 1]; pCPUTensor[(height * width * 2) + pixelInd] = (float)pData[i]; } } // else if() // TODO: for Gray8 else { std::cerr << "Unsupported pixelFormat"; } return tf; } TensorFloat LoadInputImageFromGPU(SoftwareBitmap softwareBitmap, const std::wstring& modelPixelFormat) { softwareBitmap = SoftwareBitmap::Convert(softwareBitmap, BitmapPixelFormat::Bgra8); BYTE* pData = nullptr; UINT32 size = 0; BitmapBuffer spBitmapBuffer(softwareBitmap.LockBuffer(wgi::BitmapBufferAccessMode::Read)); wf::IMemoryBufferReference reference = spBitmapBuffer.CreateReference(); com_ptr<::Windows::Foundation::IMemoryBufferByteAccess> spByteAccess = reference.as<::Windows::Foundation::IMemoryBufferByteAccess>(); spByteAccess->GetBuffer(&pData, &size); std::vector shape = {1, 3, softwareBitmap.PixelHeight(), softwareBitmap.PixelWidth()}; float* pCPUTensor; uint32_t uCapacity; // CPU tensor initialization TensorFloat tf = TensorFloat::Create(shape); com_ptr itn = tf.as(); itn->GetBuffer(reinterpret_cast(&pCPUTensor), &uCapacity); uint32_t height = softwareBitmap.PixelHeight(); uint32_t width = softwareBitmap.PixelWidth(); if (BitmapPixelFormat::Bgra8 == GetPixelFormat(modelPixelFormat)) { // loop condition is i < size - 2 to avoid potential for extending past the memory buffer for (UINT32 i = 0; i < size - 2; i += 4) { UINT32 pixelInd = i / 4; pCPUTensor[pixelInd] = (float)pData[i]; pCPUTensor[(height * width) + pixelInd] = (float)pData[i + 1]; pCPUTensor[(height * width * 2) + pixelInd] = (float)pData[i + 2]; } } else if (BitmapPixelFormat::Rgba8 == GetPixelFormat(modelPixelFormat)) { for (UINT32 i = 0; i < size - 2; i += 4) { UINT32 pixelInd = i / 4; pCPUTensor[pixelInd] = (float)pData[i + 2]; pCPUTensor[(height * width) + pixelInd] = (float)pData[i + 1]; pCPUTensor[(height * width * 2) + pixelInd] = (float)pData[i]; } } // else if() // TODO: for Gray8 else { std::cerr << "unsupported pixelFormat"; } // create the d3d device. com_ptr pD3D12Device = nullptr; WINML_EXPECT_NO_THROW(D3D12CreateDevice( nullptr, D3D_FEATURE_LEVEL::D3D_FEATURE_LEVEL_11_0, __uuidof(ID3D12Device), reinterpret_cast(&pD3D12Device) )); // create the command queue. com_ptr dxQueue = nullptr; D3D12_COMMAND_QUEUE_DESC commandQueueDesc = {}; commandQueueDesc.Type = D3D12_COMMAND_LIST_TYPE_DIRECT; pD3D12Device->CreateCommandQueue(&commandQueueDesc, __uuidof(ID3D12CommandQueue), reinterpret_cast(&dxQueue)); com_ptr devicefactory = get_activation_factory(); com_ptr tensorfactory = get_activation_factory(); com_ptr<::IUnknown> spUnk; devicefactory->CreateFromD3D12CommandQueue(dxQueue.get(), spUnk.put()); // Create ID3D12GraphicsCommandList and Allocator D3D12_COMMAND_LIST_TYPE queuetype = dxQueue->GetDesc().Type; com_ptr alloctor; com_ptr cmdList; pD3D12Device->CreateCommandAllocator(queuetype, winrt::guid_of(), alloctor.put_void()); pD3D12Device->CreateCommandList( 0, queuetype, alloctor.get(), nullptr, winrt::guid_of(), cmdList.put_void() ); // Create Committed Resource // 3 is number of channels we use. R G B without alpha. UINT64 bufferbytesize = 3 * sizeof(float) * softwareBitmap.PixelWidth() * softwareBitmap.PixelHeight(); D3D12_HEAP_PROPERTIES heapProperties = { D3D12_HEAP_TYPE_DEFAULT, D3D12_CPU_PAGE_PROPERTY_UNKNOWN, D3D12_MEMORY_POOL_UNKNOWN, 0, 0 }; D3D12_RESOURCE_DESC resourceDesc = { D3D12_RESOURCE_DIMENSION_BUFFER, 0, bufferbytesize, 1, 1, 1, DXGI_FORMAT_UNKNOWN, {1, 0}, D3D12_TEXTURE_LAYOUT_ROW_MAJOR, D3D12_RESOURCE_FLAG_ALLOW_UNORDERED_ACCESS }; com_ptr pGPUResource = nullptr; com_ptr imageUploadHeap; pD3D12Device->CreateCommittedResource( &heapProperties, D3D12_HEAP_FLAG_NONE, &resourceDesc, D3D12_RESOURCE_STATE_COMMON, nullptr, __uuidof(ID3D12Resource), pGPUResource.put_void() ); // Create the GPU upload buffer. auto heap_properties = CD3DX12_HEAP_PROPERTIES(D3D12_HEAP_TYPE_UPLOAD); auto buffer_desc = CD3DX12_RESOURCE_DESC::Buffer(bufferbytesize); WINML_EXPECT_NO_THROW(pD3D12Device->CreateCommittedResource( &heap_properties, D3D12_HEAP_FLAG_NONE, &buffer_desc, D3D12_RESOURCE_STATE_GENERIC_READ, nullptr, __uuidof(ID3D12Resource), imageUploadHeap.put_void() )); // Copy from Cpu to GPU D3D12_SUBRESOURCE_DATA CPUData = {}; CPUData.pData = reinterpret_cast(pCPUTensor); CPUData.RowPitch = static_cast(bufferbytesize); CPUData.SlicePitch = static_cast(bufferbytesize); UpdateSubresources(cmdList.get(), pGPUResource.get(), imageUploadHeap.get(), 0, 0, 1, &CPUData); // Close the command list and execute it to begin the initial GPU setup. WINML_EXPECT_NO_THROW(cmdList->Close()); ID3D12CommandList* ppCommandLists[] = {cmdList.get()}; dxQueue->ExecuteCommandLists(_countof(ppCommandLists), ppCommandLists); //Create Event HANDLE directEvent = CreateEvent(nullptr, FALSE, FALSE, nullptr); wil::unique_event hDirectEvent(directEvent); //Create Fence ::Microsoft::WRL::ComPtr spDirectFence = nullptr; WINML_EXPECT_HRESULT_SUCCEEDED( pD3D12Device->CreateFence(0, D3D12_FENCE_FLAG_NONE, IID_PPV_ARGS(spDirectFence.ReleaseAndGetAddressOf())) ); //Adds fence to queue WINML_EXPECT_HRESULT_SUCCEEDED(dxQueue->Signal(spDirectFence.Get(), FENCE_SIGNAL_VALUE)); WINML_EXPECT_HRESULT_SUCCEEDED(spDirectFence->SetEventOnCompletion(FENCE_SIGNAL_VALUE, hDirectEvent.get())); //Wait for signal DWORD retVal = WaitForSingleObject(hDirectEvent.get(), INFINITE); if (retVal != WAIT_OBJECT_0) { WINML_EXPECT_HRESULT_SUCCEEDED(E_UNEXPECTED); } // GPU tensorize com_ptr<::IUnknown> spUnkTensor; TensorFloat input1imagetensor(nullptr); int64_t shapes[4] = {1, 3, softwareBitmap.PixelWidth(), softwareBitmap.PixelHeight()}; tensorfactory->CreateFromD3D12Resource(pGPUResource.get(), shapes, 4, spUnkTensor.put()); spUnkTensor.try_as(input1imagetensor); return input1imagetensor; } bool VerifyHelper(VideoFrame actual, VideoFrame expected) { // Verify two input ImageFeatureValues are identified. auto softwareBitmapActual = actual.SoftwareBitmap(); auto softwareBitmapExpected = expected.SoftwareBitmap(); WINML_EXPECT_TRUE(softwareBitmapActual.PixelHeight() == softwareBitmapExpected.PixelHeight()); WINML_EXPECT_TRUE(softwareBitmapActual.PixelWidth() == softwareBitmapExpected.PixelWidth()); WINML_EXPECT_TRUE(softwareBitmapActual.BitmapPixelFormat() == softwareBitmapExpected.BitmapPixelFormat()); uint32_t size = 4 * softwareBitmapActual.PixelHeight() * softwareBitmapActual.PixelWidth(); ws::Streams::Buffer actualOutputBuffer(size); ws::Streams::Buffer expectedOutputBuffer(size); softwareBitmapActual.CopyToBuffer(actualOutputBuffer); softwareBitmapExpected.CopyToBuffer(expectedOutputBuffer); byte* actualBytes; actualOutputBuffer.try_as<::Windows::Storage::Streams::IBufferByteAccess>()->Buffer(&actualBytes); byte* expectedBytes; expectedOutputBuffer.try_as<::Windows::Storage::Streams::IBufferByteAccess>()->Buffer(&expectedBytes); byte* pActualByte = actualBytes; byte* pExpectedByte = expectedBytes; // hard code, might need to be modified later. const float cMaxErrorRate = 0.4f; int8_t epsilon = 20; // Even given two same ImageFeatureValues, the comparison cannot exactly match. // So we use error rate. UINT errors = 0; for (uint32_t i = 0; i < size; i++, pActualByte++, pExpectedByte++) { // Only the check the first three channels, which are (B, G, R) if ((i + 1) % 4 == 0) continue; auto diff = (*pActualByte - *pExpectedByte); if (diff > epsilon) { errors++; } } std::cerr << "total errors is " << errors << "/" << size << ", errors rate is " << (float)errors / size << std::endl; return (float)errors / size < cMaxErrorRate; } } // namespace ImageTestHelper