mirror of
https://github.com/saymrwulf/onnxruntime.git
synced 2026-05-16 21:00:14 +00:00
# Motivation Currently, ORT minimal builds use kernel def hashes to map from nodes to kernels to execute when loading the model. As the kernel def hashes must be known ahead of time, this works for statically registered kernels. This works well for the CPU EP. For this approach to work, the kernel def hashes must also be known at ORT format model conversion time, which means the EP with statically registered kernels must also be enabled then. This is not an issue for the always-available CPU EP. However, we do not want to require that any EP which statically registers kernels is always available too. Consequently, we explore another approach to match nodes to kernels that does not rely on kernel def hashes. An added benefit of this is the possibility of moving away from kernel def hashes completely, which would eliminate the maintenance burden of keeping the hashes stable. # Approach In a full build, ORT uses some information from the ONNX op schema to match a node to a kernel. We want to avoid including the ONNX op schema in a minimal build to reduce binary size. Essentially, we take the necessary information from the ONNX op schema and make it available in a minimal build. We decouple the ONNX op schema from the kernel matching logic. The kernel matching logic instead relies on per-op information which can either be obtained from the ONNX op schema or another source. This per-op information must be available in a minimal build when there are no ONNX op schemas. We put it in the ORT format model. Existing uses of kernel def hashes to look up kernels are replaced with the updated kernel matching logic. We no longer store kernel def hashes in the ORT format model’s session state and runtime optimization representations. We no longer keep the logic to generate and ensure stability of kernel def hashes.
170 lines
5.6 KiB
Text
170 lines
5.6 KiB
Text
// Copyright (c) Microsoft Corporation. All rights reserved.
|
|
// Licensed under the MIT License.
|
|
|
|
#import "MNISTDataHandler.h"
|
|
#import "OnnxruntimeModule.h"
|
|
#import "TensorHelper.h"
|
|
#import <Foundation/Foundation.h>
|
|
#import <React/RCTLog.h>
|
|
|
|
NS_ASSUME_NONNULL_BEGIN
|
|
|
|
@implementation MNISTDataHandler
|
|
|
|
RCT_EXPORT_MODULE(MNISTDataHandler)
|
|
|
|
// It returns mode path in local device,
|
|
// so that onnxruntime is able to load a model using a given path.
|
|
RCT_EXPORT_METHOD(getLocalModelPath : (RCTPromiseResolveBlock)resolve rejecter : (RCTPromiseRejectBlock)reject) {
|
|
@try {
|
|
NSString *modelPath = [[NSBundle mainBundle] pathForResource:@"mnist" ofType:@"ort"];
|
|
NSFileManager *fileManager = [NSFileManager defaultManager];
|
|
if ([fileManager fileExistsAtPath:modelPath]) {
|
|
resolve(modelPath);
|
|
} else {
|
|
reject(@"mnist", @"no such a model", nil);
|
|
}
|
|
} @catch (NSException *exception) {
|
|
reject(@"mnist", @"no such a model", nil);
|
|
}
|
|
}
|
|
|
|
// It returns image path.
|
|
RCT_EXPORT_METHOD(getImagePath : (RCTPromiseResolveBlock)resolve reject : (RCTPromiseRejectBlock)reject) {
|
|
@try {
|
|
NSString *imagePath = [[NSBundle mainBundle] pathForResource:@"3" ofType:@"jpg"];
|
|
NSFileManager *fileManager = [NSFileManager defaultManager];
|
|
if ([fileManager fileExistsAtPath:imagePath]) {
|
|
resolve(imagePath);
|
|
} else {
|
|
reject(@"mnist", @"no such an image", nil);
|
|
}
|
|
} @catch (NSException *exception) {
|
|
reject(@"mnist", @"no such an image", nil);
|
|
}
|
|
}
|
|
|
|
// It gets raw input data, which can be uri or byte array and others,
|
|
// returns cooked data formatted as input of a model.
|
|
RCT_EXPORT_METHOD(preprocess
|
|
: (NSString *)uri resolve
|
|
: (RCTPromiseResolveBlock)resolve reject
|
|
: (RCTPromiseRejectBlock)reject) {
|
|
@try {
|
|
NSDictionary *inputDataMap = [self preprocess:uri];
|
|
resolve(inputDataMap);
|
|
} @catch (NSException *exception) {
|
|
reject(@"mnist", @"can't load an image", nil);
|
|
}
|
|
}
|
|
|
|
// It gets a result from onnxruntime and a duration of session time for input data,
|
|
// returns output data formatted as React Native map.
|
|
RCT_EXPORT_METHOD(postprocess
|
|
: (NSDictionary *)result resolve
|
|
: (RCTPromiseResolveBlock)resolve reject
|
|
: (RCTPromiseRejectBlock)reject) {
|
|
@try {
|
|
NSDictionary *cookedMap = [self postprocess:result];
|
|
resolve(cookedMap);
|
|
} @catch (NSException *exception) {
|
|
reject(@"mnist", @"can't pose-process an image", nil);
|
|
}
|
|
}
|
|
|
|
- (NSDictionary *)preprocess:(NSString *)uri {
|
|
UIImage *image = [UIImage imageNamed:@"3.jpg"];
|
|
|
|
CGSize scale = CGSizeMake(28, 28);
|
|
UIGraphicsBeginImageContextWithOptions(scale, NO, 1.0);
|
|
[image drawInRect:CGRectMake(0, 0, scale.width, scale.height)];
|
|
UIImage *scaledImage = UIGraphicsGetImageFromCurrentImageContext();
|
|
UIGraphicsEndImageContext();
|
|
|
|
CGImageRef imageRef = [scaledImage CGImage];
|
|
NSUInteger width = CGImageGetWidth(imageRef);
|
|
NSUInteger height = CGImageGetHeight(imageRef);
|
|
CGColorSpaceRef colorSpace = CGColorSpaceCreateDeviceRGB();
|
|
|
|
const NSUInteger rawDataSize = height * width * 4;
|
|
std::vector<unsigned char> rawData(rawDataSize);
|
|
NSUInteger bytesPerPixel = 4;
|
|
NSUInteger bytesPerRow = bytesPerPixel * width;
|
|
CGContextRef context = CGBitmapContextCreate(rawData.data(), width, height, 8, bytesPerRow, colorSpace,
|
|
kCGImageAlphaPremultipliedLast | kCGImageByteOrder32Big);
|
|
CGColorSpaceRelease(colorSpace);
|
|
CGContextSetBlendMode(context, kCGBlendModeCopy);
|
|
CGContextDrawImage(context, CGRectMake(0, 0, width, height), imageRef);
|
|
CGContextRelease(context);
|
|
|
|
const NSInteger dimSize = height * width;
|
|
const NSInteger byteBufferSize = dimSize * sizeof(float);
|
|
|
|
unsigned char *byteBuffer = static_cast<unsigned char *>(malloc(byteBufferSize));
|
|
NSData *byteBufferRef = [NSData dataWithBytesNoCopy:byteBuffer length:byteBufferSize];
|
|
float *floatPtr = (float *)[byteBufferRef bytes];
|
|
for (NSUInteger h = 0; h < height; ++h) {
|
|
for (NSUInteger w = 0; w < width; ++w) {
|
|
NSUInteger byteIndex = (bytesPerRow * h) + w * bytesPerPixel;
|
|
*floatPtr++ = rawData[byteIndex];
|
|
}
|
|
}
|
|
floatPtr = (float *)[byteBufferRef bytes];
|
|
|
|
NSMutableDictionary *inputDataMap = [NSMutableDictionary dictionary];
|
|
|
|
NSMutableDictionary *inputTensorMap = [NSMutableDictionary dictionary];
|
|
|
|
// dims
|
|
NSArray *dims = @[
|
|
[NSNumber numberWithInt:1],
|
|
[NSNumber numberWithInt:1],
|
|
[NSNumber numberWithInt:static_cast<int>(height)],
|
|
[NSNumber numberWithInt:static_cast<int>(width)]
|
|
];
|
|
inputTensorMap[@"dims"] = dims;
|
|
|
|
// type
|
|
inputTensorMap[@"type"] = JsTensorTypeFloat;
|
|
|
|
// encoded data
|
|
NSString *data = [byteBufferRef base64EncodedStringWithOptions:0];
|
|
inputTensorMap[@"data"] = data;
|
|
|
|
inputDataMap[@"Input3"] = inputTensorMap;
|
|
|
|
return inputDataMap;
|
|
}
|
|
|
|
- (NSDictionary *)postprocess:(NSDictionary *)result {
|
|
NSMutableString *detectionResult = [NSMutableString string];
|
|
|
|
NSDictionary *outputTensor = [result objectForKey:@"Plus214_Output_0"];
|
|
|
|
NSString *data = [outputTensor objectForKey:@"data"];
|
|
NSData *buffer = [[NSData alloc] initWithBase64EncodedString:data options:0];
|
|
float *values = (float *)[buffer bytes];
|
|
int count = (int)[buffer length] / 4;
|
|
|
|
int argmax = 0;
|
|
float maxValue = 0.0f;
|
|
for (int i = 0; i < count; ++i) {
|
|
if (values[i] > maxValue) {
|
|
maxValue = values[i];
|
|
argmax = i;
|
|
}
|
|
}
|
|
|
|
if (maxValue == 0.0f) {
|
|
detectionResult = [NSMutableString stringWithString:@"No match"];
|
|
} else {
|
|
detectionResult = [NSMutableString stringWithFormat:@"%d", argmax];
|
|
}
|
|
|
|
NSDictionary *cookedMap = @{@"result" : detectionResult};
|
|
return cookedMap;
|
|
}
|
|
|
|
@end
|
|
|
|
NS_ASSUME_NONNULL_END
|