onnxruntime/csharp/sample/InferenceSample/Microsoft.ML.OnnxRuntime.InferenceSample/InferenceSampleApi.cs
Dmitri Smirnov bd4d011142
[C#] Rename unreleased API, add utilities (#16806)
### Description
1. rename OrtValue.FillStringTensorElement to StringTensorSetElementAt .
To the API user I think we're conceptually setting the string at an
offset in the tensor with is roughly equivalent to `List<string> list
... list[index] = "value"`.
2. While working on new inference examples, I noticed that I am still
inclined to use `DenseTensor` for N-D indexing. Added `GetStrides()` and
`GetIndex()` from strides for long dims, so the user can obtain strides
and translate N-D indices into a flat index to operate directly on the
native `OrtValue` buffers. Expose these functions to the user.
3. Make sure we generate docs for C# public static  functions.
2023-08-02 10:06:42 -07:00

145 lines
5.5 KiB
C#

using Microsoft.ML.OnnxRuntime.Tensors;
using System;
using System.Collections.Generic;
using System.Diagnostics;
using System.IO;
namespace Microsoft.ML.OnnxRuntime.InferenceSample
{
public class InferenceSampleApi : IDisposable
{
public InferenceSampleApi()
{
_model = LoadModelFromEmbeddedResource("TestData.squeezenet.onnx");
// this is the data for only one input tensor for this model
var inputData = LoadTensorFromEmbeddedResource("TestData.bench.in");
// create default session with default session options
// Creating an InferenceSession and loading the model is an expensive operation, so generally you would
// do this once. InferenceSession.Run can be called multiple times, and concurrently.
CreateInferenceSession();
// setup sample input data
var inputMeta = _inferenceSession.InputMetadata;
_inputData = new List<OrtValue>(inputMeta.Count);
_orderedInputNames = new List<string>(inputMeta.Count);
foreach (var name in inputMeta.Keys)
{
// We create an OrtValue in this case over the buffer of potentially different shapes.
// It is Okay as long as the specified shape does not exceed the actual length of the buffer
var shape = Array.ConvertAll<int, long>(inputMeta[name].Dimensions, Convert.ToInt64);
Debug.Assert(ShapeUtils.GetSizeForShape(shape) <= inputData.LongLength);
var ortValue = OrtValue.CreateTensorValueFromMemory(inputData, shape);
_inputData.Add(ortValue);
_orderedInputNames.Add(name);
}
}
public void CreateInferenceSession(SessionOptions options = null)
{
// Optional : Create session options and set any relevant values.
// If an additional execution provider is needed it should be added to the SessionOptions prior to
// creating the InferenceSession. The CPU Execution Provider is always added by default.
if (options == null)
{
options = new SessionOptions { LogId = "Sample" };
}
_inferenceSession = new InferenceSession(_model, options);
}
public void Execute()
{
// Run the inference
// 'results' is an IDisposableReadOnlyCollection<OrtValue> container
using (var results = _inferenceSession.Run(null, _orderedInputNames, _inputData, _inferenceSession.OutputNames))
{
// dump the results
for (int i = 0; i < results.Count; ++i)
{
var name = _inferenceSession.OutputNames[i];
Console.WriteLine("Output for {0}", name);
// We can now access the native buffer directly from the OrtValue, no copy is involved.
// Spans are structs and are stack allocated. They do not add any GC pressure.
ReadOnlySpan<float> span = results[i].GetTensorDataAsSpan<float>();
Console.Write($"Input {i} results:");
for(int k = 0; k < span.Length; ++k)
{
Console.Write($" {span[k]}");
}
Console.WriteLine();
}
}
}
protected virtual void Dispose(bool disposing)
{
if (disposing && !_disposed)
{
_inferenceSession?.Dispose();
_inferenceSession = null;
if (_inputData != null)
foreach(var v in _inputData)
{
v?.Dispose();
}
_disposed = true;
}
}
public void Dispose()
{
Dispose(true);
GC.SuppressFinalize(this);
}
static float[] LoadTensorFromEmbeddedResource(string path)
{
var tensorData = new List<float>();
var assembly = typeof(InferenceSampleApi).Assembly;
using (var inputFile =
new StreamReader(assembly.GetManifestResourceStream($"{assembly.GetName().Name}.{path}")))
{
inputFile.ReadLine(); // skip the input name
string[] dataStr = inputFile.ReadLine().Split(new char[] { ',', '[', ']' },
StringSplitOptions.RemoveEmptyEntries);
for (int i = 0; i < dataStr.Length; i++)
{
tensorData.Add(Single.Parse(dataStr[i]));
}
}
return tensorData.ToArray();
}
static byte[] LoadModelFromEmbeddedResource(string path)
{
var assembly = typeof(InferenceSampleApi).Assembly;
byte[] model = null;
using (Stream stream = assembly.GetManifestResourceStream($"{assembly.GetName().Name}.{path}"))
{
using (MemoryStream memoryStream = new MemoryStream())
{
stream.CopyTo(memoryStream);
model = memoryStream.ToArray();
}
}
return model;
}
private bool _disposed = false;
private readonly byte[] _model;
private readonly List<string> _orderedInputNames;
private readonly List<OrtValue> _inputData;
private InferenceSession _inferenceSession;
}
}