mirror of
https://github.com/saymrwulf/onnxruntime.git
synced 2026-05-14 20:48:00 +00:00
### Description <!-- Describe your changes. --> Switch to setImmediate to avoid starving the Node.js event loop There should really be a true async version though, running computationally intensive things on the event loop will stop everything else from happening while it is running, e.g. a web server from answering requests. This can be done by wrapping `RunAsync` behind a [`napi::Promise`](https://github.com/nodejs/node-addon-api/blob/main/doc/promises.md) to run on the onnxruntime thread pool or [`AsyncWorker`]( https://github.com/nodejs/node-addon-api/blob/main/doc/async_worker.md) for the Node.js/libuv thread pool. ### Motivation and Context <!-- - Why is this change required? What problem does it solve? - If it fixes an open issue, please link to the issue here. --> Without this, if you run inference in a tight loop, without anything else in between that is async/deferred, `process.nextTick` will lead to starving the event loop and not letting anything else run, `setImmediate` at least lets the event loop spin between calls to `run`. See https://dev.to/ynmanware/setimmediate-settimeout-and-process-nexttick-3mfd Contributed on behalf of [Swimm](https://swimm.io/)
72 lines
2.2 KiB
TypeScript
72 lines
2.2 KiB
TypeScript
// Copyright (c) Microsoft Corporation. All rights reserved.
|
|
// Licensed under the MIT License.
|
|
|
|
import {Backend, InferenceSession, InferenceSessionHandler, SessionHandler} from 'onnxruntime-common';
|
|
|
|
import {Binding, binding} from './binding';
|
|
|
|
class OnnxruntimeSessionHandler implements InferenceSessionHandler {
|
|
#inferenceSession: Binding.InferenceSession;
|
|
|
|
constructor(pathOrBuffer: string|Uint8Array, options: InferenceSession.SessionOptions) {
|
|
this.#inferenceSession = new binding.InferenceSession();
|
|
if (typeof pathOrBuffer === 'string') {
|
|
this.#inferenceSession.loadModel(pathOrBuffer, options);
|
|
} else {
|
|
this.#inferenceSession.loadModel(pathOrBuffer.buffer, pathOrBuffer.byteOffset, pathOrBuffer.byteLength, options);
|
|
}
|
|
this.inputNames = this.#inferenceSession.inputNames;
|
|
this.outputNames = this.#inferenceSession.outputNames;
|
|
}
|
|
|
|
async dispose(): Promise<void> {
|
|
this.#inferenceSession.dispose();
|
|
}
|
|
|
|
readonly inputNames: string[];
|
|
readonly outputNames: string[];
|
|
|
|
startProfiling(): void {
|
|
// TODO: implement profiling
|
|
}
|
|
endProfiling(): void {
|
|
// TODO: implement profiling
|
|
}
|
|
|
|
async run(feeds: SessionHandler.FeedsType, fetches: SessionHandler.FetchesType, options: InferenceSession.RunOptions):
|
|
Promise<SessionHandler.ReturnType> {
|
|
return new Promise((resolve, reject) => {
|
|
setImmediate(() => {
|
|
try {
|
|
resolve(this.#inferenceSession.run(feeds, fetches, options));
|
|
} catch (e) {
|
|
// reject if any error is thrown
|
|
reject(e);
|
|
}
|
|
});
|
|
});
|
|
}
|
|
}
|
|
|
|
class OnnxruntimeBackend implements Backend {
|
|
async init(): Promise<void> {
|
|
return Promise.resolve();
|
|
}
|
|
|
|
async createInferenceSessionHandler(pathOrBuffer: string|Uint8Array, options?: InferenceSession.SessionOptions):
|
|
Promise<InferenceSessionHandler> {
|
|
return new Promise((resolve, reject) => {
|
|
setImmediate(() => {
|
|
try {
|
|
resolve(new OnnxruntimeSessionHandler(pathOrBuffer, options || {}));
|
|
} catch (e) {
|
|
// reject if any error is thrown
|
|
reject(e);
|
|
}
|
|
});
|
|
});
|
|
}
|
|
}
|
|
|
|
export const onnxruntimeBackend = new OnnxruntimeBackend();
|
|
export const listSupportedBackends = binding.listSupportedBackends;
|