mirror of
https://github.com/saymrwulf/onnxruntime.git
synced 2026-05-17 21:10:43 +00:00
* working on re-organizing js code for ortweb * remove dup files * move folder * fix common references * fix common es5 * add webpack to common * split interfact/impl * use cjs for node * add npmignore for common * update sourcemap config for common * update node * adjust folder/path in CI and build * update folder * nit: readme * add bundle for dev * correct nodejs paths * enable ORT_API_MANUAL_INIT * set name for umd library * correct name for commonjs export * add priority into registerBackend() * fix npm ci pwd * update eslintrc * revise code * revert package-lock lockfileVersion 2->1 * update prebuild * resolve comments * update document * revise eslint config * update eslint for typescript rules * revert changes by mistake in backend.ts * add env * resolve comments
26 lines
1,003 B
TypeScript
26 lines
1,003 B
TypeScript
// Copyright (c) Microsoft Corporation. All rights reserved.
|
|
// Licensed under the MIT License.
|
|
|
|
import {InferenceSession} from 'onnxruntime-common';
|
|
import * as path from 'path';
|
|
|
|
import {Tensor} from '../../lib';
|
|
import {SQUEEZENET_INPUT0_DATA, SQUEEZENET_OUTPUT0_DATA, TEST_DATA_ROOT} from '../test-utils';
|
|
import {assertTensorEqual} from '../test-utils';
|
|
|
|
describe('E2E Tests - InferenceSession.run()', async () => {
|
|
let session: InferenceSession;
|
|
const input0 = new Tensor('float32', SQUEEZENET_INPUT0_DATA, [1, 3, 224, 224]);
|
|
const expectedOutput0 = new Tensor('float32', SQUEEZENET_OUTPUT0_DATA, [1, 1000, 1, 1]);
|
|
|
|
before(async () => {
|
|
session = await InferenceSession.create(path.join(TEST_DATA_ROOT, 'squeezenet.onnx'));
|
|
});
|
|
|
|
it('multiple run() calls', async () => {
|
|
for (let i = 0; i < 1000; i++) {
|
|
const result = await session!.run({'data_0': input0}, ['softmaxout_1']);
|
|
assertTensorEqual(result.softmaxout_1, expectedOutput0);
|
|
}
|
|
}).timeout('120s');
|
|
});
|