onnxruntime/js/web
Arthur Islamov ccf14e891e
[js/web] JSEP node assignment optimization (#17128)
### Description
Since WebGPU supports only float32 and int32, having Gather, Reshape,
Shape, Squeeze and Unsqueeze ops with other data types create additional
MemCpy ops and slow down the overall execution as all other OPs with
other tensor types will be done on CPU.

Before this patch SD Unet had these numbers:
Node(s) placed on [CPUExecutionProvider]. Number of nodes: 1141
Node(s) placed on [JsExecutionProvider]. Number of nodes: 4025
memcpy tokens: 2001

After patch:
Node(s) placed on [CPUExecutionProvider]. Number of nodes: 1735
Node(s) placed on [JsExecutionProvider]. Number of nodes: 2243
memcpu tokens: 813

It also gives more than 5X performance benefit. From 12sec for one Unet
step to 2.2sec on RTX 3090 Ti, so we are almost getting to native
performance.

UPD: with latest changes from main branch and multi-threading it went
down to 1.6sec. Will try re-exporting my model to onnx with maximum
optimizations, like using MultiHeadAttention to decrease node count.
Maybe after implementing that it can go in less than 1 sec
2023-08-15 18:58:05 -07:00
..
docs [JS/WebGPU] Support Log operator (#17045) 2023-08-14 18:04:12 -07:00
lib [js/web] JSEP node assignment optimization (#17128) 2023-08-15 18:58:05 -07:00
script [js/web] enable webgpu in browser unit test (#16310) 2023-08-08 11:45:04 -07:00
test [JS/WebGPU] Support Log operator (#17045) 2023-08-14 18:04:12 -07:00
.gitignore [js/web] add target ort.webgpu.min.js (#15780) 2023-05-04 10:05:39 -07:00
.npmignore [js/web] add target ort.webgpu.min.js (#15780) 2023-05-04 10:05:39 -07:00
karma.conf.js [js/web] enable webgpu in browser unit test (#16310) 2023-08-08 11:45:04 -07:00
package-lock.json [js/common] add unit tests for onnxruntime-common (#16812) 2023-07-25 14:37:41 -07:00
package.json [js/common] add unit tests for onnxruntime-common (#16812) 2023-07-25 14:37:41 -07:00
README.md [js] enable formatter for more file types (#16888) 2023-07-28 15:46:58 -07:00
tsconfig.json [js/common] add unit tests for onnxruntime-common (#16812) 2023-07-25 14:37:41 -07:00
types.d.ts [js/web] add target ort.webgpu.min.js (#15780) 2023-05-04 10:05:39 -07:00
webpack.config.js [js] enable formatter for more file types (#16888) 2023-07-28 15:46:58 -07:00

ONNX Runtime Web

ONNX Runtime Web is a Javascript library for running ONNX models on browsers and on Node.js.

ONNX Runtime Web has adopted WebAssembly and WebGL technologies for providing an optimized ONNX model inference runtime for both CPUs and GPUs.

Why ONNX models

The Open Neural Network Exchange (ONNX) is an open standard for representing machine learning models. The biggest advantage of ONNX is that it allows interoperability across different open source AI frameworks, which itself offers more flexibility for AI frameworks adoption.

Why ONNX Runtime Web

With ONNX Runtime Web, web developers can score models directly on browsers with various benefits including reducing server-client communication and protecting user privacy, as well as offering install-free and cross-platform in-browser ML experience.

ONNX Runtime Web can run on both CPU and GPU. On CPU side, WebAssembly is adopted to execute the model at near-native speed. ONNX Runtime Web complies the native ONNX Runtime CPU engine into WebAssembly backend by using Emscripten, so it supports most functionalities native ONNX Runtime offers, including full ONNX operator coverage, multi-threading, ONNX Runtime Quantization as well as ONNX Runtime Mobile. For performance acceleration with GPUs, ONNX Runtime Web leverages WebGL, a popular standard for accessing GPU capabilities. We are keeping improving op coverage and optimizing performance in WebGL backend.

See Compatibility and Operators Supported for a list of platforms and operators ONNX Runtime Web currently supports.

Usage

Refer to ONNX Runtime JavaScript examples for samples and tutorials.

Documents

Developement

Refer to the following links for development information:

Compatibility

OS/Browser Chrome Edge Safari Electron Node.js
Windows 10 wasm, webgl wasm, webgl - wasm, webgl wasm
macOS wasm, webgl wasm, webgl wasm, webgl wasm, webgl wasm
Ubuntu LTS 18.04 wasm, webgl wasm, webgl - wasm, webgl wasm
iOS wasm, webgl wasm, webgl wasm, webgl - -
Android wasm, webgl wasm, webgl - - -

Operators

WebAssembly backend

ONNX Runtime Web currently support all operators in ai.onnx and ai.onnx.ml.

WebGL backend

ONNX Runtime Web currently supports a subset of operators in ai.onnx operator set. See webgl-operators.md for a complete, detailed list of which ONNX operators are supported by WebGL backend.

WebGPU backend

WebGPU backend is still an experimental feature. See webgpu-operators.md for a detailed list of which ONNX operators are supported by WebGPU backend.

License

License information can be found here.