"use strict"; // Copyright (c) Microsoft Corporation. All rights reserved. // Licensed under the MIT License. var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) { if (k2 === undefined) k2 = k; var desc = Object.getOwnPropertyDescriptor(m, k); if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) { desc = { enumerable: true, get: function() { return m[k]; } }; } Object.defineProperty(o, k2, desc); }) : (function(o, m, k, k2) { if (k2 === undefined) k2 = k; o[k2] = m[k]; })); var __exportStar = (this && this.__exportStar) || function(m, exports) { for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p); }; Object.defineProperty(exports, "__esModule", { value: true }); /** * # ONNX Runtime JavaScript API * * ONNX Runtime JavaScript API is a unified API for all JavaScript usages, including the following NPM packages: * * - [onnxruntime-node](https://www.npmjs.com/package/onnxruntime-node) * - [onnxruntime-web](https://www.npmjs.com/package/onnxruntime-web) * - [onnxruntime-react-native](https://www.npmjs.com/package/onnxruntime-react-native) * * See also: * - [Get Started](https://onnxruntime.ai/docs/get-started/with-javascript/) * - [Inference examples](https://github.com/microsoft/onnxruntime-inference-examples/tree/main/js) * * @packageDocumentation */ __exportStar(require("./backend.js"), exports); __exportStar(require("./env.js"), exports); __exportStar(require("./inference-session.js"), exports); __exportStar(require("./tensor.js"), exports); __exportStar(require("./tensor-conversion.js"), exports); __exportStar(require("./tensor-factory.js"), exports); __exportStar(require("./trace.js"), exports); __exportStar(require("./onnx-model.js"), exports); __exportStar(require("./onnx-value.js"), exports); //# sourceMappingURL=index.js.map