Files
voice_recognition/whispervad/node_modules/onnxruntime-web/lib/onnxjs/session-handler-inference.ts

61 lines
1.9 KiB
TypeScript

// Copyright (c) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License.
import { InferenceSession, InferenceSessionHandler, SessionHandler, Tensor } from 'onnxruntime-common';
import { Session } from './session';
import { Tensor as OnnxjsTensor } from './tensor';
export class OnnxjsSessionHandler implements InferenceSessionHandler {
constructor(private session: Session) {
this.inputNames = this.session.inputNames;
this.outputNames = this.session.outputNames;
}
get inputMetadata(): readonly InferenceSession.ValueMetadata[] {
throw new Error('Getting model metadata is not supported in webgl backend.');
}
get outputMetadata(): readonly InferenceSession.ValueMetadata[] {
throw new Error('Getting model metadata is not supported in webgl backend.');
}
async dispose(): Promise<void> {}
inputNames: readonly string[];
outputNames: readonly string[];
async run(
feeds: SessionHandler.FeedsType,
_fetches: SessionHandler.FetchesType,
_options: InferenceSession.RunOptions,
): Promise<SessionHandler.ReturnType> {
const inputMap = new Map<string, OnnxjsTensor>();
for (const name in feeds) {
if (Object.hasOwnProperty.call(feeds, name)) {
const feed = feeds[name];
inputMap.set(
name,
new OnnxjsTensor(
feed.dims,
feed.type as OnnxjsTensor.DataType,
undefined,
undefined,
feed.data as OnnxjsTensor.NumberType,
),
);
}
}
const outputMap = await this.session.run(inputMap);
const output: SessionHandler.ReturnType = {};
outputMap.forEach((tensor, name) => {
output[name] = new Tensor(tensor.type, tensor.data, tensor.dims);
});
return output;
}
startProfiling(): void {
this.session.startProfiling();
}
endProfiling(): void {
this.session.endProfiling();
}
}