Skip to content

Commit

Permalink
[js/webgpu] Add unifroms support to concat op
Browse files Browse the repository at this point in the history
  • Loading branch information
axinging committed Nov 3, 2023
1 parent d8d7952 commit 0ecdc6c
Showing 1 changed file with 34 additions and 12 deletions.
46 changes: 34 additions & 12 deletions js/web/lib/wasm/jsep/webgpu/ops/concat.ts
Original file line number Diff line number Diff line change
Expand Up @@ -4,9 +4,9 @@
import {TensorView} from '../../tensor-view';
import {ShapeUtil} from '../../util';
import {AttributeWithCacheKey, createAttributeWithCacheKey} from '../attribute-with-cache-key';
import {ComputeContext, ProgramInfo} from '../types';
import {ComputeContext, ProgramInfo, ProgramInputTensorInfoDependency, ProgramUniform} from '../types';

import {IndicesHelper, inputVariable, outputVariable, ShaderHelper} from './common';
import {createTensorShapeVariables, enableShapesUniforms, IndicesHelper, inputVariable, outputVariable, ShaderHelper} from './common';

export interface ConcatAttributes extends AttributeWithCacheKey {
readonly axis: number;
Expand Down Expand Up @@ -35,8 +35,8 @@ const validateInputs = (inputs: readonly TensorView[]): void => {

const calculateInputIndexImpl = (numberOfTensors: number): string => `
fn calculateInputIndex(index: u32) -> u32 {
for (var i: u32 = 0u; i < ${numberOfTensors}u; i += 1u ) {
if (index < sizeInConcatAxis[i]) {
for (var i: u32 = 0u; i < ${numberOfTensors}; i += 1u ) {
if (index < uniforms.sizeInConcatAxis[i]) {
return i;
}
}
Expand Down Expand Up @@ -92,40 +92,62 @@ const createConcatProgramInfo = (inputs: readonly TensorView[], axis: number): P
const dataType = inputs[0].dataType;

let previousSum = 0;
const inputDependencies = [];
const inputShapeOrRanks = [];
const enableInputShapesUniforms = [];
let programUniforms: ProgramUniform[] = [{type: 'uint32', data: outputSize}];
for (let i = 0; i < inputs.length; ++i) {
previousSum += inputs[i].dims[adjustedAxis];
sizeInConcatAxis[i] = previousSum;
enableInputShapesUniforms.push(enableShapesUniforms(inputs[i].dims.length));
inputShapeOrRanks.push(enableInputShapesUniforms[i] ? inputs[i].dims.length : inputs[i].dims);
inputVars[i] = inputVariable(`input${i}`, dataType, inputShapeOrRanks[i]);
inputDependencies.push('rank');
}
programUniforms.push({type: 'uint32', data: sizeInConcatAxis});
for (let i = 0; i < inputs.length; ++i) {
if (enableInputShapesUniforms[i]) {
programUniforms.push(...createTensorShapeVariables(inputs[i].dims));
}
}

inputVars[i] = inputVariable(`input${i}`, dataType, inputs[i].dims);
const enableOutputShapesUniforms = enableShapesUniforms(outputShape.length);
if (enableOutputShapesUniforms) {
programUniforms.push(...createTensorShapeVariables(outputShape));
}

const output = outputVariable('output', dataType, outputShape);
const outputShapeOrRank = enableOutputShapesUniforms ? outputShape.length : outputShape;
const output = outputVariable('output', dataType, outputShapeOrRank);

const indicesAxis = output.indicesGet('indices', adjustedAxis);
const getShaderSource = (shaderHelper: ShaderHelper) => `
${shaderHelper.declareVariables(...inputVars, output)}
${
shaderHelper.registerUniform('outputSize', 'u32')
.registerUniform(`sizeInConcatAxis`, `vec${inputs.length}<u32>`)
.declareVariables(...inputVars, output)}
const sizeInConcatAxis = array<u32, ${sizeInConcatAxis.length}>(${sizeInConcatAxis.map(i => `${i}u`).join(',')});
${calculateInputIndexImpl(sizeInConcatAxis.length)}
${shaderHelper.mainStart()}
${shaderHelper.guardAgainstOutOfBoundsWorkgroupSizes(outputSize)}
${shaderHelper.guardAgainstOutOfBoundsWorkgroupSizes('uniforms.outputSize')}
var indices = ${output.offsetToIndices('global_idx')};
let inputIndex = calculateInputIndex(${indicesAxis});
if (inputIndex != 0u) {
${indicesAxis} -= sizeInConcatAxis[inputIndex - 1u];
${indicesAxis} -= uniforms.sizeInConcatAxis[inputIndex - 1u];
}
${assignOutputData(inputVars, output)}
}`;

return {
name: 'Concat',
shaderCache: {hint: `${axis}`},
shaderCache: {hint: `${axis}`, inputDependencies: inputDependencies as ProgramInputTensorInfoDependency[]},
getRunData: () => ({
outputs: [{dims: outputShape, dataType: inputs[0].dataType}],
dispatchGroup: {x: Math.ceil(outputSize / 64 /* workgroup size */)}
dispatchGroup: {x: Math.ceil(outputSize / 64 /* workgroup size */)},
programUniforms: programUniforms as ProgramUniform[],
}),
getShaderSource,
};
Expand Down

0 comments on commit 0ecdc6c

Please sign in to comment.