Operator List#

namekindstagesincelabelsdescription
_conv_depthwise2dNeuralNetworkBeta2.2aten, ConvolutionA depthwise convolution for the conv2d neural network function.
_functional_sym_constrain_range_for_sizeTensorBeta5.0aten, KernelGenA low-level function used in symbolic shape analysis to restrict the possible numerical range (min/max) of an unbacked symbolic integer.
_unique2TensorStable2.1atenReturns the unique elements of the input tensor. This is an internal PyTorch function.
_upsample_bicubic2d_aaNeuralNetworkStable2.2aten, ReductionA variant of upsample() that has mode set to bicubic.
_upsample_nearest_exact1dNeuralNetworkBeta5.0aten, ReductionIncreases the length of a 1D tensor using nearest-neighbor interpolation, ensuring the output aligns with library-standard algorithms like PIL.
absMathStable1.0aten, pointwiseComputes the absolute value of each element in input. This is a simple wrapper of the existing torch abs operator.
abs_MathStable2.2aten, pointwiseThe in-place version of abs(), which is a simple wrapper of the Torch abs operator.
absoluteMathBeta5.0aaten, KernelGenThis is an alias for abs() with the low-level operations implemented by invoking low-level Torch operators.
acosMathStable5.0aten, pointwiseReturns a new tensor with the arccosine (in radians) of each element in input.
addMathStable1.0aten, pointwiseAdd a scalar or tensor to self tensor. If both alpha and other are specified, each element of other is scaled by alpha before being used.
add_MathStable2.2aten, pointwiseThe in-place version of add().
addcdivLinearAlgStable4.0aten, pointwisePerforms the element-wise division of tensor1 by tensor2, multiplies the result by the scalar value and adds it to input.
addcmulLinearAlgStable4.0aten, pointwisePerforms the element-wise multiplication of tensor1 by tensor2, multiplies the result by the scalar value and adds it to input.
addmmBLASStable1.0atenPerforms a matrix multiplication of the matrices mat1 and mat2. The matrix input is added to the final result.
addmm_outBLASStable4.0atenPerforms a matrix multiplication of the matrices mat1 and mat2. The matrix input is added to the final result.
addmvLinearAlgStable4.0atenPerforms a matrix-vector product of the matrix mat and the vector vec. The vector input is added to the final result.
addmv_outLinearAlgStable4.0atenPerforms a matrix-vector product of the matrix mat and the vector vec. The vector input is added to the final result.
addrLinearAlgStable4.0atenPerforms the outer-product of vectors vec1 and vec2 and adds it to the matrix input.
alias_copyTensorBeta5.0aten, KernelGenCreates a new tensor that shares the same storage data as the original tensor, but without preserving the original tensor's metadata (like shape or strides) in a way that links future mutations.
alias_copy_outTensorBeta5.0aten, KernelGenA variant of alias_copy() that assigns the output to the out tensor.
allMathStable2.0aten, ReductionTests if all elements in input evaluate to True.
all_dimMathStable2.0aten, ReductionFor each row of input in the given dimension dim, returns True if all elements in the row evaluate to True and False otherwise.
all_dimsMathStable2.0aten, ReductionA variant of all.
allcloseMathStable2.1atenThis function checks if input and other satisfy a condition specified via atol and rtol elementwise, for all elements of input and other.
amaxLinearAlgStable2.0aten, ReductionReturns the maximum value of each slice of the input tensor in the given dimension(s) dim.
angleMathStable3.0aten, pointwiseComputes the element-wise angle (in radians) of the given input tensor.
anyMathStable2.0aten, ReductionTests if any element in input evaluates to True.
any_dimMathStable2.0aten, ReductionFor each row of input in the given dimension dim, returns True if any element in the row evaluate to True and False otherwise.
any_dimsMathStable2.0aten, ReductionFor each row of input in the given dimensions in dims, returns True if any element in the row evaluate to True and False otherwise. The dims contains tuple of ints indicating the dimensions to reduce.
apply_repetition_penaltiesNeuralNetworkStable5.0fused, vLLMModifies logit tensors in place to penalize tokens that have already appeared in the generated sequence.
apply_rotary_pos_embNeuralNetworkStable2.0fusedA method to incorporate positional information into the Transformer architecture. Rotary Positional Embedding (RoPE) applies position-dependent rotation to the query (Q) and key (K) vectors before computing the attention score.
arangeTensorStable2.1atenReturns a 1-D tensor of size ceiling((end−start)/step) with values from the interval [start, end) taken with common difference step beginning from start.
arange_starttensorStable2.1atenA variant of arange, with start and/or step specified.
arcsinhMathBeta5.0aten, KernelGenPerforms an element-wise inverse hyperbolic sine computation on the given tensor.
arcsinh_MathBeta5.0aten, KernelGenThe in-place version of arcsinh().
arcsinh_outMathBeta5.0aten, KernelGenA variant of arcsinh that allows the output to be assigned to the out tensor.
arctanh_MathBeta5.0aten, KernelGenComputes the element-wise inverse hyperbolic tangent of a given input tensor. This is an in-place version.
argmaxLinearAlgStable2.0aten, ReductionReturns the indices of the maximum value of all elements in the input tensor.
argminLinearAlgStable2.2aten, ReductionReturns the indices of the minimum value(s) of the flattened tensor or along a dimension.
asinh_MathBeta5.0aten, KernelGenComputes the inverse hyperbolic sine for each element of a tensor in-place.
atanMathStable4.0aten, pointwiseReturns a new tensor with the arctangent of the elements (in radians) in the input tensor.
atan_MathStable4.0aten, pointwiseThe in-place version of atan().
avg_pool2dNeuralNetworkStable4.1nn.functionalApplies 2D average-pooling operation in kH \mul kW regions by step size sH \mul sW steps. The number of output features is equal to the number of input planes.
avg_pool2d_backwardNeuralNetworkStable4.1atenThe backward version of avg_pool2d().
baddbmmBLASStable4.1atenPerforms a batch matrix-matrix product of matrices in batch1 and batch2. input is added to the final result. batch1 and batch2 must be 3-D tensors each containing the same number of matrices.
batch_normNeuralNetworkStable3.0atenAn internal operator used for implementing the BatchNorm functionality.
batch_norm_backwardNeuralNetworkStable3.0atenThe backward version of batch_norm().
bincountReductionStable5.0aten, pointwiseCount the frequency of each value in an array of non-negative integers.
bitwise_and_scalarMathStable2.0aten, pointwiseComputes the bitwise AND of input and other scalar.
bitwise_and_scalar_MathStable2.2aten, pointwiseThe in-place, scalar version of bitwise_and().
bitwise_and_scalar_tensorMathStable2.0aten, pointwiseA variant of bitwise_and().
bitwise_and_tensorMathStable2.0aten, pointwiseThe Tensor method version of bitwise_and().
bitwise_and_tensor_MathStable2.2aten, pointwiseThe in-place, Tensor method version of bitwise_and().
bitwise_left_shiftMathStable4.0aten, pointwiseComputes the left arithmetic shift of input by other bits.
bitwise_notMathStable2.0aten, pointwiseComputes the bitwise NOT of the given input tensor.
bitwise_not_MathStable2.2aten, pointwiseThe in-place version of bitwise_not().
bitwise_or_scalarMathStable2.0aten, pointwiseComputes the bitwise OR of scalars input and other.
bitwise_or_scalar_MathStable2.2atenThe in-place version of bitwise_or_scalar.
bitwise_or_scalar_tensorMathStable2.0aten, pointwiseComputes the bitwise OR of input and other.
bitwise_or_tensorMathStable2.0aten, pointwiseComputes the bitwise OR of input and other, this is the Tensor method variant.
bitwise_or_tensor_MathStable2.2aten, pointwiseThe in-place version of bitwise_or_tensor().
bitwise_right_shiftMathStable4.0aten, pointwiseComputes the right arithmetic shift of input by other bits.
bmmBLASStable1.0atenPerforms a batch matrix-matrix product of matrices stored in input and mat2.
bmm_outBLASStable5.0atenPerforms a batch matrix-matrix product of matrices stored in input and mat2. This is a variant of bmm with out specified.
catTensorStable2.2atenConcatenates the given sequence of tensors in tensors in the given dimension.
ceilMathStable5.0aten, pointwiseReturns a new tensor with the ceil of the elements of input, the smallest integer greater than or equal to each element.
ceil_MathStable5.0aten, pointwiseThe in-place version of ceil().
ceil_outMathStable5.0aten, pointwiseA variant of ceil() with out specified.
celuNeuralNetworkStable4.0aten, nn.functional, pointwiseApplies the quantized CELU (Continuously Differentiable Exponential Linear Unit) activation function element-wise.
celu_NeuralNetworkStable4.0aten, nn.functional, pointwiseThe in-place version of celu().
chunk_gated_delta_rule_fwdAttentionAlpha5.0fused, FLAThe forward case for ChunkGatedDeltaRuleFunction with Flash Linear Attention (FLA).
clampMathStable2.0aten, pointwiseClamps all elements in input into the range [min, max].
clamp_MathStable2.2aten, pointwiseThe in-place version of clamp().
clamp_minMathStable4.0aten, pointwiseA variant of clamp() with min set to min.
clamp_min_MathStable4.0aten, pointwiseThe in-place version of clamp_().
clamp_tensorMathStable2.0aten, pointwiseThe tensor version of clamp().
clamp_tensor_MathStable2.2aten, pointwiseThe in-place, tensor version of clamp().
concat_and_cache_mlaAttentionBeta3.0fused, MLAWrites the latent and RoPE value into KV cache for Multi-head Latent Attention forward case.
constant_pad_ndNeuralNetworkStable2.2aten, IRPads the input tensor boundaries with a constant value. This is an IR representation, not a public API.
contiguousTensorBeta4.1atenReturns a contiguous in memory tensor containing the same data as self tensor. Introduced in v2.2 and removed from exposed list in v4.1.
conv1dConvolutionStable4.2atenApplies a 1D convolution over a quantized 1D input composed of several input planes.
conv2dConvolutionStable4.2atenApplies a 2D convolution over a quantized 2D input composed of several input planes.
conv3dConvolutionStable4.2atenApplies a 3D convolution over a quantized 3D input composed of several input planes.
copyTensorBeta4.2aten, pointwiseAs a wrapper of copy_, this operator copies elements from src to out using given template for shapes.
copy_TensorStable4.1aten, pointwiseCopies the elements from src into self tensor and returns self`.
cosMathStable2.0aten, pointwiseReturns a new tensor with the cosine of the elements of input given in radians.
cos_MathStable2.2aten, pointwiseThe in-place version of cos().
count_nonzeroDataStable2.2aten, ReductionCounts the number of non-zero values in the tensor input along the given dim. If no dim is specified then all non-zeros in the tensor are counted.
cross_entropy_lossNeuralNetworkRemoved3.0fused, ReductionComputes the cross entropy loss between input logits and target.
cummaxMathStable3.0aten, ReductionReturns a named tuple (values, indices) where values is the cumulative maximum of elements of input in the dimension dim. And indices is the index location of each maximum value found in the dimension dim.
cumminMathStable2.2aten, ReductionReturns a named tuple (values, indices) where values is the cumulative minimum of elements of input in the dimension dim. And indices is the index location of each minimum value found in the dimension dim.
cumsumLinearAlgStable1.0aten
cumsum_outReductionStable3.0aten
cutlass_scaled_mmLinearAlgBeta5.0fused, vLLM
dgegluNeuralNetworkAlpha4.2fused, TransformerGaussian Error Gated Linear Unit with GELU activation instead of sigmoid function. This is for the backward case.
diagTensorStable2.2aten
  • If input is a vector (1-D tensor), then returns a 2-D square tensor with the elements of input as the diagonal.
  • If input is a matrix (2-D tensor), then returns a 1-D tensor with the diagonal elements of input.
diag_embedTensorStable2.2aten, pointwiseCreates a tensor whose diagonals of certain 2D planes (specified by dim1 and dim2) are filled by input. To facilitate creating batched diagonal matrices, the 2D planes formed by the last two dimensions of the returned tensor are chosen by default.
diagonal_backwardLinearAlgStable2.2aten, pointwiseA diagonal operation returns a partial view of input with the its diagonal elements with respect to dim1 and dim2 appended as a dimension at the end of the shape. This is the backward case for diagonal().
digamma_MathBeta5.0aten, KernelGenComputes the in-place digamma function, which is the logarithmic derivative of the Gamma function.
dispatch_fused_moe_kernelMoEBeta5.0fused, Activation, vLLMAccelerates neural network training by combining token routing (dispatch/all-to-all communication), expert computation (GEMM), and result aggregation into a single GPU kernel.
div_modeMathStable1.0aten, pointwiseDivides each element of the input by the corresponding element of other. An optional rounding_mode can be specified.
div_mode_MathStable2.2aten, pointwiseThe in-place version of div_mode().
dotBLASStable3.0atenComputes the dot product of two 1D tensors.
dregluNeuralNetworkAlpha4.2fused, TransformerRectified Gated Linear Unit is a variant of GLU that uses ReLU instead of the sigmoid function for gating. This is the backward case. This operator was introduced in v4.2 but not exposed.
dropoutNeuralNetworkStable1.0aten, nn.functionalAn internal IR for implementing torch.nn.functional.dropout.
dropout_backwardNeuralNetworkStable3.0aten, nn.functionalThe backward case of dropout().
dswigluNeuralNetworkAlpha5.0fused, TransformerSwish-Gated Linear Unit, a variant of GLU with the Swish activation function. This is for the backward case.
eluNeuralNetworkStable2.2aten, nn.functional, pointwiseApply the Exponential Linear Unit (ELU) function element-wise.
elu_NeuralNetworkStable4.0aten, pointwiseThe in-place version of elu().
elu_backwardNeuralNetworkStable4.0aten, pointwiseThe backward version of elu().
embeddingNeuralNetworkStable2.1aten, nn.functionalGenerate a simple lookup table that looks up embeddings in a fixed dictionary and size. Note that the parameter sequence differs from torch.nn.functional.embedding.
embedding_backwardNeuralNetworkStable3.0atenThe backward version of embedding().
embedding_dense_backwardNeuralNetworkStable5.0atenCalculates the gradient of the weight matrix for a dense embedding layer during backpropagation.
eqMathStable2.0aten, pointwiseComputes element-wise equality.
eq_scalarMathStable2.0aten, pointwiseComputes equality between scalars.
equalMathStable5.0aten, ReductionReturns True if two tensors have the same size and elements, False otherwise.
erfScienceStable2.1atenComputes the error function of input.
erf_ScienceStable2.2aten, pointwiseThe in-place version of erf().
expMathStable1.0aten, pointwiseReturns a new tensor with the exponential of the elements of the input tensor input.
exp_MathStable2.2aten, pointwiseThe in-place version of exp().
exp_outMathStable4.1aten, pointwiseA variant of exp2(), with out specified.
exp2MathStable4.0aten, pointwiseComputes the base two exponential function of input.
exp2_MathStable4.0aten, pointwiseThe in-place version of exp2().
exponential_DistributionStable2.1atenFills self tensor with elements drawn from a PDF (probability density function).
eyeLinearAlgStable3.0aten, ReductionReturns a 2-D tensor with ones on the diagonal and zeros elsewhere.
eye_mLinearAlgStable3.0aten, ReductionTriton-based implementation of torch.eye_m(n, m), using 2D tiles to split the matrix into blocks.
fill_scalarTensorStable2.2aten, pointwiseFills a scalar with the specified value.
fill_scalar_TensorStable2.2aten, pointwiseThe in-place version of fill_scalar().
fill_scalar_outTensorStable5.0aten, pointwiseA variant of fill_scalar() that assigns the output to an out tensor.
fill_tensorTensorStable2.2aten, pointwiseFills a tensor with the specified value.
fill_tensor_TensorStable2.2aten, pointwiseThe in-place version of fill_tensor().
fill_tensor_outTensorStable5.0aten, pointwiseA variant of fill_tensor() that assigns the output to an out tensor.
flash_attention_forwardNeuralNetworkStable3.0aten
flash_attn_varlen_funcNeuralNetworkStable3.1aten, Attention, FlashAttentionCompute attention for sequences of variable lengths within a single batch. Eliminating the need for padding.
flash_mlaNeuralNetworkStable3.0fused, Attention, vLLMA variant of Multi-head Latent Attention (MLA).
flipTensorStable2.1aten, pointwiseReverse the order of an n-D tensor along given axis in dims.
floor_MathBeta5.0aten, KernelGenPerforms an in-place element-wise floor operation, rounding each element of a tensor down to the nearest integer less than or equal to itself.
floor_divideMathStable2.1atenComputes input divided by other, elementwise, and floors the result.
floor_divide_MathStable2.2atenComputes input divided by other, elementwise, and floors the result.
fminMathBeta5.0aten, KernelGenComputes the element-wise minimum of two tensors, specially handling NaN values by prioritizing the numerical value. Unlike minimum(), if one input is NaN and the other is a number, fmin() returns the number. It supports broadcasting, type promotion, and operates on both CPU and GPU.
fmin_outMathBeta5.0aten, KernelGenA variant of fmin() that assigns the output to the out tensor.
fullTensorStable2.1aten, pointwiseCreates a tensor of size size filled with fill_value. The tensor’s dtype is inferred from fill_value.
full_likeTensorStable2.1aten, pointwiseReturns a tensor with the same size as input filled with fill_value.
fused_add_rms_normNeuralNetworkStable2.0fused, Normalization
fused_recurrent_gated_delta_rule_fwdAttentionAlpha5.0fused, FLAThe forward case for fused_recurrent_gated_delta_rule used in Flash Linear Attention (FLA).
gatherTensorStable2.2aten, ReductionGathers values along an axis specified by dim.
gather_backwardTensorStable2.2aten, ReductionThe backward version of gather().
geMathStable2.0aten, pointwiseComputes input is greater or equal to other element-wise.
ge_scalarMathStable2.0aten, pointwiseThe scalar version of ge().
gegluNeuralNetworkAlpha4.2fused, Activation, TransformerGaussian Error Gated Linear Unit with GELU activation instead of sigmoid function.
geluNeuralNetworkStable1.0aten, pointwise, Activation, nn.functionalApply Cumulative Distribution Function for Gaussian Distribution function element-wise.
gelu_NeuralNetworkStable2.2aten, Activation, pointwiseThe in-place version of gelu().
gelu_and_mulNeuralNetworkStable2.0fused, pointwise, ActivationAn activation function for GeGLU.
gelu_backwardNeuralNetworkStable3.0aten, Activation, pointwiseThe backward version of gelu().
get_scheduler_metadataAttentionStable4.0vLLMComputes scheduling metadata for attention work partitioning so that CPU computations can be routed to ISA-specific kernel implmentations. The metadata is stored in a tensor.
gluNeuralNetworkStable3.0aten, Activation, pointwiseGated Linear Unit activation for modulating the output of a linear transformation with a gate.
glu_backwardNeuralNetworkStable4.0aten, Activation, pointwiseThe backward version of glu().
group_normNeuralNetworkStable2.0aten, ReductionAn internal IR for applying Group Normalization for last certain number of dimensions.
group_norm_backwardNeuralNetworkStable3.0aten, ReductionThe backward case for group_norm().
grouped_topkMoEStable5.0fused, vLLMA specialized routing mechanism used in Mixture-of-Experts (MoE) models (like DeepSeek-V3/R1) to select top-k experts by first grouping them, rather than selecting globally.
gtMathStable2.0aten, pointwiseComputes that input is greater than other element-wise.
gt_scalarMathStable2.0aten, pointwiseThe scalar version of gt().
hardsigmoidNeuralNetworkBeta5.0aten, pointwise, nn.functional, Activation, KernelGenAn activation function that provides a piecewise linear approximation of the standard sigmoid function, mapping inputs to a range between 0 and 1.
hardsigmoid_outNeuralNetworkBeta5.0aten, pointwise, nn.functional, Activation, KernelGenA variant of hardsigmoid that supports an output tensor to receive the result.
hardswish_NeuralNetworkBeta5.0aten, pointwise, KernelGen, ActivationApplies the Hard Swish activation function, commonly used in models like MobileNetV3 to improve accuracy while reducing computational cost compared to traditional Swish. This is an in-place version.
hstackTensorStable2.2atenStack tensors in sequence horizontally (column wise). This is equivalent to concatenation along the first axis for 1-D tensors, and along the second axis for all other tensors.
hypotMathBeta5.0aten, KernelGenGiven the legs of a right triangle, return its hypotenuse. The shapes of both input tensors must be broadcastable.
hypot_outMathBeta5.0aten, KernelGenGiven the legs of a right triangle, return its hypotenuse. The shapes of both input tensors must be broadcastable. This is a variant of hypot that allows the output to be a different tensor.
i0MathBeta5.0aten, KernelGenComputes the modified Bessel function of the first kind of order zero element-wise for a given input tensor.
i0_MathBeta5.0aten, KernelGenThe inplace version of i0.
i0_outMathBeta5.0aten, KernelGenA variant of i0 that assigns the output to the out tensor.
indexReductionStable4.2atenExtract, access or modify specific elements, slices, or subsets of data within a tensor. The location of data is specified for each dimension, starting from index 0.
index_addTensorStable2.2atenAccumulate the elements of alpha times source into the input tensor by adding to the indices in the order given in index.
index_add_TensorStable4.0atenThe in-place version of index_add().
index_putTensorStable2.2atenPuts values from the tensor values into the tensor input using the indices specified in indices (which is a tuple of Tensors).
index_put_TensorStable3.0atenThe in-place version of index_put().
index_selectTensorStable2.1atenReturns a new tensor which indexes the input tensor along dimension dim using the entries in index.
inplace_fused_expertsMoEBeta5.0fused, Activation, vLLMThis operator writes output directly to hidden_states.
instance_normNeuralNetworkRemoved3.0fusedApply Instance Normalization independently for each channel in every data sample within a batch.
iscloseMathStable2.1aten, pointwiseReturns a new tensor with boolean elements representing if each element of input is "close" to the corresponding element of other. The closeness is defined with rtol and atol.
isfiniteMathStable2.1aten, pointwiseReturns a new tensor with boolean elements representing if each element is finite or not.
isinDataStable2.2atenTests if each element of elements is in test_elements. Returns a boolean tensor of the same shape as elements that is True for elements in test_elements and False otherwise.
isinfMathStable2.0aten, pointwiseTests if each element of input is infinite (positive or negative infinity) or not.
isnanMathStable2.0aten, pointwiseReturns a new tensor with boolean elements representing if each element of input is NaN or not.
kronLinearAlgStable2.2atenComputes the Kronecker product of input and other.
layer_normNeuralNetworkStable1.0atenAn internal IR for applying Layer Normalization for last certain number of dimensions.
layer_norm_backwardReductionStable3.0atenThe backward case for layer_norm().
leMathStable2.0aten, pointwiseComputes that input is less than or equal to other element-wise.
le_scalarMathStable2.0atenThe scalar version of le().
lerp_scalarLinearAlgStable3.0aten, pointwiseThe scalar version of lerp().
lerp_scalar_LinearAlgStable3.0aten, pointwiseThe in-place, scalar version of lerp().
lerp_tensorLinearAlgStable3.0aten, pointwisePerforms a linear interpolation of two tensors start (given by input) and end based on a scalar or tensor weight and returns the resulting out tensor.
lerp_tensor_LinearAlgStable3.0aten, pointwiseThe in-place version of lerp().
lift_fresh_copyTensorBeta5.0aten, KernelGenCreates a new, independent copy of a tensor within a compiled graph.
linspaceDataStable2.2atenCreates a one-dimensional tensor of size steps whose values are evenly spaced from start to end, inclusive.
logMathStable2.2aten, pointwiseReturns a new tensor with the natural logarithm of the elements of input.
log_sigmoidNeuralNetworkStable2.2aten, pointwise, nn.functionalApplies the Logsigmoid function element-wise.
log_softmaxNeuralNetworkStable3.0aten, ReductionAn internal IR for applying a softmax followed by a logarithm.
log_softmax_backwardNeuralNetworkStable3.0aten, ReductionThe backward case for log_softmax().
log1p_MathBeta5.0aten, KernelGenComputes the natural logarithm of 1+x(y_i=log_e(x_i+1)) for each element in the input tensor in-place.
logaddexpMathBeta5.0aten, pointwise, KernelGenComputes the element-wise logarithm of the sum of the exponentials of two input tensors.
logaddexp_outMathBeta5.0aten, pointwise, KernelGenA variant of logaddexp that allows the output to be assigned to an out tensor.
logical_andMathStable2.2aten, pointwiseComputes the element-wise logical AND of the given input tensors. Zeros are treated as False and nonzeros are treated as True.
logical_and_MathStable5.0aten, pointwiseThe in-place version of logical_and().
logical_notMathStable2.2aten, pointwiseComputes the element-wise logical NOT of the given input tensor.
logical_orMathStable2.2aten, pointwiseComputes the element-wise logical OR of the given input tensors.
logical_or_MathStable5.0aten, pointwiseThe in-place version of logical_or().
logical_xorMathStable2.2aten, pointwiseComputes the element-wise logical XOR of the given input tensors.
logitLinearAlgBeta5.0aten, pointwise, KernelGenReturns a new tensor with the logit of the elements of input. input is clamped to [eps, 1-eps] when eps is not None. When eps is None and input<0 or input>1, the function will yield NaN.
logit_LinearAlgBeta5.0aten, pointwise, KernelGenThe in-place version of logit().
logit_outLinearAlgBeta5.0aten, pointwise, KernelGenA variant of logit that allows the output to be assigned to another tensor.
logspacetensorStable4.0atenCreates a one-dimensional tensor of size steps whose values are evenly spaced from base^start to base^end, inclusive, on a logarithmic scale with base base.
ltMathStable2.0aten, pointwiseComputes that input is less than other element-wise.
lt_scalarMathStable2.0aten, pointwiseThe scalar version of lt.
masked_fillTensorStable2.2aten, pointwiseFills elements of given tensor with value where mask is True.
masked_fill_tensorStable2.2aten, pointwiseThe in-place version of masked_fill().
masked_scattertensorStable4.2atenCopies elements from source into the given tensor at positions where the mask is True.
masked_scatter_tensorStable4.2atenThe in-place version of masked_scatter().
masked_selectTensorStable2.1atenReturns a new 1-D tensor which indexes the input tensor according to the boolean mask mask which is a BoolTensor.
maxLinearAlgStable2.0aten, ReductionReturns the maximum value of all elements in the input tensor.
max_dimLinearAlgStable2.0aten, ReductionReturns a namedtuple (values, indices) where values is the maximum value of each row of the input tensor in the given dimension dim. And indices is the index location of each maximum value found (argmax).
max_pool2d_backwardIRStable4.0atenApplies a 2D max pooling over an input signal composed of several input planes. This is an IR representation rather than a public API and it is for the backward step.
max_pool2d_with_indicesIRStable4.0atenApplies a 2D max pooling over an input signal composed of several input planes. This is an IR representation rather than a public API.
maximumMathStable2.1aten, pointwiseComputes the element-wise maximum of input and other.
meanLinearAlgStable1.0aten, ReductionReturns the mean value of all elements in the input tensor. Input must be floating point or complex.
mean_dimReductionStable2.0atenReturns the mean value of each row of the input tensor in the given dimension dim. If dim is a list of dimensions, reduce over all of them.
minLinearAlgStable2.0aten, ReductionReturns the minimum value of all elements in the input tensor.
min_dimLinearAlgStable2.0aten, ReductionReturns a namedtuple (values, indices) where values is the minimum value of each row of the input tensor in the given dimension dim. And indices is the index location of each minimum value found (argmin).
minimumMathStable2.1aten, pointwiseComputes the element-wise minimum of input and other.
mmBLASStable1.0atenPerforms a matrix multiplication of the two input matrices.
mm_outBLASStable3.0atenA variant of mm() with out specified.
moe_align_block_sizeMoEStable4.0fused, Reduction, vLLMAligns the token distribution across experts to be compatible with block size for matrix multiplication.
moe_align_block_size_tritonMoEStable4.2fused, Reduction, vLLMAligns the token distribution across experts to be compatible with block size for matrix multiplication. This is the Triton version.
moe_sumMoERemoved5.0fused, Reduction, vLLMAn implementation of Mixture of Experts (MoE) with sum-based aggregation instead of the more common weighted average.
mse_lossNeuralNetworkStable2.2aten, pointwise, nn.functionalCompute the element-wise mean squared error, with optional weighting.
mulMathStable1.0aten, pointwiseMultiplies input by other.
mul_MathStable2.2aten, pointwiseThe in-place version of mul().
multinomialDistributionStable2.1atenReturns a tensor where each row contains num_samples indices sampled from the multinomial probability distribution located in the corresponding row of tensor input.
mvBLASStable2.0atenPerforms a matrix-vector product of the matrix input and the vector vec.
nan_to_numMathStable3.0aten, pointwiseReplaces NaN, positive infinity, and negative infinity values in input with the values specified by nan, posinf, and neginf, respectively.
neMathStable2.0aten, pointwiseComputes that input is not equal to other element-wise.
ne_scalarMathStable2.0aten, pointwiseThe scalar version of ne().
negMathStable2.0aten, pointwiseReturns a new tensor with the negative of the elements of input.
neg_MathStable2.2aten, pointwiseThe in-place version of neg().
nll_loss2d_backwardNeuralNetworkStable2.2aten, IRAn internal IR for supporting torch.nn.NLLLoss2d, which has been deprecated and is now integrated into the standard torch.nn.NLLLoss. This is the backward case.
nll_loss2d_forwardNeuralNetworkStable2.2aten, IRAn internal IR for supporting torch.nn.NLLLoss2d, which has been deprecated and is now integrated into the standard torch.nn.NLLLoss. This is the forward case.
nll_loss_backwardNeuralNetworkStable2.2aten, IRCompute the negative log likelihood loss. This is the backward case.
nll_loss_forwardNeuralNetworkStable2.2aten, IRCompute the negative log likelihood loss. This is the forward case.
nll_loss_nd_backwardNeuralNetworkStable5.0atenMeasures the performance of a classification model by penalizing low probabilities for correct classe.s This computes the gradients of this loss with respect to model parameters using automatic differentiation.
nll_loss_nd_forwardNeuralNetworkStable5.0atenMeasures the performance of a classification model by calculating the negative log probability of the true class. This defines the computation flow, transforming input data through layers to produce output predictions.
nonzeroDataStable2.1atenReturns a 2-D tensor where each row is the index for a nonzero value. When as_tuple is explicitly set to True, this returns a tuple of 1-D index tensors, allowing for advanced indexing of all nonzero values.
normal_DistributionStable5.0aten, pointwiseReturns a tensor of random numbers drawn from separate normal distributions whose mean and standard deviation are given. This is one of the variants that takes a float mean and a float std.
normal_float_tensorDistributionStable2.1aten, pointwiseReturns a tensor of random numbers drawn from separate normal distributions whose mean and standard deviation are given. This is one of the variants that takes a float mean and a tensor std.
normal_tensor_floatDistributionStable2.1aten, pointwiseReturns a tensor of random numbers drawn from separate normal distributions whose mean and standard deviation are given. This is one of the variants that takes a tensor mean and a float std.
normal_tensor_tensorDistributionStable2.1aten, pointwiseReturns a tensor of random numbers drawn from separate normal distributions whose mean and standard deviation are given. This is one of the variants that takes a tensor mean and a tensor std.
normed_cumsumReductionStable2.1atenGet the normalized cumulative sum where each step is divided by the total sum of the dataset, resulting in values ranging from 0 to 1. Internally used by the multinomial operator.
one_hotNeuralNetworkStable5.0aten, nn.functionalTakes LongTensor with index values of shape (*) and returns a tensor of shape (*, num_classes) that have zeros everywhere except where the index of last dimension matches the corresponding value of the input tensor, in which case it will be 1.
onesTensorStable2.1atenReturns a tensor filled with the scalar value 1, with the shape defined by the variable argument size.
ones_likeTensorStable2.1atenReturns a tensor filled with the scalar value 1, with the same size as input.
outerBLASRemoved3.0fusedComputes outer product of self and the input vector. If the self tensor is a vector of size n and the input tensor is a vector of size m, the out tensor (if specified) must be a matrix of size n * m.
outplace_fused_expertsMoEBeta5.0fused, Activation, vLLMThis operator allocates and returns a new output tensor.
padNeuralNetworkStable2.1aten, pointwise, nn.functionalThis pads a tenor using the specified mode.
per_token_group_quant_fp8QuantizationAlpha4.2vLLMFunction to perform per-token-group quantization on an input tensor x. It converts the tensor values into signed float8 values and returns the quantized tensor along with the scaling factor used for quantization.
pixel_unshuffleNeuralNetworkBeta5.0aten, KernelGenRearranges elements from a low-resolution feature map with many channels into a higher-resolution feature map with fewer channels.
pixel_unshuffle_outNeuralNetworkBeta5.0aten, KernelGenA variant of pixel_unshuffle that assigns the output to the out tensor.
polarMathStable3.0aten, pointwiseConstructs a complex tensor whose elements are Cartesian coordinates corresponding to the polar coordinates with absolute value abs and angle angle.
pow_scalarMathStable1.0atenTakes the power of each element in input with exponent and returns a tensor with the result. The input is a single float, while the exponent is a tensor.
pow_tensor_scalarMathStable1.0aten, pointwiseTakes the power of each element in input with exponent and returns a tensor with the result. The input is a tensor, while the exponent is a float.
pow_tensor_scalar_MathStable2.2aten, pointwiseThis is the in-place version of pow_tensor_scalar().
pow_tensor_tensorMathStable1.0aten, pointwiseTakes the power of each element in input with exponent and returns a tensor with the result. The input is a tensor, while the exponent is also a tensor.
pow_tensor_tensor_MathStable2.2aten, pointwiseThis is the in-place version of pow_tensor_tensor().
preluNeuralNetworkBeta5.0aten, Activation, pointwise, nn.functional, KernelGenAn activation function used in neural networks that improves upon ReLU (Rectified Linear Unit) by allowing the network to learn the slope of negative inputs. It performs an element-wise operation that keeps positive values and scales negative values by a learnable parameter.
prodLinearAlgStable2.0aten, ReductionReturns the product of all elements in the input tensor.
prod_dimReductionStable2.0atenReturns the product of each row of the input tensor in the given dimension dim.
quantileDataStable2.2atenComputes the q-th quantiles of each row of the input tensor along the dimension dim.
randDistributionStable2.1atenReturns a tensor filled with random numbers from a uniform distribution on the interval [0,1).
rand_likeDistributionStable2.1atenReturns a tensor with the same size as input that is filled with random numbers from a uniform distribution on the interval [0,1).
randnDistributionStable2.1atenReturns a tensor filled with random numbers from a normal distribution with mean 0 and variance 1 (also called the standard normal distribution).
randn_likeDistributionStable2.1atenReturns a tensor with the same size as input that is filled with random numbers from a normal distribution with mean 0 and variance 1.
randpermDistributionStable2.2atenReturns a random permutation of integers from 0 to n - 1.
reciprocalMathStable1.0aten, pointwiseReturns a new tensor with the reciprocal of the elements of input.
reciprocal_MathStable2.2aten, pointwiseThis is the in-place version of reciprocal().
reflection_pad1dNeuralNetworkBeta5.0aten, pointwise, KernelGenPads the input 3D or 2D tensor (typically representing signals or sequences) by reflecting the boundary values at the edges.
reflection_pad1d_outNeuralNetworkBeta5.0aten, pointwise, KernelGenA variant of reflection_pad1d that assigns the output to out tensor.
reflection_pad2dNeuralNetworkBeta5.0aten, pointwise, KernelGenPads the input 3D or 2D tensor (typically representing signals or sequences) by reflecting the boundary values at the both edges.
reflection_pad2d_outNeuralNetworkBeta5.0aten, pointwise, KernelGenA variant of reflection_pad2d that assigns the output to out tensor.
regluNeuralNetworkAlpha4.2fused, TransformerRectified Gated Linear Unit is a variant of GLU that uses ReLU instead of the sigmoid function for gating. This operator was introduced in v4.2 release but not exposed.
reluNeuralNetworkStable1.0aten, Activation, pointwise, nn.functionalApply the RELU (Rectified Linear Unit) activation function element-wise.
relu_NeuralNetworkStable2.2aten, pointwise, ActivationThis is the in-place version of relu().
relu6NeuralNetworkBeta5.0aten, pointwise, Activation, KernelGenApplies the element-wise function f(x)=min(max(0,x),6). This is a variation of the standard ReLU activation function that "caps" its output at a maximum value of 6.
remainderMathStable2.2atenComputes Python’s modulus operation entrywise. The result has the same sign as the divisor other and its absolute value is less than that of other.
remainder_MathStable2.2atenThis is the in-place version of remainder().
repeatTensorStable2.1atenRepeats this tensor along the specified dimensions.
repeat_interleave_self_intTensorStable2.2aten, pointwiseRepeats elements of a tensor. The number of repetitions is specified as an integer repeats.
repeat_interleave_self_tensorTensorStable2.2aten, pointwiseRepeats elements of a tensor. The number of repetitions is specified as a tensor repeats. repeats is broadcasted to fit the shape of the given axis.
repeat_interleave_tensorTensorStable2.2aten, pointwiseRepeats 0 repeats[0] times, 1 repeats[1] times, 2 repeats[2] times, etc.
replication_pad1dTensorBeta5.0aten, KernelGenPads the edge of a 1D input tensor by repeating the boundary values.
replication_pad1d_outTensorBeta5.0aten, KernelGenA variant of replication_pad1d that assigns the output to the out tensor.
replication_pad3dNeuralNetworkAlpha5.0atenPads the edge of a 3D input tensor by repeating the boundary values.
reshape_and_cacheAttentionStable3.0fused, vLLMStore the key/value token states into the pre-allcated kv_cache buffers of paged attention.
reshape_and_cache_flashAttentionStable3.0fusedStore the key/value token states into the pre-allcated kv_cache buffers of paged attention.
resolve_conjScienceStable2.1atenReturns a new tensor with materialized conjugation if input's conjugate bit is set to True, else returns input. The output tensor will always have its conjugate bit set to False.
resolve_negScienceStable2.1atenReturns a new tensor with materialized negation if input's negative bit is set to True, else returns input. The output tensor will always have its negative bit set to False.
rms_normNeuralNetworkStable2.0aten, nn.functional, ReductionApply Root Mean Square Layer Normalization over a mini-batch of inputs.
rms_norm_backwardNeuralNetworkStable2.0aten, nn.functional, ReductionThis is the backward case for rms_norm.
rms_norm_forwardNeuralNetworkStable2.0aten, nn.functional, Reducito0nThis is the forward case for rms_norm.
rotary_embeddingAttentionStable3.0vLLMApply rotary positional embeddings.
rrelu_with_noise_backwardNeuralNetworkBeta5.0aten, KernelGenComputes the gradient of the Randomized Leaky ReLU (RReLU) activation function with respect to its input during backpropagation. It uses the noise tensor generated in the forward pass to correctly apply the slope to negative input values.
rsqrtMathStable1.0aten, pointwiseReturns a new tensor with the reciprocal of the square-root of each of the elements of input.
rsqrt_MathStable2.2aten, pointwiseThe in-place version of rsqrt().
rwkv_ka_fusionRWKVStable4.1fusedMerges, aligns, and enhances features from different data sources or spatial directions using the efficient, linear-time RWKV framework.
rwkv_mm_sparsityRWKVStable4.1fusedOptimized, lossless sparse matrix multiplication in RWKV-7 models.
scaled_dot_product_attentionNeuralNetworkStable2.2nn.functional, AttentionComputes scaled dot product attention on query, key and value tensors, using an optional attention mask if passed and applying dropout if a probability greater than 0.0 is specified. The optional scale argument can only be specified as a keyword argument.
scaled_dot_product_attention_backwardNeuralNetworkStable2.2nn.functional, AttentionThe backward case for scaled_dot_product_attention.
scaled_dot_product_attention_forwardNeuralNetworkStable2.2nn.functional, AttentionThe forward case for scaled_dot_product_attention.
scaled_softmax_backwardReductionStable4.2atenThe backward pass for a scaled softmax function, commonly used in Scaled Dot-Product Attention (SDPA) within Transformer models, computes the gradient of the loss with respect to the input logits, incorporating a scaling factor to stabilize training.
scaled_softmax_forwardReductionStable4.2atenThe backward pass for a scaled softmax function, commonly used in Scaled Dot-Product Attention (SDPA) within Transformer models, computes the gradient of the loss with respect to the input logits, incorporating a scaling factor to stabilize training.
scatterTensorStable2.2atenWrites all values from the tensor src into provided tensor at the indices specified in the index tensor. For each value in src, its output index is specified by its index in src for dimension != dim and by the corresponding value in index for dimension = dim. The optional reduce argument allows specification of an optional reduction operation, which is applied to all values in the tensor src into the tensor at the indices specified in the index.
scatter_TensorStable3.0atenThis is the in-place version of scatter().
scatter_add_TensorStable4.2atenAdds all values from the tensor src into self at the indices specified in the index tensor in a similar fashion as scatter_(). For each value in src, it is added to an index in self which is specified by its index in src for dimension != dim and by the corresponding value in index for dimension = dim.
select_scatterTensorStable2.2atenEmbeds the values of the src tensor into input at the given index. This function returns a tensor with fresh storage; it does not create a view.
seluNeuralNetworkBeta5.0aten, pointwise, nn.functional, Activation, KernelGenApplies an element-wise activation function that induces self-normalizing properties in neural networks. It scales the Exponential Linear Unit (ELU) to ensure activations remain close to zero mean and unit variance.
selu_NeuralNetworkBeta5.0aten, pointwise, Activation, KernelGenThis is the in-place version of selu.
sgn_MathBeta5.0aten, KernelGenComputes the sign of each element in the self tensor, element-wise. This function is an extension of sign() designed to handle complex tensors in addition to real-valued ones.
sigmoidNeuralNetworkStable2.0aten, pointwiseComputes the expit (also known as the logistic sigmoid function) of the elements of input.
sigmoid_NeuralNetworkStable2.2aten, pointwiseThe in-place version of sigmoid().
sigmoid_backwardNeuralNetworkStable3.0aten, pointwiseThe backward version of sigmoid().
siluNeuralNetworkStable1.0aten, pointwise, nn.functionalSiLU (Sigmoid Linear Unit), a simple approximation of ReLU but without any discontinuity of the first derivative.
silu_NeuralNetworkStable2.2aten, nn.functional, pointwiseThe in-place version of silu().
silu_and_mulActivationStable2.0fused, pointwise, vLLMA custom operator in vLLM as activation function for SwiGLU.
silu_and_mul_outActivationStable2.0fused, pointwise, vLLMA variant of silu_and_mul with an extra out argument.
silu_backwardNeuralNetworkStable3.0aten, pointwiseA variant of silu() for backward case.
sinMathStable2.0aten, pointwiseReturns a new tensor with the sine of the elements in the input tensor, where each value in this input tensor is in radians.
sin_MathStable2.2aten, pointwiseThe in-place version of sin().
sinh_MathBeta5.0aten, KernelGenComputes the hyperbolic sine (e^x-e^{-x})/2 of each element in a tensor. This is an in-place version.
skip_layer_normNeuralNetworkStable2.0fused, TransformerAn optimized operation used in Transformer models to improve performance by combining residual connection (skip connection) addition and Layer Normalization (LayerNorm) into a single kernel.
slice_backwardNeuralNetworkStable5.0atenAn automatic differentiation (autograd) function that computes the gradient of a tensor slicing operation (tensor[start:end]) during backpropagation.
slice_scatterTensorStable2.2atenEmbeds the values of the src tensor into input at the given dimension. This function returns a tensor with fresh storage; it does not create a view.
softmaxNeuralNetworkStable1.0aten, nn.functionalApply a softmax function.
softmax_backwardReductionStable3.0aten, nn.functionalThe in-place version of softmax().
softplusNeuralNetworkStable4.0aten, nn.functional, pointwiseApplies element-wise, the function Softplus.
softshrinkNeuralNetworkBeta5.0aten, nn.functional, Activation, KernelGenApplies the soft shrinkage function element-wise to an input tensor. It is an activation function often used in signal processing and sparse representation, such as image denoising.
softshrink_outNeuralNetworkBeta5.0aten, nn.functional, Activation, KernelGenThis is a variant of softshrink that supports an output tensor.
sortDataStable2.2atenSorts the elements of the input tensor along a given dimension in ascending order by value.
sort_stableDataStable3.0atenSorts the elements of the input tensor along a given dimension in ascending order by value. This is a variant of sort() where stable is set to True to preserve the order of equivalent elements.
sparse_mla_fwd_interfaceDSAStable5.0fused
special_i1MathBeta5.0aten, pointwise, KernelGenComputes the modified Bessel function of the first kind of order 1 (I_1(x)) for each element in the input tensor, designed for special mathematical functions.
special_i1_outMathBeta5.0aten, pointwise, KernelGenA variant of special_i1 that allows the output to be assigned to another tensor.
sqrtMathStable4.0aten, pointwiseReturns a new tensor with the square-root of the elements of input.
sqrt_MathStable4.0aten, pointwiseThis is the in-place version of sqrt().
stackTensorStable2.2atenConcatenates a sequence of tensors along a new dimension.
stdReductionStable4.0atenCalculates the standard deviation over the dimensions specified by dim. dim can be a single dimension, list of dimensions, or None to reduce over all dimensions.
subMathStable1.0aten, pointwiseSubtracts other, scaled by alpha, from the input tensor.
sub_MathStable2.2aten, pointwiseSubtracts other, scaled by alpha, from the input tensor. This is the in-place version.
sumLinearAlgStable2.0aten, ReductionReturns the sum of all elements in the input tensor.
sum_dimLinearAlgStable2.0aten, ReductionReturns the sum of each row of the input tensor in the given dimension dim. dim is a list of dimensions, reduce over all of them.
sum_dim_outLinearAlgStable3.0aten, ReductionA variant of sum_dim() with the out argument.
sum_outLinearAlgStable3.0aten, ReductionA variant of sum() with the out argument.
swigluNeuralNetworkStable5.0fused, TransformerSwish-Gated Linear Unit, a variant of GLU with the Swish activation function.
t_copyTensorBeta5.0aten, KernelGenTranspose a 2D tensor into a new tensor with contiguous memory layout.
t_copy_outTensorBeta5.0aten, KernelGenA variant of t_copy() that allows the output to be assigned to the out tensor.
tanNeuralNetworkStable4.1aten, pointwiseReturns a new tensor with the tangent of the elements in the input tensor, where each value in this input tensor is in radians.
tan_Stable4.1aten, pointwiseThis is the in-place version of tan().
tanhMathStable2.0aten, pointwiseReturns a new tensor with the hyperbolic tangent of the elements of input.
tanh_MathStable2.2aten, pointwiseThis is the in-place version of tanh().
tanh_backwardMathStable3.0aten, pointwiseThis is the backward case for tanh().
thresholdNeuralNetworkStable3.0aten, nn.functional, pointwiseApply a threshold to each element of the input Tensor.
threshold_backwardNeuralNetworkStable3.0aten, nn.functional, pointwiseThis is the backward version for threshold.
tileTensorStable2.1atenConstructs a tensor by repeating the elements of input. The dims argument specifies the number of repetitions in each dimension.
to_copyTensorBeta4.1aten, pointwise
topkTensorStable2.1atenReturns the k largest elements of the given input tensor along a given dimension. If dim is not given, the last dimension of the input is chosen. If largest is False then the k smallest elements are returned.
topk_softmaxMoEStable4.0fused, vLLMSelects the k most likely next-token candicates, sets all others to zero, and renormalize the prbabilities of these top candidates.
traceReductionStable4.0atenReturns the sum of the elements of the diagonal of the input 2-D matrix.
trilBLASBeta5.0aten, KernelGenReturns the lower triangular part of an input matrix (or a batch of matrices) and sets all other elements to zero.
triuBLASStable1.0atenReturns the upper triangular part of a matrix (2-D tensor) or batch of matrices input, the other elements of the result tensor out are set to 0.
triu_NeuralNetworkStable5.0atenThe in-place version of triu().
true_divideMathStable2.1aten, pointwiseDivides each element of the input input by the corresponding element of other. Note that torch.divide() is an alias of torch.div() and torch.true_divide() is an alias of torch.div() with rounding_mode=None.
true_divide_MathStable2.1atenThis is the in-place version of true_divide().
true_divide_outMathStable4.2atenThis is an variant of true_divide() with an out argument.
trunc_divideMathStable2.1atenThe div function with rounding_mode set to trunc.
unfold_backwardNeuralNetworkStable5.0aten, nn.functionalAn operator for calculating the gradient of the unfold operation during backpropagation. It takes the gradient of the unfolded output and accumulates it back into the original input shape, reversing sliding local block extraction and resolving overlaps.
uniform_DistributionStable2.1atenFills self tensor with numbers sampled from the continuous uniform distribution.
upsample_bicubic2dNeuralNetworkStable5.0aten, ReductionA variant of upsample() that has mode set to bicubic.
upsample_linear1dNeuralNetworkStable5.0atenUpsamples the input, using linear mode. The input has to be 3 dimensional, and the output_size is an optional tuple of ints.
upsample_nearest1dNeuralNetworkStable5.0atenUpsamples the input, using nearest neighbours' pixel values. The input has to be 3 dimensional, and the output_size is an optional tuple of ints.
upsample_nearest2dNeuralNetworkStable2.2atenUpsamples the input, using nearest neighbours' pixel values. The input has to be 4 dimensional. The scales can be provided with scales_h and scales_w.
upsample_nearest3dNeuralNetworkStable5.0atenPerforms 3D nearest-neighbor interpolation to increase the spatial size of volumetric data, such as 5D tensors. It scales up inputs by copying values from the nearest pixel/voxel, without calculating new values through linear interpolation.
var_meanLinearAlgStable2.0aten, ReductionCalculates the variance and mean over the dimensions specified by dim. dim can be a single dimension, list of dimensions, or None to reduce over all dimensions.
vdotBLASStable2.2atenComputes the dot product of two 1D vectors along a dimension.
vector_normLinearAlg NeuralNetworkStable2.0aten, ReductionComputes a vector norm.
vstackTensorStable2.2atenStack tensors in sequence vertically (row wise).
weight_normNeuralNetworkStable3.0fusedReparameterizes a module's weight tensor by decoupling its magnitude (g) from its direction (v). It is a hook that compute the actual weight before each forward pass.
weight_norm_interfaceNeuralNetworkStable2.2aten, fusedApply weight normalization to neural network layers, decoupling the magnitued of a weight tensor from its direction. It is used to stabilize training, particularly for models with small batch sizes.
weight_norm_interface_backwardNeuralNetworkStable3.0aten, fusedComputes the gradients for weight normalization during the backward pass. It calculates the necessary derivatives for updating both the magnitude (g) and direction (v) parameters of a weight-normalized layer, based on gradients received from the previous operation.
where_scalar_otherTensorStable2.1aten, pointwiseReturn a tensor of elements selected from either input or other, depending on condition.
where_scalar_selfTensorStable2.1aten, pointwiseReturns a tensor of elements selected from either input or other, where input is a tensor while other is a scalar.
where_selfTensorStable2.1aten, pointwiseReturns a LongTensor. This operation is identical to torch.nonzero(condition, as_tuple=True).
where_self_outTensorStable2.2aten, pointwiseThis is a variant of where_self() with an argument out.
zeroTensorBeta5.0aten, KernelGenFills tensor with zeros.
zero_TensorStable5.0atenFills self tensor with zeros.
zero_outTensorBeta5.0aten, KernelGenFills tensor with zeros but assign the output to the out tensor.
zerosTensorStable2.1atenReturns a tensor filled with the scalar value 0, with the shape defined by the variable argument size.
zeros_likeTensorStable2.1atenReturns a tensor filled with the scalar value 0, with the same size as input.