/* Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #pragma once #include "paddle/phi/common/int_array.h" #include "paddle/phi/common/scalar.h" #include "paddle/phi/core/meta_tensor.h" namespace phi { // Common InferMeta Functions for multiary operators, The format like: // // 1. The number of input MetaTensor is more than 3: // void [FunctionDesc|OpName]InferMeta(const MetaTensor& x, // const MetaTensor& y, // const MetaTensor& z, // const MetaTensor& w, // ..., // MetaTensor* out) {} // // 2. There are `const vector&` in params: // void [FunctionDesc|OpName]InferMeta(const vector& x, // ..., // MetaTensor* out) {} // // NOTE: The InferMeta Functions in this file are arranged in alphabetic order. std::vector GetMetaTensorsDim( const std::vector& tensors); void AdadeltaInferMeta(const MetaTensor& param, const MetaTensor& grad, const MetaTensor& avg_squared_grad, const MetaTensor& avg_squared_update, const MetaTensor& learning_rate, const MetaTensor& master_param, float rho, float epsilon, bool multi_precision, MetaTensor* param_out, MetaTensor* avg_squared_grad_out, MetaTensor* avg_squared_update_out, MetaTensor* master_param_outs); void AdagradInferMeta(const MetaTensor& param, const MetaTensor& grad, const MetaTensor& moment, const MetaTensor& learning_rate, const MetaTensor& master_param, float epsilon, bool multi_precision, MetaTensor* param_out, MetaTensor* moment_out, MetaTensor* master_param_out); void AdamaxInferMeta(const MetaTensor& param, const MetaTensor& grad, const MetaTensor& learning_rate, const MetaTensor& moment, const MetaTensor& inf_norm, const MetaTensor& beta1_pow, const MetaTensor& master_param, float beta1, float beta2, float epsilon, bool multi_precision, MetaTensor* param_out, MetaTensor* moment_out, MetaTensor* inf_norm_out, MetaTensor* master_param_outs); void AdamInferMeta(const MetaTensor& param, const MetaTensor& grad, const MetaTensor& learning_rate, const MetaTensor& moment1, const MetaTensor& moment2, const MetaTensor& beta1_pow, const MetaTensor& beta2_pow, const MetaTensor& master_param, const MetaTensor& skip_update, const Scalar& beta1, const Scalar& beta2, const Scalar& epsilon, bool lazy_mode, int64_t min_row_size_to_use_multithread, bool multi_precision, bool use_global_beta_pow, MetaTensor* param_out, MetaTensor* moment1_out, MetaTensor* moment2_out, MetaTensor* beta1_pow_out, MetaTensor* beta2_pow_out, MetaTensor* master_param_outs); void AdamwInferMeta(const MetaTensor& param, const MetaTensor& grad, const MetaTensor& learning_rate, const MetaTensor& moment1, const MetaTensor& moment2, const MetaTensor& beta1_pow, const MetaTensor& beta2_pow, const MetaTensor& master_param, const MetaTensor& skip_update, const Scalar& beta1, const Scalar& beta2, const Scalar& epsilon, float lr_ratio, float coeff, bool with_decay, bool lazy_mode, int64_t min_row_size_to_use_multithread, bool multi_precision, bool use_global_beta_pow, MetaTensor* param_out, MetaTensor* moment1_out, MetaTensor* moment2_out, MetaTensor* beta1_pow_out, MetaTensor* beta2_pow_out, MetaTensor* master_param_outs); void AddNInferMeta(const std::vector& x, MetaTensor* out, MetaConfig config = MetaConfig()); void AddNTensorArrayInferMeta(const std::vector& x, MetaTensor* out, MetaConfig config); void AucInferMeta(const MetaTensor& input, const MetaTensor& label, const MetaTensor& stat_pos, const MetaTensor& stat_neg, const MetaTensor& ins_tag_weight, const std::string& curve, int num_thresholds, int slide_steps, MetaTensor* auc, MetaTensor* stat_pos_out, MetaTensor* stat_neg_out, MetaConfig config = MetaConfig()); void AverageAccumulatesInferMeta(const MetaTensor& param, const MetaTensor& in_sum_1, const MetaTensor& in_sum_2, const MetaTensor& in_sum_3, const MetaTensor& in_num_accumulates, const MetaTensor& in_old_num_accumulates, const MetaTensor& in_num_updates, float average_window, int64_t max_average_window, int64_t min_average_window, MetaTensor* out_sum_1, MetaTensor* out_sum_2, MetaTensor* out_sum_3, MetaTensor* out_num_accumulates, MetaTensor* out_old_num_accumulates, MetaTensor* out_num_updates); void BatchNormInferMeta(const MetaTensor& x, const MetaTensor& mean, const MetaTensor& variance, const MetaTensor& scale, const MetaTensor& bias, bool is_test, float momentum, float epsilon, const std::string& data_layout, bool use_global_stats, bool trainable_statistics, MetaTensor* y, MetaTensor* mean_out, MetaTensor* variance_out, MetaTensor* saved_mean, MetaTensor* saved_variance, MetaTensor* reserve_space, MetaConfig config = MetaConfig()); void BatchNormInferInferMeta(const MetaTensor& x, const MetaTensor& mean, const MetaTensor& variance, const MetaTensor& scale, const MetaTensor& bias, float momentum, float epsilon, const std::string& data_layout, MetaTensor* y, MetaTensor* mean_out, MetaTensor* variance_out, MetaConfig config = MetaConfig()); void BilinearInferMeta(const MetaTensor& x, const MetaTensor& y, const MetaTensor& weight, const MetaTensor& bias, MetaTensor* out, MetaConfig config = MetaConfig()); void BroadcastTensorsInferMeta(const std::vector& x, std::vector out); void CheckFiniteAndUnscaleInferMeta(const std::vector& xs, const MetaTensor& scale, std::vector outs, MetaTensor* found_infinite); void CoalesceTensorInferMeta(const std::vector& input, DataType dtype, bool copy_data, bool set_constant, bool persist_output, float constant, bool use_align, int align_size, int size_of_dtype, const std::vector& concated_shapes, const std::vector& concated_ranks, std::vector output, MetaTensor* fused_output, MetaConfig config = MetaConfig()); void CheckMemoryContinueInferMeta(const std::vector& input, MetaTensor* output, std::vector xout, MetaConfig config = MetaConfig()); void ConcatInferMeta(const std::vector& x, const Scalar& axis_scalar, MetaTensor* out, MetaConfig config = MetaConfig()); void CudnnLSTMInferMeta( const MetaTensor& x, const MetaTensor& init_h, const MetaTensor& init_c, const MetaTensor& w, const paddle::optional>& weight_list, const MetaTensor& sequence_length, float dropout_prob, bool is_bidirec, int hidden_size, int num_layers, bool is_test, int seed, MetaTensor* out, MetaTensor* last_h, MetaTensor* last_c, MetaTensor* reserve, MetaTensor* state_out); void DecayedAdagradInferMeta(const MetaTensor& param, const MetaTensor& grad, const MetaTensor& moment, const MetaTensor& learning_rate, float decay, float epsilon, MetaTensor* param_out, MetaTensor* moment_out); void DeformableConvInferMeta(const MetaTensor& x, const MetaTensor& offset, const MetaTensor& filter, const MetaTensor& mask, const std::vector& strides, const std::vector& paddings, const std::vector& dilations, int deformable_groups, int groups, int im2col_step, MetaTensor* out, MetaConfig config = MetaConfig()); void DGCMomentumInferMeta(const MetaTensor& param, const MetaTensor& grad, const MetaTensor& velocity, const MetaTensor& learning_rate, const MetaTensor& master_param, const MetaTensor& current_step_tensor, const MetaTensor& nranks_tensor, float mu, bool use_nesterov, const std::string& regularization_method, float regularization_coeff, bool multi_precision, float rescale_grad, float rampup_begin_step, MetaTensor* param_out, MetaTensor* velocity_out, MetaTensor* master_param_out, MetaTensor* grad_out); void EditDistanceInferMeta(const MetaTensor& hyps, const MetaTensor& refs, const MetaTensor& hypslength, const MetaTensor& refslength, bool normalized, MetaTensor* sequencenum, MetaTensor* out); void FusedBatchNormActInferMeta(const MetaTensor& x, const MetaTensor& scale, const MetaTensor& bias, const MetaTensor& mean, const MetaTensor& variance, MetaTensor* y, MetaTensor* mean_out, MetaTensor* variance_out, MetaTensor* saved_mean, MetaTensor* saved_variance, MetaTensor* reserve_space); void FusedBiasActInferMeta(const MetaTensor& x, const MetaTensor& bias, const MetaTensor& dequant_scales, const MetaTensor& shift, const MetaTensor& smooth, const std::string& act_method, const std::string& compute_dtype, float quant_scale, int quant_round_type, float quant_max_bound, float quant_min_bound, MetaTensor* out, MetaConfig config = MetaConfig()); void FusedLayerNormInferMeta(const MetaTensor& x, const MetaTensor& bias, const MetaTensor& residual, const MetaTensor& norm_weight, const MetaTensor& norm_bias, const float epsilon, const float residual_alpha, const int begin_norm_axis, const float quant_scale, const int quant_round_type, const float quant_max_bound, const float quant_min_bound, MetaTensor* out, MetaTensor* residual_out, MetaTensor* mean, MetaTensor* variance); void FusedLinearParamGradAddInferMeta(const MetaTensor& x, const MetaTensor& dout, const MetaTensor& dweight, const MetaTensor& dbias, bool multi_precision, bool has_bias, MetaTensor* dweight_out, MetaTensor* dbias_out); void FusionGroupInferMeta(const std::vector& ins, const std::vector& outs_dtype, const std::vector& inputs_dtype, const std::string& func_name, int type, std::vector outs); void GenerateProposalsV2InferMeta(const MetaTensor& scores, const MetaTensor& bbox_deltas, const MetaTensor& im_shape, const MetaTensor& anchors, const MetaTensor& variances, int pre_nms_top_n, int post_nms_top_n, float nms_thresh, float min_size, float eta, bool pixel_offset, MetaTensor* rpn_rois, MetaTensor* rpn_roi_probs, MetaTensor* rpn_rois_num); void GraphReindexInferMeta(const MetaTensor& x, const MetaTensor& neighbors, const MetaTensor& count, const MetaTensor& hashtable_value, const MetaTensor& hashtable_index, MetaTensor* reindex_src, MetaTensor* reindex_dst, MetaTensor* out_nodes); void GraphSampleNeighborsInferMeta(const MetaTensor& row, const MetaTensor& col_ptr, const MetaTensor& x, const MetaTensor& eids, const MetaTensor& perm_buffer, int sample_size, bool return_eids, bool flag_perm_buffer, MetaTensor* out, MetaTensor* out_count, MetaTensor* out_eids); void HSigmoidLossInferMeta(const MetaTensor& x, const MetaTensor& label, const MetaTensor& w, const MetaTensor& bias, const MetaTensor& path, const MetaTensor& code, int num_classes, bool is_sparse, MetaTensor* out, MetaTensor* pre_out, MetaTensor* w_out); void InterpolateInferMeta( const MetaTensor& x, const MetaTensor& out_size, const paddle::optional>& size_tensor, const MetaTensor& scale_tensor, const std::string& data_layout, int out_d, int out_h, int out_w, const std::vector& scale, const std::string& interp_method, bool align_corners, int align_mode, MetaTensor* output, MetaConfig config = MetaConfig()); void IndexPutInferMeta(const MetaTensor& x, const std::vector& indices, const MetaTensor& value, bool accumulate, MetaTensor* out); void LambInferMeta(const MetaTensor& param, const MetaTensor& grad, const MetaTensor& learning_rate, const MetaTensor& moment1, const MetaTensor& moment2, const MetaTensor& beta1_pow, const MetaTensor& beta2_pow, const MetaTensor& master_param, const MetaTensor& skip_update, float weight_decay, float beta1, float beta2, float epsilon, bool always_adapt, bool multi_precision, MetaTensor* param_out, MetaTensor* moment1_out, MetaTensor* moment2_out, MetaTensor* beta1_pow_out, MetaTensor* beta2_pow_out, MetaTensor* master_param_outs); void LLMInt8LinearInferMeta(const MetaTensor& x, const MetaTensor& weight, const MetaTensor& bias, const MetaTensor& weight_scale, const float threshold, MetaTensor* out); void LogspaceInferMeta(const MetaTensor& start, const MetaTensor& stop, const MetaTensor& number, const MetaTensor& base, DataType dtype, MetaTensor* out); void MergedAdamInferMeta( const std::vector& param, const std::vector& grad, const std::vector& learning_rate, const std::vector& moment1, const std::vector& moment2, const std::vector& beta1_pow, const std::vector& beta2_pow, const paddle::optional>& master_param, const Scalar& beta1, const Scalar& beta2, const Scalar& epsilon, bool multi_precision, bool use_global_beta_pow, std::vector param_out, std::vector moment1_out, std::vector moment2_out, std::vector beta1_pow_out, std::vector beta2_pow_out, std::vector master_param_out); void MergedMomentumInferMeta( const std::vector& param, const std::vector& grad, const std::vector& velocity, const std::vector& learning_rate, const paddle::optional>& master_param, float mu, bool use_nesterov, const std::vector& regularization_method, const std::vector& regularization_coeff, bool multi_precision, float rescale_grad, std::vector param_out, std::vector velocity_out, std::vector master_param_out); void MemoryEfficientAttentionInferMeta(const MetaTensor& query, const MetaTensor& key, const MetaTensor& value, const MetaTensor& bias, const MetaTensor& cu_seqlens_q, const MetaTensor& cu_seqlens_k, const MetaTensor& causal_diagonal, const MetaTensor& seqlen_k, const Scalar& max_seqlen_q, const Scalar& max_seqlen_k, const bool causal, const double dropout_p, const float scale, const bool is_test, MetaTensor* output, MetaTensor* logsumexp, MetaTensor* seed_and_offset); void VariableLengthMemoryEfficientAttentionInferMeta( const MetaTensor& query, const MetaTensor& key, const MetaTensor& value, const MetaTensor& seq_lens, const MetaTensor& kv_seq_lens, const MetaTensor& mask, float scale, bool causal, MetaTensor* out); void MeshgridInferMeta(const std::vector& inputs, std::vector outputs); void MomentumInferMeta(const MetaTensor& param, const MetaTensor& grad, const MetaTensor& velocity, const MetaTensor& learning_rate, const MetaTensor& master_param, float mu, bool use_nesterov, const std::string& regularization_method, float regularization_coeff, bool multi_precision, float rescale_grad, MetaTensor* param_out, MetaTensor* velocity_out, MetaTensor* master_param_out); void MultiDotInferMeta(const std::vector& x, MetaTensor* out); void MultiplexInferMeta(const std::vector& ins, const MetaTensor& ids, MetaTensor* out); void PsroiPoolInferMeta(const MetaTensor& x, const MetaTensor& rois, const MetaTensor& rois_num, int pooled_height, int pooled_width, int output_channels, float spatial_scale, MetaTensor* out); void RmsNormInferMeta(const MetaTensor& x, const MetaTensor& bias, const MetaTensor& residual, const MetaTensor& norm_weight, const MetaTensor& norm_bias, const float epsilon, const int begin_norm_axis, const float quant_scale, const int quant_round_type, const float quant_max_bound, const float quant_min_bound, MetaTensor* out, MetaTensor* residual_out); void RmspropInferMeta(const MetaTensor& param, const MetaTensor& mean_square, const MetaTensor& grad, const MetaTensor& moment, const MetaTensor& learning_rate, const MetaTensor& mean_grad, const MetaTensor& master_param, float epsilon, float decay, float momentum, bool centered, bool multi_precision, MetaTensor* param_out, MetaTensor* moment_out, MetaTensor* mean_square_out, MetaTensor* mean_grad_out, MetaTensor* master_param_outs); void RnnInferMeta(const MetaTensor& x, const std::vector& pre_state, const std::vector& weight_list, const MetaTensor& sequence_length, float dropout_prob, bool is_bidirec, int input_size, int hidden_size, int num_layers, const std::string& mode, int seed, bool is_test, MetaTensor* out, MetaTensor* dropout_state, std::vector state, MetaTensor* reserve); void SendUERecvInferMeta(const MetaTensor& x, const MetaTensor& y, const MetaTensor& src_index, const MetaTensor& dst_index, const std::string& message_op, const std::string& reduce_op, const IntArray& out_size, MetaTensor* out, MetaTensor* dst_count); void SendUVInferMeta(const MetaTensor& x, const MetaTensor& y, const MetaTensor& src_index, const MetaTensor& dst_index, const std::string& message_op, MetaTensor* out); void SgdInferMeta(const MetaTensor& param, const MetaTensor& learning_rate, const MetaTensor& grad, const MetaTensor& master_param, bool multi_precision, MetaTensor* param_out, MetaTensor* master_param_out); void SigmoidCrossEntropyWithLogitsInferMeta(const MetaTensor& x, const MetaTensor& label, const MetaTensor& pos_weight, bool normalize, int ignore_index, MetaTensor* out, MetaConfig config = MetaConfig()); void StackInferMeta(const std::vector& x, int axis, MetaTensor* out, MetaConfig config = MetaConfig()); void UnchangedMultiInferMeta(const std::vector& x, std::vector out); void ShareBufferInferMeta(const std::vector& x, const std::vector& share_dims_and_dtype, std::vector out, std::vector xout); void UpdateLossScalingInferMeta(const std::vector& xs, const MetaTensor& found_infinite, const MetaTensor& prev_loss_scaling, const MetaTensor& in_good_steps, const MetaTensor& in_bad_steps, std::vector outs, MetaTensor* loss_scaling, MetaTensor* out_good_steps, MetaTensor* out_bad_steps); void WarpctcInferMeta(const MetaTensor& logits, const MetaTensor& label, const MetaTensor& logits_length, const MetaTensor& labels_length, int blank, bool norm_by_times, MetaTensor* loss, MetaTensor* warpctcgrad); void WarprnntInferMeta(const MetaTensor& input, const MetaTensor& label, const MetaTensor& input_lengths, const MetaTensor& label_lengths, int blank, float fastemit_lambda, MetaTensor* loss, MetaTensor* warpctcgrad); void WeightOnlyLinearInferMeta(const MetaTensor& x, const MetaTensor& weight, const MetaTensor& bias, const MetaTensor& weight_scale, const std::string& weight_dtype, MetaTensor* out); void WeightedSampleNeighborsInferMeta(const MetaTensor& row, const MetaTensor& col_ptr, const MetaTensor& edge_weight, const MetaTensor& x, const MetaTensor& eids, int sample_size, bool return_eids, MetaTensor* out, MetaTensor* out_count, MetaTensor* out_eids); void WhereInferMeta(const MetaTensor& condition, const MetaTensor& x, const MetaTensor& y, MetaTensor* out); void YoloLossInferMeta(const MetaTensor& x, const MetaTensor& gt_box, const MetaTensor& gt_label, const MetaTensor& gt_score, const std::vector& anchors, const std::vector& anchor_mask, int class_num, float ignore_thresh, int downsample_ratio, bool use_label_smooth, float scale_x_y, MetaTensor* loss, MetaTensor* objectness_mask, MetaTensor* gt_match_mask); void FusedAdamInferMeta( const std::vector& params, const std::vector& grads, const MetaTensor& learning_rate, const std::vector& moments1, const std::vector& moments2, const std::vector& beta1_pows, const std::vector& beta2_pows, const paddle::optional>& master_params, const MetaTensor& skip_update, const Scalar& beta1, const Scalar& beta2, const Scalar& epsilon, int chunk_size, float weight_decay, bool use_adamw, bool multi_precision, bool use_global_beta_pow, std::vector params_out, std::vector moments1_out, std::vector moments2_out, std::vector beta1_pows_out, std::vector beta2_pows_out, std::vector master_params_out); void FusedConvInferMeta(const MetaTensor& input, const MetaTensor& filter, const MetaTensor& bias, const MetaTensor& residual_param, const std::vector& strides, const std::vector& paddings, const std::string& padding_algorithm, const std::vector& dilations, int groups, const std::string& data_format, const std::string& mkldnn_data_type, const std::string& fuse_activation, bool fuse_residual_conn, bool force_fp32_output, MetaTensor* out, MetaConfig config); void MoeInferMeta(const MetaTensor& x, const MetaTensor& gate, const MetaTensor& bmm0, const MetaTensor& bias0, const MetaTensor& bmm1, const MetaTensor& bias1, const std::string& act_type, MetaTensor* out); void FusedMultiHeadAttentionInferMeta(const MetaTensor& query, const MetaTensor& key, const MetaTensor& value, const MetaTensor& mask, float scale, bool causal, MetaTensor* out); void FusedMultiHeadAttentionVariableInferMeta(const MetaTensor& query, const MetaTensor& key, const MetaTensor& value, const MetaTensor& seq_lens, const MetaTensor& mask, float scale, bool causal, MetaTensor* out); void FusedRopeInferMeta(const MetaTensor& q, const MetaTensor& k, const MetaTensor& v, const MetaTensor& sin, const MetaTensor& cos, MetaTensor* out_q, MetaTensor* out_k, MetaTensor* out_v); void MaskedMultiheadAttentionInferMeta(const MetaTensor& x, const MetaTensor& cache_kv, const MetaTensor& bias, const MetaTensor& src_mask, const MetaTensor& cum_offsets, const MetaTensor& sequence_lengths, const MetaTensor& rotary_tensor, const MetaTensor& beam_cache_offset, const MetaTensor& qkv_out_scale, const MetaTensor& out_shift, const MetaTensor& out_smooth, int seq_len, int rotary_emb_dims, const bool use_neox_rotary_style, const std::string& compute_dtype, const float out_scale, const int quant_round_type, const float quant_max_bound, const float quant_min_bound, MetaTensor* out, MetaTensor* cache_kv_out, MetaTensor* beam_cache_offset_out); void FullWithTensorInferMeta(const MetaTensor& shape, DataType dtype, MetaTensor* out); } // namespace phi