diff --git a/paddle/CMakeLists.txt b/paddle/CMakeLists.txt index 25848620326c8485492880fd81a2b6fc382639ed..924362253850e3e40e035553f6bbb8f038389804 100644 --- a/paddle/CMakeLists.txt +++ b/paddle/CMakeLists.txt @@ -33,7 +33,7 @@ add_subdirectory(ir) # Next, (to be discusssed) # (1) move all source files to same folder, -# (2) naturally, and and configure tests in only one `CMakeLists.txt`, +# (2) naturally, and configure tests in only one `CMakeLists.txt`, # (3) cc tests support linking pre-built dynamic libraries. For example, use the dynamic # library in the installed paddle by `pip`. diff --git a/paddle/fluid/framework/data_set.h b/paddle/fluid/framework/data_set.h index e37f44f3f8b5e0a9b25ba87f60ae1e05a4f24176..22bb5b703944da95ffae78e2086ad6415f4a67f2 100644 --- a/paddle/fluid/framework/data_set.h +++ b/paddle/fluid/framework/data_set.h @@ -156,7 +156,7 @@ class Dataset { virtual void DestroyPreLoadReaders() = 0; // set preload thread num virtual void SetPreLoadThreadNum(int thread_num) = 0; - // seperate train thread and dataset thread + // separate train thread and dataset thread virtual void DynamicAdjustChannelNum(int channel_num, bool discard_remaining_ins = false) = 0; virtual void DynamicAdjustReadersNum(int thread_num) = 0; diff --git a/paddle/fluid/framework/op_desc.h b/paddle/fluid/framework/op_desc.h index 16d97b6c1e4c23d9e01a950228c267beabdbfcaa..6fab6f06cb0b224746042a53c54daded959afeb2 100644 --- a/paddle/fluid/framework/op_desc.h +++ b/paddle/fluid/framework/op_desc.h @@ -219,7 +219,7 @@ class OpDesc { return ret_val; } - // it it really needed? or just maintain a ptr from block? + // Is it really needed? Or just maintain a ptr from the block? proto::OpDesc desc_; BlockDesc *block_{nullptr}; // not_own // input arg name => input variable names diff --git a/paddle/fluid/framework/operator.cc b/paddle/fluid/framework/operator.cc index d7aa0bdaa6dd43cfb073ffccd2e93cf8a5054668..7550854709ac80c9f76d332cf80f2d0a3b1d54fb 100644 --- a/paddle/fluid/framework/operator.cc +++ b/paddle/fluid/framework/operator.cc @@ -3502,7 +3502,7 @@ void OperatorWithKernel::BuildPhiKernelContext( // we try to add these Attrs to the RuntimeAttrs, but these OpDesc will lose // the RuntimeAttrs information in the process of converting the Graph to // the Program, so additional record configuration will be introduced, - // which increases the The cost of development and understanding, so we + // which increases the cost of development and understanding, so we // still use Attrs to get and the attributes set by these passes from Attrs // for the time being. In the future, it is necessary to clarify the // positioning of RuntimeAttrs and expand related functions. diff --git a/paddle/fluid/framework/var_desc.h b/paddle/fluid/framework/var_desc.h index ab60dfa56bd23f157d471898fcf688539a05c7cb..b6b62c03df6a2830cb6358cc38119910f5e3be72 100644 --- a/paddle/fluid/framework/var_desc.h +++ b/paddle/fluid/framework/var_desc.h @@ -210,7 +210,7 @@ class VarDesc { proto::VarType::TensorDesc *mutable_tensor_desc(); std::vector mutable_tensor_descs(); - // it it really needed? or just mantain a ptr from block? + // Is it really needed? Or just mantain a ptr from the block? proto::VarDesc desc_; AttributeMap attrs_; diff --git a/paddle/fluid/inference/tensorrt/convert/c_allreduce_op.cc b/paddle/fluid/inference/tensorrt/convert/c_allreduce_op.cc index 5f9dca9a0d26fe1c5ad6e219375ff64d979acafe..48aa3c7188b5b2d24ee381d7221dbd4c064b8fc0 100644 --- a/paddle/fluid/inference/tensorrt/convert/c_allreduce_op.cc +++ b/paddle/fluid/inference/tensorrt/convert/c_allreduce_op.cc @@ -56,7 +56,7 @@ class CAllReduceOpConverter : public OpConverter { output_num, 1UL, platform::errors::InvalidArgument( - "The ouput Out's size must equal to 1 in TRT c_allreduce op. " + "The output Out's size must equal to 1 in TRT c_allreduce op. " "But received Out's size %u.", output_num)); // Get attrs diff --git a/paddle/fluid/operators/cinn/cinn_launch_context.cc b/paddle/fluid/operators/cinn/cinn_launch_context.cc index 3be173859491fa3f5331729052a6528df956f9cd..78f9b9bf2951d280573d13ee35c01eec4c2ff4ad 100644 --- a/paddle/fluid/operators/cinn/cinn_launch_context.cc +++ b/paddle/fluid/operators/cinn/cinn_launch_context.cc @@ -237,7 +237,8 @@ std::unordered_set CinnLaunchContext::ExtractInternalVarNames( }; VLOG(1) << "Input var list: " << string::join_strings(input_var_names, ", "); - VLOG(1) << "Ouput var list: " << string::join_strings(output_var_names, ", "); + VLOG(1) << "Output var list: " + << string::join_strings(output_var_names, ", "); std::for_each( input_var_names.begin(), input_var_names.end(), exclude_names_fn); std::for_each( diff --git a/paddle/fluid/operators/dlnne/dlnne_engine_op.h b/paddle/fluid/operators/dlnne/dlnne_engine_op.h index 28f9b4d73a18fddac1d06e321bd72ad8ffd7935f..7d5203b650e7aa6a0cab9afc5022ee417fee479f 100644 --- a/paddle/fluid/operators/dlnne/dlnne_engine_op.h +++ b/paddle/fluid/operators/dlnne/dlnne_engine_op.h @@ -277,7 +277,7 @@ class DlnneEngineOp : public framework::OperatorBase { std::string rlym_file_name = subgraph_root_path_ + "/" + engine_key_ + ".rlym"; - // quantize don't support set quantized ouput model path now, + // quantize don't support set quantized output model path now, // the quantized model file is in current dir std::string quantized_rlym_file_name = engine_key_ + ".quantized.rlym"; diff --git a/paddle/fluid/operators/fused/attn_gemm_int8.h b/paddle/fluid/operators/fused/attn_gemm_int8.h index e26273b7452603b69a2e8fb6f0948e5283c350d7..c61a7f60d43599cfcc09a2dc647f934e52f7a88f 100644 --- a/paddle/fluid/operators/fused/attn_gemm_int8.h +++ b/paddle/fluid/operators/fused/attn_gemm_int8.h @@ -90,7 +90,7 @@ class AttnMatmulINT8 { PADDLE_ENFORCE_EQ(cudaGetLastError(), cudaSuccess, platform::errors::Fatal( - "cuda error occured after computing bias. " + "cuda error occurred after computing bias. " "But it does not mean this error is caused by " "bias computing")); } @@ -144,7 +144,7 @@ class AttnMatmulINT8 { PADDLE_ENFORCE_EQ(cudaGetLastError(), cudaSuccess, platform::errors::Fatal( - "cuda error occured after computing bias. " + "cuda error occurred after computing bias. " "But it does not mean this error is caused by " "bias computing")); } diff --git a/paddle/fluid/operators/fused/fused_gate_attention.h b/paddle/fluid/operators/fused/fused_gate_attention.h index b7611eff765d2a7b9f29837dd839dba57bcbea00..c8c4733df2e2e53aa8a5e7afdb3120bde8a66d7a 100644 --- a/paddle/fluid/operators/fused/fused_gate_attention.h +++ b/paddle/fluid/operators/fused/fused_gate_attention.h @@ -715,7 +715,7 @@ class FMHAGateRef { softmax_out->dims(), platform::errors::InvalidArgument( "The shape of qk_out_grad and softmax_out is " - "expected to be the same. But recieved qk_out_grad's " + "expected to be the same. But received qk_out_grad's " "shape = %s, softmax_out's shape = %s.", qk_out_grad->dims(), softmax_out->dims())); diff --git a/paddle/fluid/operators/fused/fused_gate_attention_op.cu b/paddle/fluid/operators/fused/fused_gate_attention_op.cu index 8ca6cdb46ccd935af4d85911600ec04da3a1f07a..b4fb69a58226ca1cf304e8f633f83db1e33b8899 100644 --- a/paddle/fluid/operators/fused/fused_gate_attention_op.cu +++ b/paddle/fluid/operators/fused/fused_gate_attention_op.cu @@ -376,7 +376,7 @@ class FusedGateAttentionOpKernel : public framework::OpKernel { true, platform::errors::InvalidArgument( "key is expected to be nullptr or the same as " - "query, but recieved key=%p, query=%p.", + "query, but received key=%p, query=%p.", key, query)); diff --git a/paddle/fluid/operators/kldiv_loss_op.cc b/paddle/fluid/operators/kldiv_loss_op.cc index e45e686dd0eeace04e9f241b69b0fb840f006490..b1a8e81a68add1ae3f589edf781294988d6d2c3f 100644 --- a/paddle/fluid/operators/kldiv_loss_op.cc +++ b/paddle/fluid/operators/kldiv_loss_op.cc @@ -71,7 +71,7 @@ class KLDivLossOpMaker : public framework::OpProtoAndCheckerMaker { While :attr:`reduction` is :attr:`none`, output loss is in the same shape as Input(X), loss in each point is calculated - seperately and no reduction is applied. + separately and no reduction is applied. While :attr:`reduction` is :attr:`mean`, output loss is in shape of [1] and loss value is the mean value of all losses. diff --git a/paddle/fluid/operators/linear_chain_crf_op.cc b/paddle/fluid/operators/linear_chain_crf_op.cc index 26f90851d566a5cde730ceda468c2ed25a43b9ab..ebaa61416f0cd6a43126cc2fe88f858e3da2a799 100644 --- a/paddle/fluid/operators/linear_chain_crf_op.cc +++ b/paddle/fluid/operators/linear_chain_crf_op.cc @@ -52,8 +52,8 @@ class LinearChainCRFOpMaker : public framework::OpProtoAndCheckerMaker { "The forward vectors for the entire batch. Denote it as $\alpha$. " "$\alpha$ is a memo table used to calculate the normalization " "factor in CRF. $\alpha[k, v]$ stores the unnormalized " - "probabilites of all possible unfinished sequences of tags that end at " - "position $k$ with tag $v$. For each $k$, " + "probabilities of all possible unfinished sequences of tags that end " + "at position $k$ with tag $v$. For each $k$, " "$\alpha[k, v]$ is a vector of length $D$ with a component for " "each tag value $v$. This vector is called a forward vecotr and " "will also be used in backward computations.") diff --git a/paddle/phi/infermeta/binary.cc b/paddle/phi/infermeta/binary.cc index 03b20c535af978f66fda28ef62ab0b11c950c84c..6544ec983fa7544e3194dd584180e3ede2b97915 100644 --- a/paddle/phi/infermeta/binary.cc +++ b/paddle/phi/infermeta/binary.cc @@ -1099,7 +1099,7 @@ void DropoutNdInferMeta(const MetaTensor& x, x_dims.size(), phi::errors::InvalidArgument( "The length of axis is expected to be less than or equal to the " - "dimension size of x. But recieved the length of axis is %d, the " + "dimension size of x. But received the length of axis is %d, the " "dimension size of x is %d, x's shape is {%s}.", axis.size(), x_dims.size(), @@ -1111,7 +1111,7 @@ void DropoutNdInferMeta(const MetaTensor& x, phi::errors::InvalidArgument( "The %d-th value of axis is expected to be greater ot " "equal to 0 and less than the dimensions of x. But " - "recieved axis is {%s}, the dimension size of x is %d.", + "received axis is {%s}, the dimension size of x is %d.", i, phi::make_ddim(axis), x_dims.size())); diff --git a/paddle/phi/infermeta/nullary.cc b/paddle/phi/infermeta/nullary.cc index fe83d869ce13892251167f36e1a1ddae428aa926..299f3f0a50c8110fbc2167005877fe163df1f380 100644 --- a/paddle/phi/infermeta/nullary.cc +++ b/paddle/phi/infermeta/nullary.cc @@ -32,7 +32,7 @@ void CreateInferMeta(const IntArray& shape, DataType dtype, MetaTensor* out) { 0, phi::errors::InvalidArgument( "Each value of attribute 'shape' is expected to be no less " - "than 0. But recieved: shape[%u] = %d; shape = [%s].", + "than 0. But received: shape[%u] = %d; shape = [%s].", i, data[i], phi::make_ddim(data))); diff --git a/paddle/phi/kernels/cpu/gather_tree_kernel.cc b/paddle/phi/kernels/cpu/gather_tree_kernel.cc index 250ee1b1e8a2e5a0e8e94f34a146f6e7f6db0022..157496d005bea6455a103e7000a78770109e324f 100644 --- a/paddle/phi/kernels/cpu/gather_tree_kernel.cc +++ b/paddle/phi/kernels/cpu/gather_tree_kernel.cc @@ -54,7 +54,7 @@ void GatherTreeKernel(const Context &dev_ctx, parent, beam_size, phi::errors::InvalidArgument( - "The parents must be less than beam size, but recieved" + "The parents must be less than beam size, but received" "parents %d is greater than or equal to beam size %d. ", parent, beam_size)); diff --git a/paddle/phi/kernels/funcs/broadcast_function.h b/paddle/phi/kernels/funcs/broadcast_function.h index 2c0aad3a6623b2906f8fc9784f8748b49ad2637c..f96a1764c24a5e63db1fbf8e40e79535a5c4f309 100644 --- a/paddle/phi/kernels/funcs/broadcast_function.h +++ b/paddle/phi/kernels/funcs/broadcast_function.h @@ -862,7 +862,7 @@ void BroadcastKernelForDifferentVecSize( kArity, phi::errors::InvalidArgument("The number of inputs is expected to be " "equal to the " - "arity of functor. But recieved: the " + "arity of functor. But received: the " "number of inputs " "is %d, the arity of functor is %d.", ins.size(), diff --git a/paddle/phi/kernels/gpu/gather_tree_kernel.cu b/paddle/phi/kernels/gpu/gather_tree_kernel.cu index 585376ee6688d56b17a66b40a67f193f80bed39d..3ae71992d24236b81c367408edd85884fc18e5c2 100644 --- a/paddle/phi/kernels/gpu/gather_tree_kernel.cu +++ b/paddle/phi/kernels/gpu/gather_tree_kernel.cu @@ -37,7 +37,7 @@ __global__ void GatherTree(const T *ids_data, auto parent = parents_data[idx]; for (int step = max_length - 2; step >= 0; step--) { PADDLE_ENFORCE((parent < beam_size), - "The parents must be less than beam size, but recieved" + "The parents must be less than beam size, but received" "parents %ld is greater than or equal to beam size %ld. ", parent, beam_size); diff --git a/paddle/scripts/README.md b/paddle/scripts/README.md index e9ec3d46a23daba9c671322c0fff85d4a8204864..2c92d0ee1ae8f3ca8d93db938c9728491f167ac3 100644 --- a/paddle/scripts/README.md +++ b/paddle/scripts/README.md @@ -35,7 +35,7 @@ A principle here is that source code lies on the development computer (host) so ### Build Environments -The lastest pre-built build environment images are: +The latest pre-built build environment images are: | Image | Tag | | ----- | --- | diff --git a/python/paddle/distributed/auto_parallel/interface.py b/python/paddle/distributed/auto_parallel/interface.py index 3d11d8cab38b072284ea8d8ed121425a3aabec3e..333ef1d5b162e551b87282c2538979f2ea1d7c15 100644 --- a/python/paddle/distributed/auto_parallel/interface.py +++ b/python/paddle/distributed/auto_parallel/interface.py @@ -119,13 +119,13 @@ def shard_op(op, process_mesh=None, in_shard_specs=None, out_shard_specs=None): in_shard_specs (list of list, optional): a list of list to describe the sharding specifications for the inputs. Each item of `in_shard_specs` is a `shard_spec` between the correspoinding input and `process_mesh`. If one item is None, the cooresponding input is replicated across all processes - If it is None, all inputs are replicated accross all processes. Note that the lenght of the + If it is None, all inputs are replicated across all processes. Note that the lenght of the `in_shard_specs` should be equal to the actual number of inputs when calling this operation. Default: None. out_shard_specs (list of list, optional): a list of list to describe the sharding specifications for the outputs. Each item of `out_shard_specs` is a `shard_spec` between the correspoinding output and `process_mesh`. If one item is None, the cooresponding output is replicated across all processes - If it is None, all outputs are replicated accross all processes. Note that the lenght of the + If it is None, all outputs are replicated across all processes. Note that the lenght of the `in_shard_specs` should be equal to the actual number of inputs when calling this operation. Default: None. Default: None. diff --git a/python/paddle/fluid/dygraph/layers.py b/python/paddle/fluid/dygraph/layers.py index cf6c000ba3f0f58f414be34f7b13fc3c4518ef71..bae77ae54eb774de807e7ee07b48ba5b269bf6b7 100644 --- a/python/paddle/fluid/dygraph/layers.py +++ b/python/paddle/fluid/dygraph/layers.py @@ -67,7 +67,7 @@ def _scope_dist2single(dist_scope): "column_parallel_linear": "linear", "vocab_parallel_embedding": "embedding", # "parallel_cross_entropy": "cross_entropy", while mp_layer has parallel_cross_entropy, - # but there is no parameters so the mapping of parallel_cross_entropy is not neccessary. + # but there is no parameters so the mapping of parallel_cross_entropy is not necessary. } return mapping.get(dist_scope, dist_scope) diff --git a/python/paddle/fluid/tests/unittests/ir/inference/auto_scan_test.py b/python/paddle/fluid/tests/unittests/ir/inference/auto_scan_test.py index 3f3af3be444a9c31f27f1e13742b099f830eb660..6b9669dd1928b5289d86f75dd407584e4954e3f7 100755 --- a/python/paddle/fluid/tests/unittests/ir/inference/auto_scan_test.py +++ b/python/paddle/fluid/tests/unittests/ir/inference/auto_scan_test.py @@ -653,7 +653,7 @@ class TrtLayerAutoScanTest(AutoScanTest): os.getenv('TEST_NUM_PERCENT_CASES', default='1.0') ) - # Use a seperate random generator for skipping tests + # Use a separate random generator for skipping tests self.skip_rng = np.random.default_rng(int(time.strftime("%W"))) def create_inference_config(self, use_trt=True) -> paddle_infer.Config: diff --git a/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py b/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py index ee2b93d3371ee93f3298f5b0252091865c11bbf8..1d7f23a9398290ac44d403405db18ab9e764aeba 100644 --- a/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py +++ b/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py @@ -338,7 +338,7 @@ class TestFusedGateAttentionOp(OpTest): # matmul(x, y, transpose_x=False, transpose_y=True). With different # transpose_x and transpose_y, cublas will launch different kernels # and the result cannot be exactly equal. - # Because the arguments of matmul in einsum is the the same as + # Because the arguments of matmul in einsum are the same as # that in fused ops, check_equal is set to False and we use allclose # to check the correctness. check_equal = False diff --git a/python/paddle/incubate/autograd/primreg.py b/python/paddle/incubate/autograd/primreg.py index 05b7ea78126e4deb8716a64b6964aada3e532952..00e6833639bce60fe1771873c994595890a471cd 100644 --- a/python/paddle/incubate/autograd/primreg.py +++ b/python/paddle/incubate/autograd/primreg.py @@ -147,7 +147,7 @@ def REGISTER_FN(op_type, *position_argnames): Args: op_type(str): The op name - position_argnames(list[str]): Input and ouput names of the op + position_argnames(list[str]): Input and output names of the op Returns: wrapper: Inner wrapper function diff --git a/python/paddle/incubate/nn/layer/fused_transformer.py b/python/paddle/incubate/nn/layer/fused_transformer.py index bd0aaa61212694d537410dc64c3e2ad14c528b79..cbfe95df2062a401830ba1c016a81906b2474903 100644 --- a/python/paddle/incubate/nn/layer/fused_transformer.py +++ b/python/paddle/incubate/nn/layer/fused_transformer.py @@ -121,7 +121,7 @@ class FusedBiasDropoutResidualLayerNorm(Layer): super().__init__() assert embed_dim > 0, ( "Expected embed_dim to be greater than 0, " - "but recieved {}".format(embed_dim) + "but received {}".format(embed_dim) ) self._dtype = self._helper.get_default_dtype() self._bias_attr = bias_attr diff --git a/python/paddle/optimizer/lr.py b/python/paddle/optimizer/lr.py index 9a3a2f72d7bac57b8987f5664ccbb29637d77eda..34ba32947cf075f955ba356c357c434e0686a4a2 100644 --- a/python/paddle/optimizer/lr.py +++ b/python/paddle/optimizer/lr.py @@ -108,7 +108,7 @@ class LRScheduler: def __call__(self): """ - Return lastest computed learning rate on current epoch. + Return latest computed learning rate on current epoch. """ return self.last_lr diff --git a/python/paddle/static/nn/common.py b/python/paddle/static/nn/common.py index 41a2d944d25b846b2c3186712e8ce9fa27bc9474..c7e0794430ef499131425fb553045a1d78c344a9 100644 --- a/python/paddle/static/nn/common.py +++ b/python/paddle/static/nn/common.py @@ -1816,11 +1816,11 @@ def conv3d_transpose( W^\prime_{out} &= (W_{in} − 1) * strides[2] + dilations[2] * (W_f − 1) + 1 If `output_size` is None, :math:`D_{out} = D^\prime_{out}, :math:`H_{out} = \ - H^\prime_{out}, W_{out} = W^\prime_{out}`; else, the specified `output_size_depth` (the depth of the ouput feature layer) :math:`D_{out}` + H^\prime_{out}, W_{out} = W^\prime_{out}`; else, the specified `output_size_depth` (the depth of the output feature layer) :math:`D_{out}` must between :math:`D^\prime_{out}` and :math:`D^\prime_{out} + strides[0]`(not including :math:`D^\prime_{out} + strides[0]`), - the specified `output_size_height` (the height of the ouput feature layer) :math:`H_{out}` must between :math:`H^\prime_{out}` + the specified `output_size_height` (the height of the output feature layer) :math:`H_{out}` must between :math:`H^\prime_{out}` and :math:`H^\prime_{out} + strides[1]`(not including :math:`H^\prime_{out} + strides[1]`), - and the the specified `output_size_width` (the width of the ouput feature layer) :math:`W_{out}` must + and the specified `output_size_width` (the width of the output feature layer) :math:`W_{out}` must between :math:`W^\prime_{out}` and :math:`W^\prime_{out} + strides[2]`(not including :math:`W^\prime_{out} + strides[2]`). Since transposed convolution can be treated as the inverse of convolution, diff --git a/tools/get_single_test_cov.py b/tools/get_single_test_cov.py index 8c900655665c98edebd007e6cce36fa8679e92f7..415d43e0b0109900c8b13073bb1ecbf8c5c2a38f 100644 --- a/tools/get_single_test_cov.py +++ b/tools/get_single_test_cov.py @@ -37,7 +37,7 @@ def getFNDAFile(rootPath, test): os.system('touch %s' % fn_filename) try: f = open(filename) - print("oepn %s succesfully" % filename) + print("oepn %s successfully" % filename) except FileNotFoundError: print("%s is not found." % filename) return @@ -84,7 +84,7 @@ def analysisFNDAFile(rootPath, test): notrelated_ut_map_file ): print( - "make %s and %s succesfully" + "make %s and %s successfully" % (related_ut_map_file, related_ut_map_file) ) else: @@ -96,7 +96,7 @@ def analysisFNDAFile(rootPath, test): fn_filename = '%s/build/ut_map/%s/fnda.tmp' % (rootPath, test) try: f = open(fn_filename) - print("oepn %s succesfully" % fn_filename) + print("oepn %s successfully" % fn_filename) except FileNotFoundError: print("%s is not found." % fn_filename) return @@ -148,7 +148,7 @@ def getBaseFnda(rootPath, test): filename = '%s/build/ut_map/%s/coverage.info.tmp' % (rootPath, test) try: f = open(filename) - print("oepn %s succesfully" % filename) + print("oepn %s successfully" % filename) except FileNotFoundError: print("%s is not found." % filename) symbol_fnda = {} @@ -228,7 +228,7 @@ def getCovinfo(rootPath, test): ) return else: - print("get coverage.info of %s succesfully" % ut_map_path) + print("get coverage.info of %s successfully" % ut_map_path) os.system( "cd %s && lcov --extract coverage.info '/paddle/paddle/phi/*' '/paddle/paddle/utils/*' '/paddle/paddle/fluid/*' '/paddle/build/*' -o coverage.info.tmp --rc lcov_branch_coverage=0 > /dev/null 2>&1" % ut_map_path @@ -239,7 +239,7 @@ def getCovinfo(rootPath, test): print("coverage.info.tmp is empty,collect coverage rate failed") return else: - print("get coverage.info.tmp succesfully") + print("get coverage.info.tmp successfully") os.system('rm -rf %s/paddle' % ut_map_path) os.system('rm -rf %s/coverage.info' % ut_map_path)