diff --git a/cmake/third_party.cmake b/cmake/third_party.cmake index 9c8a9e0af1c126b462973cee7a078b7d66f7c1ce..837babea020392a1de6b43e458823f4b8628ee6a 100644 --- a/cmake/third_party.cmake +++ b/cmake/third_party.cmake @@ -210,7 +210,7 @@ list(APPEND third_party_deps extern_eigen3 extern_gflags extern_glog extern_boos list(APPEND third_party_deps extern_zlib extern_dlpack extern_warpctc extern_threadpool) # download file -set(CUDAERROR_URL "https://paddlepaddledeps.bj.bcebos.com/cudaErrorMessage.tar.gz" CACHE STRING "" FORCE) +set(CUDAERROR_URL "http://paddlepaddledeps.bj.bcebos.com/cudaErrorMessage.tar.gz" CACHE STRING "" FORCE) file_download_and_uncompress(${CUDAERROR_URL} "cudaerror") if(WITH_AMD_GPU) diff --git a/python/paddle/fluid/clip.py b/python/paddle/fluid/clip.py index 23a983d845bf26fabff10933f84525d0598f9571..0b9abdf38d82647739b3823482990130f1c94cb6 100644 --- a/python/paddle/fluid/clip.py +++ b/python/paddle/fluid/clip.py @@ -843,6 +843,8 @@ def append_gradient_clip_ops(param_grads): # change wrong mapping relation between param & grad in clip op +# Note: This function is sensitive to the time cost of the network with gradient clipping +# and should not be changed easily. If you must change, please test the time cost. def _correct_clip_op_role_var(params_grads, param_new_grad_name_dict): block_id_list = [] if len(param_new_grad_name_dict) == 0: