Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

support build with Ninja on Linux #44210

Merged
merged 11 commits into from
Aug 1, 2022

Conversation

gglin001
Copy link
Contributor

@gglin001 gglin001 commented Jul 11, 2022

PR types

Others

PR changes

Others

Describe

LInux 支持 ninja 构建

相关的 issue

部分修改来源于

Note

只测试了部分 cmake 配置下可以正常使用 Ninja, 修改不影响继续使用 make 工具, 使用 Ninja 构建可以极大的加速 rebuild 的时间, 提高效率

@paddle-bot paddle-bot bot added contributor External developers status: proposed labels Jul 11, 2022
@paddle-bot
Copy link

paddle-bot bot commented Jul 11, 2022

你的PR提交成功,感谢你对开源项目的贡献!
请关注后续CI自动化测试结果,详情请参考Paddle-CI手册
Your PR has been submitted. Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@gglin001 gglin001 changed the title support build with ninja support build with ninja on Linux Jul 11, 2022
@gglin001 gglin001 changed the title support build with ninja on Linux support build with Ninja on Linux Jul 11, 2022
@Aganlengzi Aganlengzi requested a review from zhwesky2010 July 13, 2022 06:49
@@ -33,7 +33,7 @@ ExternalProject_Add(
URL_MD5 "94e6fa1bc97169d0e1aad44570fe3251"
PREFIX "${DGC_PREFIX_DIR}"
CONFIGURE_COMMAND ""
BUILD_COMMAND make -j $(nproc)
BUILD_COMMAND make -j${NPROC}
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这个make -j 的命令在ninja时是否能编过?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

ninja 下是ok的, 因为 ${NPROC} 会被替换成 数字, build.ninja 里面不会出现 $

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这个编dgc时是会用 make -j${NPROC} 来跑

@@ -61,6 +61,12 @@ else()
CACHE FILEPATH "mkldnn library." FORCE)
endif()

if(UNIX AND NOT APPLE)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

目前Windows、Linux都支持ninja,这里副产物是不是除了mac都适用?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这里是因为 windows 的 CI 有问题, mac 下 我没测试, 所以只在 linux 下设置

详细的log:
https://xly.bce.baidu.com/paddlepaddle/paddle/newipipe/detail/6075099/job/16531188

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

还是改成if(NOT MAC),然后处理下Windows的情形,因为Windows现在也支持ninja

Copy link
Contributor Author

@gglin001 gglin001 Jul 22, 2022

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这里修改的细节见下面的回复, 根据 CI 的结果, windows 下不设置 BUILD_BYPRODUCTS 也是可以正常编译的, 反而设置了该变量后会出现问题

可以在后面对 第三方库的 cmake 配置做统一的修改, 实现:

  1. 不区分平台的统一配置(目前有些区分了 win 和 非 win 平台)
  2. 抛弃 make -j 的写法(使用 cmake 的标准配置)
  3. 更加统一规范的写法, 比如对 ExternalProject_Add() 的使用

一些良好的参考, 比如
https://fuchsia.googlesource.com/cobalt/+/refs/tags/v0.1.3/CMakeLists.txt

@@ -53,7 +53,7 @@ if(NOT WIN32)
PREFIX ${CBLAS_PREFIX_DIR}
INSTALL_DIR ${CBLAS_INSTALL_DIR}
BUILD_IN_SOURCE 1
BUILD_COMMAND make -j$(nproc) ${COMMON_ARGS} ${OPTIONAL_ARGS}
BUILD_COMMAND make -j${NPROC} ${COMMON_ARGS} ${OPTIONAL_ARGS}
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

make 命令 在ninja时能否跑通

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

同上, 我这里测试时是 ok的, 这里和一些其他的一些写死使用 make的 三方库可以在后面慢慢更新成 配置使用 cmake 而不是 make

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这里编openblas,按道理调用了make -j 应该会出错呀

Copy link
Contributor

@tiancaishaonvjituizi tiancaishaonvjituizi Jul 22, 2022

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

为什么会出错呢,这里 paddle 只需要用编译产物,具体编译过程是用 make 还是 ninja 还是裸调用 gcc 都是不关心的

Copy link
Contributor

@zhwesky2010 zhwesky2010 Jul 29, 2022

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

为什么会出错呢,这里 paddle 只需要用编译产物,具体编译过程是用 make 还是 ninja 还是裸调用 gcc 都是不关心的

我是说应该让只安装了ninja 没有安装make的用户能编过,才算完整支持ninja。既然已经设置了CMAKE -GNinja,却还要背后里依赖其他的Generator是不合理的

@tiancaishaonvjituizi
Copy link
Contributor

这个工作真是太伟大了

if(UNIX AND NOT APPLE)
set(BUILD_BYPRODUCTS_OPTION ${MKLDNN_LIB})
else()
set(BUILD_BYPRODUCTS_OPTION "")
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

建议换掉 _OPTION 后缀?option 在 cmake 里有另外的含义

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

fixed, -> BUILD_BYPRODUCTS_ARGS

set(BUILD_COMMAND $(MAKE) --silent)
set(INSTALL_COMMAND $(MAKE) install)
set(BUILD_COMMAND ${CMAKE_MAKE_PROGRAM})
set(INSTALL_COMMAND ${CMAKE_MAKE_PROGRAM} install)
Copy link
Contributor

@tiancaishaonvjituizi tiancaishaonvjituizi Jul 14, 2022

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这里可以用 ${CMAKE_COMMAND} --build .${CMAKE_COMMAND} --install . 吗,因为不能保证所有的工具都是像 make 和 ninja 一样在不带参数执行的时候表示 “build”,也许有某个奇怪的工具一定要 xxx build 才是 build

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

是的, 使用 ${CMAKE_COMMAND} 会是更好的选择, 但是我这边测试时发现使用

  set(BUILD_COMMAND ${CMAKE_COMMAND} --build .)
  set(INSTALL_COMMAND ${CMAKE_COMMAND} --build --target install .)

会报错

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

另外, 还有更多的3rd 库还是在直接使用 make, 对 ninja/其他工具 的支持 需要更多修改/测试

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

上面的 命令顺序写错了, 使用

  set(BUILD_COMMAND ${CMAKE_COMMAND} --build .)
  set(INSTALL_COMMAND ${CMAKE_COMMAND} --build . --target install)

是 ok 的, 其实这里只用 INSTALL_COMMAND 也是足够的

@gglin001
Copy link
Contributor Author

@zhouwei25 hi, 请问这个PR还有其他问题/需要修改的吗?

@zhwesky2010
Copy link
Contributor

@zhouwei25 hi, 请问这个PR还有其他问题/需要修改的吗?

make -j 这种写死了的是不是改成cmake的通用命令会好一些,为啥第三方库能编过呢,可能是既用了ninja也用了make,没有跟随主编译器

@@ -83,7 +89,8 @@ ExternalProject_Add(
-DCMAKE_POSITION_INDEPENDENT_CODE=ON
-DDNNL_BUILD_TESTS=OFF
-DDNNL_BUILD_EXAMPLES=OFF
CMAKE_CACHE_ARGS -DCMAKE_INSTALL_PREFIX:PATH=${MKLDNN_INSTALL_DIR})
CMAKE_CACHE_ARGS -DCMAKE_INSTALL_PREFIX:PATH=${MKLDNN_INSTALL_DIR}
BUILD_BYPRODUCTS ${BUILD_BYPRODUCTS_ARGS})
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这个BUILD_BYPRODUCTS是不是就是 ${MKLDNN_LIB},不用再设置新的变量了

Copy link
Contributor Author

@gglin001 gglin001 Jul 21, 2022

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@gglin001
[fix mkldnn on windows](https://github.com/PaddlePaddle/Paddle/pull/44210/commits/f24ec6e490544b2890346c877be6865270f77e12)
[f24ec6e](https://github.com/PaddlePaddle/Paddle/pull/44210/commits/f24ec6e490544b2890346c877be6865270f77e12)
@gglin001
[fix mkldnn on windows up1](https://github.com/PaddlePaddle/Paddle/pull/44210/commits/644ce589b8d1f53c2a337981227450456bd8f16a)
[644ce58](https://github.com/PaddlePaddle/Paddle/pull/44210/commits/644ce589b8d1f53c2a337981227450456bd8f16a)
@gglin001
[up2](https://github.com/PaddlePaddle/Paddle/pull/44210/commits/f6f206f5aab4da411978bd3627156cc188baa0c8)
[f6f206f](https://github.com/PaddlePaddle/Paddle/pull/44210/commits/f6f206f5aab4da411978bd3627156cc188baa0c8)
@gglin001
[up3](https://github.com/PaddlePaddle/Paddle/pull/44210/commits/0b3402908313bf5f757266673b190c20a8c3e9fa)

可以参考下这几次提交的 CI 失败结果, 这里的 BUILD_BYPRODUCTS 设置成 ${MKLDNN_LIB} 时, windows 的 CI 会失败, 这也是这个修改的原因

@gglin001
Copy link
Contributor Author

@zhouwei25 hi, 请问这个PR还有其他问题/需要修改的吗?

make -j 这种写死了的是不是改成cmake的通用命令会好一些,为啥第三方库能编过呢,可能是既用了ninja也用了make,没有跟随主编译器

是这样, 目前由于一些 第三方库 是写死使用 make -j 的, 会使用 ninja 混合 make

make -j 这种写死了的是不是改成cmake的通用命令会好一些

当然这样是最理想的结果, 我觉得这部分可以放在新的PR里面进行, 由 baidu 内部的开发者来做更加合适一点(因为如果涉及到修改CI/paddle_build.sh, 外部的人做会比较麻烦)

@gglin001
Copy link
Contributor Author

在 mac 下(x86 的cpu) 使用 ninja 可以成功构建, log 文件:

ninja_build.log

@gglin001
Copy link
Contributor Author

gglin001 commented Jul 22, 2022

APPEND: 这个PR 只是基本解决了目前的cmake 配置在使用 ninja 构建的一些小 bug, 对 ninja/其他工具 的更加完善的支持(比如一些第三方库抛弃 make, 某些配置下可能还只是支持 make ) 需要更多修改/测试, 不是这个PR考虑的问题

@paddle-bot-old
Copy link

Sorry to inform you that d689d16's CIs have passed for more than 7 days. To prevent PR conflicts, you need to re-run all CIs manually.

@@ -61,6 +61,12 @@ else()
CACHE FILEPATH "mkldnn library." FORCE)
endif()

if(UNIX AND NOT APPLE)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这个写 if(LINUX)吧,标准一些。

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

fixed

Copy link
Contributor

@zhwesky2010 zhwesky2010 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@zhwesky2010 zhwesky2010 merged commit 1d79f1f into PaddlePaddle:develop Aug 1, 2022
@gglin001 gglin001 deleted the support_build_with_ninja branch August 1, 2022 06:16
xuewujiao added a commit to xuewujiao/Paddle that referenced this pull request Aug 4, 2022
* fix python3.10 compile bug on window (PaddlePaddle#44330)

* Fix random seed for several unit tests (PaddlePaddle#44135)

* Fix test_functional_conv2d_transpose random seed

* Fix random seed and use np.testing

* Fix random seed for test_lu_unpack_op

* Fix test_autograd_functional_dynamic random seed

* Remove boost library (PaddlePaddle#44092)

* add fused token prune op and plugin (PaddlePaddle#44281)

* add fused token prune op and plugin

* Fix run inference bug for standalone executor (PaddlePaddle#44340)

* xpu-paddlepaddle-33 [任务] matmul单测 timeout (PaddlePaddle#44333)

test=kunlun

* [IPU] add custom-op UTs 0/N (PaddlePaddle#44328)

* add custom-op UTs 0

* add authors

Co-authored-by: Allen Guo <[email protected]>
Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>

Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>

* [IPU] add custom-op UTs 1/N (PaddlePaddle#44329)

* add custom-op UTs 1

* add authors

Co-authored-by: Allen Guo <[email protected]>
Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>

* update url

Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>

* support KL2 multi-card training, *test=kunlun (PaddlePaddle#43889)

* update xccl lib
    * use separate streams for compute/comm on XPU
    * add broadcast op to xpu2_op_list

* Remove auto to_pascal_case for args in op generator (PaddlePaddle#44350)

* remove auto to_pascal_case for args in op generator

* fix yaml config

* Standard sparse conv name (PaddlePaddle#44353)

* [Eager] eager variable back sync (PaddlePaddle#44343)

* eager variable back sync

* [ Phi Kernel ] Transfer as_real to phi. (PaddlePaddle#44263)

* transfer as_real to phi

* fix erros

* blocking: True -> False

* [Eager]Fix assert statement (PaddlePaddle#43492)

* Not rename pb file to avoid re-compile (PaddlePaddle#44370)

* [Phi] Migrate solve kernel to phi (PaddlePaddle#44363)

* draft version

* draft version

* draft version

* migrate solve kernel to phi

* polish

* polish

* re useless header file, fix a bug in grad_kernel_impl

* add header file in need

* [auto parallel] remove comm init control (PaddlePaddle#44385)

* [CustomDevice] remove unused file (PaddlePaddle#44358)

* [Paddle-TRT] reshape fill_constant (PaddlePaddle#44314)

* reshape fill_constant

* commit

* commit

* set seed for uts (PaddlePaddle#44372)

* [Paddle-TRT] remove useless code in fc (PaddlePaddle#44382)

* remove useless code in fc

* [Paddle-TRT] Fix cast (PaddlePaddle#44312)

* fix_cast

* fix_cast

* commit

* Polish jit layer cmakelists to hide some message (PaddlePaddle#44351)

* Enable inference multi stream ci test (PaddlePaddle#44275)

* test

* update

* fix bug of old pp (PaddlePaddle#44361)

* add xpu resnet_unit (PaddlePaddle#44297)

* add xpu resnet_unit
*test=kunlun

* tmp
*test=kunlun

* add blacklist in prim2orig interface (PaddlePaddle#44383)

* [Plugin] Fix Custom device in eager mode, test=develop (PaddlePaddle#43952)

* [Plugin] Fix Custom device in eager mode, test=develop

* update test case, test=develop

* update ut for coverage, test=develop

* add ipu support for standalone executor.  (PaddlePaddle#44342)

* fix typos in template for codegen of operators (PaddlePaddle#44364)

* fix duplicate slice logic in _grad (PaddlePaddle#44396)

* [MLU] fix mlu ctest final. (PaddlePaddle#44404)

* fix data transform bug of interpolate op (PaddlePaddle#44401)

* [Sparse] Add sparse matmul kernel(coo*dense->dense) (PaddlePaddle#44346)

* fix new autodiff api docs (PaddlePaddle#44341)

* fix build error in low arch (PaddlePaddle#44391)

* [new api] add new api paddle.vision.ops.distribute_fpn_proposals (PaddlePaddle#43736)

* add distribute_fpn_proposals

* change to new dygraph

* fix doc and example code

* change fluid impl to current version

* update (PaddlePaddle#44418)

* [Paddle-TRT] Shape sum fix scale (PaddlePaddle#44394)

* shape sum

* add shape, sum trt layer

* [Phi] Migrate infermeta and add yaml for solve op (PaddlePaddle#44379)

* migrate solve kernel to phi

* re useless header file, fix a bug in grad_kernel_impl

* add header file in need

* add yaml for solve op

* fix solve_sig.cc ArgumentMapping and update tests case

* disable legacy dygraph check in op_test

* rm solve_op.cc / solve_sig.cc and migrate yaml config

* Update op_test.py

disable legacy dygraph check when check_eager is True

* add labels for infer ut (PaddlePaddle#44279)

* add labels for infer ut

* add RUN_TYPE=INFER for cpp ut

* fix formaterror

* update

* Add mfence for XPU2 KP (PaddlePaddle#44258)

* remove include of all.h in resnet_basic_block_op_xpu.cc, test=kunlun (PaddlePaddle#44423)

* Rename BOOST_GET macros (PaddlePaddle#44368)

* Rename BOOST_GET macros

* Fix conflicts

* [new API] add paddle.vision.ops.generate_proposals (PaddlePaddle#43611)

* add generate_proposals into paddle.vision

* remove class api

* im_info -> img_size

* change fluid impl to current version

* Accelerate inference period in op Cache method (PaddlePaddle#43857)

* Added pad3d and pad2d FP32 FWD oneDNN kernels (PaddlePaddle#43990)

* Piotrek's changes for pad3d

* my changes

* first version of pad3d, single copy, unnecessary reads

* optimized pad3d kernel

* test upadte

* removed magic numbers

* added support for pad2d

* reverted two files

* reverted one old change

* added support for Paddings tensor

* CI fix

* CI fix

* fixed timeout of tests

* fixed typo

* changes to GetKernelTypeForVar

* Revert "changes to GetKernelTypeForVar"

This reverts commit 4691061.

* added AsExtra() to pad2d

Co-authored-by: Piotr Paturej <[email protected]>

* add save_cache/patch (PaddlePaddle#44420)

* add save_cache/patch

* add pybind

* remove pybind

* remove const_cast

* add fleet

* Standard name of sparse pool (PaddlePaddle#44344)

* move eig operator from fluid to phi (PaddlePaddle#44398)

* move eig operator from fluid to phi

* add eig_grad unitest, upgrade IsComplexType() from fluid to phi

* [Phi]Move angle op to phi (PaddlePaddle#44393)

* Move angle op to phi

* Replace mutable_data using Alloc

* Remove some include

* Try to fix windows ci error

* include math.h to fix windows ci error

* Fix kernel name

* Move angle_grad infershape

* [Eager]release gil when run backward (PaddlePaddle#44433)

* release gil when run backward

* compile phi/backends into one static library (PaddlePaddle#44373)

* compile into one static library

* fix xpu compile

* fix xpu compile

* fix inference compile

* fix inference compile

* add custom test

* revert one file

* [IPU] Add more Ops (PaddlePaddle#44414)

* [IPU] Add more Ops

* update boost API

* Clean CI_SKIP_CPP_TEST (PaddlePaddle#44412)

* Add dependency for read op in standalone executor (PaddlePaddle#44362)

* Add dependency for read op in standalone executor

* Fix CI errors

* Add UT

* add_dependency -> dependency_utils

* Fix CI errors

* Add distro in ci docker (PaddlePaddle#44332)

* add distro zstd

* test

* test

* add pip3.8

* [Phi] migrate as_complex kernel to phi (PaddlePaddle#44438)

* migrate as_complex kernel to phi

* support as_complex and as_real in phi

* rm GetExpectedKernelType for AsRealOp

* [GPUPS]FleetWrapper initialize (PaddlePaddle#44441)

* fix FleetWrapper initialize

* [XPU][NPU] (1) add device_guard. (2) add support for LoDTensorArray of sum op. (PaddlePaddle#44367)

* device_guard support xpu. test=kunlun

* sum op of xpu support LoDTensorArray. add test for while op of xpu. test=kunlun.

* [IPU] add Op uts (PaddlePaddle#44415)

* transfer block_id to CreateVarNode in multi_devices_graph_pass (PaddlePaddle#44366)

* fix CreateVarNode in multi_devices_graph_pass

* Revert "Fix var duplication bug for graph_to_program_pass (PaddlePaddle#44278)"

This reverts commit a2c4c86.

* 【GPUPS】Adam accessor (PaddlePaddle#43919)

* add adam/sharedadam optimzier for gpups;edit optimizer struct;test=develop

* [Phi] migrate sync_batch_norm to phi (PaddlePaddle#44369)

* [GPUPS]Fix psgpuwrapper initialization (PaddlePaddle#44468)

* Update ps_gpu_wrapper.h

* Update ps_gpu_wrapper.h

* Update ps_gpu_wrapper.cc

* [Phi] migrate exponential kernel to phi (PaddlePaddle#44376)

* [Phi] migrate exponential kernel to phi

* fix comment

* fix CI

* [PHI] move diag_embed op to phi. (PaddlePaddle#44408)

* move diag_embed to phi.

* [MLU] set_value performance optimizing (PaddlePaddle#44390)

* Update api changing approve members (PaddlePaddle#44463)

* update api approve members, test=document_fix

* add qingqnig into list, test=document_fix

* fix bug,test=document_fix (PaddlePaddle#44478)

* [Phi] migrate clip_by_norm to phi (PaddlePaddle#44458)

* add eigen3 dependency for phi_backends (PaddlePaddle#44479)

* remove fleet_13 ut in parallel_UT_rule.py; test=develop (PaddlePaddle#44477)

* [PHI]Seperate xshape kernel from normal kernel (PaddlePaddle#44315)

* seperate xshape kernel from normal kernel

* fix bugs in infermeta

* fix compile bugs

* fix compile bugs

* [AutoParallel] fix unittest with paddle.distributed.launch (PaddlePaddle#44439)

* fix unittest

* fix log_dir

* _enable_legacy_dygraph

* [Phi] add temporal_shift yaml (PaddlePaddle#44409)

* add temporal_shift yaml and unittest

* [Paddle inference] Add conv_fusion_fp16 (PaddlePaddle#44435)

* convfusionfp16

* convfusionfp16

* convfusionfp16

* fix some convert error found in tipc. (PaddlePaddle#44457)

* fix some error found in tipc.

* update

* [BugFix]Fix randint_like bugs when save program that don't need use tensor's value (PaddlePaddle#44446)

* fix bugs of random

* fix unittest error

* fix unittest bugs

* add adaptive pool and softmax with cross entropy supports different axis, * test = kunlun  (PaddlePaddle#44428)

* add xpu pnorm op and fix pool op, *test=kunlun

* add adaptive pool, and softmax with cross entropy supports different axis, *test=kunlun

* add slot attr for push sparse op (PaddlePaddle#44422)

* add slot attr for push sparse op

* add pybind

* remove fleet

* add unittest

* fix

* [Dy2Sta]Fix Segment Fault while training multi-card if params have no grad (PaddlePaddle#44485)

* [Dy2Sta]Fix Segment Fault while training multi-card if params have no grad

* fix unittest

* fix tensor stream error in custom op (PaddlePaddle#44500)

* Replace with dygraph op calling method. (PaddlePaddle#44331)

* Replace with dygraph op calling method.

* [JitLayer]Pybind PEFunction and call phi api in layer_test (PaddlePaddle#44465)

* Support predictor function in JitLayer

* Pybind PEFunction

* Pybind PEFunction and call phi api in layer_test

* Call sqrt phi API

* Polish flags

* Fix comments

* [Sparse] Add sparse addmm kernel (dense+coo*dense->dense,dense+csr*dense->dense) (PaddlePaddle#44451)

* [Eager] bilinear_tensor_product yaml (PaddlePaddle#44459)

* bilinear_tensor_product yaml

* [ Phi ] svd transfer (PaddlePaddle#44392)

* svd cpu forward

* svd gpu forward

* transfer the backward of svd

* remove cusolver in svd_grad

* svd kernel bug fix

* fix bugs

* fix bugs.

* fix bug

* [Paddle-TRT] fix_fill_constant (PaddlePaddle#44481)

* fix_fill_constant

* fix_fill_constant

* fix_ernie

* [MLU] transpose avg_pool2d to NHWC for better performance. (PaddlePaddle#44475)

* [jit] jit support property.proto (PaddlePaddle#44337)

* add property.proto, can compiled

* property get and deserilize

* support get float

* format code

* format code

* add unittest

* add more set method

* fix grammar error

* Update paddle/fluid/jit/property.h

Co-authored-by: Aurelius84 <[email protected]>

* Update paddle/fluid/jit/property.cc

Co-authored-by: Aurelius84 <[email protected]>

* Update paddle/fluid/jit/property.cc

Co-authored-by: Aurelius84 <[email protected]>

* Update paddle/fluid/jit/property.cc

Co-authored-by: Aurelius84 <[email protected]>

* fix comment

* fix error throw

* fix property save unit test

* fix error info

* fix copyright and header import

* reorder jit property tensor datatype

Co-authored-by: Aurelius84 <[email protected]>

* [ Dy2static ] infer_program may be incorrect in amp mode. (PaddlePaddle#44487)

* fix the outputs of net is x,x

* add unittest for duplicate output

* fix

* fix _infer_program use the original program not the amp program.

* get _***program_id back and avoid duplicate cache
ing

* fix

* Fc fp16 (PaddlePaddle#44505)

* fc support fp16

* add a ‘,’ on paddle_pass_builder.cc

* fc support fp16 on non-cuda.

* add batch stream (PaddlePaddle#44524)

* shufflechannelfix (PaddlePaddle#44516)

* fix arg_max to select first index (PaddlePaddle#44521)

* [MLU] add floor kernel and grid_sampler kernel (PaddlePaddle#44498)

* commit (PaddlePaddle#44534)

* [CustomDevice] register Copy for custom device (PaddlePaddle#44200)

* [CustomDevice] register Copy for custom device

* [CustomDevice] register Copy for custom device

* [CustomDevice] register Copy for custom device

* merge and add uts

* merge and add uts

* fix for blocking and unittests coverage

* (modified) fc support fp16 (PaddlePaddle#44540)

* Add code of occupancy computing on DCU and avoid threadID bug for DCU profiler (PaddlePaddle#44520)

* add xpu lars_momentum/pow2_decay (PaddlePaddle#44448)

*test=kunlun

* [phi] move inverse op from fluid to phi (PaddlePaddle#44471)

* move inverse from fluid to phi with unitest bug

* fix bug, add eager op yaml

* support send_partial, recv_partial and allgather_partial in ProcessGroupNCCL (PaddlePaddle#44444)

* [Sparse]add sparse unary api(expm1/deg2rad/rad2deg/relu6/leaky_relu) (PaddlePaddle#44432)

* Fc fp16 (PaddlePaddle#44558)

* (modified) fc support fp16

* __CUDA_ARCH__ version

* delete half

* delete half

* Fix bug of amp code-gen (PaddlePaddle#44570)

* fix bug of amp code_gen

* fix bug

* [JitLayer]Fix jit.save error when save params combined (PaddlePaddle#44504)

* Fix jit.save error when save params combined

* Change dict_value to list

* [Phi] Migrate squared_l2_norm_op to phi (PaddlePaddle#44492)

* add swish  using TensorRT layer (PaddlePaddle#44561)

* update

* empty commit

* update

* update

* update

* Phi gird sampler migration (PaddlePaddle#44562)

* add_ymal_utest for phi grid_sampler op

* skip dist test cases if mlu card number only one, test=develop (PaddlePaddle#44549)

* [dy2st]Add ProgramHelper to polish build program logic in autoparallel.Engine (PaddlePaddle#44513)

* [dy2st]Add ProgramHelper to polish build program logic in autoparallel.Engine

* refine code

* 【Hackathon No.21】为 Paddle 新增 SoftMarginLoss (PaddlePaddle#42364)

* 2022-04-28

* 2022-04-28_V2

* 2022-04-30

* 2022-04-30_V2

* 2022-05-01

* 2022-05-02

* 2022-05-02_V2

* 2022-05-05_V1

* 2022-05-06_V1

* 2022-05-07_V1

* Update loss.py

* 2022-05-07_V2

* 2022-05-13_V1

* Update test_soft_margin_loss.py

* Update loss.py

* Update loss.py

* 2022-05-16_V1

* 2022-05-19_V1

* 2022-05-20_V1

* Update test_soft_margin_loss.py

* 2022-06-01_V1

* 2022-06-05

* 2022-06-07

* 2022-06-07

* 2022-06-08

* 2022-06-08_V2

* 2022-06-17-code_style

* Modify python

* 2022-06-20

* for

* for CI;test=document_fix

Co-authored-by: Ligoml <[email protected]>

* [MLU]transpose convbpf output to HWCN for better performance (PaddlePaddle#44552)

* Fc fp16 (PaddlePaddle#44578)

* (modified) fc support fp16

* __CUDA_ARCH__ version

* delete half

* delete half

* add half support

* add half support

* add half support

* [Auto Parallel] Add dist op cost (PaddlePaddle#44146)

* update comp cost

* add dist default op cost

* add dist fill constant batch size like op cost

* add elewise op cost

* add fill_constant_batch_size_like op cost unittest

* add unittest and remove fill_constant_batch_size_like grad op cost

* add to cmakelist

* fix unittest bug

* Improve CI unittest parallel execution strategy (PaddlePaddle#44334)

* paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=parallel_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test pre_test_bak

* test cfs

* test_cfs,test=paralle_test_daily

* test_cfs,test=paralle_test_daily

* fix nightly test name,test=paralle_test_daily

* fix nightly test name,test=paralle_test_daily

* test ci parallel speed

* refine parallel rule,test=paralle_test_daily

* Move bmm OP from fluid to phi (PaddlePaddle#44496)

* [PHI]Move slogdeterminant op to phi (PaddlePaddle#44547)

* Move slogdeterminant op to phi

* Add yaml and unit test for slogdeterminant

* Rename pybind_boost_header.h (PaddlePaddle#44592)

* unify data type and property enum value (PaddlePaddle#44585)

* inference multi stream support handle lazy init. (PaddlePaddle#44563)

* multi stream support handle lazy init.

* support eigen lazy init

* update

* fix ci problem

* Remove ControlDepVar in GraphToBlock (PaddlePaddle#44591)

* transfer the svd infer into phi infermeta (PaddlePaddle#44528)

* transfer the svd infer into phi infermeta

* remove the svd.h

* modify svd api

* fix svd error by insert optional

* Einsum grad complex (PaddlePaddle#44598)

* add complex for einsum grad kernel

* pass the ci

* add reverse yaml (PaddlePaddle#44518)

* add reverse yaml

* Set more attrs in ReplaceScaleLossGradOp (PaddlePaddle#44576)

* Set more attrs in ReplaceScaleLossGradOp

* Fix typos

* Fix CI errors

* Add UT

* [Phi] Migrate box coder to phi. (PaddlePaddle#44550)

* fix behavior of device_id=None in Tensor.cuda (PaddlePaddle#44515)

* fix behavior of device_id=None in Tensor.cuda

* fix CI

* fix windows cuda11.7 bug (PaddlePaddle#44601)

* add  horizontal federation learning ps feature (PaddlePaddle#44327)

* back fl

* delete ssl cert

* .

* make warning

* .

* unittest paral degree

* solve unittest

* heter & multi cloud commm ready

* .

* .

* fl-ps v1.0

* .

* support N + N mode

* .

* .

* .

* .

* delete print

* .

* .

* .

* .

* fix bug

* .

* .

* fl-ps with coordinator ready

* merge dev

* update message parse only

* update fl client scheduler

* fix bug

* update multithreads sync

* fix ci errors

* update role_maker.py

* update role_maker.py

* fix ci error: windows py import error

* fix ci error: windows py import error

* fix windows ci pylib import error

* add dump fields & params

* try to fix windows import fleet error

* fix ps FLAGS error

* [MLU] rollback cntoolkit vetsion to 2.8.5 (PaddlePaddle#44595)

* [CustomDevice] add blas_axpby api for gradient_accumulator (PaddlePaddle#44584)

* add sin,cos,exp primitive operators (PaddlePaddle#44345)

* Optimize sparse convolution (PaddlePaddle#43576)

* Merge kProgramDescs in GraphToProgram (PaddlePaddle#44526)

* [Eager] Add warpctc yaml (PaddlePaddle#44617)

* Add a feed op before each input parameter var. (PaddlePaddle#44499)

* Add a feed op before each input parameter var.

* Fix some issues about the unit test build_cinn_pass_test.

* fix record event for operator type in new dygraph (PaddlePaddle#44582)

* fix new dygraph record event for op

* update unit test

* fix bug of elementwise_add_grad, *test=kunlun (PaddlePaddle#44545)

* fix bug of elementwise_add_grad, *test=kunlun

* fix bug, *test=kunlun

* rm pooling_t, *test=kunlun

* fix bug of ew_add_grad when inplace, *test=kunlun

* [IPU] small bug fix (PaddlePaddle#44473)

* sync misc changes

* add authors

Co-authored-by: Zhaorui Chen <[email protected]>

* up x

* Revert "up x"

This reverts commit f3fde45.

* add guarg for ipu

Co-authored-by: Zhaorui Chen <[email protected]>

* support auto fallback to  cpu kernel for cusom device (PaddlePaddle#44639)

* fix dygraph bugs in broadcast_to api. (PaddlePaddle#44612)

* add set_dtype for inverse_op (PaddlePaddle#44618)

* refine overalls.cmake (PaddlePaddle#44623)

* [PHI]Add yaml and unittest for bmm op (PaddlePaddle#44625)

Add yaml and unittest for bmm op

* Phi average accumulates migration (PaddlePaddle#44554)

* move average_accumulates op to phi kernel

* new exe not support pg (PaddlePaddle#44628)

* [CustomDevice]fix phi kernel header (PaddlePaddle#44637)

* [CustomDevice] add process_group_xccl ut (PaddlePaddle#44632)

* [CustomDevice] add process_group_xccl ut

* update

* Fix conv api name (PaddlePaddle#44636)

* [DCU] Fix NAN problem when training BERT on DUC platform (PaddlePaddle#44643)

* [JitLayer]Remove include fluid head files in JitLayer (PaddlePaddle#44597)

* Remove include fluid head files in JitLayer

* Format code

* Remove const to fix ci error

* Fix param error

* Polish jit layer include and cp some headers to python/include

* Fix comment

* [jit]  jit.save support property serialization (PaddlePaddle#44581)

* jit.save support peropty serilization

* extract set property function

* fix property test file name

* fix typing error

* fix typing error

* fix test coverage

* Replaced add_custom_command with add_custom_target in xpu_kp_cmake (PaddlePaddle#44619)

* Replaced add_custom_command with add_custom_target in xpu_kp_cmake

* add adagrad and rmsprop yaml (PaddlePaddle#44631)

* [phi] move crop_tensor kernel from fluid to phi (PaddlePaddle#44574)

* move crop_tensor from fluid to phi

* delete fluid header files

* fix crop_tensor_op dygraph_mode bug

* modify header files, add out tensor check

* fix RemoveIntermediateOut in fuse_elewise_add_act_pass while converting graph to program (PaddlePaddle#44593)

* fix RemoveNode in fuse_elewise_add_act_pass

* fix

* change pointer to share_ptr

* fix

* fix

* fix format

* fix

* fix graph_safe_remove_nodes

* fix UTs on physical ipu (PaddlePaddle#44647)

* [IPU] add more loss ops  (PaddlePaddle#44646)

* add more loss ops

* add authors

Co-authored-by: Zhaorui Chen <[email protected]>

Co-authored-by: Zhaorui Chen <[email protected]>

* add g_ipuplace_pytype (PaddlePaddle#44648)

* Strided slice fp16 (PaddlePaddle#44653)

* [MLU]fix sync_batch_norm and concat_grad op (PaddlePaddle#44586)

* retain dist op returns (PaddlePaddle#44634)

* xpu unittest grad compute supports more types, *test=kunlun (PaddlePaddle#44606)

* [Eager] Add hierarchical_sigmoid yaml (PaddlePaddle#44638)

* add matrix_nms in python/paddle/vision/ops.py (PaddlePaddle#44357)

* [auto parallel] bug fix for op has sub_block attr created with copy_from (PaddlePaddle#44664)

* Change the way to set attributes for grad op maker (PaddlePaddle#44514)

* fix typos in template for codegen of operators
* change the way to set attributes for grad op maker

* [XPU] add top_k op (PaddlePaddle#44656)

* [XPU] add top_k op. test=kunlun

* [XPU] add top_k op. test=kunlun

* use PADDLE_ENFORCE_XDNN_NOT_NULL to check pointer. test=kunlun

* Support broadcast tensor in phi system (PaddlePaddle#44590)

* [PHI] Move spectral_norm to phi (PaddlePaddle#44577)

* Add kernel declarations

* Copy kernel implementation code

* Transfer implementation code

* Fix: Move out_grad to first

* Register new kernels

* Remove old kernels

* Move out_grad to last

* Fix bugs

* Transfer infermeta

* Add yaml files

* Add blank line

* Fix code style

* Optimize directory structure

Co-authored-by: Bobholamovic <[email protected]>

* Complete the dtypes for all_gather, add all_gather_object api (PaddlePaddle#44417)

* [Eager] refactor general_grad and fix some bugs (PaddlePaddle#44611)

* refactor general_grad and fix some bugs

* add TODO: support prune logic deeper

* support log_grad op, *test=kunlun (PaddlePaddle#44662)

* [LAUNCH] add distributed launch check tools (PaddlePaddle#44495)

* add launch test

* launch test for cpu

* bs 1

* Move api(lgamma) from legacy_api.yaml to api.yaml (PaddlePaddle#44355)

* Move api(lgamma) from legacy_api.yaml to api.yaml

* Move api(lgamma) from legacy_api.yaml to api.yaml

* Move api(lgamma) from legacy_api.yaml to api.yaml

* modify code style

* add x to X mapping

* add definition of lgamma

* delete redundant lgamma definitions

* Modify code comments

* Modify ops.py code format

* add lgamma  single test and lgamma api in fluid

* Optimized lgamma unittest

* Move frame kernel to phi (PaddlePaddle#44615)

* Move frame OP to phi、add frame OP yaml config and supplement single test

* add Header file of in_dygraph_mode

* Modify variable name and FrameGradInferMeta multiplex UnchangedInferMeta

* move seq2col to phi

* delete elementwise pow in xpu_kp_list (PaddlePaddle#44661)

* [MLU] fix log_softmax mode selection. (PaddlePaddle#44669)

* adapt for resnet (PaddlePaddle#44685)

* Fix some problem of kernel fallback in C++ API (PaddlePaddle#44681)

* support auto fallback to  cpu kernel for cusom device

* fix some problem of kernel fallback

* fix bugs of lstsq (PaddlePaddle#44689)

* migrate dirichlet kernel to phi (PaddlePaddle#44434)

* migrate dirichlet op kernel to phi

* fix dirichlet sample memory leak

* [phi]move softsign from fluid to phi (PaddlePaddle#44616)

* test_activation_op unitest error, yaml & activation.py in_dygraph_mode incomplete

* fix test_activation_op unitest error, add yaml and dygraph test

* fix code style with pre-commit

* try to fix namespace error of abs in activation_functor.h

* fix namespace error of abs

* [Paddle Inference] Support depthwise_conv2d fp16. (PaddlePaddle#44642)

* depthwise_fp16

* depthwise_fp16

* depthwise_fp16

* depthwise_fp16

* fix logging debug level (PaddlePaddle#44684)

* back fl

* delete ssl cert

* .

* make warning

* .

* unittest paral degree

* solve unittest

* heter & multi cloud commm ready

* .

* .

* fl-ps v1.0

* .

* support N + N mode

* .

* .

* .

* .

* delete print

* .

* .

* .

* .

* fix bug

* .

* .

* fl-ps with coordinator ready

* merge dev

* update message parse only

* update fl client scheduler

* fix bug

* update multithreads sync

* fix ci errors

* update role_maker.py

* update role_maker.py

* fix ci error: windows py import error

* fix ci error: windows py import error

* fix windows ci pylib import error

* add dump fields & params

* try to fix windows import fleet error

* fix ps FLAGS error

* fix logging risk

* fix logging possible risk

* Skip CUDA Graph case for standalone executor (PaddlePaddle#44693)

* [Eager] fix lerp grad kernel logic (PaddlePaddle#44705)

* clone ort_predictor reuse session (PaddlePaddle#44703)

* [XPU] add sampling_id op, add top_k op, update xdnn api. test=kunlun (PaddlePaddle#44704)

* fused_fc_elementwise_layernorm_op support fp16 (PaddlePaddle#44710)

* fused_fc_elementwise_layernorm support fp16

* fused_fc_elementwise_layernorm support double

* [Phi] Add yaml for assign_value (PaddlePaddle#44596)

* [Phi] Add yaml for assign_value

* [Phi] Fix the bug of the assign api and modify the unittest

* [Phi] Fix the bug when the tensor does not have the backend info

* [Phi] Replace the functional-style cast init by the brace-init

* [Phi] Cast the data explicitly

* [PHI] Move lu to phi  (PaddlePaddle#44605)

* Add kernel declarations

* Copy kernel implementation code

* Transfer implementation code

* Register new kernels

* Remove old kernels

* Fix code style

* Fix bugs

* mutable_data->HostAlloc

* Transfer infermeta

* Add yaml and update python api

* Add PADDLE_WITH_HIP check

* Update unittests

* Fix bugs

* Fix bugs

* Optimize directory structure

* Add output checks

* lu_impl.h->lu_kernel_impl.h

Co-authored-by: Bobholamovic <[email protected]>

* [MLU] add pytest for mlu strided_slice kernel (PaddlePaddle#44523)

* Support backward final hook (PaddlePaddle#44686)

* update to sdk2.6.0 (PaddlePaddle#44673)

* move CUDAStream to phi (PaddlePaddle#44529)

* init

* move CUDAStream to phi

* fix compilation

* merge develop

* add stream_owned_ member

* split cuda_stream.h

* fix cpu compile

* fix constructor

* fix bug

* fix windows compile

* fix inference test_levit

* fix windows tests

* [Auto parallel] Optimization Tuning (PaddlePaddle#43782)

* fixed bug for pass & engine

* fixed bug for benchmark GPT-3

* add tuner & profiler

* add algorithms & config

* skip cast trt convert when input dtype is bool (PaddlePaddle#44716)

* skip cast trt convert when input dtype is bool

* [LAUNCH] fix set args bug (PaddlePaddle#44717)

* Phi softplus migration (PaddlePaddle#44542)

* add yaml and utests of phi softplus

add yaml of softplus

fix softplus bug in phi

* update utests

* bug fix

* bug fix for test_layers

* layer api match

* match def and doc in ops.py

* doc polish

* fix unwanted modified of thresholded_relu

* style imporve

* 【PaddlePaddle Hackathon 3 No.15】为 Paddle 新增 count_nonzero (PaddlePaddle#44169)

* add count_nonzero api

* remove grad test

* [WIP] Matmul v1 & v2 unification -- part 1 (PaddlePaddle#44640)

* - Unit tests to be debugged

- fix

- refactor

- diagnostic

- more diagnostic

- fix

- Fix number two

- fix

- fix

- fix

- alpha added

- more fixes

- compilation fix

- removed diagnostic code

- cosmetic fixes

* lint

* add FLAGS_enable_api_kernel_fallback (PaddlePaddle#44706)

* add FLAGS_enable_api_kernel_fallback

* deal with more cases

* add ut for coverage

* phi_multiclass_nms3 (PaddlePaddle#44613)

* add some fp16 op for kunlun resnet50 model (PaddlePaddle#44672)

* add some fp16 op for kunlun resnet50 model
*test=kunlun

* tmp
*test=kunlun

* add dist op costs (PaddlePaddle#44701)

* [API/OP] Migrate Lstsq op into phi (PaddlePaddle#44318)

* migrate lstsq op

* update

* fix bugs for CIs

* update

* fix bugs

* add uts

* update

* update

* update

* fix bugs of jip

* fix bugs of hip

* update

* update according to review

* update

* update

* update

* update

* Add sparse SyncBatchNorm (PaddlePaddle#43520)

* add sparse SyncBatchNorm

* unify fluid::CUDADeviceContext and phi::GpuContext (PaddlePaddle#44723)

* remove cudaDeviceContext

* remove more template

* fix rocm compile

* 【PaddlePaddle Hackathon 3 No.12】为 Paddle 新增 pairwise_distance (PaddlePaddle#44161)

* add paddle.nn.functional.pairwise_distance (cattidea#273)
* remove the test case for undefined behavior

Co-authored-by: SigureMo <[email protected]>

* Phi prior box (PaddlePaddle#44431)

* phi_prior_box

* add float[] support

* phi_prior_box_optest

* update

* ort backend support output mutable data (PaddlePaddle#44724)

* [PHI] Move lu_unpack to phi (PaddlePaddle#44674)

* Add kernel declarations

* Copy kernel implementation code

* Transfer implementation code

* Register new kernels

* Remove old kernels

* Fix code style

* Fix bugs

* mutable_data->HostAlloc

* Transfer infermeta

* Add yaml and update python api

* Add PADDLE_WITH_HIP check

* Update unittests

* Add kernel declarations

* Copy kernel implementation code

* Transfer kernel implementation code

* Register new kernels

* Remove old kernels

* Add lu_unpack_sig

* Fix bugs

* Fix bugs

* Fix bugs

* Optimize directory structure

* Add output checks

* Update include files

* lu_impl.h->lu_kernel_impl.h

* Transfer infermeta

* Add yaml and update python api

* Add check_eager

Co-authored-by: Bobholamovic <[email protected]>

* update document of quantile and nanquantile; test=document_fix (PaddlePaddle#42413)

* migrate reduce_amin,reduce_amax kernel to phi (PaddlePaddle#44698)

* [Paddle Inference] add varlen_token_prune plugin, pass, convert (PaddlePaddle#44733)

* add varlen_token_prune plugin, pass, convert

* support build with Ninja on Linux (PaddlePaddle#44210)

* support ninja

* fix mkldnn on windows

* fix mkldnn on windows up1

* up2

* up3

* fix gflags

* BUILD_BYPRODUCTS_OPTION -> BUILD_BYPRODUCTS_ARGS

* use CMAKE_COMMAND

* up x

* migrate overlap_add and overlap_add_grad op (PaddlePaddle#44739)

* update code format

* add ymal and test

* update for comments

* Fix to CI (PaddlePaddle#44744)

* - fix

* - another fix

* lint

* infer context fix place error. (PaddlePaddle#44726)

* infer context fix place error.

* update

* update

* [operator migration] Migrate unstack_op and nms_op (PaddlePaddle#44424)

* update unstack_op

* update unstack_op

* update unstack_op

* fix unstack test

* update unstack

* update with remote

* fix unstack_test.py

* temp_save_change_nms_op

* add nms test

* update nms fix

* update unstack_op

* temp save change

* finish fix nms_op

* pass nms test

* fix CI

* fix ops test

* save change

* fix code style

* fix code style

* fix ci and codestyle

* fix ci

Co-authored-by: ShiningZhang <[email protected]>

* Update linalg.py (PaddlePaddle#44347)

* Fix test and doc (PaddlePaddle#44735)

* fix test and doc

* fix all_gather_object with various length, test=allcases (PaddlePaddle#44718)

* update manipulation.py paddle.moveaxis (PaddlePaddle#44191)

* [CI] CI for Distributed (PaddlePaddle#44085)

* generate_unify_header supports excludes (PaddlePaddle#44761)

* [JitLayer]Polish PEFuntion to speed up JitLayer and fix memory leak (PaddlePaddle#44738)

* Polish PEFuntion to speed up JitLayer

* Polish PEFunction code

* Fix comments

* paddle2onnx update version to 1.0.0rc2 (PaddlePaddle#44759)

* set parallel_job according to CUDA memory in Windows CI unittest (PaddlePaddle#44695)

* set parallel_job according to CUDA memory

* fix bug: add whitespace between conten and [] or condition wont work

* [Sparse] optimize sparse attention (PaddlePaddle#44743)

* GPUGraph merge to develop (PaddlePaddle#44594)

Co-authored-by: seemingwang <[email protected]>
Co-authored-by: DesmonDay <[email protected]>
Co-authored-by: seemingwang <[email protected]>
Co-authored-by: Thunderbrook <[email protected]>
Co-authored-by: xuewujiao <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: Thunderbrook <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: huwei02 <[email protected]>
Co-authored-by: yaoxuefeng <[email protected]>
Co-authored-by: lxsbupt <[email protected]>
Co-authored-by: miaoli06 <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: chao9527 <[email protected]>
Co-authored-by: qingshui <[email protected]>
Co-authored-by: yangjunchao <[email protected]>

* Revert for cmake static library errors on XPU KP PaddlePaddle#44762

* unify gpu context (PaddlePaddle#44740)

* remove cudaDeviceContext

* remove more template

* fix rocm compile

* remove alias name CUDADeviceContext

* fix compile

* fix tests

* revert changes

* API doc(en) Bugs fix in 第四期体验评估 (PaddlePaddle#44749)

* fix docs(en) bugs;test=document_fix

* update paddle.add docs;test=document_fix

* update paddle.where docs;test=document_fix

* for ci;test=document_fix

* Update manipulation.py

* update paddle.where;test=document_fix

Co-authored-by: Ligoml <[email protected]>

* Modify the output result annotation under the lerp function (PaddlePaddle#44035)

* Refactor build_op_downstream_map for standalone executor (PaddlePaddle#44729)

* Refactor build_op_downstream_map for standalone executor

* Add some comments

* update xpu.cmake to 20220731, test=kunlun (PaddlePaddle#44767)

* fix ut new_group_api (PaddlePaddle#44764)

* support beam_search operator on xpu. test=kunlun (PaddlePaddle#44720)

* support beam_search operator on xpu. test=kunlun

* support beam_search operator on xpu. test=kunlun

* support beam_search operator on xpu. test=kunlun

* support beam_search operator on xpu. test=kunlun

* support beam_search operator on xpu. test=kunlun

* [phi] add yolov3_loss yaml and unittest (PaddlePaddle#44476)

* add yaml and unittest

* update yaml

* update backward yaml and unittest

* update yaml

* add Yolov3LossGradInferMeta

* update yolov3_loss_op.cc

* fix bug

* code format

* Update manipulation.py for rot90() (PaddlePaddle#44038)

* fix compile error;test=develop

* fix compile error;test=develop

* fix compile;test=develop

Co-authored-by: Sing_chan <[email protected]>
Co-authored-by: zlsh80826 <[email protected]>
Co-authored-by: Ruibiao Chen <[email protected]>
Co-authored-by: RichardWooSJTU <[email protected]>
Co-authored-by: taixiurong <[email protected]>
Co-authored-by: Allen Guo <[email protected]>
Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>
Co-authored-by: zhangxiaoci <[email protected]>
Co-authored-by: zyfncg <[email protected]>
Co-authored-by: zhangkaihuo <[email protected]>
Co-authored-by: wanghuancoder <[email protected]>
Co-authored-by: xiongkun <[email protected]>
Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: Leo Chen <[email protected]>
Co-authored-by: Weilong Wu <[email protected]>
Co-authored-by: caozhou <[email protected]>
Co-authored-by: ronnywang <[email protected]>
Co-authored-by: zhoutianzi666 <[email protected]>
Co-authored-by: Haohongxiang <[email protected]>
Co-authored-by: WangZhen <[email protected]>
Co-authored-by: Wilber <[email protected]>
Co-authored-by: ShenLiang <[email protected]>
Co-authored-by: QingshuChen <[email protected]>
Co-authored-by: levi131 <[email protected]>
Co-authored-by: Qi Li <[email protected]>
Co-authored-by: 王明冬 <[email protected]>
Co-authored-by: Feiyu Chan <[email protected]>
Co-authored-by: Xiaoxu Chen <[email protected]>
Co-authored-by: Chenxiao Niu <[email protected]>
Co-authored-by: Zhou Wei <[email protected]>
Co-authored-by: JYChen <[email protected]>
Co-authored-by: YUNSHEN XIE <[email protected]>
Co-authored-by: niuliling123 <[email protected]>
Co-authored-by: zhangyikun02 <[email protected]>
Co-authored-by: huzhiqiang <[email protected]>
Co-authored-by: jakpiase <[email protected]>
Co-authored-by: Piotr Paturej <[email protected]>
Co-authored-by: zhaocaibei123 <[email protected]>
Co-authored-by: freeliuzc <[email protected]>
Co-authored-by: tianshuo78520a <[email protected]>
Co-authored-by: zmxdream <[email protected]>
Co-authored-by: houj04 <[email protected]>
Co-authored-by: pangyoki <[email protected]>
Co-authored-by: lyq <[email protected]>
Co-authored-by: Zhong Hui <[email protected]>
Co-authored-by: fuyou765 <[email protected]>
Co-authored-by: Chen Weihang <[email protected]>
Co-authored-by: YuanRisheng <[email protected]>
Co-authored-by: zhaoyingli <[email protected]>
Co-authored-by: ccrrong <[email protected]>
Co-authored-by: xiaoxiaohehe001 <[email protected]>
Co-authored-by: ykkk2333 <[email protected]>
Co-authored-by: Li Min <[email protected]>
Co-authored-by: Hui Zhang <[email protected]>
Co-authored-by: ming1753 <[email protected]>
Co-authored-by: cifar10 <[email protected]>
Co-authored-by: fwenguang <[email protected]>
Co-authored-by: Aganlengzi <[email protected]>
Co-authored-by: yuguo <[email protected]>
Co-authored-by: Zhang Jun <[email protected]>
Co-authored-by: Wang Bojun <[email protected]>
Co-authored-by: yangguohao <[email protected]>
Co-authored-by: Ligoml <[email protected]>
Co-authored-by: Lux et Veritas <[email protected]>
Co-authored-by: zhangbo9674 <[email protected]>
Co-authored-by: BiynXu <[email protected]>
Co-authored-by: ziyoujiyi <[email protected]>
Co-authored-by: Zhen Wang <[email protected]>
Co-authored-by: chenjian <[email protected]>
Co-authored-by: helen88 <[email protected]>
Co-authored-by: Yuang Liu <[email protected]>
Co-authored-by: qipengh <[email protected]>
Co-authored-by: shangliang Xu <[email protected]>
Co-authored-by: Jiabin Yang <[email protected]>
Co-authored-by: Lin Manhui <[email protected]>
Co-authored-by: Bobholamovic <[email protected]>
Co-authored-by: LiYuRio <[email protected]>
Co-authored-by: kuizhiqing <[email protected]>
Co-authored-by: Charles-hit <[email protected]>
Co-authored-by: HongyuJia <[email protected]>
Co-authored-by: heliqi <[email protected]>
Co-authored-by: Yulong Ao <[email protected]>
Co-authored-by: JZ-LIANG <[email protected]>
Co-authored-by: thunder95 <[email protected]>
Co-authored-by: Jacek Czaja <[email protected]>
Co-authored-by: zhiboniu <[email protected]>
Co-authored-by: Ainavo <[email protected]>
Co-authored-by: SigureMo <[email protected]>
Co-authored-by: Asthestarsfalll <[email protected]>
Co-authored-by: Wangzheee <[email protected]>
Co-authored-by: Thomas Young <[email protected]>
Co-authored-by: ShiningZhang <[email protected]>
Co-authored-by: OccupyMars2025 <[email protected]>
Co-authored-by: mrcangye <[email protected]>
Co-authored-by: Roc <[email protected]>
Co-authored-by: seemingwang <[email protected]>
Co-authored-by: DesmonDay <[email protected]>
Co-authored-by: seemingwang <[email protected]>
Co-authored-by: Thunderbrook <[email protected]>
Co-authored-by: xuewujiao <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: Thunderbrook <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: huwei02 <[email protected]>
Co-authored-by: yaoxuefeng <[email protected]>
Co-authored-by: lxsbupt <[email protected]>
Co-authored-by: miaoli06 <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: chao9527 <[email protected]>
Co-authored-by: qingshui <[email protected]>
Co-authored-by: yangjunchao <[email protected]>
Co-authored-by: yang131313 <[email protected]>
Co-authored-by: mengqingchun02 <[email protected]>
Co-authored-by: 熊峻峰 <[email protected]>
@paddle-bot-old paddle-bot-old bot removed the contributor External developers label Oct 17, 2022
@zlsh80826
Copy link
Collaborator

太棒了,期盼支持 ninja 兩年多了,感謝 @gglin001 的支持和修改!
也感謝 @phlrain 有在後續對 ninja 編譯做了許多修復和優化!

lxsbupt added a commit to lxsbupt/Paddle that referenced this pull request Dec 17, 2022
* fix python3.10 compile bug on window (PaddlePaddle#44330)

* Fix random seed for several unit tests (PaddlePaddle#44135)

* Fix test_functional_conv2d_transpose random seed

* Fix random seed and use np.testing

* Fix random seed for test_lu_unpack_op

* Fix test_autograd_functional_dynamic random seed

* Remove boost library (PaddlePaddle#44092)

* add fused token prune op and plugin (PaddlePaddle#44281)

* add fused token prune op and plugin

* Fix run inference bug for standalone executor (PaddlePaddle#44340)

* xpu-paddlepaddle-33 [任务] matmul单测 timeout (PaddlePaddle#44333)

test=kunlun

* [IPU] add custom-op UTs 0/N (PaddlePaddle#44328)

* add custom-op UTs 0

* add authors

Co-authored-by: Allen Guo <[email protected]>
Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>

Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>

* [IPU] add custom-op UTs 1/N (PaddlePaddle#44329)

* add custom-op UTs 1

* add authors

Co-authored-by: Allen Guo <[email protected]>
Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>

* update url

Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>

* support KL2 multi-card training, *test=kunlun (PaddlePaddle#43889)

* update xccl lib
    * use separate streams for compute/comm on XPU
    * add broadcast op to xpu2_op_list

* Remove auto to_pascal_case for args in op generator (PaddlePaddle#44350)

* remove auto to_pascal_case for args in op generator

* fix yaml config

* Standard sparse conv name (PaddlePaddle#44353)

* [Eager] eager variable back sync (PaddlePaddle#44343)

* eager variable back sync

* [ Phi Kernel ] Transfer as_real to phi. (PaddlePaddle#44263)

* transfer as_real to phi

* fix erros

* blocking: True -> False

* [Eager]Fix assert statement (PaddlePaddle#43492)

* Not rename pb file to avoid re-compile (PaddlePaddle#44370)

* [Phi] Migrate solve kernel to phi (PaddlePaddle#44363)

* draft version

* draft version

* draft version

* migrate solve kernel to phi

* polish

* polish

* re useless header file, fix a bug in grad_kernel_impl

* add header file in need

* [auto parallel] remove comm init control (PaddlePaddle#44385)

* [CustomDevice] remove unused file (PaddlePaddle#44358)

* [Paddle-TRT] reshape fill_constant (PaddlePaddle#44314)

* reshape fill_constant

* commit

* commit

* set seed for uts (PaddlePaddle#44372)

* [Paddle-TRT] remove useless code in fc (PaddlePaddle#44382)

* remove useless code in fc

* [Paddle-TRT] Fix cast (PaddlePaddle#44312)

* fix_cast

* fix_cast

* commit

* Polish jit layer cmakelists to hide some message (PaddlePaddle#44351)

* Enable inference multi stream ci test (PaddlePaddle#44275)

* test

* update

* fix bug of old pp (PaddlePaddle#44361)

* add xpu resnet_unit (PaddlePaddle#44297)

* add xpu resnet_unit
*test=kunlun

* tmp
*test=kunlun

* add blacklist in prim2orig interface (PaddlePaddle#44383)

* [Plugin] Fix Custom device in eager mode, test=develop (PaddlePaddle#43952)

* [Plugin] Fix Custom device in eager mode, test=develop

* update test case, test=develop

* update ut for coverage, test=develop

* add ipu support for standalone executor.  (PaddlePaddle#44342)

* fix typos in template for codegen of operators (PaddlePaddle#44364)

* fix duplicate slice logic in _grad (PaddlePaddle#44396)

* [MLU] fix mlu ctest final. (PaddlePaddle#44404)

* fix data transform bug of interpolate op (PaddlePaddle#44401)

* [Sparse] Add sparse matmul kernel(coo*dense->dense) (PaddlePaddle#44346)

* fix new autodiff api docs (PaddlePaddle#44341)

* fix build error in low arch (PaddlePaddle#44391)

* [new api] add new api paddle.vision.ops.distribute_fpn_proposals (PaddlePaddle#43736)

* add distribute_fpn_proposals

* change to new dygraph

* fix doc and example code

* change fluid impl to current version

* update (PaddlePaddle#44418)

* [Paddle-TRT] Shape sum fix scale (PaddlePaddle#44394)

* shape sum

* add shape, sum trt layer

* [Phi] Migrate infermeta and add yaml for solve op (PaddlePaddle#44379)

* migrate solve kernel to phi

* re useless header file, fix a bug in grad_kernel_impl

* add header file in need

* add yaml for solve op

* fix solve_sig.cc ArgumentMapping and update tests case

* disable legacy dygraph check in op_test

* rm solve_op.cc / solve_sig.cc and migrate yaml config

* Update op_test.py

disable legacy dygraph check when check_eager is True

* add labels for infer ut (PaddlePaddle#44279)

* add labels for infer ut

* add RUN_TYPE=INFER for cpp ut

* fix formaterror

* update

* Add mfence for XPU2 KP (PaddlePaddle#44258)

* remove include of all.h in resnet_basic_block_op_xpu.cc, test=kunlun (PaddlePaddle#44423)

* Rename BOOST_GET macros (PaddlePaddle#44368)

* Rename BOOST_GET macros

* Fix conflicts

* [new API] add paddle.vision.ops.generate_proposals (PaddlePaddle#43611)

* add generate_proposals into paddle.vision

* remove class api

* im_info -> img_size

* change fluid impl to current version

* Accelerate inference period in op Cache method (PaddlePaddle#43857)

* Added pad3d and pad2d FP32 FWD oneDNN kernels (PaddlePaddle#43990)

* Piotrek's changes for pad3d

* my changes

* first version of pad3d, single copy, unnecessary reads

* optimized pad3d kernel

* test upadte

* removed magic numbers

* added support for pad2d

* reverted two files

* reverted one old change

* added support for Paddings tensor

* CI fix

* CI fix

* fixed timeout of tests

* fixed typo

* changes to GetKernelTypeForVar

* Revert "changes to GetKernelTypeForVar"

This reverts commit 4691061.

* added AsExtra() to pad2d

Co-authored-by: Piotr Paturej <[email protected]>

* add save_cache/patch (PaddlePaddle#44420)

* add save_cache/patch

* add pybind

* remove pybind

* remove const_cast

* add fleet

* Standard name of sparse pool (PaddlePaddle#44344)

* move eig operator from fluid to phi (PaddlePaddle#44398)

* move eig operator from fluid to phi

* add eig_grad unitest, upgrade IsComplexType() from fluid to phi

* [Phi]Move angle op to phi (PaddlePaddle#44393)

* Move angle op to phi

* Replace mutable_data using Alloc

* Remove some include

* Try to fix windows ci error

* include math.h to fix windows ci error

* Fix kernel name

* Move angle_grad infershape

* [Eager]release gil when run backward (PaddlePaddle#44433)

* release gil when run backward

* compile phi/backends into one static library (PaddlePaddle#44373)

* compile into one static library

* fix xpu compile

* fix xpu compile

* fix inference compile

* fix inference compile

* add custom test

* revert one file

* [IPU] Add more Ops (PaddlePaddle#44414)

* [IPU] Add more Ops

* update boost API

* Clean CI_SKIP_CPP_TEST (PaddlePaddle#44412)

* Add dependency for read op in standalone executor (PaddlePaddle#44362)

* Add dependency for read op in standalone executor

* Fix CI errors

* Add UT

* add_dependency -> dependency_utils

* Fix CI errors

* Add distro in ci docker (PaddlePaddle#44332)

* add distro zstd

* test

* test

* add pip3.8

* [Phi] migrate as_complex kernel to phi (PaddlePaddle#44438)

* migrate as_complex kernel to phi

* support as_complex and as_real in phi

* rm GetExpectedKernelType for AsRealOp

* [GPUPS]FleetWrapper initialize (PaddlePaddle#44441)

* fix FleetWrapper initialize

* [XPU][NPU] (1) add device_guard. (2) add support for LoDTensorArray of sum op. (PaddlePaddle#44367)

* device_guard support xpu. test=kunlun

* sum op of xpu support LoDTensorArray. add test for while op of xpu. test=kunlun.

* [IPU] add Op uts (PaddlePaddle#44415)

* transfer block_id to CreateVarNode in multi_devices_graph_pass (PaddlePaddle#44366)

* fix CreateVarNode in multi_devices_graph_pass

* Revert "Fix var duplication bug for graph_to_program_pass (PaddlePaddle#44278)"

This reverts commit a2c4c86.

* 【GPUPS】Adam accessor (PaddlePaddle#43919)

* add adam/sharedadam optimzier for gpups;edit optimizer struct;test=develop

* [Phi] migrate sync_batch_norm to phi (PaddlePaddle#44369)

* [GPUPS]Fix psgpuwrapper initialization (PaddlePaddle#44468)

* Update ps_gpu_wrapper.h

* Update ps_gpu_wrapper.h

* Update ps_gpu_wrapper.cc

* [Phi] migrate exponential kernel to phi (PaddlePaddle#44376)

* [Phi] migrate exponential kernel to phi

* fix comment

* fix CI

* [PHI] move diag_embed op to phi. (PaddlePaddle#44408)

* move diag_embed to phi.

* [MLU] set_value performance optimizing (PaddlePaddle#44390)

* Update api changing approve members (PaddlePaddle#44463)

* update api approve members, test=document_fix

* add qingqnig into list, test=document_fix

* fix bug,test=document_fix (PaddlePaddle#44478)

* [Phi] migrate clip_by_norm to phi (PaddlePaddle#44458)

* add eigen3 dependency for phi_backends (PaddlePaddle#44479)

* remove fleet_13 ut in parallel_UT_rule.py; test=develop (PaddlePaddle#44477)

* [PHI]Seperate xshape kernel from normal kernel (PaddlePaddle#44315)

* seperate xshape kernel from normal kernel

* fix bugs in infermeta

* fix compile bugs

* fix compile bugs

* [AutoParallel] fix unittest with paddle.distributed.launch (PaddlePaddle#44439)

* fix unittest

* fix log_dir

* _enable_legacy_dygraph

* [Phi] add temporal_shift yaml (PaddlePaddle#44409)

* add temporal_shift yaml and unittest

* [Paddle inference] Add conv_fusion_fp16 (PaddlePaddle#44435)

* convfusionfp16

* convfusionfp16

* convfusionfp16

* fix some convert error found in tipc. (PaddlePaddle#44457)

* fix some error found in tipc.

* update

* [BugFix]Fix randint_like bugs when save program that don't need use tensor's value (PaddlePaddle#44446)

* fix bugs of random

* fix unittest error

* fix unittest bugs

* add adaptive pool and softmax with cross entropy supports different axis, * test = kunlun  (PaddlePaddle#44428)

* add xpu pnorm op and fix pool op, *test=kunlun

* add adaptive pool, and softmax with cross entropy supports different axis, *test=kunlun

* add slot attr for push sparse op (PaddlePaddle#44422)

* add slot attr for push sparse op

* add pybind

* remove fleet

* add unittest

* fix

* [Dy2Sta]Fix Segment Fault while training multi-card if params have no grad (PaddlePaddle#44485)

* [Dy2Sta]Fix Segment Fault while training multi-card if params have no grad

* fix unittest

* fix tensor stream error in custom op (PaddlePaddle#44500)

* Replace with dygraph op calling method. (PaddlePaddle#44331)

* Replace with dygraph op calling method.

* [JitLayer]Pybind PEFunction and call phi api in layer_test (PaddlePaddle#44465)

* Support predictor function in JitLayer

* Pybind PEFunction

* Pybind PEFunction and call phi api in layer_test

* Call sqrt phi API

* Polish flags

* Fix comments

* [Sparse] Add sparse addmm kernel (dense+coo*dense->dense,dense+csr*dense->dense) (PaddlePaddle#44451)

* [Eager] bilinear_tensor_product yaml (PaddlePaddle#44459)

* bilinear_tensor_product yaml

* [ Phi ] svd transfer (PaddlePaddle#44392)

* svd cpu forward

* svd gpu forward

* transfer the backward of svd

* remove cusolver in svd_grad

* svd kernel bug fix

* fix bugs

* fix bugs.

* fix bug

* [Paddle-TRT] fix_fill_constant (PaddlePaddle#44481)

* fix_fill_constant

* fix_fill_constant

* fix_ernie

* [MLU] transpose avg_pool2d to NHWC for better performance. (PaddlePaddle#44475)

* [jit] jit support property.proto (PaddlePaddle#44337)

* add property.proto, can compiled

* property get and deserilize

* support get float

* format code

* format code

* add unittest

* add more set method

* fix grammar error

* Update paddle/fluid/jit/property.h

Co-authored-by: Aurelius84 <[email protected]>

* Update paddle/fluid/jit/property.cc

Co-authored-by: Aurelius84 <[email protected]>

* Update paddle/fluid/jit/property.cc

Co-authored-by: Aurelius84 <[email protected]>

* Update paddle/fluid/jit/property.cc

Co-authored-by: Aurelius84 <[email protected]>

* fix comment

* fix error throw

* fix property save unit test

* fix error info

* fix copyright and header import

* reorder jit property tensor datatype

Co-authored-by: Aurelius84 <[email protected]>

* [ Dy2static ] infer_program may be incorrect in amp mode. (PaddlePaddle#44487)

* fix the outputs of net is x,x

* add unittest for duplicate output

* fix

* fix _infer_program use the original program not the amp program.

* get _***program_id back and avoid duplicate cache
ing

* fix

* Fc fp16 (PaddlePaddle#44505)

* fc support fp16

* add a ‘,’ on paddle_pass_builder.cc

* fc support fp16 on non-cuda.

* add batch stream (PaddlePaddle#44524)

* shufflechannelfix (PaddlePaddle#44516)

* fix arg_max to select first index (PaddlePaddle#44521)

* [MLU] add floor kernel and grid_sampler kernel (PaddlePaddle#44498)

* commit (PaddlePaddle#44534)

* [CustomDevice] register Copy for custom device (PaddlePaddle#44200)

* [CustomDevice] register Copy for custom device

* [CustomDevice] register Copy for custom device

* [CustomDevice] register Copy for custom device

* merge and add uts

* merge and add uts

* fix for blocking and unittests coverage

* (modified) fc support fp16 (PaddlePaddle#44540)

* Add code of occupancy computing on DCU and avoid threadID bug for DCU profiler (PaddlePaddle#44520)

* add xpu lars_momentum/pow2_decay (PaddlePaddle#44448)

*test=kunlun

* [phi] move inverse op from fluid to phi (PaddlePaddle#44471)

* move inverse from fluid to phi with unitest bug

* fix bug, add eager op yaml

* support send_partial, recv_partial and allgather_partial in ProcessGroupNCCL (PaddlePaddle#44444)

* [Sparse]add sparse unary api(expm1/deg2rad/rad2deg/relu6/leaky_relu) (PaddlePaddle#44432)

* Fc fp16 (PaddlePaddle#44558)

* (modified) fc support fp16

* __CUDA_ARCH__ version

* delete half

* delete half

* Fix bug of amp code-gen (PaddlePaddle#44570)

* fix bug of amp code_gen

* fix bug

* [JitLayer]Fix jit.save error when save params combined (PaddlePaddle#44504)

* Fix jit.save error when save params combined

* Change dict_value to list

* [Phi] Migrate squared_l2_norm_op to phi (PaddlePaddle#44492)

* add swish  using TensorRT layer (PaddlePaddle#44561)

* update

* empty commit

* update

* update

* update

* Phi gird sampler migration (PaddlePaddle#44562)

* add_ymal_utest for phi grid_sampler op

* skip dist test cases if mlu card number only one, test=develop (PaddlePaddle#44549)

* [dy2st]Add ProgramHelper to polish build program logic in autoparallel.Engine (PaddlePaddle#44513)

* [dy2st]Add ProgramHelper to polish build program logic in autoparallel.Engine

* refine code

* 【Hackathon No.21】为 Paddle 新增 SoftMarginLoss (PaddlePaddle#42364)

* 2022-04-28

* 2022-04-28_V2

* 2022-04-30

* 2022-04-30_V2

* 2022-05-01

* 2022-05-02

* 2022-05-02_V2

* 2022-05-05_V1

* 2022-05-06_V1

* 2022-05-07_V1

* Update loss.py

* 2022-05-07_V2

* 2022-05-13_V1

* Update test_soft_margin_loss.py

* Update loss.py

* Update loss.py

* 2022-05-16_V1

* 2022-05-19_V1

* 2022-05-20_V1

* Update test_soft_margin_loss.py

* 2022-06-01_V1

* 2022-06-05

* 2022-06-07

* 2022-06-07

* 2022-06-08

* 2022-06-08_V2

* 2022-06-17-code_style

* Modify python

* 2022-06-20

* for

* for CI;test=document_fix

Co-authored-by: Ligoml <[email protected]>

* [MLU]transpose convbpf output to HWCN for better performance (PaddlePaddle#44552)

* Fc fp16 (PaddlePaddle#44578)

* (modified) fc support fp16

* __CUDA_ARCH__ version

* delete half

* delete half

* add half support

* add half support

* add half support

* [Auto Parallel] Add dist op cost (PaddlePaddle#44146)

* update comp cost

* add dist default op cost

* add dist fill constant batch size like op cost

* add elewise op cost

* add fill_constant_batch_size_like op cost unittest

* add unittest and remove fill_constant_batch_size_like grad op cost

* add to cmakelist

* fix unittest bug

* Improve CI unittest parallel execution strategy (PaddlePaddle#44334)

* paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=parallel_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test=paralle_test_daily

* test pre_test_bak

* test cfs

* test_cfs,test=paralle_test_daily

* test_cfs,test=paralle_test_daily

* fix nightly test name,test=paralle_test_daily

* fix nightly test name,test=paralle_test_daily

* test ci parallel speed

* refine parallel rule,test=paralle_test_daily

* Move bmm OP from fluid to phi (PaddlePaddle#44496)

* [PHI]Move slogdeterminant op to phi (PaddlePaddle#44547)

* Move slogdeterminant op to phi

* Add yaml and unit test for slogdeterminant

* Rename pybind_boost_header.h (PaddlePaddle#44592)

* unify data type and property enum value (PaddlePaddle#44585)

* inference multi stream support handle lazy init. (PaddlePaddle#44563)

* multi stream support handle lazy init.

* support eigen lazy init

* update

* fix ci problem

* Remove ControlDepVar in GraphToBlock (PaddlePaddle#44591)

* transfer the svd infer into phi infermeta (PaddlePaddle#44528)

* transfer the svd infer into phi infermeta

* remove the svd.h

* modify svd api

* fix svd error by insert optional

* Einsum grad complex (PaddlePaddle#44598)

* add complex for einsum grad kernel

* pass the ci

* add reverse yaml (PaddlePaddle#44518)

* add reverse yaml

* Set more attrs in ReplaceScaleLossGradOp (PaddlePaddle#44576)

* Set more attrs in ReplaceScaleLossGradOp

* Fix typos

* Fix CI errors

* Add UT

* [Phi] Migrate box coder to phi. (PaddlePaddle#44550)

* fix behavior of device_id=None in Tensor.cuda (PaddlePaddle#44515)

* fix behavior of device_id=None in Tensor.cuda

* fix CI

* fix windows cuda11.7 bug (PaddlePaddle#44601)

* add  horizontal federation learning ps feature (PaddlePaddle#44327)

* back fl

* delete ssl cert

* .

* make warning

* .

* unittest paral degree

* solve unittest

* heter & multi cloud commm ready

* .

* .

* fl-ps v1.0

* .

* support N + N mode

* .

* .

* .

* .

* delete print

* .

* .

* .

* .

* fix bug

* .

* .

* fl-ps with coordinator ready

* merge dev

* update message parse only

* update fl client scheduler

* fix bug

* update multithreads sync

* fix ci errors

* update role_maker.py

* update role_maker.py

* fix ci error: windows py import error

* fix ci error: windows py import error

* fix windows ci pylib import error

* add dump fields & params

* try to fix windows import fleet error

* fix ps FLAGS error

* [MLU] rollback cntoolkit vetsion to 2.8.5 (PaddlePaddle#44595)

* [CustomDevice] add blas_axpby api for gradient_accumulator (PaddlePaddle#44584)

* add sin,cos,exp primitive operators (PaddlePaddle#44345)

* Optimize sparse convolution (PaddlePaddle#43576)

* Merge kProgramDescs in GraphToProgram (PaddlePaddle#44526)

* [Eager] Add warpctc yaml (PaddlePaddle#44617)

* Add a feed op before each input parameter var. (PaddlePaddle#44499)

* Add a feed op before each input parameter var.

* Fix some issues about the unit test build_cinn_pass_test.

* fix record event for operator type in new dygraph (PaddlePaddle#44582)

* fix new dygraph record event for op

* update unit test

* fix bug of elementwise_add_grad, *test=kunlun (PaddlePaddle#44545)

* fix bug of elementwise_add_grad, *test=kunlun

* fix bug, *test=kunlun

* rm pooling_t, *test=kunlun

* fix bug of ew_add_grad when inplace, *test=kunlun

* [IPU] small bug fix (PaddlePaddle#44473)

* sync misc changes

* add authors

Co-authored-by: Zhaorui Chen <[email protected]>

* up x

* Revert "up x"

This reverts commit f3fde45.

* add guarg for ipu

Co-authored-by: Zhaorui Chen <[email protected]>

* support auto fallback to  cpu kernel for cusom device (PaddlePaddle#44639)

* fix dygraph bugs in broadcast_to api. (PaddlePaddle#44612)

* add set_dtype for inverse_op (PaddlePaddle#44618)

* refine overalls.cmake (PaddlePaddle#44623)

* [PHI]Add yaml and unittest for bmm op (PaddlePaddle#44625)

Add yaml and unittest for bmm op

* Phi average accumulates migration (PaddlePaddle#44554)

* move average_accumulates op to phi kernel

* new exe not support pg (PaddlePaddle#44628)

* [CustomDevice]fix phi kernel header (PaddlePaddle#44637)

* [CustomDevice] add process_group_xccl ut (PaddlePaddle#44632)

* [CustomDevice] add process_group_xccl ut

* update

* Fix conv api name (PaddlePaddle#44636)

* [DCU] Fix NAN problem when training BERT on DUC platform (PaddlePaddle#44643)

* [JitLayer]Remove include fluid head files in JitLayer (PaddlePaddle#44597)

* Remove include fluid head files in JitLayer

* Format code

* Remove const to fix ci error

* Fix param error

* Polish jit layer include and cp some headers to python/include

* Fix comment

* [jit]  jit.save support property serialization (PaddlePaddle#44581)

* jit.save support peropty serilization

* extract set property function

* fix property test file name

* fix typing error

* fix typing error

* fix test coverage

* Replaced add_custom_command with add_custom_target in xpu_kp_cmake (PaddlePaddle#44619)

* Replaced add_custom_command with add_custom_target in xpu_kp_cmake

* add adagrad and rmsprop yaml (PaddlePaddle#44631)

* [phi] move crop_tensor kernel from fluid to phi (PaddlePaddle#44574)

* move crop_tensor from fluid to phi

* delete fluid header files

* fix crop_tensor_op dygraph_mode bug

* modify header files, add out tensor check

* fix RemoveIntermediateOut in fuse_elewise_add_act_pass while converting graph to program (PaddlePaddle#44593)

* fix RemoveNode in fuse_elewise_add_act_pass

* fix

* change pointer to share_ptr

* fix

* fix

* fix format

* fix

* fix graph_safe_remove_nodes

* fix UTs on physical ipu (PaddlePaddle#44647)

* [IPU] add more loss ops  (PaddlePaddle#44646)

* add more loss ops

* add authors

Co-authored-by: Zhaorui Chen <[email protected]>

Co-authored-by: Zhaorui Chen <[email protected]>

* add g_ipuplace_pytype (PaddlePaddle#44648)

* Strided slice fp16 (PaddlePaddle#44653)

* [MLU]fix sync_batch_norm and concat_grad op (PaddlePaddle#44586)

* retain dist op returns (PaddlePaddle#44634)

* xpu unittest grad compute supports more types, *test=kunlun (PaddlePaddle#44606)

* [Eager] Add hierarchical_sigmoid yaml (PaddlePaddle#44638)

* add matrix_nms in python/paddle/vision/ops.py (PaddlePaddle#44357)

* [auto parallel] bug fix for op has sub_block attr created with copy_from (PaddlePaddle#44664)

* Change the way to set attributes for grad op maker (PaddlePaddle#44514)

* fix typos in template for codegen of operators
* change the way to set attributes for grad op maker

* [XPU] add top_k op (PaddlePaddle#44656)

* [XPU] add top_k op. test=kunlun

* [XPU] add top_k op. test=kunlun

* use PADDLE_ENFORCE_XDNN_NOT_NULL to check pointer. test=kunlun

* Support broadcast tensor in phi system (PaddlePaddle#44590)

* [PHI] Move spectral_norm to phi (PaddlePaddle#44577)

* Add kernel declarations

* Copy kernel implementation code

* Transfer implementation code

* Fix: Move out_grad to first

* Register new kernels

* Remove old kernels

* Move out_grad to last

* Fix bugs

* Transfer infermeta

* Add yaml files

* Add blank line

* Fix code style

* Optimize directory structure

Co-authored-by: Bobholamovic <[email protected]>

* Complete the dtypes for all_gather, add all_gather_object api (PaddlePaddle#44417)

* [Eager] refactor general_grad and fix some bugs (PaddlePaddle#44611)

* refactor general_grad and fix some bugs

* add TODO: support prune logic deeper

* support log_grad op, *test=kunlun (PaddlePaddle#44662)

* [LAUNCH] add distributed launch check tools (PaddlePaddle#44495)

* add launch test

* launch test for cpu

* bs 1

* Move api(lgamma) from legacy_api.yaml to api.yaml (PaddlePaddle#44355)

* Move api(lgamma) from legacy_api.yaml to api.yaml

* Move api(lgamma) from legacy_api.yaml to api.yaml

* Move api(lgamma) from legacy_api.yaml to api.yaml

* modify code style

* add x to X mapping

* add definition of lgamma

* delete redundant lgamma definitions

* Modify code comments

* Modify ops.py code format

* add lgamma  single test and lgamma api in fluid

* Optimized lgamma unittest

* Move frame kernel to phi (PaddlePaddle#44615)

* Move frame OP to phi、add frame OP yaml config and supplement single test

* add Header file of in_dygraph_mode

* Modify variable name and FrameGradInferMeta multiplex UnchangedInferMeta

* move seq2col to phi

* delete elementwise pow in xpu_kp_list (PaddlePaddle#44661)

* [MLU] fix log_softmax mode selection. (PaddlePaddle#44669)

* adapt for resnet (PaddlePaddle#44685)

* Fix some problem of kernel fallback in C++ API (PaddlePaddle#44681)

* support auto fallback to  cpu kernel for cusom device

* fix some problem of kernel fallback

* fix bugs of lstsq (PaddlePaddle#44689)

* migrate dirichlet kernel to phi (PaddlePaddle#44434)

* migrate dirichlet op kernel to phi

* fix dirichlet sample memory leak

* [phi]move softsign from fluid to phi (PaddlePaddle#44616)

* test_activation_op unitest error, yaml & activation.py in_dygraph_mode incomplete

* fix test_activation_op unitest error, add yaml and dygraph test

* fix code style with pre-commit

* try to fix namespace error of abs in activation_functor.h

* fix namespace error of abs

* [Paddle Inference] Support depthwise_conv2d fp16. (PaddlePaddle#44642)

* depthwise_fp16

* depthwise_fp16

* depthwise_fp16

* depthwise_fp16

* fix logging debug level (PaddlePaddle#44684)

* back fl

* delete ssl cert

* .

* make warning

* .

* unittest paral degree

* solve unittest

* heter & multi cloud commm ready

* .

* .

* fl-ps v1.0

* .

* support N + N mode

* .

* .

* .

* .

* delete print

* .

* .

* .

* .

* fix bug

* .

* .

* fl-ps with coordinator ready

* merge dev

* update message parse only

* update fl client scheduler

* fix bug

* update multithreads sync

* fix ci errors

* update role_maker.py

* update role_maker.py

* fix ci error: windows py import error

* fix ci error: windows py import error

* fix windows ci pylib import error

* add dump fields & params

* try to fix windows import fleet error

* fix ps FLAGS error

* fix logging risk

* fix logging possible risk

* Skip CUDA Graph case for standalone executor (PaddlePaddle#44693)

* [Eager] fix lerp grad kernel logic (PaddlePaddle#44705)

* clone ort_predictor reuse session (PaddlePaddle#44703)

* [XPU] add sampling_id op, add top_k op, update xdnn api. test=kunlun (PaddlePaddle#44704)

* fused_fc_elementwise_layernorm_op support fp16 (PaddlePaddle#44710)

* fused_fc_elementwise_layernorm support fp16

* fused_fc_elementwise_layernorm support double

* [Phi] Add yaml for assign_value (PaddlePaddle#44596)

* [Phi] Add yaml for assign_value

* [Phi] Fix the bug of the assign api and modify the unittest

* [Phi] Fix the bug when the tensor does not have the backend info

* [Phi] Replace the functional-style cast init by the brace-init

* [Phi] Cast the data explicitly

* [PHI] Move lu to phi  (PaddlePaddle#44605)

* Add kernel declarations

* Copy kernel implementation code

* Transfer implementation code

* Register new kernels

* Remove old kernels

* Fix code style

* Fix bugs

* mutable_data->HostAlloc

* Transfer infermeta

* Add yaml and update python api

* Add PADDLE_WITH_HIP check

* Update unittests

* Fix bugs

* Fix bugs

* Optimize directory structure

* Add output checks

* lu_impl.h->lu_kernel_impl.h

Co-authored-by: Bobholamovic <[email protected]>

* [MLU] add pytest for mlu strided_slice kernel (PaddlePaddle#44523)

* Support backward final hook (PaddlePaddle#44686)

* update to sdk2.6.0 (PaddlePaddle#44673)

* move CUDAStream to phi (PaddlePaddle#44529)

* init

* move CUDAStream to phi

* fix compilation

* merge develop

* add stream_owned_ member

* split cuda_stream.h

* fix cpu compile

* fix constructor

* fix bug

* fix windows compile

* fix inference test_levit

* fix windows tests

* [Auto parallel] Optimization Tuning (PaddlePaddle#43782)

* fixed bug for pass & engine

* fixed bug for benchmark GPT-3

* add tuner & profiler

* add algorithms & config

* skip cast trt convert when input dtype is bool (PaddlePaddle#44716)

* skip cast trt convert when input dtype is bool

* [LAUNCH] fix set args bug (PaddlePaddle#44717)

* Phi softplus migration (PaddlePaddle#44542)

* add yaml and utests of phi softplus

add yaml of softplus

fix softplus bug in phi

* update utests

* bug fix

* bug fix for test_layers

* layer api match

* match def and doc in ops.py

* doc polish

* fix unwanted modified of thresholded_relu

* style imporve

* 【PaddlePaddle Hackathon 3 No.15】为 Paddle 新增 count_nonzero (PaddlePaddle#44169)

* add count_nonzero api

* remove grad test

* [WIP] Matmul v1 & v2 unification -- part 1 (PaddlePaddle#44640)

* - Unit tests to be debugged

- fix

- refactor

- diagnostic

- more diagnostic

- fix

- Fix number two

- fix

- fix

- fix

- alpha added

- more fixes

- compilation fix

- removed diagnostic code

- cosmetic fixes

* lint

* add FLAGS_enable_api_kernel_fallback (PaddlePaddle#44706)

* add FLAGS_enable_api_kernel_fallback

* deal with more cases

* add ut for coverage

* phi_multiclass_nms3 (PaddlePaddle#44613)

* add some fp16 op for kunlun resnet50 model (PaddlePaddle#44672)

* add some fp16 op for kunlun resnet50 model
*test=kunlun

* tmp
*test=kunlun

* add dist op costs (PaddlePaddle#44701)

* [API/OP] Migrate Lstsq op into phi (PaddlePaddle#44318)

* migrate lstsq op

* update

* fix bugs for CIs

* update

* fix bugs

* add uts

* update

* update

* update

* fix bugs of jip

* fix bugs of hip

* update

* update according to review

* update

* update

* update

* update

* Add sparse SyncBatchNorm (PaddlePaddle#43520)

* add sparse SyncBatchNorm

* unify fluid::CUDADeviceContext and phi::GpuContext (PaddlePaddle#44723)

* remove cudaDeviceContext

* remove more template

* fix rocm compile

* 【PaddlePaddle Hackathon 3 No.12】为 Paddle 新增 pairwise_distance (PaddlePaddle#44161)

* add paddle.nn.functional.pairwise_distance (cattidea#273)
* remove the test case for undefined behavior

Co-authored-by: SigureMo <[email protected]>

* Phi prior box (PaddlePaddle#44431)

* phi_prior_box

* add float[] support

* phi_prior_box_optest

* update

* ort backend support output mutable data (PaddlePaddle#44724)

* [PHI] Move lu_unpack to phi (PaddlePaddle#44674)

* Add kernel declarations

* Copy kernel implementation code

* Transfer implementation code

* Register new kernels

* Remove old kernels

* Fix code style

* Fix bugs

* mutable_data->HostAlloc

* Transfer infermeta

* Add yaml and update python api

* Add PADDLE_WITH_HIP check

* Update unittests

* Add kernel declarations

* Copy kernel implementation code

* Transfer kernel implementation code

* Register new kernels

* Remove old kernels

* Add lu_unpack_sig

* Fix bugs

* Fix bugs

* Fix bugs

* Optimize directory structure

* Add output checks

* Update include files

* lu_impl.h->lu_kernel_impl.h

* Transfer infermeta

* Add yaml and update python api

* Add check_eager

Co-authored-by: Bobholamovic <[email protected]>

* update document of quantile and nanquantile; test=document_fix (PaddlePaddle#42413)

* migrate reduce_amin,reduce_amax kernel to phi (PaddlePaddle#44698)

* [Paddle Inference] add varlen_token_prune plugin, pass, convert (PaddlePaddle#44733)

* add varlen_token_prune plugin, pass, convert

* support build with Ninja on Linux (PaddlePaddle#44210)

* support ninja

* fix mkldnn on windows

* fix mkldnn on windows up1

* up2

* up3

* fix gflags

* BUILD_BYPRODUCTS_OPTION -> BUILD_BYPRODUCTS_ARGS

* use CMAKE_COMMAND

* up x

* migrate overlap_add and overlap_add_grad op (PaddlePaddle#44739)

* update code format

* add ymal and test

* update for comments

* Fix to CI (PaddlePaddle#44744)

* - fix

* - another fix

* lint

* infer context fix place error. (PaddlePaddle#44726)

* infer context fix place error.

* update

* update

* [operator migration] Migrate unstack_op and nms_op (PaddlePaddle#44424)

* update unstack_op

* update unstack_op

* update unstack_op

* fix unstack test

* update unstack

* update with remote

* fix unstack_test.py

* temp_save_change_nms_op

* add nms test

* update nms fix

* update unstack_op

* temp save change

* finish fix nms_op

* pass nms test

* fix CI

* fix ops test

* save change

* fix code style

* fix code style

* fix ci and codestyle

* fix ci

Co-authored-by: ShiningZhang <[email protected]>

* Update linalg.py (PaddlePaddle#44347)

* Fix test and doc (PaddlePaddle#44735)

* fix test and doc

* fix all_gather_object with various length, test=allcases (PaddlePaddle#44718)

* update manipulation.py paddle.moveaxis (PaddlePaddle#44191)

* [CI] CI for Distributed (PaddlePaddle#44085)

* generate_unify_header supports excludes (PaddlePaddle#44761)

* [JitLayer]Polish PEFuntion to speed up JitLayer and fix memory leak (PaddlePaddle#44738)

* Polish PEFuntion to speed up JitLayer

* Polish PEFunction code

* Fix comments

* paddle2onnx update version to 1.0.0rc2 (PaddlePaddle#44759)

* set parallel_job according to CUDA memory in Windows CI unittest (PaddlePaddle#44695)

* set parallel_job according to CUDA memory

* fix bug: add whitespace between conten and [] or condition wont work

* [Sparse] optimize sparse attention (PaddlePaddle#44743)

* GPUGraph merge to develop (PaddlePaddle#44594)

Co-authored-by: seemingwang <[email protected]>
Co-authored-by: DesmonDay <[email protected]>
Co-authored-by: seemingwang <[email protected]>
Co-authored-by: Thunderbrook <[email protected]>
Co-authored-by: xuewujiao <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: Thunderbrook <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: huwei02 <[email protected]>
Co-authored-by: yaoxuefeng <[email protected]>
Co-authored-by: lxsbupt <[email protected]>
Co-authored-by: miaoli06 <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: chao9527 <[email protected]>
Co-authored-by: qingshui <[email protected]>
Co-authored-by: yangjunchao <[email protected]>

* Revert for cmake static library errors on XPU KP PaddlePaddle#44762

* unify gpu context (PaddlePaddle#44740)

* remove cudaDeviceContext

* remove more template

* fix rocm compile

* remove alias name CUDADeviceContext

* fix compile

* fix tests

* revert changes

* API doc(en) Bugs fix in 第四期体验评估 (PaddlePaddle#44749)

* fix docs(en) bugs;test=document_fix

* update paddle.add docs;test=document_fix

* update paddle.where docs;test=document_fix

* for ci;test=document_fix

* Update manipulation.py

* update paddle.where;test=document_fix

Co-authored-by: Ligoml <[email protected]>

* Modify the output result annotation under the lerp function (PaddlePaddle#44035)

* Refactor build_op_downstream_map for standalone executor (PaddlePaddle#44729)

* Refactor build_op_downstream_map for standalone executor

* Add some comments

* update xpu.cmake to 20220731, test=kunlun (PaddlePaddle#44767)

* fix ut new_group_api (PaddlePaddle#44764)

* support beam_search operator on xpu. test=kunlun (PaddlePaddle#44720)

* support beam_search operator on xpu. test=kunlun

* support beam_search operator on xpu. test=kunlun

* support beam_search operator on xpu. test=kunlun

* support beam_search operator on xpu. test=kunlun

* support beam_search operator on xpu. test=kunlun

* [phi] add yolov3_loss yaml and unittest (PaddlePaddle#44476)

* add yaml and unittest

* update yaml

* update backward yaml and unittest

* update yaml

* add Yolov3LossGradInferMeta

* update yolov3_loss_op.cc

* fix bug

* code format

* Update manipulation.py for rot90() (PaddlePaddle#44038)

* fix compile error;test=develop

* fix compile error;test=develop

* fix compile;test=develop

Co-authored-by: Sing_chan <[email protected]>
Co-authored-by: zlsh80826 <[email protected]>
Co-authored-by: Ruibiao Chen <[email protected]>
Co-authored-by: RichardWooSJTU <[email protected]>
Co-authored-by: taixiurong <[email protected]>
Co-authored-by: Allen Guo <[email protected]>
Co-authored-by: Zhixin Yao <[email protected]>
Co-authored-by: Zhaorui Chen <[email protected]>
Co-authored-by: zhangxiaoci <[email protected]>
Co-authored-by: zyfncg <[email protected]>
Co-authored-by: zhangkaihuo <[email protected]>
Co-authored-by: wanghuancoder <[email protected]>
Co-authored-by: xiongkun <[email protected]>
Co-authored-by: Aurelius84 <[email protected]>
Co-authored-by: Leo Chen <[email protected]>
Co-authored-by: Weilong Wu <[email protected]>
Co-authored-by: caozhou <[email protected]>
Co-authored-by: ronnywang <[email protected]>
Co-authored-by: zhoutianzi666 <[email protected]>
Co-authored-by: Haohongxiang <[email protected]>
Co-authored-by: WangZhen <[email protected]>
Co-authored-by: Wilber <[email protected]>
Co-authored-by: ShenLiang <[email protected]>
Co-authored-by: QingshuChen <[email protected]>
Co-authored-by: levi131 <[email protected]>
Co-authored-by: Qi Li <[email protected]>
Co-authored-by: 王明冬 <[email protected]>
Co-authored-by: Feiyu Chan <[email protected]>
Co-authored-by: Xiaoxu Chen <[email protected]>
Co-authored-by: Chenxiao Niu <[email protected]>
Co-authored-by: Zhou Wei <[email protected]>
Co-authored-by: JYChen <[email protected]>
Co-authored-by: YUNSHEN XIE <[email protected]>
Co-authored-by: niuliling123 <[email protected]>
Co-authored-by: zhangyikun02 <[email protected]>
Co-authored-by: huzhiqiang <[email protected]>
Co-authored-by: jakpiase <[email protected]>
Co-authored-by: Piotr Paturej <[email protected]>
Co-authored-by: zhaocaibei123 <[email protected]>
Co-authored-by: freeliuzc <[email protected]>
Co-authored-by: tianshuo78520a <[email protected]>
Co-authored-by: zmxdream <[email protected]>
Co-authored-by: houj04 <[email protected]>
Co-authored-by: pangyoki <[email protected]>
Co-authored-by: lyq <[email protected]>
Co-authored-by: Zhong Hui <[email protected]>
Co-authored-by: fuyou765 <[email protected]>
Co-authored-by: Chen Weihang <[email protected]>
Co-authored-by: YuanRisheng <[email protected]>
Co-authored-by: zhaoyingli <[email protected]>
Co-authored-by: ccrrong <[email protected]>
Co-authored-by: xiaoxiaohehe001 <[email protected]>
Co-authored-by: ykkk2333 <[email protected]>
Co-authored-by: Li Min <[email protected]>
Co-authored-by: Hui Zhang <[email protected]>
Co-authored-by: ming1753 <[email protected]>
Co-authored-by: cifar10 <[email protected]>
Co-authored-by: fwenguang <[email protected]>
Co-authored-by: Aganlengzi <[email protected]>
Co-authored-by: yuguo <[email protected]>
Co-authored-by: Zhang Jun <[email protected]>
Co-authored-by: Wang Bojun <[email protected]>
Co-authored-by: yangguohao <[email protected]>
Co-authored-by: Ligoml <[email protected]>
Co-authored-by: Lux et Veritas <[email protected]>
Co-authored-by: zhangbo9674 <[email protected]>
Co-authored-by: BiynXu <[email protected]>
Co-authored-by: ziyoujiyi <[email protected]>
Co-authored-by: Zhen Wang <[email protected]>
Co-authored-by: chenjian <[email protected]>
Co-authored-by: helen88 <[email protected]>
Co-authored-by: Yuang Liu <[email protected]>
Co-authored-by: qipengh <[email protected]>
Co-authored-by: shangliang Xu <[email protected]>
Co-authored-by: Jiabin Yang <[email protected]>
Co-authored-by: Lin Manhui <[email protected]>
Co-authored-by: Bobholamovic <[email protected]>
Co-authored-by: LiYuRio <[email protected]>
Co-authored-by: kuizhiqing <[email protected]>
Co-authored-by: Charles-hit <[email protected]>
Co-authored-by: HongyuJia <[email protected]>
Co-authored-by: heliqi <[email protected]>
Co-authored-by: Yulong Ao <[email protected]>
Co-authored-by: JZ-LIANG <[email protected]>
Co-authored-by: thunder95 <[email protected]>
Co-authored-by: Jacek Czaja <[email protected]>
Co-authored-by: zhiboniu <[email protected]>
Co-authored-by: Ainavo <[email protected]>
Co-authored-by: SigureMo <[email protected]>
Co-authored-by: Asthestarsfalll <[email protected]>
Co-authored-by: Wangzheee <[email protected]>
Co-authored-by: Thomas Young <[email protected]>
Co-authored-by: ShiningZhang <[email protected]>
Co-authored-by: OccupyMars2025 <[email protected]>
Co-authored-by: mrcangye <[email protected]>
Co-authored-by: Roc <[email protected]>
Co-authored-by: seemingwang <[email protected]>
Co-authored-by: DesmonDay <[email protected]>
Co-authored-by: seemingwang <[email protected]>
Co-authored-by: Thunderbrook <[email protected]>
Co-authored-by: xuewujiao <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: Thunderbrook <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: huwei02 <[email protected]>
Co-authored-by: yaoxuefeng <[email protected]>
Co-authored-by: lxsbupt <[email protected]>
Co-authored-by: miaoli06 <[email protected]>
Co-authored-by: root <[email protected]>
Co-authored-by: chao9527 <[email protected]>
Co-authored-by: qingshui <[email protected]>
Co-authored-by: yangjunchao <[email protected]>
Co-authored-by: yang131313 <[email protected]>
Co-authored-by: mengqingchun02 <[email protected]>
Co-authored-by: 熊峻峰 <[email protected]>
@zlsh80826 zlsh80826 mentioned this pull request Dec 18, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants