Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[cherry-pick2.4]docs fix #47668

Closed
wants to merge 211 commits into from
Closed

[cherry-pick2.4]docs fix #47668

wants to merge 211 commits into from

Commits on Sep 13, 2022

  1. [cherry-pick] Allow manaully set py_reader name in standalone executor (

    #45898) (#45931)
    
    * Allow manaully set py_reader name in standalone executor
    
    * Fix CI errors
    From00 authored Sep 13, 2022
    Configuration menu
    Copy the full SHA
    29c44eb View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    0903020 View commit details
    Browse the repository at this point in the history

Commits on Sep 14, 2022

  1. Configuration menu
    Copy the full SHA
    e223cf7 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    9d5003d View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    20d168d View commit details
    Browse the repository at this point in the history
  4. [chery-pick] Fix namespace error (#45925) (#46029)

    修复cuda11.7编译出错的问题
    engineer1109 authored Sep 14, 2022
    Configuration menu
    Copy the full SHA
    925e84b View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    2fac8ab View commit details
    Browse the repository at this point in the history
  6. cherry pick delay tensorrt log (#45958)

    * cherry pick delay tensorrt log
    * Update trt_plugin.h
    JZZ-NOTE authored Sep 14, 2022
    Configuration menu
    Copy the full SHA
    2ca6590 View commit details
    Browse the repository at this point in the history
  7. merge python lib (#46013)

    JZZ-NOTE authored Sep 14, 2022
    Configuration menu
    Copy the full SHA
    5130b0a View commit details
    Browse the repository at this point in the history

Commits on Sep 15, 2022

  1. fix trt multiclass_nms3 (#45166) (#46034)

    * Support dynamic shape in multiclass_nms3 Plugin for Paddle-TensorRT.
    zhangjun authored Sep 15, 2022
    Configuration menu
    Copy the full SHA
    61a3e30 View commit details
    Browse the repository at this point in the history
  2. Fix arm fp16 compile error (#45991) (#46048)

    * fix arm fp16 compile error
    
    * polish macro impl
    chenwhql authored Sep 15, 2022
    Configuration menu
    Copy the full SHA
    91677eb View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    526e032 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    9012e8b View commit details
    Browse the repository at this point in the history
  5. Delete eigen header in data_type.h (#46036) (#46066)

    * delete eigen header in data_type.h
    
    * fix complie bug
    
    * refactor
    zyfncg authored Sep 15, 2022
    Configuration menu
    Copy the full SHA
    2680a71 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    0793311 View commit details
    Browse the repository at this point in the history
  7. Support 0 shapes input Tensor for MKL slice (#45930) (#46072)

    Support 0 shapes input Tensor for MKL slice kernel
    0x45f authored Sep 15, 2022
    Configuration menu
    Copy the full SHA
    903c87b View commit details
    Browse the repository at this point in the history
  8. [ Dy2Static ] Fix bugs when select inputs meeting different shape or …

    …undefined-var (#45916) (#46020)
    
    * fix select_input with different shape errors:
    1. select_input_with_buildin_type directly return non-undefinedvar branch when meeting undefined var
    2. the output shape of select_input is inferred from inputs.
    
    * reverse the logic in select_input
    2742195759 authored Sep 15, 2022
    Configuration menu
    Copy the full SHA
    0048695 View commit details
    Browse the repository at this point in the history

Commits on Sep 16, 2022

  1. [Cherry-pick] Normalize yaml name and label (#46052)

    * normalize yaml file name (#45894)
    
    * Clear extra attributes of activation op in OpMaker (#45772)
    
    * clear extra attr of activation op in opmaker
    
    * fix syntax bug
    
    * fix mkldnn kernel
    
    * fix merge conflict
    
    * fix bug
    
    * [PHI] Normalize yaml op label (#45976)
    
    * normalize yaml op label
    
    * revert op_compat yaml change
    
    * fix prelu and rnn compat problem
    
    * replace api by op
    
    * support assign op backward refuse forward (#45879)
    
    * normize yaml backward op label (#46028)
    
    Co-authored-by: zyfncg <zhangyunfei07@baidu.com>
    Co-authored-by: Charles-hit <56987902+Charles-hit@users.noreply.github.com>
    3 people authored Sep 16, 2022
    Configuration menu
    Copy the full SHA
    8caaf85 View commit details
    Browse the repository at this point in the history
  2. [cherry-pick][jit] Jit skip forward (#45926)

    * skip forward save
    
    * fix bug
    
    * more ci for jit skip forward
    zh794390558 authored Sep 16, 2022
    Configuration menu
    Copy the full SHA
    e25e947 View commit details
    Browse the repository at this point in the history
  3. (cherry-pick)Fix split infershape in static mode and add convert rule…

    …s for fill_any_like op (#46079)
    
    * Fix split bug in static mode (#45906)
    
    * fix split bug in static mode
    
    * modify code style
    
    * modify code style
    
    * add unit test for split
    
    * add convert rules for fill_any_like op in paddle science (#45985)
    
    * add convert rules for fill_any_like op in paddle science
    
    * add unit test for fill_any_like op in paddle science
    
    * modify fill_any_like convert rule
    
    * modify fill_any_like convert rule dtype
    Charles-hit authored Sep 16, 2022
    Configuration menu
    Copy the full SHA
    4e09e40 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    2c2cd79 View commit details
    Browse the repository at this point in the history

Commits on Sep 17, 2022

  1. V2.4 - cherry-pick (#46126)

    * back fl
    
    * delete ssl cert
    
    * .
    
    * make warning
    
    * .
    
    * unittest paral degree
    
    * solve unittest
    
    * heter & multi cloud commm ready
    
    * .
    
    * .
    
    * fix gloo compile warning
    
    * adapt for nn fl-ps
    ziyoujiyi authored Sep 17, 2022
    Configuration menu
    Copy the full SHA
    a76fa41 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    f6dd201 View commit details
    Browse the repository at this point in the history

Commits on Sep 19, 2022

  1. Unify core avx and core_noavx to libpaddle (#46095) (#46113)

    * unify  core_avx and core_noavx
    
    * fix except error
    
    * revert mac compile logic
    
    * revert dylib to so
    
    * add core_noavx branch
    
    * remove core_noavx
    
    * replace paddle_core by lib paddle
    
    * polish var name
    
    * replace paddle_core by libpaddle
    
    * update custom device commit
    
    * polish code by comments
    chenwhql authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    4261ae3 View commit details
    Browse the repository at this point in the history
  2. [Cherry-pick] fix bug for TransformedDistribution (#46157)

    fix bug for TransformedDistribution
    MayYouBeProsperous authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    a5d4f57 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    7a6db0a View commit details
    Browse the repository at this point in the history
  4. Revert "Simplify size op impl (#45808)" (#46168)

    This reverts commit c252b1d.
    chenwhql authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    dabb8f2 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    f4ec156 View commit details
    Browse the repository at this point in the history
  6. [Eager] Optimize log (#45783) (#46133)

    * make eager log readable
    
    * fix compile error
    
    * recover test
    
    * invoke ci again
    JiabinYang authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    e468e93 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    860f607 View commit details
    Browse the repository at this point in the history
  8. [Cherry-pick][Auto Parallel] Improve the APIs (#46164)

    * [AutoParallel] adapt gradient merge pass (#45915)
    
    * adapt gradient merge
    
    * fix op_role
    
    * fix strategy
    
    * [Auto Parallel] Gradient Fuse Allreduce (#45643)
    
    * bugfix (#45332)
    
    * dist embedding support lookup table v1
    
    * add unitest
    
    * customize wait_comm
    
    * group gradients
    
    * bugfix
    
    * update program
    
    * [Auto Parallel] Improve the APIs (#45776)
    
    * [Auto Parallel] Use c++ dist attr in the completion process
    
    * [Auto Parallel] Add minor changes
    
    * [Auto Parallel] Use c++ dist attr in the completion process
    
    * [Auto Parallel] Add minor changes
    
    * [Auto Parallel] Add the serialization process for dist attrs
    
    * [Auto Parallel] Remove unnecessary comments
    
    * [Auto Parallel] Fix some bugs
    
    * [Auto Parallel] Fix the code style
    
    * [Auto Parallel] Remove unnecessary impls
    
    * [Auto Parallel] Fix the importing error
    
    * [Auto Parallel] Fix the copy from bugs of op dist attr
    
    * [Auto Parallel] Replace the use of constexpr if
    
    * [Auto Parallel] Redesign the shard_tensor, shard_op and ProcessMesh
    
    * [Auto Parallel] Change API of the completion unittest
    
    * [Auto Parallel] Fix the bug when set_attr an int
    
    * [Auto Parallel] Add the unittest for the serialization
    
    * [Auto Parallel] Add some unit tests
    
    * [Auto Paralle] Unify the strategy
    
    * [Auto Parallel] Improve the engine api
    
    * [Auto Parallel] Reset the changes made to the framework
    
    * [Auto Parallel] Change the engine unittest
    
    * [Auto Parallel] Update API of the completion and partitioner
    
    * [Auto Parallel] Update unit tests using engine api
    
    * update shard annotation
    
    * [Auto Parallel] Remove the modifications of other modules
    
    * [Auto Parallel] Add docs for APIs
    
    * add new strategy
    
    * [Auto Parallel] Replace the logger
    
    * [Auto Parallel] Restore the test_program.py
    
    * [Auto Parallel] Change the import rules
    
    * [Auto Parallel] Add the examples for Engine
    
    * [Auto Parallel] Do some minor changes
    
    * [Auto Parallel] Remove yaml dependency
    
    * [Auto Parallel] Fix the unittests
    
    * add valid after train
    
    * bug fix
    
    Co-authored-by: zhaoyingli <zhaoyingli@baidu.com>
    Co-authored-by: caozhou <caozhou@radi.ac.cn>
    Co-authored-by: caozhou <48191911+Caozhou1995@users.noreply.github.com>
    
    * [Auto Parallel] Bugfix allreduce fuse for MP (#46086)
    
    * bugfix
    
    * bugfix
    
    * typos fixed
    
    * update strategy (#46138)
    
    Co-authored-by: zhaoyingli <86812880+zhaoyinglia@users.noreply.github.com>
    Co-authored-by: JZ-LIANG <jianzhongliang10@gmail.com>
    Co-authored-by: zhaoyingli <zhaoyingli@baidu.com>
    Co-authored-by: caozhou <caozhou@radi.ac.cn>
    Co-authored-by: caozhou <48191911+Caozhou1995@users.noreply.github.com>
    6 people authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    c5cc427 View commit details
    Browse the repository at this point in the history
  9. refactor mp. (#45803) (#46121)

    * refactor mp.
    
    * update setup.py.
    
    * update mp_layers.py for compatibility.
    
    * add documents for mp_layers.py
    
    * update init.py
    
    * update collective.py.
    
    * update.
    
    * update mp_ops.py
    
    * update.
    
    * update code style.
    
    * update code style.
    wuhuachaocoding authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    e5dc9d6 View commit details
    Browse the repository at this point in the history
  10. Add INT8 support for fused_multi_transformer_op (#45284) (#46169)

    Co-authored-by: RichardWooSJTU <37864677+RichardWooSJTU@users.noreply.github.com>
    minghaoBD and RichardWooSJTU authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    db368d5 View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    6b59a07 View commit details
    Browse the repository at this point in the history
  12. fix (#46125)

    zhaocaibei123 authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    855fdde View commit details
    Browse the repository at this point in the history
  13. [Cherry-pick] Support bmm and bmm_grad in xpu (#45887) (#46132)

    * [PHI] Support bmm and bmm_grad in xpu (#45887)
    
    * support bmm and bmm_grad in xpu
    
    * add error removal
    
    * test=kunlun
    
    * refactor code for better structure
    
    * test=kunlun
    
    * add fp16 kernel for bmm
    
    * test=kunlun
    
    * test=kunlun
    JiabinYang authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    1c7e95c View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    7f0c1f0 View commit details
    Browse the repository at this point in the history
  15. (cherry-pick)support some op backward refuse forward (#46201)

    * add unit test for sum higher level op (#45961)
    
    * support slice op backward refuse forward and add high level unit test (#45960)
    
    * support tile op backward refuse forward (#45942)
    
    * support expand_v2 op backward refuse forward (#45941)
    
    * support concat backward refuse forward (#45940)
    Charles-hit authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    adab3c5 View commit details
    Browse the repository at this point in the history
  16. cherry-pick 46152 (#46183)

    jiweibo authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    707d838 View commit details
    Browse the repository at this point in the history
  17. Configuration menu
    Copy the full SHA
    a056601 View commit details
    Browse the repository at this point in the history
  18. [cherry-pick] [dy2static] support user to use decorator in their prog…

    …ram (#46194)
    
    * [dy2static] support user to use decorator in their program (#45768)
    
    * support deco
    
    * fix deco ast type
    
    * arg_str
    
    * 1
    
    * support callable deco
    
    * code style
    
    * codestyle
    
    * test_error
    
    * fix decos in another file
    
    * recover conflict codes
    
    * [BugFix] fixed a bug in decorator transformer, it can not analyze decorator with params correctly (#46055)
    
    * fix deco call
    
    * add raise
    
    * add test
    
    * add warn, fix paddle api
    
    * fix error type
    
    * fix coverage
    feifei-111 authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    d1ce974 View commit details
    Browse the repository at this point in the history
  19. Configuration menu
    Copy the full SHA
    45a3c65 View commit details
    Browse the repository at this point in the history
  20. [cherry-pick] add abs,mean,sum,ge,gt,pow,etc higher-order differentia…

    …tion operators (#46184)
    
    * [cherry-pick] extend reduce_sum,reduce_sum,eq,ne,ge,abs,pow,etc higher order operators
    
    * add reduce_mean,reduce_sum primitive ops
    * add ne_p gt_p primitive operators
    * add ge_p abs_p primitive oparators
    * add cast primitive operators
    * add pow,square prim2oirg rules
    * add elementwise_div orig2prim rule
    
    * [cherry-pick] add mean,sum,ge,gt,ne,abs,etc higher-order differentiation operators(#45888)
    
    * add reduce_mean,reduce_sum primitive ops
    
    * add ne_p gt_p primitive operators
    
    * add ge_p abs_p primitive oparators
    cxxly authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    ad8beaa View commit details
    Browse the repository at this point in the history
  21. [vision.ops.nms] Fix return order error and duplicate results with sp…

    …ecific inputs (#46148) (#46193)
    
    * fix return order error and duplicate results with specific inputs
    RichardWooSJTU authored Sep 19, 2022
    Configuration menu
    Copy the full SHA
    be84cac View commit details
    Browse the repository at this point in the history
  22. Configuration menu
    Copy the full SHA
    4bced24 View commit details
    Browse the repository at this point in the history

Commits on Sep 20, 2022

  1. Fix wrong eigen header include (#46082) (#46202)

    * fix wrong eigen header include
    
    * fix complie bug
    
    * fix nan_inf_utils_detail
    
    * fix resource_manager
    
    * fix conv_miopen_helper
    zyfncg authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    ac8cce2 View commit details
    Browse the repository at this point in the history
  2. [Cherry-pick] Fix static check (#46253)

    * fix static_check error when compile twice (#46140)
    
    * [CI] fix static check in build_pr_dev (#46192)
    
    Co-authored-by: Zhou Wei <1183042833@qq.com>
    sljlp and zhwesky2010 authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    7712ce1 View commit details
    Browse the repository at this point in the history
  3. [Cherry-pick] Update layoutautotune for inplace (#45826) (#46226)

    cherry-pick from #45826
    LayoutAutotune 支持 inplace 类型的OP
     根据 Add eager layout autotune #45409 修改意见调整UseAutotune
    将LayoutAutotune判断放到controller中,与AMP 判断保持一致
    AnnaTrainingG authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    c0324e8 View commit details
    Browse the repository at this point in the history
  4. Run_program_op add scope cache & reuse (#45813) (#46223)

    * add scope cache & reuse
    
    * add gc scope for end of each train step
    
    * del scope reuse for jit
    
    * refine code
    
    * test
    zhangbo9674 authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    4f28a4c View commit details
    Browse the repository at this point in the history
  5. [Inference] fix preln_residual_bias_fuse_pass bug in TNT_small model (#…

    …46178) (#46260)
    
    * fix preln_residual_bias_fuse_pass bug in TNT_small model
    zhoutianzi666 authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    c384b00 View commit details
    Browse the repository at this point in the history
  6. (cherry-pick)Support some op refuse forward and fix some bugs (#46211)

    * support cast op backward refuse forward and fix some bugs (#46173)
    
    * support cast op backward refuse forward
    
    * Fix the bug of high order unit test framework
    
    * support sign op backward refuse forward (#46002)
    Charles-hit authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    bc92d5f View commit details
    Browse the repository at this point in the history
  7. [Eager] Fix linspace error in amp (#46088) (#46206)

    * fix linspace error in amp
    
    * fix log
    
    * fix amp error
    JiabinYang authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    38c0fd0 View commit details
    Browse the repository at this point in the history
  8. [Cherry-pick] Sparse add InferMeta (#46235)

    cherry-pick : #46016, #46021, #45974
    
    * [Sparse]Sparse add support gpu (#45974)
    
    * [Sparse]Remove unused code (#46021)
    
    * [Sparse] Add infer meta (#46016)
    zhangkaihuo authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    fd8ec4a View commit details
    Browse the repository at this point in the history
  9. Fix TransDataBackend Error when call unsqueeze using MKL Tensor (#46094

    …) (#46186)
    
    * Fix TransDataBackend Error when call unsqueeze using MKL Tensor
    
    * Add UT
    
    * Refine UT
    0x45f authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    5034030 View commit details
    Browse the repository at this point in the history
  10. [cherry-pick] Refine thread pool config of interpretercore (#46219)

    * add config
    
    * add config
    
    * follow comments
    
    * fix serial run
    zhiqiu authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    1418a71 View commit details
    Browse the repository at this point in the history
  11. [Paddle-TRT] Support matmul_v2 in Paddle-TensorRT (#46177)

    * Support matmul_v2 in Paddle-TensorRT converter.
    zhoutianzi666 authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    654807c View commit details
    Browse the repository at this point in the history
  12. [Release/2.4][Cherry-pick] Fix bug of reduce_sum op (#46160)

    * Fix bug of reduce_sum op. When input.numel() > INT32_MAX,
    its result is wrong.
    
    * Cherry-pick of PR 46045
    
    * Fix bug of reduce_sum kp op.
    
    * Fix bug of reduce_sum kp operator compilation.
    If compilation device is XPU, eigen kernel should be ignored.
    GhostScreaming authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    759736d View commit details
    Browse the repository at this point in the history
  13. [Cherry-pick] Fix amp error cp (#46272)

    * [Eager] Fix ocr (#46124)
    
    * fix linspace error in amp
    
    * fix log
    
    * fix amp error
    
    * fix ocr error which caused by amp
    
    * add more check
    
    * rename dtype ns
    
    * [Eager Bug fix]Fix Detection (#46147)
    
    * fix linspace error in amp
    
    * fix log
    
    * fix amp error
    
    * Revert "Simplify size op impl (#45808)"
    
    This reverts commit c252b1d.
    
    * fix_seg
    
    * fix detection
    
    Co-authored-by: Chen Weihang <sunny_cwh@163.com>
    
    Co-authored-by: Chen Weihang <sunny_cwh@163.com>
    JiabinYang and chenwhql authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    da173c4 View commit details
    Browse the repository at this point in the history
  14. [Cherry-Pick][AutoParallel] change import way and fix strategy (#46270)

    * [Auto Parallel] Change the import way of Auto Parallel (#46115)
    
    * fix strategy (#46256)
    
    * [Auto Parallel] performance improvement for Sharding-DP hybrid parallelism (#46180)
    
    * remove no need grad allreduce communication when sharding-dp
    
    * remove no need grad allreduce communication when sharding-dp
    
    * bugfix
    
    * bugfix
    
    * bugfix
    
    Co-authored-by: Yulong Ao <aoyulong@baidu.com>
    Co-authored-by: JZ-LIANG <jianzhongliang10@gmail.com>
    3 people authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    c43ebfc View commit details
    Browse the repository at this point in the history
  15. [PolishComments] Polish some code comments (#46032) (#46261)

    * polish code comments
    
    * polish data_device_transform.cc
    jiahy0825 authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    42e56f6 View commit details
    Browse the repository at this point in the history
  16. [Paddle-TRT] Full support for ops with persistable input (#45545) (#4…

    …6280)
    
    * Move ITensor construction for Weight (persistable variable) from OpConvert to TensorRTEngine.
    zhoutianzi666 authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    adb2f5e View commit details
    Browse the repository at this point in the history
  17. [cherry-pick][xpu] update xdnn activations (#46282)

    * [XPU] update xdnn activations. (#46246)
    
    * [XPU] update xpu cmake. test=kunlun
    houj04 authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    a43f960 View commit details
    Browse the repository at this point in the history
  18. [Paddle-TRT][Cherry-Pick]Fix cast bug (#46293)

    * fix cast bug
    zhoutianzi666 authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    230b9a8 View commit details
    Browse the repository at this point in the history
  19. Configuration menu
    Copy the full SHA
    bab1109 View commit details
    Browse the repository at this point in the history
  20. Configuration menu
    Copy the full SHA
    461099c View commit details
    Browse the repository at this point in the history
  21. cherry-pick V2.4 (#46294)

    * back fl
    
    * delete ssl cert
    
    * .
    
    * make warning
    
    * .
    
    * unittest paral degree
    
    * solve unittest
    
    * heter & multi cloud commm ready
    
    * .
    
    * .
    
    * fix gloo compile warning
    
    * adapt for nn fl-ps
    
    * flps del fake-init op
    
    * add learning_rate_0 intializer op
    ziyoujiyi authored Sep 20, 2022
    Configuration menu
    Copy the full SHA
    3e8b322 View commit details
    Browse the repository at this point in the history

Commits on Sep 21, 2022

  1. Configuration menu
    Copy the full SHA
    b027652 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    311eb87 View commit details
    Browse the repository at this point in the history
  3. fix CI sccache install dir (#46267) (#46327)

    Co-authored-by: Zhou Wei <1183042833@qq.com>
    sljlp and zhwesky2010 authored Sep 21, 2022
    Configuration menu
    Copy the full SHA
    f3d5c85 View commit details
    Browse the repository at this point in the history
  4. [JitLayer]Erase out vars in scope to avoid data rewritinig (#46249) (#…

    …46273)
    
    * [JitLayer]Erase out vars to avoid data rewrittinig
    
    * Fix code comments
    0x45f authored Sep 21, 2022
    Configuration menu
    Copy the full SHA
    b74c092 View commit details
    Browse the repository at this point in the history
  5. [Cherry-pick][BugFix]Fix pooling output_size bug if encounter list[Te…

    …nsor] (#46360)
    
    * [Check]Enhance pooling output_size type check
    
    * add unittest
    Aurelius84 authored Sep 21, 2022
    Configuration menu
    Copy the full SHA
    cc3e7cd View commit details
    Browse the repository at this point in the history

Commits on Sep 22, 2022

  1. Configuration menu
    Copy the full SHA
    083853c View commit details
    Browse the repository at this point in the history
  2. [Dygraph] Fix bugs of mp in eager mode (#46303) (#46396)

    * fix bugs of mp
    
    * fix bugs of mp
    
    * update
    
    * update
    
    * fix bug
    haohongxiang authored Sep 22, 2022
    Configuration menu
    Copy the full SHA
    372505b View commit details
    Browse the repository at this point in the history
  3. logger manager (#45909) (#46087)

    uniform logger manager in FleetAPI.
    hidde API under distributed/utils which users don't need.
    sljlp authored Sep 22, 2022
    Configuration menu
    Copy the full SHA
    7eb046c View commit details
    Browse the repository at this point in the history

Commits on Sep 23, 2022

  1. [Cherry-Pick][BugFix]Fix reduce_mean/min/sum/prod, cumsum grad_op inf…

    …ershape bug (#46409)
    
    * [BugFix]Fix reduce_mean/min/sum/prod, cumsum grad_op infershape bug
    
    * fix typo
    
    * fix typo
    Aurelius84 authored Sep 23, 2022
    Configuration menu
    Copy the full SHA
    484377c View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    6a50833 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    980292c View commit details
    Browse the repository at this point in the history
  4. [BugFix] fixed a bug that deco_name can't be parsed corrected (#46297) (

    #46366)
    
    * use re replace judge by case
    
    * simplify re
    feifei-111 authored Sep 23, 2022
    Configuration menu
    Copy the full SHA
    cbf3f4b View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    55f73ba View commit details
    Browse the repository at this point in the history

Commits on Sep 24, 2022

  1. Remove audio ParameterError (#46316) (#46468)

    * unexpose audio ParameterError
    
    * clean audio utils api
    SmileGoat authored Sep 24, 2022
    Configuration menu
    Copy the full SHA
    0ab2672 View commit details
    Browse the repository at this point in the history

Commits on Sep 26, 2022

  1. [cherrypick] Fix elementwise_sub sign reverse for mkldnn (#46107)

    * fix sub sign reverse for mkldnn
    
    * refactor code as comment
    
    * remove useless
    zh794390558 authored Sep 26, 2022
    Configuration menu
    Copy the full SHA
    6990edf View commit details
    Browse the repository at this point in the history
  2. fix conflict (#46388)

    feifei-111 authored Sep 26, 2022
    Configuration menu
    Copy the full SHA
    4a8aa6d View commit details
    Browse the repository at this point in the history
  3. cherry-pick V2.4 (#46358)

    * back fl
    
    * delete ssl cert
    
    * .
    
    * make warning
    
    * .
    
    * unittest paral degree
    
    * solve unittest
    
    * heter & multi cloud commm ready
    
    * .
    
    * .
    
    * fix gloo compile warning
    
    * adapt for nn fl-ps
    
    * flps del fake-init op
    
    * add learning_rate_0 intializer op
    
    * bug fix
    
    * .
    
    * .
    ziyoujiyi authored Sep 26, 2022
    Configuration menu
    Copy the full SHA
    536d9d8 View commit details
    Browse the repository at this point in the history

Commits on Sep 27, 2022

  1. Configuration menu
    Copy the full SHA
    8089a1f View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    dc3a3f1 View commit details
    Browse the repository at this point in the history
  3. [cherry-pick] clear extra attrs of some ops in OpMaker (#45845, #45984,…

    … 46060) (#46218)
    
    * Clear extra attrs of elementwise op in OpMaker (#45845)
    
    * clear extra attrs of elementwise op in opmaker
    
    * fix op_debug_string_test
    
    * fix bug of grad_add
    
    * fix sort of runtime attrs
    
    * Clear extra attrs of scale in OpMaker (#45984)
    
    * clear extra attr of scale in opmaker
    
    * fix sum bug
    
    * fix merge conflict
    
    * fix minus
    
    * Clear extra attributes of some Op in OpMaker (Part4) (#46060)
    
    * clear extra attr of some ops in opmaker
    
    * revert clear use_cudnn for pool
    
    * fix test_operator_desc
    
    * fix Attr interface of OperatorBase
    
    * fix code stype
    zyfncg authored Sep 27, 2022
    Configuration menu
    Copy the full SHA
    0cc2251 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    5711bbe View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    5dab0b0 View commit details
    Browse the repository at this point in the history

Commits on Sep 28, 2022

  1. Configuration menu
    Copy the full SHA
    a77a6f6 View commit details
    Browse the repository at this point in the history
  2. [cherry-pick] Clear extra attrs of some ops in OpMaker (#46150, #46321,

    #46418, #46451, #46457) (#46553)
    
    * Clear extra attributes of some Op in OpMaker (Part4) (#46060)
    
    * clear extra attr of some ops in opmaker
    
    * revert clear use_cudnn for pool
    
    * fix test_operator_desc
    
    * fix Attr interface of OperatorBase
    
    * clear extra attrs of condition op in opmaker (#46150)
    
    * Clear extra attrs of lookup_table_v2 in OpMaker (#46321)
    
    * clear extra attrs of look_up_table_v2 in opmaker
    
    * fix bug
    
    * clear extra attrs of quantize op in opmaker (#46418)
    
    * delete repeated item
    
    * clear extra attrs of distribute op in opmaker (#46451)
    
    * clear extra atts of sequence_softmax in opmaker (#46457)
    zyfncg authored Sep 28, 2022
    Configuration menu
    Copy the full SHA
    b2e4211 View commit details
    Browse the repository at this point in the history
  3. Fix libpaddle soname mismatch error (#46344) (#46576)

    * fix libpaddle soname mismatch error
    
    * fix windows failed
    
    * polish linux and windows make impl
    
    * unify winddows lib name
    
    * fix windows error
    
    * revert copy dst change
    
    * revert naming change
    
    * revert windows change
    
    * fix gpups compile failed
    chenwhql authored Sep 28, 2022
    Configuration menu
    Copy the full SHA
    1c22ed7 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    3f35e63 View commit details
    Browse the repository at this point in the history

Commits on Sep 29, 2022

  1. Configuration menu
    Copy the full SHA
    f5956be View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    d90db9b View commit details
    Browse the repository at this point in the history
  3. [cherry-pick] Open the clip_extra flag in save_inference_model (#46577)

    * set flag of clip_extra in save_inference_model to true (#46151)
    
    * open the clip_extra flag in paddle.static.save_inference_model, test=allcase (#46456)
    
    * Open the clip_extra flag in TracedLayer.save_inference_model (#46473)
    
    * open the clip_extra flag in paddle.static.save_inference_model, test=allcase
    
    * set the defalut value of clip_extra in TracedLayer from False to True, test=allcase
    
    * update english doc of paddle.static.save_inference_model, test=document_fix (#46484)
    
    * Fix clip_extra logic in remove_training_info (#46534)
    
    * fix clip_extra code in remove_training_info
    
    * revert rnn opmaker clear
    zyfncg authored Sep 29, 2022
    Configuration menu
    Copy the full SHA
    d67da3d View commit details
    Browse the repository at this point in the history
  4. [cherry-pick] Add FP16 support for uniform in dygraph mode on Nvidia …

    …GPU (#46641)
    
    Add FP16 support for uniform in dygraph mode on Nvidia GPU
    Dev PR link PR46212
    Courtesy-Xs authored Sep 29, 2022
    Configuration menu
    Copy the full SHA
    a58663f View commit details
    Browse the repository at this point in the history

Commits on Oct 9, 2022

  1. [Dy2Static] refactor the return transformer (#45900) (#46205)

    * 1. refactor the return transformer.
    2. fix some bugs in return transformer.
    
    * support raise error while return stmt's father is For or while
    
    * fix ci error.
    
    * fix ci error and add some unittest
    
    * code format
    
    * fix ci error
    2742195759 authored Oct 9, 2022
    Configuration menu
    Copy the full SHA
    4282af6 View commit details
    Browse the repository at this point in the history

Commits on Oct 10, 2022

  1. [PHI] Shape op migration (#46051) (#46724)

    * First approach
    
    * Shape kernel corrected
    
    * Compilation error fixed
    
    * Resize corrected
    
    * Registered types added
    
    * Mistake corrected & types added
    
    * sum kernel deleted
    
    Co-authored-by: Paulina Gacek <paulina.gacek.pl@gmail.com>
    Silv3S and paulinagacek authored Oct 10, 2022
    Configuration menu
    Copy the full SHA
    3cc3f60 View commit details
    Browse the repository at this point in the history
  2. [PHI] migrate softmax_grad kernel (#46257) (#46725)

    * init
    
    * remove softmaxop
    
    * merge dev
    
    * correct dir
    
    * style
    Silv3S authored Oct 10, 2022
    Configuration menu
    Copy the full SHA
    44ecae6 View commit details
    Browse the repository at this point in the history
  3. [PHI] Migrate slice, slice_grad, split, pad and pad3d oneDNN kernels (#…

    …46101) (#46726)
    
    * Convert split, pad and pad3d kernels
    
    * Convert slice+grad oneDNN fluid kernels to PHI
    
    * change out->mutable_data to dev_ctx.Alloc
    
    Co-authored-by: Piotr Paturej <48731682+piotrekobi@users.noreply.github.com>
    Silv3S and piotrekobi authored Oct 10, 2022
    Configuration menu
    Copy the full SHA
    51a91fe View commit details
    Browse the repository at this point in the history
  4. [cherry-pick] [PHI] Migrate sgd and stack oneDNN kernels (#46374) (#4…

    …6729)
    
    * [PHI] Migrate sgd and stack oneDNN kernels (#46374)
    
    * Convert slice+grad oneDNN fluid kernels to PHI
    
    * Change mutable_data to Alloc
    
    * Refactor licences
    
    * update dependencies
    
    Co-authored-by: Piotr Paturej <48731682+piotrekobi@users.noreply.github.com>
    Silv3S and piotrekobi authored Oct 10, 2022
    Configuration menu
    Copy the full SHA
    25d61cd View commit details
    Browse the repository at this point in the history
  5. [cherry-pick] [PHI] Migrate concat+grad, expand+grad, fill_constant ……

    … oneDNN kernels (#45863) (#46727)
    
    * [PHI] Migrate concat+grad, expand+grad, fill_constant, nearest_interp and bilinear_interp oneDNN kernels (#45863)
    
    * Migrate concat+grad, expand+grad, fill_constant, nearest_interp_v2 and bilinear_interp_v2 oneDNN kernels to PHI
    
    * Remove old namespace variable
    
    * Fix invalid out dims error
    
    * Add mutable_data method to concat output
    
    * Add check for -1 dim before computing out_dims
    
    * Capitalize oneDNNGetDataType function name
    
    * Change fill_constant kernel to correct PHI kernel
    
    * Attempt to fix dims error
    
    * Fix fill_constant (full) kernel
    
    * update dependencies
    
    Co-authored-by: Piotr Paturej <48731682+piotrekobi@users.noreply.github.com>
    Silv3S and piotrekobi authored Oct 10, 2022
    Configuration menu
    Copy the full SHA
    fdd0d6d View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    d8daf64 View commit details
    Browse the repository at this point in the history
  7. Fix gather op convert for Paddle-TensorRT (#46779) (#46825)

    * fix gather op convert to only support int32 index as input.
    * add ut
    feng_shuai authored Oct 10, 2022
    Configuration menu
    Copy the full SHA
    a0e0341 View commit details
    Browse the repository at this point in the history

Commits on Oct 11, 2022

  1. Configuration menu
    Copy the full SHA
    d091d1b View commit details
    Browse the repository at this point in the history
  2. [BugFix]Fix concat bugs when call onednn kernel (#46518) (#46845)

    * fix concat bug
    
    * fix ci bugs
    
    * fix ci bugs
    YuanRisheng authored Oct 11, 2022
    Configuration menu
    Copy the full SHA
    6a6c749 View commit details
    Browse the repository at this point in the history
  3. Cherry pick for dygraph pp (#46876)

    * bug fix for virtual pipeline parallel (#45922)
    
    * dont wait for send op under dygraph pp (#46209)
    
    * [interleave pp] sync recv for 1f1b (#46399)
    
    * [dygraph pp] all sync for allgather partial (#46483)
    FeixLiu authored Oct 11, 2022
    Configuration menu
    Copy the full SHA
    9cc3f69 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    f556549 View commit details
    Browse the repository at this point in the history
  5. Revert pool+grad oneDNN kernel conversion (#45989) (#46860)

    Co-authored-by: Piotr Paturej <48731682+piotrekobi@users.noreply.github.com>
    Silv3S and piotrekobi authored Oct 11, 2022
    Configuration menu
    Copy the full SHA
    7b3837e View commit details
    Browse the repository at this point in the history
  6. [cherry-pick] [PHI] relu6_grad kernel (#46501) (#46862)

    * [PHI] Migrate gelu kernels (#45596)
    
    * gaussian random
    
    * mkldnn to onednn renaming
    
    * fix merge conflicts
    
    * remove fluid code
    
    * onednn renaming
    
    * gelu fwd
    
    * sort activations
    
    * gelu gradient
    
    * remove unused macros
    
    * merge conflicts
    
    * fix merge conflicts
    
    * remove extra contraint from gelu op
    
    * [PHI] relu6_grad kernel (#46501)
    
    * Relu6
    
    * remove fluid handler
    
    * add individual kernel signature
    
    * coding style
    
    * replace bounded_relu with clip
    
    * whitespace
    
    * code style
    Silv3S authored Oct 11, 2022
    Configuration menu
    Copy the full SHA
    2bcbf8b View commit details
    Browse the repository at this point in the history
  7. hard_swish grad (#46857)

    Silv3S authored Oct 11, 2022
    Configuration menu
    Copy the full SHA
    2c6bd4a View commit details
    Browse the repository at this point in the history
  8. add seed check (#46858)

    Silv3S authored Oct 11, 2022
    Configuration menu
    Copy the full SHA
    2190da2 View commit details
    Browse the repository at this point in the history
  9. set_value_op: add support for complex types (#46885)

    Feiyu Chan authored Oct 11, 2022
    Configuration menu
    Copy the full SHA
    b051455 View commit details
    Browse the repository at this point in the history

Commits on Oct 12, 2022

  1. cherry pick pr46536 (#46901)

    cherry pick pr46536
    ronny1996 authored Oct 12, 2022
    Configuration menu
    Copy the full SHA
    08d233f View commit details
    Browse the repository at this point in the history
  2. [Cherry-pick]Update layout autotune for module with no modified (#46541

    …) (#46515) (#46880)
    
    Cherry-pick 46541
    保证Reset50 TSM deeplabv3模型零修改下实现Layout自动调优
    AnnaTrainingG authored Oct 12, 2022
    Configuration menu
    Copy the full SHA
    61273c0 View commit details
    Browse the repository at this point in the history

Commits on Oct 13, 2022

  1. [cherry-pick] [PHI] transpose2_grad op migration (#46139) (#46873)

    * Revert pool+grad oneDNN kernel conversion (#45989)
    
    * [PHI] transpose2_grad op migration (#46139)
    
    * op migrated, Copy(OneDNNContext, ...) added
    
    * mutable_data & op registration in fluid removed
    
    * refactoring
    
    * OneDNNGetDataType to uppercase
    
    * missing cpu check added, handler moved to .h file
    
    * name changed to transpose_grad
    
    * Copy changed back to TensorCopy
    
    * Resizing corrected, Copy(OneDNNContext) removed
    
    Co-authored-by: Piotr Paturej <48731682+piotrekobi@users.noreply.github.com>
    Co-authored-by: Paulina Gacek <paulina.gacek@intel.com>
    3 people authored Oct 13, 2022
    Configuration menu
    Copy the full SHA
    0280c0b View commit details
    Browse the repository at this point in the history
  2. [Cherry-pick] Add fp16 dtype support for set_value op (#46906)

    Fix set_value failure when source tensor is fp16 Dtype and destiny value is a number
    (dev PR link:#46801)
    Courtesy-Xs authored Oct 13, 2022
    Configuration menu
    Copy the full SHA
    100a075 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    d90aaa6 View commit details
    Browse the repository at this point in the history

Commits on Oct 14, 2022

  1. Configuration menu
    Copy the full SHA
    b8677c0 View commit details
    Browse the repository at this point in the history
  2. [cherry-pick 2.4][inference] fix reshape2 opteller (#46871)

    * fix reshape2 opteller;
    add elementwise min/max register for tensorrt
    zhangjun authored Oct 14, 2022
    Configuration menu
    Copy the full SHA
    535d757 View commit details
    Browse the repository at this point in the history
  3. [BUG]Fix expand_as_v2 bug while X and Y with different dtype (#46950) (

    …#46999)
    
    * [BUG]Fix expand_as_v2 bug while X and Y with different dtype
    
    * fix commit
    Aurelius84 authored Oct 14, 2022
    Configuration menu
    Copy the full SHA
    4b47265 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    2744432 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    8f1ac7c View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    84333cf View commit details
    Browse the repository at this point in the history
  7. cherry-pick 46942 (#47015)

    jiweibo authored Oct 14, 2022
    Configuration menu
    Copy the full SHA
    82db499 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    5c2bea1 View commit details
    Browse the repository at this point in the history

Commits on Oct 17, 2022

  1. [Cherry-Pick]Move valid check from python to kernel (#46980)

    为了提升性能,将label的边界检查从python端转移到kernel内,减少额外op的调用,如min、max和同步拷贝等
        当前的模板参数IgnoreIndex仅在ignore_index取值范围在[0, dim)时才生效,但是当某个label值超出了边界,ignore_index等于该label,这种情况下是应该仍然能正常计算。虽然当前的计算逻辑在结果上不会出错,但逻辑上仍是有问题的,且模板参数IgnoreIndex是没有必要的
    ZzSean authored Oct 17, 2022
    Configuration menu
    Copy the full SHA
    8bfd45a View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    a1cdbad View commit details
    Browse the repository at this point in the history
  3. [IPU] paddle-inference support custom-ops (#45235) (#46868)

    * paddle-inference support custom-ops
    
    Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai>
    
    * fix tolower
    
    Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai>
    
    Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai>
    gglin001 and Zhixin Yao authored Oct 17, 2022
    Configuration menu
    Copy the full SHA
    bd89be1 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    d913bc9 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    8c6c79a View commit details
    Browse the repository at this point in the history
  6. [cherry-pick] Fix the bug of exporting model in dygraph QAT (#47028)

    * fix dygraph new format quant
    * fix unittest
    * fix conflict
    yghstill authored Oct 17, 2022
    Configuration menu
    Copy the full SHA
    7eef05c View commit details
    Browse the repository at this point in the history
  7. Optimize performance of depthwise_conv (#46896)

    Optimize performance of depthwise_conv
    
    Config: input[2048, 1024, 4, 4], filter[1024, 1, 4, 4], stride=1, pad=0, dilation=1
    ZzSean authored Oct 17, 2022
    Configuration menu
    Copy the full SHA
    976af0d View commit details
    Browse the repository at this point in the history
  8. [cherry-pick]Sparse static graph (#46838)

    cherry-pick : #46322, #46245
    Sparse API 支持静态图
    zhangkaihuo authored Oct 17, 2022
    Configuration menu
    Copy the full SHA
    10225d2 View commit details
    Browse the repository at this point in the history
  9. [Cherry-pick] Collective communication APIs (#46922)

    * Support both use_calc_stream and sync_op in send recv APIs (#46023)
    
    * Support both use_calc_stream and sync_op in allgather API (#46295)
    
    * Support both use_calc_stream and sync_op in collective communication API (#46761)
    
    * Move group and all reduce from collective to communication (#45848)
    
    * Completes bfloat16 dtype for collective api in eager mode (#45844)
    
    * Fix collective APIs cannot be recognized when building docs (#46962)
    
    Co-authored-by: LiYuRio <63526175+LiYuRio@users.noreply.github.com>
    HermitSun and LiYuRio authored Oct 17, 2022
    Configuration menu
    Copy the full SHA
    5fba2a9 View commit details
    Browse the repository at this point in the history

Commits on Oct 18, 2022

  1. [Cherry pick] trt pool2d adaptive ifx (#47069)

    * draft with debug print
    * remove debug print
    * bug fix for ci
    wwbitejotunn authored Oct 18, 2022
    Configuration menu
    Copy the full SHA
    5f6b9f1 View commit details
    Browse the repository at this point in the history
  2. [cherry-pick] Fix perf issues of mp/pp/fuse in eager mode (#47071)

    * [Dygraph] Fix performance of pp+mp by using send/recv_calc_stream instead of send/recv (#46116)
    
    * [Dygraph] Fix Perf of FusedFeedForward and FusedAttention with AllReduce (#46780)
    
    * update
    haohongxiang authored Oct 18, 2022
    Configuration menu
    Copy the full SHA
    b84edd9 View commit details
    Browse the repository at this point in the history
  3. Cherry pick for sharding (#47061)

    * [dygraph sharding] Overlap the reduce and the caculation for sharding stage 2. (#46495)
    
    * [dygraph sharding stage 2] sharding broadcast overlap (#46656)
    
    * Multi groups for broadcast of sharding stage 2 (#46894)
    FeixLiu authored Oct 18, 2022
    Configuration menu
    Copy the full SHA
    5b64214 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    5a44c12 View commit details
    Browse the repository at this point in the history
  5. [cherry-pick 2.4] add sparse api transpose/reshape/is_same_shape (#47076

    )
    
    新增sparse.is_same_shape、sparse.reshape、sparse.transpose 三个API
    zhwesky2010 authored Oct 18, 2022
    Configuration menu
    Copy the full SHA
    5fef043 View commit details
    Browse the repository at this point in the history
  6. Add symbolic shape deduction function for unfold, scatter_nd_add, p_n…

    …orm, grid_sampler, pad3d, etc (#46291) (#47003)
    weishengying authored Oct 18, 2022
    Configuration menu
    Copy the full SHA
    2cc8797 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    de6f15b View commit details
    Browse the repository at this point in the history

Commits on Oct 19, 2022

  1. [ cherrypick] Construct exec and ctx only once in cond op to speed up (

    …#47012)
    
    Construct exec and ctx only once in cond op to speed up
    zh794390558 authored Oct 19, 2022
    Configuration menu
    Copy the full SHA
    fcb9c0b View commit details
    Browse the repository at this point in the history
  2. update audio api examples (#46938) (#47126)

    * update audio api examples
    
    * fix format
    
    * format
    
    * fix
    
    * test api
    
    * fix format
    
    * fix static check error
    
    * fix doc error
    
    * fix ci
    
    * fix api error
    
    * update api.spec
    
    * fix ci
    
    * fix typo in window gaussian
    SmileGoat authored Oct 19, 2022
    Configuration menu
    Copy the full SHA
    f08c104 View commit details
    Browse the repository at this point in the history
  3. [Dy2St]Fix recurrent op eager deletion pass error in dy2st (#47105) (#…

    …47134)
    
    [CherryPick][Dy2St]Fix recurrent op eager deletion pass error in dy2st
    0x45f authored Oct 19, 2022
    Configuration menu
    Copy the full SHA
    69515e9 View commit details
    Browse the repository at this point in the history
  4. Add enable_partial_send_recv switch in pipeline_configs (#46992) (#47083

    )
    
    * Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result
    is wrong.
    
    * Support allow_partial switch, which can be configure in
    pipeline_configs. If sent tensor are not the same from
    different hosts, they shouldn't been sent partially and
    then concated as a whole tensor.
    
    * Change name allow_partial to enable_partial_send_recv.
    
    * Add global variable _enable_partial_send_recv
    GhostScreaming authored Oct 19, 2022
    Configuration menu
    Copy the full SHA
    1d015f1 View commit details
    Browse the repository at this point in the history
  5. [CherryPick] Support TypeHint for function decorated by @to_static (#…

    …47147)
    
    * [Dy2Static] Support TypeHint for function decorated by @to_static (#47121)
    
    * Add TypeHint Transformer
    
    * add unittest for typehint transformer
    
    * [Dy2Static] Remove GradTransformer (#47063)
    
    * [Dy2Static] Remove GradTransformer
    1. fix einsum infershape bugs.
    2. remove grad_transformer and unify paddle.grad and paddle.static.gradient.
    3. add dygraph_and_dy2static_only decorator for dy2static.
    
    * fix bugs
    
    * rename
    2742195759 authored Oct 19, 2022
    Configuration menu
    Copy the full SHA
    247ef47 View commit details
    Browse the repository at this point in the history
  6. [Dy2Stat]Polish @to_static temporary file directory to speed up trans…

    …formation (#47102) (#47144)
    
    Polish @to_static temporary file directory to speed up transformation
    Aurelius84 authored Oct 19, 2022
    Configuration menu
    Copy the full SHA
    5a9befe View commit details
    Browse the repository at this point in the history
  7. Add unsigned int8 scale propagation (#46378) (#47156)

    * Add unsigned int8 propagation
    
    * Add or modify unit tests
    
    * Correct concat scale checking
    
    * Apply review suggestions
    
    * Corrections
    
    Co-authored-by: joanna.wozna.intel <joanna.wozna@intel.com>
    yeliang2258 and wozna authored Oct 19, 2022
    Configuration menu
    Copy the full SHA
    66dccd7 View commit details
    Browse the repository at this point in the history
  8. [cherry-pick] strided_slice grad add fp16 support (#47159)

    * strided_slice grad add fp16 support
    Zhang Ting authored Oct 19, 2022
    Configuration menu
    Copy the full SHA
    23f2a4e View commit details
    Browse the repository at this point in the history
  9. [Cherry-Pick][AutoParallel] auto_parallel cherry-pick to release2.4 (#…

    …47145)
    
    * [Auto Parallel] Make Engine class callable (#46416)
    
    * [Auto Parallel] Imporve the user-defined fetches and logging
    
    * [Auto Parallel] Make Engine class callable
    
    * [Auto Parallel] Update the data loading of tuner
    
    * Print IPS in auto parallel Engine (#46554)
    
    * [AutoParallel] fix dist_split (#46505)
    
    * [AutoParallel] fix dist_split
    
    * add unittest
    
    * update cmakelist
    
    * [AutoParallel] fix sharding (#46572)
    
    * [AutoParallel] fix process_mesh (#46583)
    
    * [AutoParallel] fix reshard when train with eval (#46605)
    
    * [AutoParallel] fix reshard when train with eval
    
    * fix mppp
    
    * [AutoParallel] fix amp when predict (#46637)
    
    * [Auto Parallel]Update comp cost and completion for gpt auto search (#46387)
    
    * update comp cost and completion for gpt auto search
    
    * add unittest
    
    * [Auto Parallel] Fix bugs caused by the inconsistent outputs of Engine API (#46633)
    
    * [Auto Parallel] Unify the logger and outputs of Engine API
    
    * [Auto Parallel] Fix the bugs of to_static
    
    * [Auto Parallel] Adjust the test_to_static.py
    
    * [Auto Parallel] Improve the fine-grained APIs (#46552)
    
    * [Auto Parallel] Suppport different dataloaders
    
    * [Auto Parallel] Add num_shards config for dataset
    
    * [Auto Parallel] Unify the logger and outputs of Engine API
    
    * [Auto Parallel] Fix the bugs of to_static
    
    * [Auto Parallel] Adjust the test_to_static.py
    
    * [Auto Parallel] Add the prepare API and replace __call__ with run
    
    * [Auto Parallel] Improve the private implementations of Engine
    
    * [Auto Parallel] Set capacity of dataloader for opt tuning
    
    * [Auto Parallel] [WIP] Change the fine-grained API
    
    * [Auto Parallel] Improve APIs to support different user cases
    
    * [Auto Parallel] Add removed config
    
    * [Auto Parallel] Add imports
    
    * [Auto Parallel] Fix bugs for to_static
    
    * [Auto Parallel] Remove unnecessary imports
    
    * bugfix (#46921)
    
    * [Auto Parallel] Fix the bug for None labels (#46987)
    
    * [AutoParallel] adapt for gpt-gen (#46771)
    
    * for gpt-gen
    
    * fix reshard
    
    * adapt assign and shape op
    
    * add dist_assign & unittest
    
    * add conditional block unittest
    
    * rename unittest
    
    * [Auto Parallel] Fix the bug of completion (#47056)
    
    * [Auto Parallel] Fix the bug for None labels
    
    * [Auto Parallel] Fix the completion bug
    
    * [AutoParallel] add callbacks (#47014)
    
    * [AutoParallel] add callbacks
    
    * fix unittest
    
    * fix dist_context
    
    * fix engine
    
    * fix cmakelist
    
    * fix unittest's returns
    
    * fix cmakelist
    
    * [Auto Parallel] Add cost interface (#47043)
    
    * add cost interface
    
    * update inferface and add unittest
    
    * update unittest
    
    * update inferface
    
    * [Auto Parallel]Add parallel tuner (#46189)
    
    * add parallel tuner
    
    * add unittest
    
    * fix unittest
    
    * set timeout of unittest
    
    * set unittest timeout
    
    * fix auto_mode setting
    
    * update unittest
    
    * sync from develop and update unittest
    
    * remove unused import
    
    * update unittest
    
    * update cmakelist
    
    * add unittests
    
    Co-authored-by: Yulong Ao <aoyulong@baidu.com>
    Co-authored-by: Ruibiao Chen <chenruibiao@baidu.com>
    Co-authored-by: caozhou <48191911+Caozhou1995@users.noreply.github.com>
    Co-authored-by: JZ-LIANG <jianzhongliang10@gmail.com>
    5 people authored Oct 19, 2022
    Configuration menu
    Copy the full SHA
    90b3179 View commit details
    Browse the repository at this point in the history

Commits on Oct 20, 2022

  1. Fix cannot import paddle.distributed in python 3.6 on release/2.4 (#…

    …47141)
    
    * fix: fix incorrect import
    
    * fix: fix incorrect usage
    HermitSun authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    c894d91 View commit details
    Browse the repository at this point in the history
  2. [Cherry-pick][Release/2.4] support pure bfloat16 for more ops

    support pure bfloat16 for more ops
    sneaxiy authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    da7d2f2 View commit details
    Browse the repository at this point in the history
  3. [Cherry-pick][Release/2.4] Fix some operators when the tensor.numel()…

    … > INT32_MAX (#47191)
    
    Fix some operators when the tensor.numel() > INT32_MAX
    sneaxiy authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    c74bf01 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    4c92524 View commit details
    Browse the repository at this point in the history
  5. [Cherry-pick] layernorm shift partation enhance (#47086)

    * Enhance the layernorm shift partation fuse op when shift size > 0 (roll shifting)
    * fix cherry-pick test
    wwbitejotunn authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    9ed1454 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    09b1923 View commit details
    Browse the repository at this point in the history
  7. [Paddle-TRT][Cherry-Pick]Rewrite strided_slice converter using shape …

    …tensor (#47153)
    
    * stride_to_24
    
    * fix CI failing
    zhoutianzi666 authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    68c4ac3 View commit details
    Browse the repository at this point in the history
  8. [cherry-pick] Fix quantize model deploy bug in MKLDNN (#47119)

    * Fix quantize model deploy bugs when using MKLDNN (#45920)
    
    * fix immutable op quantize bugs
    
    * fix
    
    * fix build bug
    
    * fix test
    
    * notest,test=inference
    
    * fix ppyoloe acc drop bugs
    
    * fix test
    
    * fix test
    
    * add test
    
    * fix
    
    * fix
    
    * fix test
    
    * fix refined name bug
    
    * fix test
    
    * bias fix
    
    * fix matmul weight dequant bug
    
    * re-ci
    
    * fix tester
    
    * fix test
    
    * fix tester
    
    * update weight dequantize func
    
    * update code
    
    * update test for converage
    
    * update test
    
    * update cmake
    
    * update cmakelist
    
    * update code
    
    * rerun ci
    
    * remove useless code
    
    * re-ci
    
    * update code
    
    * update code
    
    * fix header
    
    * update code for log
    yeliang2258 authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    c2d344d View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    3d647b1 View commit details
    Browse the repository at this point in the history
  10. Add value check & error message for gather_tree (#47051) (#47221)

    Add value check & error message for gather_tree
    cherry-pick #47051
    FrostML authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    6712e26 View commit details
    Browse the repository at this point in the history
  11. [cherry pick] Add FusedMultiTransformer fuse pass for GPT3 (#47150)

    * add fused_attention_pass. test=develop
    
    * support fp16. test=develop
    
    * fix format. test=develop
    heavengate authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    396427a View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    50d4fa5 View commit details
    Browse the repository at this point in the history
  13. [Cherry-pick] Simplify conv codes and fix cache and autotune bugs. (#…

    …47197)
    
    * Simplify the codes of conv. (#45966)
    
    * Enable to record whether the conv algo is got by exhaustive search to fix autotune cache bug. (#47065)
    Xreki authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    c0ed872 View commit details
    Browse the repository at this point in the history
  14. Add paddle audio dataset & backend (#45939) (#47230)

    * add audio feature dataset
    
    * fix coding style
    
    * fix coding style2
    
    * rm librosa
    
    * rm voxceleb
    
    * rm librosa in test
    
    * add scipy fftpack
    
    * add functional
    
    * fix setup
    
    * fix setup2
    
    * rm colorlog
    
    * refactor dataset __init__.py
    
    * fix converage
    
    * fix librosa import error
    
    * fix windows test
    
    * fix windows ci
    
    * rm datasets
    
    * fix setup
    
    * remove testdata
    
    * add librosa in requirement
    
    * add librosa in requirement2
    
    * change librosa to 0.8.1
    
    * update ci docker
    
    * fix ci error
    
    * fix ci error2
    
    * fix ci coverage
    
    * fix converage
    
    * fix coverage
    
    * rm audio_base in test, notest,test=coverage
    
    * fix copyright
    
    * rm backend
    
    * add datast in __init__
    
    * rm compliance&&add function test
    
    * fix setup
    
    * fix windows
    
    * fix windows2
    
    * fix test timeout
    
    * add backend & datasets
    
    * fix bugs
    
    * fix ci time issue
    
    * add dataset test
    
    * rm test_audio_feature
    
    * avoid windows isssue, tmp
    
    * note windows isssue
    
    * skip windows issue
    
    * refactor dataset test
    
    * add dataset.py
    
    * fix dtype in layers.mfcc
    
    * fix ci-static-check
    
    * fix dtype in layers.mfcc && fix ci-static-check
    
    * add relative accuracy
    
    * modity API.spec
    
    * skip cuda11.2 test
    
    * skip cuda11.2 test2
    
    * skip cuda11.2
    
    * change dataset name
    
    * fix format
    
    * update api.spec
    
    * update api.spec2
    
    * fix coverage
    
    * add dataset test
    
    * rm download load dict
    
    * rm download load dict in init
    
    * update api.spec3
    
    * fix dataset coverage
    
    * fix coverage
    
    * fix coverage2
    
    * restore api.spec
    
    * restore api.spec2
    
    * fix api-spec 3
    
    * fix api-spec 4
    
    * fix api.spec
    
    * fix api.spec6
    
    * refactor init_backend
    
    * fix typo
    
    * change paddleaudio backend set
    
    * fix get_current_audio_backend()
    
    * fix format
    
    * fix format2
    
    * remove format in parameters
    
    * fix format2
    
    * add warning massage in wave_backend && remove redundant audio util
    
    * rm audio util in print_signatures
    
    * fix format3
    
    * add tess dataset license
    
    * format warning
    
    * add more info in warning msg
    
    * add paddleaudio version check
    
    * replace dataset esc50 with tess
    
    * add tess dataset && rm numpy transform in dataset.py
    
    * fix set audio backend bug
    
    * fix equal error
    
    * fix format && coverage error
    
    * add api example
    
    * fix format
    
    * fix error
    
    * fix typo
    
    * add noqa in __init__
    
    * fix backend doc example error
    
    * rm seed in dataset
    
    * update bakcend example
    
    * fix typo
    
    * fix typo
    
    * fix example err
    
    * fix typo
    
    * fix ci dataset test
    
    * fix example fil
    
    * try to fix ci
    
    * clean dataset doc
    
    * change get_current_audio_backend to get_current_backend
    
    * creplace paddle.audio.backends.info with paddle.audio.info, same with load, save
    
    * fix ci error
    
    * repalce api in test_audio_backend
    
    * fix save&&set_backend exmaple
    SmileGoat authored Oct 20, 2022
    Configuration menu
    Copy the full SHA
    29c9f02 View commit details
    Browse the repository at this point in the history

Commits on Oct 21, 2022

  1. Add infer prune function (#47047)

    * Add infer prune function
    
    * add fusion op
    JZZ-NOTE authored Oct 21, 2022
    Configuration menu
    Copy the full SHA
    8739497 View commit details
    Browse the repository at this point in the history
  2. add pdsa-2022-001, test=document_fix (#47228)

    Add PDSA-2022-001 security advisory
    VigiZhang authored Oct 21, 2022
    Configuration menu
    Copy the full SHA
    001c8a6 View commit details
    Browse the repository at this point in the history
  3. [CustomDevice] turn on WITH_CUSTOM_DEVICE when WITH_PYTHON=ON (#47165)

    cherry pick #47108
    
    原 WITH_CUSTOM_DEVICE 默认打开/关闭的策略随 ON_INFER开关,由于训练和预测共同发包,现在训练包编译时会打开 ON_INFER,导致 WITH_CUSTOM_DEVICE 默认关闭,custom device 功能不可用
    
    WITH_CUSTOM_DEVICE 默认打开/关闭的策略更改为随 WITH_PYTHON 开关
    ronny1996 authored Oct 21, 2022
    Configuration menu
    Copy the full SHA
    d1fedc5 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    281891c View commit details
    Browse the repository at this point in the history
  5. [Paddle-TRT] Fix conv2d (#47034)

    * forbid Conv2D into Paddle-TensoRT when weight is not persistable.
    zhoutianzi666 authored Oct 21, 2022
    Configuration menu
    Copy the full SHA
    d42a1dc View commit details
    Browse the repository at this point in the history

Commits on Oct 24, 2022

  1. Configuration menu
    Copy the full SHA
    82f1e1b View commit details
    Browse the repository at this point in the history
  2. Support BF16 training for sharding (#46846) (#47246)

    * Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result
    is wrong.
    
    * support pure bfloat16
    
    * support bf16 linear
    
    * update PR to pass CI
    
    * tiny fix where_grad_kernel.cu
    
    * Support bfloat16 type for reducer and sharding.
    
    * Fix some bug.
    
    * Polish code.
    
    * Polise code.
    
    * Add bfloat16 datatype in fill_grad kernels.
    
    Co-authored-by: sneaxiy <sneaxiy@126.com>
    
    Co-authored-by: sneaxiy <sneaxiy@126.com>
    GhostScreaming and sneaxiy authored Oct 24, 2022
    Configuration menu
    Copy the full SHA
    5c85f1a View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    9780eb7 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    caf2751 View commit details
    Browse the repository at this point in the history
  5. Fix hAPI bug of not compatible with LayerHook (#47001) (#47283)

    * Fix hAPI bug of not compatible with LayerHook
    parap1uie-s authored Oct 24, 2022
    Configuration menu
    Copy the full SHA
    e8d6339 View commit details
    Browse the repository at this point in the history
  6. [CodeStyle] add black config to release2.4 (#47146)

    * [CodeStyle] add black config to release2.4
    
    * empty commit, test=document_fix
    SigureMo authored Oct 24, 2022
    Configuration menu
    Copy the full SHA
    6454133 View commit details
    Browse the repository at this point in the history

Commits on Oct 25, 2022

  1. [cherry-pick] add prior_box and box_coder for paddle.vision.ops (#46786)

    * add prior_box and box_coder for paddle.vision.ops
    
    * fix UT change assertTrue to assert_allclose
    
    * fix formula format
    nemonameless authored Oct 25, 2022
    Configuration menu
    Copy the full SHA
    d5c6386 View commit details
    Browse the repository at this point in the history
  2. [geometric] fix english doc (#46485) (#47317)

    * fix geometric doc
    DesmonDay authored Oct 25, 2022
    Configuration menu
    Copy the full SHA
    99d8ba4 View commit details
    Browse the repository at this point in the history
  3. [Sparse] Fix indices (#47190) (#47226)

    当前无法从Tensor中获取到SparseTensor的sparse_dim,无法准确推断出indices的shape,所以目前先以3D点云模型为主,输入的SparseTensor的维度是5D的,其中非零元素是一维向量,所以indices是[4, -1]。
    zhangkaihuo authored Oct 25, 2022
    Configuration menu
    Copy the full SHA
    942ab42 View commit details
    Browse the repository at this point in the history

Commits on Oct 26, 2022

  1. [Cherry-pick][Release/2.4]Refine the memory usage of fused_attention …

    …and fused_feedforward ops (#47235)
    
    * fix fused_attention fused_feedforward
    
    * fix ci
    
    * fix ci
    
    * fix ci PADDLE_GET_CONST
    
    * fix ci ut
    sneaxiy authored Oct 26, 2022
    Configuration menu
    Copy the full SHA
    9a6dd8f View commit details
    Browse the repository at this point in the history
  2. Added workaround for elementwise oneDNN kernel (#47080) (#47342)

    * return proper state
    
    * fix for dims
    
    * fix
    
    Co-authored-by: jakpiase <jakpia21@gmail.com>
    yeliang2258 and jakpiase authored Oct 26, 2022
    Configuration menu
    Copy the full SHA
    7c6550a View commit details
    Browse the repository at this point in the history
  3. [Cherry-Pick][Dy2Stat]Fix module loading OSError in multiprocess (#47302

    )
    
    [Dy2Stat]Fix module loading OSError in multiprocess
    Aurelius84 authored Oct 26, 2022
    Configuration menu
    Copy the full SHA
    12e6dfc View commit details
    Browse the repository at this point in the history
  4. [Cherry-pick][audio] fix tess split fold (#47350)

    * fix tess split fold
    
    * format
    SmileGoat authored Oct 26, 2022
    Configuration menu
    Copy the full SHA
    85094bc View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    a16ef9f View commit details
    Browse the repository at this point in the history
  6. Fix inference performance problem caused by selecting cudnn kernel of…

    … softmax (#47338) (#47367)
    
    * fix inference perfermence problem caused by selecting cudnn kernel for softmax
    
    * recover use_cudnn in opmaker of softmax
    zyfncg authored Oct 26, 2022
    Configuration menu
    Copy the full SHA
    0369cd0 View commit details
    Browse the repository at this point in the history

Commits on Oct 27, 2022

  1. fix slice bug (#47349) (#47376)

    修改一处Slice的bug
    wanghuancoder authored Oct 27, 2022
    Configuration menu
    Copy the full SHA
    99cec1a View commit details
    Browse the repository at this point in the history
  2. [cherry-pick] add batch_norm_kernel (#47394)

    * cherry-pick #46359 and resolve conflict
    zhangkaihuo authored Oct 27, 2022
    Configuration menu
    Copy the full SHA
    b143e00 View commit details
    Browse the repository at this point in the history
  3. [Cherry-pick Release/2.4] Fix multi_tensor adam and momentum bug when…

    … the parameter is list of dict (#47372)
    
    * reformat file by black
    
    * fix multi_tensor adam/momentum bug
    sneaxiy authored Oct 27, 2022
    Configuration menu
    Copy the full SHA
    94240e2 View commit details
    Browse the repository at this point in the history

Commits on Oct 28, 2022

  1. Configuration menu
    Copy the full SHA
    eec93bd View commit details
    Browse the repository at this point in the history
  2. [cherry-pick]add sync_batch_norm_bn and deliver indices_dict (#47407)

    add sync_batch_norm_bn and deliver indices_dict
    zhangkaihuo authored Oct 28, 2022
    Configuration menu
    Copy the full SHA
    0fa8309 View commit details
    Browse the repository at this point in the history
  3. [Cherry-pick][JIT] Add Predictor for JITLayer (#47379) (#47419)

    * [JIT] Add Predictor for JITLayer (#47379)
    
    * add predictor_engine
    
    * add predictor_engine
    
    * fix zero shape
    
    * fix lodTensor
    
    * fix unittest
    
    * fix code style
    
    * update CmakeList
    
    * fix new executor
    Aurelius84 authored Oct 28, 2022
    Configuration menu
    Copy the full SHA
    c42929c View commit details
    Browse the repository at this point in the history
  4. [Dy2St]Fix abnormal growth of memory in train mode and no_grad for Dy…

    …2St (#47398) (#47414)
    
    * [Dy2St]Fix abnormal growth of memory in train mode and no_grad for Dy2St
    0x45f authored Oct 28, 2022
    Configuration menu
    Copy the full SHA
    7618cbd View commit details
    Browse the repository at this point in the history

Commits on Oct 29, 2022

  1. [JITLayer]Enable OneDNN on CPU and Fix zero shape (#47428) (#47436)

    * [JITLayer]Enable OneDNN on CPU and Fix zero shape
    Aurelius84 authored Oct 29, 2022
    Configuration menu
    Copy the full SHA
    f478844 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    26465cd View commit details
    Browse the repository at this point in the history
  3. [Cherry-pick][Release/2.4]Add fused_allreduce_gradients_with_group fo…

    …r PPFleetX (#47458)
    
    * reformat hybrid_parallel_util.py by black
    
    * add fused_allreduce_gradients_with_group
    
    * add scale
    
    * fix ci
    sneaxiy authored Oct 29, 2022
    Configuration menu
    Copy the full SHA
    df64e79 View commit details
    Browse the repository at this point in the history

Commits on Oct 31, 2022

  1. [cherry-pick] update dygraph PTQ export_model api (#47415)

    * update dygraph PTQ export_model api
    
    * remove postprocess
    yghstill authored Oct 31, 2022
    Configuration menu
    Copy the full SHA
    12b9b03 View commit details
    Browse the repository at this point in the history
  2. [Cherry-pick][audio] rm kaiser window in audio get_window function &&…

    … rm audio utils(#47469) (#47479)
    
    * [audio] rm kaiser window in audio get_window function && rm audio utils (#47469)
    
    * rm kaiser window in audio window function
    
    * rm paddle audio utils which is redundant
    
    * rm kaiser in test_audio_functions.py
    Conflicts:
    	python/paddle/audio/utils/error.py
    	python/paddle/tests/test_audio_functions.py
    
    * format
    SmileGoat authored Oct 31, 2022
    Configuration menu
    Copy the full SHA
    f93e9a5 View commit details
    Browse the repository at this point in the history
  3. 2.4/fix engine build (#47462)

    * update codestyle
    
    * [AutoParallel] fix fp16 for subblock (#47189)
    
    * [AutoParallel] fix fp16 for subblock
    
    * fix engine
    
    * fix comment
    
    * [AutoParallel] fix engine _build and cost method (#47263)
    
    * fix engine build method
    
    * fix import
    
    * update engine cost
    
    * update raise error
    
    * update cmakelist
    
    * revert optimizer
    
    * revert optimizer
    
    * fix unittest
    
    * fix unittest
    
    Co-authored-by: caozhou <caozhou@radi.ac.cn>
    
    Co-authored-by: caozhou <caozhou@radi.ac.cn>
    zhaoyinglia and Caozhou1995 authored Oct 31, 2022
    Configuration menu
    Copy the full SHA
    4b3589f View commit details
    Browse the repository at this point in the history

Commits on Nov 1, 2022

  1. Configuration menu
    Copy the full SHA
    0201ccc View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    5ffd4af View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    23c05f2 View commit details
    Browse the repository at this point in the history
  4. [cherry-pick][code-gen] Support code-gen for opmaker of sparse op (#4…

    …6993) (#47417)
    
    * support generating code of opmaker for backward op invoke forward op (#46912)
    
    * [code-gen] Support code-gen for opmaker of sparse op (#46993)
    
    * support generating code of opmaker for backward op invoke forward op
    
    * gsupport code-gen of opmaker for sparse op
    
    * refind logic of choose phi kernrel
    
    * fix complie budg
    
    * fix code_gen bug
    
    * fix bug
    
    * fix kernel signature code-gen
    
    * fix complie bug of VarType
    
    * fix complie bug of VarType
    
    * fix test_sparse_conv_op
    
    * fix test_sparse_norm_op
    
    * [Phi] Refactor logic of judging whether having a phi kernrel (#46920)
    
    * refind logic of choose phi kernrel
    
    * fix complie budg
    
    * update cmake
    zyfncg authored Nov 1, 2022
    Configuration menu
    Copy the full SHA
    601626a View commit details
    Browse the repository at this point in the history
  5. [cherry-pick]Fix english documents of sparse api (#47496)

    Fix english documents of sparse api
    zhangkaihuo authored Nov 1, 2022
    Configuration menu
    Copy the full SHA
    61953b9 View commit details
    Browse the repository at this point in the history

Commits on Nov 2, 2022

  1. Configuration menu
    Copy the full SHA
    7a1cf27 View commit details
    Browse the repository at this point in the history

Commits on Nov 3, 2022

  1. [cherry pick] fix memory copy in prepare_data of FusedMultiTransforme…

    …r pass (#47308)
    
    * fix memory copy in prepare_data. test=develop
    
    * add cache_kv fp16 support. test=develop
    
    * fit for simplify_with_basic_ops_pass. test=develop
    heavengate authored Nov 3, 2022
    Configuration menu
    Copy the full SHA
    ba4fbe7 View commit details
    Browse the repository at this point in the history
  2. support unbalanced data for pipeline (#47199) (#47569)

    * add unbalanced data
    
    * fix utest
    ForFishes authored Nov 3, 2022
    Configuration menu
    Copy the full SHA
    d4bf8b1 View commit details
    Browse the repository at this point in the history
  3. [Sparse] Unified api args name (#47529) (#47627)

    Unified api args name
    zhangkaihuo authored Nov 3, 2022
    Configuration menu
    Copy the full SHA
    75088bb View commit details
    Browse the repository at this point in the history
  4. Fix ComputePropagateScalesMkldnnPass of MKLDNN (#47574) (#47639)

    * add constant_folding_pass pass for mkldnn int8
    
    * update UpdateScaleOpInOutScales
    yeliang2258 authored Nov 3, 2022
    Configuration menu
    Copy the full SHA
    559b975 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    99c872f View commit details
    Browse the repository at this point in the history

Commits on Nov 4, 2022

  1. [cherry-pick2.4]for CodeStyle (#47608)

    * only run pre-commit
    
    * only run pre-commit
    Ligoml authored Nov 4, 2022
    Configuration menu
    Copy the full SHA
    cfee9c1 View commit details
    Browse the repository at this point in the history
  2. #46165

    Liyulingyue authored and Ligoml committed Nov 4, 2022
    Configuration menu
    Copy the full SHA
    1ccd3ea View commit details
    Browse the repository at this point in the history