-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[cherry-pick2.4]docs fix #47668
[cherry-pick2.4]docs fix #47668
Commits on Sep 13, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 29c44eb - Browse repository at this point
Copy the full SHA 29c44ebView commit details -
Configuration menu - View commit details
-
Copy full SHA for 0903020 - Browse repository at this point
Copy the full SHA 0903020View commit details
Commits on Sep 14, 2022
-
Configuration menu - View commit details
-
Copy full SHA for e223cf7 - Browse repository at this point
Copy the full SHA e223cf7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9d5003d - Browse repository at this point
Copy the full SHA 9d5003dView commit details -
Configuration menu - View commit details
-
Copy full SHA for 20d168d - Browse repository at this point
Copy the full SHA 20d168dView commit details -
[chery-pick] Fix namespace error (#45925) (#46029)
修复cuda11.7编译出错的问题
Configuration menu - View commit details
-
Copy full SHA for 925e84b - Browse repository at this point
Copy the full SHA 925e84bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 2fac8ab - Browse repository at this point
Copy the full SHA 2fac8abView commit details -
cherry pick delay tensorrt log (#45958)
* cherry pick delay tensorrt log * Update trt_plugin.h
Configuration menu - View commit details
-
Copy full SHA for 2ca6590 - Browse repository at this point
Copy the full SHA 2ca6590View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5130b0a - Browse repository at this point
Copy the full SHA 5130b0aView commit details
Commits on Sep 15, 2022
-
fix trt multiclass_nms3 (#45166) (#46034)
* Support dynamic shape in multiclass_nms3 Plugin for Paddle-TensorRT.
Configuration menu - View commit details
-
Copy full SHA for 61a3e30 - Browse repository at this point
Copy the full SHA 61a3e30View commit details -
Fix arm fp16 compile error (#45991) (#46048)
* fix arm fp16 compile error * polish macro impl
Configuration menu - View commit details
-
Copy full SHA for 91677eb - Browse repository at this point
Copy the full SHA 91677ebView commit details -
Configuration menu - View commit details
-
Copy full SHA for 526e032 - Browse repository at this point
Copy the full SHA 526e032View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9012e8b - Browse repository at this point
Copy the full SHA 9012e8bView commit details -
Delete eigen header in data_type.h (#46036) (#46066)
* delete eigen header in data_type.h * fix complie bug * refactor
Configuration menu - View commit details
-
Copy full SHA for 2680a71 - Browse repository at this point
Copy the full SHA 2680a71View commit details -
Configuration menu - View commit details
-
Copy full SHA for 0793311 - Browse repository at this point
Copy the full SHA 0793311View commit details -
Support 0 shapes input Tensor for MKL slice (#45930) (#46072)
Support 0 shapes input Tensor for MKL slice kernel
Configuration menu - View commit details
-
Copy full SHA for 903c87b - Browse repository at this point
Copy the full SHA 903c87bView commit details -
[ Dy2Static ] Fix bugs when select inputs meeting different shape or …
Configuration menu - View commit details
-
Copy full SHA for 0048695 - Browse repository at this point
Copy the full SHA 0048695View commit details
Commits on Sep 16, 2022
-
[Cherry-pick] Normalize yaml name and label (#46052)
* normalize yaml file name (#45894) * Clear extra attributes of activation op in OpMaker (#45772) * clear extra attr of activation op in opmaker * fix syntax bug * fix mkldnn kernel * fix merge conflict * fix bug * [PHI] Normalize yaml op label (#45976) * normalize yaml op label * revert op_compat yaml change * fix prelu and rnn compat problem * replace api by op * support assign op backward refuse forward (#45879) * normize yaml backward op label (#46028) Co-authored-by: zyfncg <zhangyunfei07@baidu.com> Co-authored-by: Charles-hit <56987902+Charles-hit@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 8caaf85 - Browse repository at this point
Copy the full SHA 8caaf85View commit details -
[cherry-pick][jit] Jit skip forward (#45926)
* skip forward save * fix bug * more ci for jit skip forward
Configuration menu - View commit details
-
Copy full SHA for e25e947 - Browse repository at this point
Copy the full SHA e25e947View commit details -
(cherry-pick)Fix split infershape in static mode and add convert rule…
…s for fill_any_like op (#46079) * Fix split bug in static mode (#45906) * fix split bug in static mode * modify code style * modify code style * add unit test for split * add convert rules for fill_any_like op in paddle science (#45985) * add convert rules for fill_any_like op in paddle science * add unit test for fill_any_like op in paddle science * modify fill_any_like convert rule * modify fill_any_like convert rule dtype
Configuration menu - View commit details
-
Copy full SHA for 4e09e40 - Browse repository at this point
Copy the full SHA 4e09e40View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2c2cd79 - Browse repository at this point
Copy the full SHA 2c2cd79View commit details
Commits on Sep 17, 2022
-
* back fl * delete ssl cert * . * make warning * . * unittest paral degree * solve unittest * heter & multi cloud commm ready * . * . * fix gloo compile warning * adapt for nn fl-ps
Configuration menu - View commit details
-
Copy full SHA for a76fa41 - Browse repository at this point
Copy the full SHA a76fa41View commit details -
Configuration menu - View commit details
-
Copy full SHA for f6dd201 - Browse repository at this point
Copy the full SHA f6dd201View commit details
Commits on Sep 19, 2022
-
Unify core avx and core_noavx to libpaddle (#46095) (#46113)
* unify core_avx and core_noavx * fix except error * revert mac compile logic * revert dylib to so * add core_noavx branch * remove core_noavx * replace paddle_core by lib paddle * polish var name * replace paddle_core by libpaddle * update custom device commit * polish code by comments
Configuration menu - View commit details
-
Copy full SHA for 4261ae3 - Browse repository at this point
Copy the full SHA 4261ae3View commit details -
[Cherry-pick] fix bug for TransformedDistribution (#46157)
fix bug for TransformedDistribution
Configuration menu - View commit details
-
Copy full SHA for a5d4f57 - Browse repository at this point
Copy the full SHA a5d4f57View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7a6db0a - Browse repository at this point
Copy the full SHA 7a6db0aView commit details -
Configuration menu - View commit details
-
Copy full SHA for dabb8f2 - Browse repository at this point
Copy the full SHA dabb8f2View commit details -
Configuration menu - View commit details
-
Copy full SHA for f4ec156 - Browse repository at this point
Copy the full SHA f4ec156View commit details -
[Eager] Optimize log (#45783) (#46133)
* make eager log readable * fix compile error * recover test * invoke ci again
Configuration menu - View commit details
-
Copy full SHA for e468e93 - Browse repository at this point
Copy the full SHA e468e93View commit details -
Configuration menu - View commit details
-
Copy full SHA for 860f607 - Browse repository at this point
Copy the full SHA 860f607View commit details -
[Cherry-pick][Auto Parallel] Improve the APIs (#46164)
* [AutoParallel] adapt gradient merge pass (#45915) * adapt gradient merge * fix op_role * fix strategy * [Auto Parallel] Gradient Fuse Allreduce (#45643) * bugfix (#45332) * dist embedding support lookup table v1 * add unitest * customize wait_comm * group gradients * bugfix * update program * [Auto Parallel] Improve the APIs (#45776) * [Auto Parallel] Use c++ dist attr in the completion process * [Auto Parallel] Add minor changes * [Auto Parallel] Use c++ dist attr in the completion process * [Auto Parallel] Add minor changes * [Auto Parallel] Add the serialization process for dist attrs * [Auto Parallel] Remove unnecessary comments * [Auto Parallel] Fix some bugs * [Auto Parallel] Fix the code style * [Auto Parallel] Remove unnecessary impls * [Auto Parallel] Fix the importing error * [Auto Parallel] Fix the copy from bugs of op dist attr * [Auto Parallel] Replace the use of constexpr if * [Auto Parallel] Redesign the shard_tensor, shard_op and ProcessMesh * [Auto Parallel] Change API of the completion unittest * [Auto Parallel] Fix the bug when set_attr an int * [Auto Parallel] Add the unittest for the serialization * [Auto Parallel] Add some unit tests * [Auto Paralle] Unify the strategy * [Auto Parallel] Improve the engine api * [Auto Parallel] Reset the changes made to the framework * [Auto Parallel] Change the engine unittest * [Auto Parallel] Update API of the completion and partitioner * [Auto Parallel] Update unit tests using engine api * update shard annotation * [Auto Parallel] Remove the modifications of other modules * [Auto Parallel] Add docs for APIs * add new strategy * [Auto Parallel] Replace the logger * [Auto Parallel] Restore the test_program.py * [Auto Parallel] Change the import rules * [Auto Parallel] Add the examples for Engine * [Auto Parallel] Do some minor changes * [Auto Parallel] Remove yaml dependency * [Auto Parallel] Fix the unittests * add valid after train * bug fix Co-authored-by: zhaoyingli <zhaoyingli@baidu.com> Co-authored-by: caozhou <caozhou@radi.ac.cn> Co-authored-by: caozhou <48191911+Caozhou1995@users.noreply.github.com> * [Auto Parallel] Bugfix allreduce fuse for MP (#46086) * bugfix * bugfix * typos fixed * update strategy (#46138) Co-authored-by: zhaoyingli <86812880+zhaoyinglia@users.noreply.github.com> Co-authored-by: JZ-LIANG <jianzhongliang10@gmail.com> Co-authored-by: zhaoyingli <zhaoyingli@baidu.com> Co-authored-by: caozhou <caozhou@radi.ac.cn> Co-authored-by: caozhou <48191911+Caozhou1995@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for c5cc427 - Browse repository at this point
Copy the full SHA c5cc427View commit details -
refactor mp. (#45803) (#46121)
* refactor mp. * update setup.py. * update mp_layers.py for compatibility. * add documents for mp_layers.py * update init.py * update collective.py. * update. * update mp_ops.py * update. * update code style. * update code style.
Configuration menu - View commit details
-
Copy full SHA for e5dc9d6 - Browse repository at this point
Copy the full SHA e5dc9d6View commit details -
Add INT8 support for fused_multi_transformer_op (#45284) (#46169)
Co-authored-by: RichardWooSJTU <37864677+RichardWooSJTU@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for db368d5 - Browse repository at this point
Copy the full SHA db368d5View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6b59a07 - Browse repository at this point
Copy the full SHA 6b59a07View commit details -
Configuration menu - View commit details
-
Copy full SHA for 855fdde - Browse repository at this point
Copy the full SHA 855fddeView commit details -
[Cherry-pick] Support bmm and bmm_grad in xpu (#45887) (#46132)
* [PHI] Support bmm and bmm_grad in xpu (#45887) * support bmm and bmm_grad in xpu * add error removal * test=kunlun * refactor code for better structure * test=kunlun * add fp16 kernel for bmm * test=kunlun * test=kunlun
Configuration menu - View commit details
-
Copy full SHA for 1c7e95c - Browse repository at this point
Copy the full SHA 1c7e95cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7f0c1f0 - Browse repository at this point
Copy the full SHA 7f0c1f0View commit details -
(cherry-pick)support some op backward refuse forward (#46201)
Configuration menu - View commit details
-
Copy full SHA for adab3c5 - Browse repository at this point
Copy the full SHA adab3c5View commit details -
Configuration menu - View commit details
-
Copy full SHA for 707d838 - Browse repository at this point
Copy the full SHA 707d838View commit details -
Configuration menu - View commit details
-
Copy full SHA for a056601 - Browse repository at this point
Copy the full SHA a056601View commit details -
[cherry-pick] [dy2static] support user to use decorator in their prog…
…ram (#46194) * [dy2static] support user to use decorator in their program (#45768) * support deco * fix deco ast type * arg_str * 1 * support callable deco * code style * codestyle * test_error * fix decos in another file * recover conflict codes * [BugFix] fixed a bug in decorator transformer, it can not analyze decorator with params correctly (#46055) * fix deco call * add raise * add test * add warn, fix paddle api * fix error type * fix coverage
Configuration menu - View commit details
-
Copy full SHA for d1ce974 - Browse repository at this point
Copy the full SHA d1ce974View commit details -
Configuration menu - View commit details
-
Copy full SHA for 45a3c65 - Browse repository at this point
Copy the full SHA 45a3c65View commit details -
[cherry-pick] add abs,mean,sum,ge,gt,pow,etc higher-order differentia…
…tion operators (#46184) * [cherry-pick] extend reduce_sum,reduce_sum,eq,ne,ge,abs,pow,etc higher order operators * add reduce_mean,reduce_sum primitive ops * add ne_p gt_p primitive operators * add ge_p abs_p primitive oparators * add cast primitive operators * add pow,square prim2oirg rules * add elementwise_div orig2prim rule * [cherry-pick] add mean,sum,ge,gt,ne,abs,etc higher-order differentiation operators(#45888) * add reduce_mean,reduce_sum primitive ops * add ne_p gt_p primitive operators * add ge_p abs_p primitive oparators
Configuration menu - View commit details
-
Copy full SHA for ad8beaa - Browse repository at this point
Copy the full SHA ad8beaaView commit details -
[vision.ops.nms] Fix return order error and duplicate results with sp…
Configuration menu - View commit details
-
Copy full SHA for be84cac - Browse repository at this point
Copy the full SHA be84cacView commit details -
Configuration menu - View commit details
-
Copy full SHA for 4bced24 - Browse repository at this point
Copy the full SHA 4bced24View commit details
Commits on Sep 20, 2022
-
Fix wrong eigen header include (#46082) (#46202)
* fix wrong eigen header include * fix complie bug * fix nan_inf_utils_detail * fix resource_manager * fix conv_miopen_helper
Configuration menu - View commit details
-
Copy full SHA for ac8cce2 - Browse repository at this point
Copy the full SHA ac8cce2View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7712ce1 - Browse repository at this point
Copy the full SHA 7712ce1View commit details -
[Cherry-pick] Update layoutautotune for inplace (#45826) (#46226)
Configuration menu - View commit details
-
Copy full SHA for c0324e8 - Browse repository at this point
Copy the full SHA c0324e8View commit details -
Run_program_op add scope cache & reuse (#45813) (#46223)
* add scope cache & reuse * add gc scope for end of each train step * del scope reuse for jit * refine code * test
Configuration menu - View commit details
-
Copy full SHA for 4f28a4c - Browse repository at this point
Copy the full SHA 4f28a4cView commit details -
Configuration menu - View commit details
-
Copy full SHA for c384b00 - Browse repository at this point
Copy the full SHA c384b00View commit details -
(cherry-pick)Support some op refuse forward and fix some bugs (#46211)
Configuration menu - View commit details
-
Copy full SHA for bc92d5f - Browse repository at this point
Copy the full SHA bc92d5fView commit details -
[Eager] Fix linspace error in amp (#46088) (#46206)
* fix linspace error in amp * fix log * fix amp error
Configuration menu - View commit details
-
Copy full SHA for 38c0fd0 - Browse repository at this point
Copy the full SHA 38c0fd0View commit details -
[Cherry-pick] Sparse add InferMeta (#46235)
cherry-pick : #46016, #46021, #45974 * [Sparse]Sparse add support gpu (#45974) * [Sparse]Remove unused code (#46021) * [Sparse] Add infer meta (#46016)
zhangkaihuo authoredSep 20, 2022 Configuration menu - View commit details
-
Copy full SHA for fd8ec4a - Browse repository at this point
Copy the full SHA fd8ec4aView commit details -
Fix TransDataBackend Error when call unsqueeze using MKL Tensor (#46094…
…) (#46186) * Fix TransDataBackend Error when call unsqueeze using MKL Tensor * Add UT * Refine UT
Configuration menu - View commit details
-
Copy full SHA for 5034030 - Browse repository at this point
Copy the full SHA 5034030View commit details -
[cherry-pick] Refine thread pool config of interpretercore (#46219)
* add config * add config * follow comments * fix serial run
Configuration menu - View commit details
-
Copy full SHA for 1418a71 - Browse repository at this point
Copy the full SHA 1418a71View commit details -
[Paddle-TRT] Support matmul_v2 in Paddle-TensorRT (#46177)
* Support matmul_v2 in Paddle-TensorRT converter.
Configuration menu - View commit details
-
Copy full SHA for 654807c - Browse repository at this point
Copy the full SHA 654807cView commit details -
[Release/2.4][Cherry-pick] Fix bug of reduce_sum op (#46160)
* Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result is wrong. * Cherry-pick of PR 46045 * Fix bug of reduce_sum kp op. * Fix bug of reduce_sum kp operator compilation. If compilation device is XPU, eigen kernel should be ignored.
Configuration menu - View commit details
-
Copy full SHA for 759736d - Browse repository at this point
Copy the full SHA 759736dView commit details -
[Cherry-pick] Fix amp error cp (#46272)
* [Eager] Fix ocr (#46124) * fix linspace error in amp * fix log * fix amp error * fix ocr error which caused by amp * add more check * rename dtype ns * [Eager Bug fix]Fix Detection (#46147) * fix linspace error in amp * fix log * fix amp error * Revert "Simplify size op impl (#45808)" This reverts commit c252b1d. * fix_seg * fix detection Co-authored-by: Chen Weihang <sunny_cwh@163.com> Co-authored-by: Chen Weihang <sunny_cwh@163.com>
Configuration menu - View commit details
-
Copy full SHA for da173c4 - Browse repository at this point
Copy the full SHA da173c4View commit details -
[Cherry-Pick][AutoParallel] change import way and fix strategy (#46270)
* [Auto Parallel] Change the import way of Auto Parallel (#46115) * fix strategy (#46256) * [Auto Parallel] performance improvement for Sharding-DP hybrid parallelism (#46180) * remove no need grad allreduce communication when sharding-dp * remove no need grad allreduce communication when sharding-dp * bugfix * bugfix * bugfix Co-authored-by: Yulong Ao <aoyulong@baidu.com> Co-authored-by: JZ-LIANG <jianzhongliang10@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for c43ebfc - Browse repository at this point
Copy the full SHA c43ebfcView commit details -
[PolishComments] Polish some code comments (#46032) (#46261)
* polish code comments * polish data_device_transform.cc
Configuration menu - View commit details
-
Copy full SHA for 42e56f6 - Browse repository at this point
Copy the full SHA 42e56f6View commit details -
[Paddle-TRT] Full support for ops with persistable input (#45545) (#4…
…6280) * Move ITensor construction for Weight (persistable variable) from OpConvert to TensorRTEngine.
Configuration menu - View commit details
-
Copy full SHA for adb2f5e - Browse repository at this point
Copy the full SHA adb2f5eView commit details -
[cherry-pick][xpu] update xdnn activations (#46282)
* [XPU] update xdnn activations. (#46246) * [XPU] update xpu cmake. test=kunlun
Configuration menu - View commit details
-
Copy full SHA for a43f960 - Browse repository at this point
Copy the full SHA a43f960View commit details -
Configuration menu - View commit details
-
Copy full SHA for 230b9a8 - Browse repository at this point
Copy the full SHA 230b9a8View commit details -
Configuration menu - View commit details
-
Copy full SHA for bab1109 - Browse repository at this point
Copy the full SHA bab1109View commit details -
Configuration menu - View commit details
-
Copy full SHA for 461099c - Browse repository at this point
Copy the full SHA 461099cView commit details -
* back fl * delete ssl cert * . * make warning * . * unittest paral degree * solve unittest * heter & multi cloud commm ready * . * . * fix gloo compile warning * adapt for nn fl-ps * flps del fake-init op * add learning_rate_0 intializer op
Configuration menu - View commit details
-
Copy full SHA for 3e8b322 - Browse repository at this point
Copy the full SHA 3e8b322View commit details
Commits on Sep 21, 2022
-
Configuration menu - View commit details
-
Copy full SHA for b027652 - Browse repository at this point
Copy the full SHA b027652View commit details -
Configuration menu - View commit details
-
Copy full SHA for 311eb87 - Browse repository at this point
Copy the full SHA 311eb87View commit details -
fix CI sccache install dir (#46267) (#46327)
Co-authored-by: Zhou Wei <1183042833@qq.com>
Configuration menu - View commit details
-
Copy full SHA for f3d5c85 - Browse repository at this point
Copy the full SHA f3d5c85View commit details -
[JitLayer]Erase out vars in scope to avoid data rewritinig (#46249) (#…
…46273) * [JitLayer]Erase out vars to avoid data rewrittinig * Fix code comments
Configuration menu - View commit details
-
Copy full SHA for b74c092 - Browse repository at this point
Copy the full SHA b74c092View commit details -
[Cherry-pick][BugFix]Fix pooling output_size bug if encounter list[Te…
…nsor] (#46360) * [Check]Enhance pooling output_size type check * add unittest
Configuration menu - View commit details
-
Copy full SHA for cc3e7cd - Browse repository at this point
Copy the full SHA cc3e7cdView commit details
Commits on Sep 22, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 083853c - Browse repository at this point
Copy the full SHA 083853cView commit details -
[Dygraph] Fix bugs of mp in eager mode (#46303) (#46396)
* fix bugs of mp * fix bugs of mp * update * update * fix bug
Configuration menu - View commit details
-
Copy full SHA for 372505b - Browse repository at this point
Copy the full SHA 372505bView commit details -
logger manager (#45909) (#46087)
uniform logger manager in FleetAPI. hidde API under distributed/utils which users don't need.
Configuration menu - View commit details
-
Copy full SHA for 7eb046c - Browse repository at this point
Copy the full SHA 7eb046cView commit details
Commits on Sep 23, 2022
-
[Cherry-Pick][BugFix]Fix reduce_mean/min/sum/prod, cumsum grad_op inf…
…ershape bug (#46409) * [BugFix]Fix reduce_mean/min/sum/prod, cumsum grad_op infershape bug * fix typo * fix typo
Configuration menu - View commit details
-
Copy full SHA for 484377c - Browse repository at this point
Copy the full SHA 484377cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 6a50833 - Browse repository at this point
Copy the full SHA 6a50833View commit details -
Configuration menu - View commit details
-
Copy full SHA for 980292c - Browse repository at this point
Copy the full SHA 980292cView commit details -
[BugFix] fixed a bug that deco_name can't be parsed corrected (#46297) (
#46366) * use re replace judge by case * simplify re
Configuration menu - View commit details
-
Copy full SHA for cbf3f4b - Browse repository at this point
Copy the full SHA cbf3f4bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 55f73ba - Browse repository at this point
Copy the full SHA 55f73baView commit details
Commits on Sep 24, 2022
-
Remove audio ParameterError (#46316) (#46468)
* unexpose audio ParameterError * clean audio utils api
Configuration menu - View commit details
-
Copy full SHA for 0ab2672 - Browse repository at this point
Copy the full SHA 0ab2672View commit details
Commits on Sep 26, 2022
-
[cherrypick] Fix elementwise_sub sign reverse for mkldnn (#46107)
* fix sub sign reverse for mkldnn * refactor code as comment * remove useless
Configuration menu - View commit details
-
Copy full SHA for 6990edf - Browse repository at this point
Copy the full SHA 6990edfView commit details -
Configuration menu - View commit details
-
Copy full SHA for 4a8aa6d - Browse repository at this point
Copy the full SHA 4a8aa6dView commit details -
* back fl * delete ssl cert * . * make warning * . * unittest paral degree * solve unittest * heter & multi cloud commm ready * . * . * fix gloo compile warning * adapt for nn fl-ps * flps del fake-init op * add learning_rate_0 intializer op * bug fix * . * .
Configuration menu - View commit details
-
Copy full SHA for 536d9d8 - Browse repository at this point
Copy the full SHA 536d9d8View commit details
Commits on Sep 27, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 8089a1f - Browse repository at this point
Copy the full SHA 8089a1fView commit details -
Configuration menu - View commit details
-
Copy full SHA for dc3a3f1 - Browse repository at this point
Copy the full SHA dc3a3f1View commit details -
[cherry-pick] clear extra attrs of some ops in OpMaker (#45845, #45984,…
… 46060) (#46218) * Clear extra attrs of elementwise op in OpMaker (#45845) * clear extra attrs of elementwise op in opmaker * fix op_debug_string_test * fix bug of grad_add * fix sort of runtime attrs * Clear extra attrs of scale in OpMaker (#45984) * clear extra attr of scale in opmaker * fix sum bug * fix merge conflict * fix minus * Clear extra attributes of some Op in OpMaker (Part4) (#46060) * clear extra attr of some ops in opmaker * revert clear use_cudnn for pool * fix test_operator_desc * fix Attr interface of OperatorBase * fix code stype
Configuration menu - View commit details
-
Copy full SHA for 0cc2251 - Browse repository at this point
Copy the full SHA 0cc2251View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5711bbe - Browse repository at this point
Copy the full SHA 5711bbeView commit details -
Configuration menu - View commit details
-
Copy full SHA for 5dab0b0 - Browse repository at this point
Copy the full SHA 5dab0b0View commit details
Commits on Sep 28, 2022
-
Configuration menu - View commit details
-
Copy full SHA for a77a6f6 - Browse repository at this point
Copy the full SHA a77a6f6View commit details -
[cherry-pick] Clear extra attrs of some ops in OpMaker (#46150, #46321,
#46418, #46451, #46457) (#46553) * Clear extra attributes of some Op in OpMaker (Part4) (#46060) * clear extra attr of some ops in opmaker * revert clear use_cudnn for pool * fix test_operator_desc * fix Attr interface of OperatorBase * clear extra attrs of condition op in opmaker (#46150) * Clear extra attrs of lookup_table_v2 in OpMaker (#46321) * clear extra attrs of look_up_table_v2 in opmaker * fix bug * clear extra attrs of quantize op in opmaker (#46418) * delete repeated item * clear extra attrs of distribute op in opmaker (#46451) * clear extra atts of sequence_softmax in opmaker (#46457)
Configuration menu - View commit details
-
Copy full SHA for b2e4211 - Browse repository at this point
Copy the full SHA b2e4211View commit details -
Fix libpaddle soname mismatch error (#46344) (#46576)
* fix libpaddle soname mismatch error * fix windows failed * polish linux and windows make impl * unify winddows lib name * fix windows error * revert copy dst change * revert naming change * revert windows change * fix gpups compile failed
Configuration menu - View commit details
-
Copy full SHA for 1c22ed7 - Browse repository at this point
Copy the full SHA 1c22ed7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3f35e63 - Browse repository at this point
Copy the full SHA 3f35e63View commit details
Commits on Sep 29, 2022
-
Configuration menu - View commit details
-
Copy full SHA for f5956be - Browse repository at this point
Copy the full SHA f5956beView commit details -
Configuration menu - View commit details
-
Copy full SHA for d90db9b - Browse repository at this point
Copy the full SHA d90db9bView commit details -
[cherry-pick] Open the clip_extra flag in save_inference_model (#46577)
* set flag of clip_extra in save_inference_model to true (#46151) * open the clip_extra flag in paddle.static.save_inference_model, test=allcase (#46456) * Open the clip_extra flag in TracedLayer.save_inference_model (#46473) * open the clip_extra flag in paddle.static.save_inference_model, test=allcase * set the defalut value of clip_extra in TracedLayer from False to True, test=allcase * update english doc of paddle.static.save_inference_model, test=document_fix (#46484) * Fix clip_extra logic in remove_training_info (#46534) * fix clip_extra code in remove_training_info * revert rnn opmaker clear
Configuration menu - View commit details
-
Copy full SHA for d67da3d - Browse repository at this point
Copy the full SHA d67da3dView commit details -
[cherry-pick] Add FP16 support for uniform in dygraph mode on Nvidia …
…GPU (#46641) Add FP16 support for uniform in dygraph mode on Nvidia GPU Dev PR link PR46212
Configuration menu - View commit details
-
Copy full SHA for a58663f - Browse repository at this point
Copy the full SHA a58663fView commit details
Commits on Oct 9, 2022
-
[Dy2Static] refactor the return transformer (#45900) (#46205)
* 1. refactor the return transformer. 2. fix some bugs in return transformer. * support raise error while return stmt's father is For or while * fix ci error. * fix ci error and add some unittest * code format * fix ci error
Configuration menu - View commit details
-
Copy full SHA for 4282af6 - Browse repository at this point
Copy the full SHA 4282af6View commit details
Commits on Oct 10, 2022
-
[PHI] Shape op migration (#46051) (#46724)
* First approach * Shape kernel corrected * Compilation error fixed * Resize corrected * Registered types added * Mistake corrected & types added * sum kernel deleted Co-authored-by: Paulina Gacek <paulina.gacek.pl@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for 3cc3f60 - Browse repository at this point
Copy the full SHA 3cc3f60View commit details -
[PHI] migrate softmax_grad kernel (#46257) (#46725)
* init * remove softmaxop * merge dev * correct dir * style
Configuration menu - View commit details
-
Copy full SHA for 44ecae6 - Browse repository at this point
Copy the full SHA 44ecae6View commit details -
[PHI] Migrate slice, slice_grad, split, pad and pad3d oneDNN kernels (#…
Configuration menu - View commit details
-
Copy full SHA for 51a91fe - Browse repository at this point
Copy the full SHA 51a91feView commit details -
[cherry-pick] [PHI] Migrate sgd and stack oneDNN kernels (#46374) (#4…
Configuration menu - View commit details
-
Copy full SHA for 25d61cd - Browse repository at this point
Copy the full SHA 25d61cdView commit details -
[cherry-pick] [PHI] Migrate concat+grad, expand+grad, fill_constant ……
… oneDNN kernels (#45863) (#46727) * [PHI] Migrate concat+grad, expand+grad, fill_constant, nearest_interp and bilinear_interp oneDNN kernels (#45863) * Migrate concat+grad, expand+grad, fill_constant, nearest_interp_v2 and bilinear_interp_v2 oneDNN kernels to PHI * Remove old namespace variable * Fix invalid out dims error * Add mutable_data method to concat output * Add check for -1 dim before computing out_dims * Capitalize oneDNNGetDataType function name * Change fill_constant kernel to correct PHI kernel * Attempt to fix dims error * Fix fill_constant (full) kernel * update dependencies Co-authored-by: Piotr Paturej <48731682+piotrekobi@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for fdd0d6d - Browse repository at this point
Copy the full SHA fdd0d6dView commit details -
Configuration menu - View commit details
-
Copy full SHA for d8daf64 - Browse repository at this point
Copy the full SHA d8daf64View commit details -
Fix gather op convert for Paddle-TensorRT (#46779) (#46825)
* fix gather op convert to only support int32 index as input. * add ut
feng_shuai authoredOct 10, 2022 Configuration menu - View commit details
-
Copy full SHA for a0e0341 - Browse repository at this point
Copy the full SHA a0e0341View commit details
Commits on Oct 11, 2022
-
Configuration menu - View commit details
-
Copy full SHA for d091d1b - Browse repository at this point
Copy the full SHA d091d1bView commit details -
[BugFix]Fix concat bugs when call onednn kernel (#46518) (#46845)
* fix concat bug * fix ci bugs * fix ci bugs
Configuration menu - View commit details
-
Copy full SHA for 6a6c749 - Browse repository at this point
Copy the full SHA 6a6c749View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9cc3f69 - Browse repository at this point
Copy the full SHA 9cc3f69View commit details -
Configuration menu - View commit details
-
Copy full SHA for f556549 - Browse repository at this point
Copy the full SHA f556549View commit details -
Revert pool+grad oneDNN kernel conversion (#45989) (#46860)
Co-authored-by: Piotr Paturej <48731682+piotrekobi@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 7b3837e - Browse repository at this point
Copy the full SHA 7b3837eView commit details -
[cherry-pick] [PHI] relu6_grad kernel (#46501) (#46862)
* [PHI] Migrate gelu kernels (#45596) * gaussian random * mkldnn to onednn renaming * fix merge conflicts * remove fluid code * onednn renaming * gelu fwd * sort activations * gelu gradient * remove unused macros * merge conflicts * fix merge conflicts * remove extra contraint from gelu op * [PHI] relu6_grad kernel (#46501) * Relu6 * remove fluid handler * add individual kernel signature * coding style * replace bounded_relu with clip * whitespace * code style
Configuration menu - View commit details
-
Copy full SHA for 2bcbf8b - Browse repository at this point
Copy the full SHA 2bcbf8bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 2c6bd4a - Browse repository at this point
Copy the full SHA 2c6bd4aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 2190da2 - Browse repository at this point
Copy the full SHA 2190da2View commit details -
set_value_op: add support for complex types (#46885)
Feiyu Chan authoredOct 11, 2022 Configuration menu - View commit details
-
Copy full SHA for b051455 - Browse repository at this point
Copy the full SHA b051455View commit details
Commits on Oct 12, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 08d233f - Browse repository at this point
Copy the full SHA 08d233fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 61273c0 - Browse repository at this point
Copy the full SHA 61273c0View commit details
Commits on Oct 13, 2022
-
[cherry-pick] [PHI] transpose2_grad op migration (#46139) (#46873)
* Revert pool+grad oneDNN kernel conversion (#45989) * [PHI] transpose2_grad op migration (#46139) * op migrated, Copy(OneDNNContext, ...) added * mutable_data & op registration in fluid removed * refactoring * OneDNNGetDataType to uppercase * missing cpu check added, handler moved to .h file * name changed to transpose_grad * Copy changed back to TensorCopy * Resizing corrected, Copy(OneDNNContext) removed Co-authored-by: Piotr Paturej <48731682+piotrekobi@users.noreply.github.com> Co-authored-by: Paulina Gacek <paulina.gacek@intel.com>
Configuration menu - View commit details
-
Copy full SHA for 0280c0b - Browse repository at this point
Copy the full SHA 0280c0bView commit details -
[Cherry-pick] Add fp16 dtype support for set_value op (#46906)
Fix set_value failure when source tensor is fp16 Dtype and destiny value is a number (dev PR link:#46801)
Configuration menu - View commit details
-
Copy full SHA for 100a075 - Browse repository at this point
Copy the full SHA 100a075View commit details -
Configuration menu - View commit details
-
Copy full SHA for d90aaa6 - Browse repository at this point
Copy the full SHA d90aaa6View commit details
Commits on Oct 14, 2022
-
Configuration menu - View commit details
-
Copy full SHA for b8677c0 - Browse repository at this point
Copy the full SHA b8677c0View commit details -
[cherry-pick 2.4][inference] fix reshape2 opteller (#46871)
* fix reshape2 opteller; add elementwise min/max register for tensorrt
Configuration menu - View commit details
-
Copy full SHA for 535d757 - Browse repository at this point
Copy the full SHA 535d757View commit details -
[BUG]Fix expand_as_v2 bug while X and Y with different dtype (#46950) (…
…#46999) * [BUG]Fix expand_as_v2 bug while X and Y with different dtype * fix commit
Configuration menu - View commit details
-
Copy full SHA for 4b47265 - Browse repository at this point
Copy the full SHA 4b47265View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2744432 - Browse repository at this point
Copy the full SHA 2744432View commit details -
Configuration menu - View commit details
-
Copy full SHA for 8f1ac7c - Browse repository at this point
Copy the full SHA 8f1ac7cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 84333cf - Browse repository at this point
Copy the full SHA 84333cfView commit details -
Configuration menu - View commit details
-
Copy full SHA for 82db499 - Browse repository at this point
Copy the full SHA 82db499View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5c2bea1 - Browse repository at this point
Copy the full SHA 5c2bea1View commit details
Commits on Oct 17, 2022
-
[Cherry-Pick]Move valid check from python to kernel (#46980)
为了提升性能,将label的边界检查从python端转移到kernel内,减少额外op的调用,如min、max和同步拷贝等 当前的模板参数IgnoreIndex仅在ignore_index取值范围在[0, dim)时才生效,但是当某个label值超出了边界,ignore_index等于该label,这种情况下是应该仍然能正常计算。虽然当前的计算逻辑在结果上不会出错,但逻辑上仍是有问题的,且模板参数IgnoreIndex是没有必要的
Configuration menu - View commit details
-
Copy full SHA for 8bfd45a - Browse repository at this point
Copy the full SHA 8bfd45aView commit details -
Configuration menu - View commit details
-
Copy full SHA for a1cdbad - Browse repository at this point
Copy the full SHA a1cdbadView commit details -
[IPU] paddle-inference support custom-ops (#45235) (#46868)
* paddle-inference support custom-ops Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai> * fix tolower Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai> Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai>
Configuration menu - View commit details
-
Copy full SHA for bd89be1 - Browse repository at this point
Copy the full SHA bd89be1View commit details -
Configuration menu - View commit details
-
Copy full SHA for d913bc9 - Browse repository at this point
Copy the full SHA d913bc9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 8c6c79a - Browse repository at this point
Copy the full SHA 8c6c79aView commit details -
[cherry-pick] Fix the bug of exporting model in dygraph QAT (#47028)
* fix dygraph new format quant * fix unittest * fix conflict
Configuration menu - View commit details
-
Copy full SHA for 7eef05c - Browse repository at this point
Copy the full SHA 7eef05cView commit details -
Optimize performance of depthwise_conv (#46896)
Optimize performance of depthwise_conv Config: input[2048, 1024, 4, 4], filter[1024, 1, 4, 4], stride=1, pad=0, dilation=1
Configuration menu - View commit details
-
Copy full SHA for 976af0d - Browse repository at this point
Copy the full SHA 976af0dView commit details -
[cherry-pick]Sparse static graph (#46838)
zhangkaihuo authoredOct 17, 2022 Configuration menu - View commit details
-
Copy full SHA for 10225d2 - Browse repository at this point
Copy the full SHA 10225d2View commit details -
[Cherry-pick] Collective communication APIs (#46922)
* Support both use_calc_stream and sync_op in send recv APIs (#46023) * Support both use_calc_stream and sync_op in allgather API (#46295) * Support both use_calc_stream and sync_op in collective communication API (#46761) * Move group and all reduce from collective to communication (#45848) * Completes bfloat16 dtype for collective api in eager mode (#45844) * Fix collective APIs cannot be recognized when building docs (#46962) Co-authored-by: LiYuRio <63526175+LiYuRio@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 5fba2a9 - Browse repository at this point
Copy the full SHA 5fba2a9View commit details
Commits on Oct 18, 2022
-
[Cherry pick] trt pool2d adaptive ifx (#47069)
* draft with debug print * remove debug print * bug fix for ci
Configuration menu - View commit details
-
Copy full SHA for 5f6b9f1 - Browse repository at this point
Copy the full SHA 5f6b9f1View commit details -
[cherry-pick] Fix perf issues of mp/pp/fuse in eager mode (#47071)
Configuration menu - View commit details
-
Copy full SHA for b84edd9 - Browse repository at this point
Copy the full SHA b84edd9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5b64214 - Browse repository at this point
Copy the full SHA 5b64214View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5a44c12 - Browse repository at this point
Copy the full SHA 5a44c12View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5fef043 - Browse repository at this point
Copy the full SHA 5fef043View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2cc8797 - Browse repository at this point
Copy the full SHA 2cc8797View commit details -
Configuration menu - View commit details
-
Copy full SHA for de6f15b - Browse repository at this point
Copy the full SHA de6f15bView commit details
Commits on Oct 19, 2022
-
[ cherrypick] Construct exec and ctx only once in cond op to speed up (…
…#47012) Construct exec and ctx only once in cond op to speed up
Configuration menu - View commit details
-
Copy full SHA for fcb9c0b - Browse repository at this point
Copy the full SHA fcb9c0bView commit details -
update audio api examples (#46938) (#47126)
* update audio api examples * fix format * format * fix * test api * fix format * fix static check error * fix doc error * fix ci * fix api error * update api.spec * fix ci * fix typo in window gaussian
Configuration menu - View commit details
-
Copy full SHA for f08c104 - Browse repository at this point
Copy the full SHA f08c104View commit details -
[Dy2St]Fix recurrent op eager deletion pass error in dy2st (#47105) (#…
…47134) [CherryPick][Dy2St]Fix recurrent op eager deletion pass error in dy2st
Configuration menu - View commit details
-
Copy full SHA for 69515e9 - Browse repository at this point
Copy the full SHA 69515e9View commit details -
Add enable_partial_send_recv switch in pipeline_configs (#46992) (#47083
) * Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result is wrong. * Support allow_partial switch, which can be configure in pipeline_configs. If sent tensor are not the same from different hosts, they shouldn't been sent partially and then concated as a whole tensor. * Change name allow_partial to enable_partial_send_recv. * Add global variable _enable_partial_send_recv
Configuration menu - View commit details
-
Copy full SHA for 1d015f1 - Browse repository at this point
Copy the full SHA 1d015f1View commit details -
[CherryPick] Support TypeHint for function decorated by @to_static (#…
…47147) * [Dy2Static] Support TypeHint for function decorated by @to_static (#47121) * Add TypeHint Transformer * add unittest for typehint transformer * [Dy2Static] Remove GradTransformer (#47063) * [Dy2Static] Remove GradTransformer 1. fix einsum infershape bugs. 2. remove grad_transformer and unify paddle.grad and paddle.static.gradient. 3. add dygraph_and_dy2static_only decorator for dy2static. * fix bugs * rename
Configuration menu - View commit details
-
Copy full SHA for 247ef47 - Browse repository at this point
Copy the full SHA 247ef47View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5a9befe - Browse repository at this point
Copy the full SHA 5a9befeView commit details -
Add unsigned int8 scale propagation (#46378) (#47156)
* Add unsigned int8 propagation * Add or modify unit tests * Correct concat scale checking * Apply review suggestions * Corrections Co-authored-by: joanna.wozna.intel <joanna.wozna@intel.com>
Configuration menu - View commit details
-
Copy full SHA for 66dccd7 - Browse repository at this point
Copy the full SHA 66dccd7View commit details -
[cherry-pick] strided_slice grad add fp16 support (#47159)
* strided_slice grad add fp16 support
Zhang Ting authoredOct 19, 2022 Configuration menu - View commit details
-
Copy full SHA for 23f2a4e - Browse repository at this point
Copy the full SHA 23f2a4eView commit details -
[Cherry-Pick][AutoParallel] auto_parallel cherry-pick to release2.4 (#…
…47145) * [Auto Parallel] Make Engine class callable (#46416) * [Auto Parallel] Imporve the user-defined fetches and logging * [Auto Parallel] Make Engine class callable * [Auto Parallel] Update the data loading of tuner * Print IPS in auto parallel Engine (#46554) * [AutoParallel] fix dist_split (#46505) * [AutoParallel] fix dist_split * add unittest * update cmakelist * [AutoParallel] fix sharding (#46572) * [AutoParallel] fix process_mesh (#46583) * [AutoParallel] fix reshard when train with eval (#46605) * [AutoParallel] fix reshard when train with eval * fix mppp * [AutoParallel] fix amp when predict (#46637) * [Auto Parallel]Update comp cost and completion for gpt auto search (#46387) * update comp cost and completion for gpt auto search * add unittest * [Auto Parallel] Fix bugs caused by the inconsistent outputs of Engine API (#46633) * [Auto Parallel] Unify the logger and outputs of Engine API * [Auto Parallel] Fix the bugs of to_static * [Auto Parallel] Adjust the test_to_static.py * [Auto Parallel] Improve the fine-grained APIs (#46552) * [Auto Parallel] Suppport different dataloaders * [Auto Parallel] Add num_shards config for dataset * [Auto Parallel] Unify the logger and outputs of Engine API * [Auto Parallel] Fix the bugs of to_static * [Auto Parallel] Adjust the test_to_static.py * [Auto Parallel] Add the prepare API and replace __call__ with run * [Auto Parallel] Improve the private implementations of Engine * [Auto Parallel] Set capacity of dataloader for opt tuning * [Auto Parallel] [WIP] Change the fine-grained API * [Auto Parallel] Improve APIs to support different user cases * [Auto Parallel] Add removed config * [Auto Parallel] Add imports * [Auto Parallel] Fix bugs for to_static * [Auto Parallel] Remove unnecessary imports * bugfix (#46921) * [Auto Parallel] Fix the bug for None labels (#46987) * [AutoParallel] adapt for gpt-gen (#46771) * for gpt-gen * fix reshard * adapt assign and shape op * add dist_assign & unittest * add conditional block unittest * rename unittest * [Auto Parallel] Fix the bug of completion (#47056) * [Auto Parallel] Fix the bug for None labels * [Auto Parallel] Fix the completion bug * [AutoParallel] add callbacks (#47014) * [AutoParallel] add callbacks * fix unittest * fix dist_context * fix engine * fix cmakelist * fix unittest's returns * fix cmakelist * [Auto Parallel] Add cost interface (#47043) * add cost interface * update inferface and add unittest * update unittest * update inferface * [Auto Parallel]Add parallel tuner (#46189) * add parallel tuner * add unittest * fix unittest * set timeout of unittest * set unittest timeout * fix auto_mode setting * update unittest * sync from develop and update unittest * remove unused import * update unittest * update cmakelist * add unittests Co-authored-by: Yulong Ao <aoyulong@baidu.com> Co-authored-by: Ruibiao Chen <chenruibiao@baidu.com> Co-authored-by: caozhou <48191911+Caozhou1995@users.noreply.github.com> Co-authored-by: JZ-LIANG <jianzhongliang10@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for 90b3179 - Browse repository at this point
Copy the full SHA 90b3179View commit details
Commits on Oct 20, 2022
-
Fix cannot import
paddle.distributed
in python 3.6 on release/2.4 (#……47141) * fix: fix incorrect import * fix: fix incorrect usage
Configuration menu - View commit details
-
Copy full SHA for c894d91 - Browse repository at this point
Copy the full SHA c894d91View commit details -
[Cherry-pick][Release/2.4] support pure bfloat16 for more ops
support pure bfloat16 for more ops
Configuration menu - View commit details
-
Copy full SHA for da7d2f2 - Browse repository at this point
Copy the full SHA da7d2f2View commit details -
[Cherry-pick][Release/2.4] Fix some operators when the tensor.numel()…
… > INT32_MAX (#47191) Fix some operators when the tensor.numel() > INT32_MAX
Configuration menu - View commit details
-
Copy full SHA for c74bf01 - Browse repository at this point
Copy the full SHA c74bf01View commit details -
Configuration menu - View commit details
-
Copy full SHA for 4c92524 - Browse repository at this point
Copy the full SHA 4c92524View commit details -
[Cherry-pick] layernorm shift partation enhance (#47086)
* Enhance the layernorm shift partation fuse op when shift size > 0 (roll shifting) * fix cherry-pick test
Configuration menu - View commit details
-
Copy full SHA for 9ed1454 - Browse repository at this point
Copy the full SHA 9ed1454View commit details -
Configuration menu - View commit details
-
Copy full SHA for 09b1923 - Browse repository at this point
Copy the full SHA 09b1923View commit details -
[Paddle-TRT][Cherry-Pick]Rewrite strided_slice converter using shape …
…tensor (#47153) * stride_to_24 * fix CI failing
Configuration menu - View commit details
-
Copy full SHA for 68c4ac3 - Browse repository at this point
Copy the full SHA 68c4ac3View commit details -
[cherry-pick] Fix quantize model deploy bug in MKLDNN (#47119)
* Fix quantize model deploy bugs when using MKLDNN (#45920) * fix immutable op quantize bugs * fix * fix build bug * fix test * notest,test=inference * fix ppyoloe acc drop bugs * fix test * fix test * add test * fix * fix * fix test * fix refined name bug * fix test * bias fix * fix matmul weight dequant bug * re-ci * fix tester * fix test * fix tester * update weight dequantize func * update code * update test for converage * update test * update cmake * update cmakelist * update code * rerun ci * remove useless code * re-ci * update code * update code * fix header * update code for log
Configuration menu - View commit details
-
Copy full SHA for c2d344d - Browse repository at this point
Copy the full SHA c2d344dView commit details -
Configuration menu - View commit details
-
Copy full SHA for 3d647b1 - Browse repository at this point
Copy the full SHA 3d647b1View commit details -
Add value check & error message for gather_tree (#47051) (#47221)
Add value check & error message for gather_tree cherry-pick #47051
Configuration menu - View commit details
-
Copy full SHA for 6712e26 - Browse repository at this point
Copy the full SHA 6712e26View commit details -
[cherry pick] Add FusedMultiTransformer fuse pass for GPT3 (#47150)
* add fused_attention_pass. test=develop * support fp16. test=develop * fix format. test=develop
Configuration menu - View commit details
-
Copy full SHA for 396427a - Browse repository at this point
Copy the full SHA 396427aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 50d4fa5 - Browse repository at this point
Copy the full SHA 50d4fa5View commit details -
[Cherry-pick] Simplify conv codes and fix cache and autotune bugs. (#…
Configuration menu - View commit details
-
Copy full SHA for c0ed872 - Browse repository at this point
Copy the full SHA c0ed872View commit details -
Add paddle audio dataset & backend (#45939) (#47230)
* add audio feature dataset * fix coding style * fix coding style2 * rm librosa * rm voxceleb * rm librosa in test * add scipy fftpack * add functional * fix setup * fix setup2 * rm colorlog * refactor dataset __init__.py * fix converage * fix librosa import error * fix windows test * fix windows ci * rm datasets * fix setup * remove testdata * add librosa in requirement * add librosa in requirement2 * change librosa to 0.8.1 * update ci docker * fix ci error * fix ci error2 * fix ci coverage * fix converage * fix coverage * rm audio_base in test, notest,test=coverage * fix copyright * rm backend * add datast in __init__ * rm compliance&&add function test * fix setup * fix windows * fix windows2 * fix test timeout * add backend & datasets * fix bugs * fix ci time issue * add dataset test * rm test_audio_feature * avoid windows isssue, tmp * note windows isssue * skip windows issue * refactor dataset test * add dataset.py * fix dtype in layers.mfcc * fix ci-static-check * fix dtype in layers.mfcc && fix ci-static-check * add relative accuracy * modity API.spec * skip cuda11.2 test * skip cuda11.2 test2 * skip cuda11.2 * change dataset name * fix format * update api.spec * update api.spec2 * fix coverage * add dataset test * rm download load dict * rm download load dict in init * update api.spec3 * fix dataset coverage * fix coverage * fix coverage2 * restore api.spec * restore api.spec2 * fix api-spec 3 * fix api-spec 4 * fix api.spec * fix api.spec6 * refactor init_backend * fix typo * change paddleaudio backend set * fix get_current_audio_backend() * fix format * fix format2 * remove format in parameters * fix format2 * add warning massage in wave_backend && remove redundant audio util * rm audio util in print_signatures * fix format3 * add tess dataset license * format warning * add more info in warning msg * add paddleaudio version check * replace dataset esc50 with tess * add tess dataset && rm numpy transform in dataset.py * fix set audio backend bug * fix equal error * fix format && coverage error * add api example * fix format * fix error * fix typo * add noqa in __init__ * fix backend doc example error * rm seed in dataset * update bakcend example * fix typo * fix typo * fix example err * fix typo * fix ci dataset test * fix example fil * try to fix ci * clean dataset doc * change get_current_audio_backend to get_current_backend * creplace paddle.audio.backends.info with paddle.audio.info, same with load, save * fix ci error * repalce api in test_audio_backend * fix save&&set_backend exmaple
Configuration menu - View commit details
-
Copy full SHA for 29c9f02 - Browse repository at this point
Copy the full SHA 29c9f02View commit details
Commits on Oct 21, 2022
-
Add infer prune function (#47047)
* Add infer prune function * add fusion op
Configuration menu - View commit details
-
Copy full SHA for 8739497 - Browse repository at this point
Copy the full SHA 8739497View commit details -
add pdsa-2022-001, test=document_fix (#47228)
Add PDSA-2022-001 security advisory
Configuration menu - View commit details
-
Copy full SHA for 001c8a6 - Browse repository at this point
Copy the full SHA 001c8a6View commit details -
[CustomDevice] turn on WITH_CUSTOM_DEVICE when WITH_PYTHON=ON (#47165)
cherry pick #47108 原 WITH_CUSTOM_DEVICE 默认打开/关闭的策略随 ON_INFER开关,由于训练和预测共同发包,现在训练包编译时会打开 ON_INFER,导致 WITH_CUSTOM_DEVICE 默认关闭,custom device 功能不可用 WITH_CUSTOM_DEVICE 默认打开/关闭的策略更改为随 WITH_PYTHON 开关
Configuration menu - View commit details
-
Copy full SHA for d1fedc5 - Browse repository at this point
Copy the full SHA d1fedc5View commit details -
Configuration menu - View commit details
-
Copy full SHA for 281891c - Browse repository at this point
Copy the full SHA 281891cView commit details -
[Paddle-TRT] Fix conv2d (#47034)
* forbid Conv2D into Paddle-TensoRT when weight is not persistable.
Configuration menu - View commit details
-
Copy full SHA for d42a1dc - Browse repository at this point
Copy the full SHA d42a1dcView commit details
Commits on Oct 24, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 82f1e1b - Browse repository at this point
Copy the full SHA 82f1e1bView commit details -
Support BF16 training for sharding (#46846) (#47246)
* Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result is wrong. * support pure bfloat16 * support bf16 linear * update PR to pass CI * tiny fix where_grad_kernel.cu * Support bfloat16 type for reducer and sharding. * Fix some bug. * Polish code. * Polise code. * Add bfloat16 datatype in fill_grad kernels. Co-authored-by: sneaxiy <sneaxiy@126.com> Co-authored-by: sneaxiy <sneaxiy@126.com>
Configuration menu - View commit details
-
Copy full SHA for 5c85f1a - Browse repository at this point
Copy the full SHA 5c85f1aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 9780eb7 - Browse repository at this point
Copy the full SHA 9780eb7View commit details -
Configuration menu - View commit details
-
Copy full SHA for caf2751 - Browse repository at this point
Copy the full SHA caf2751View commit details -
Fix hAPI bug of not compatible with LayerHook (#47001) (#47283)
* Fix hAPI bug of not compatible with LayerHook
Configuration menu - View commit details
-
Copy full SHA for e8d6339 - Browse repository at this point
Copy the full SHA e8d6339View commit details -
[CodeStyle] add black config to release2.4 (#47146)
* [CodeStyle] add black config to release2.4 * empty commit, test=document_fix
Configuration menu - View commit details
-
Copy full SHA for 6454133 - Browse repository at this point
Copy the full SHA 6454133View commit details
Commits on Oct 25, 2022
-
[cherry-pick] add prior_box and box_coder for paddle.vision.ops (#46786)
* add prior_box and box_coder for paddle.vision.ops * fix UT change assertTrue to assert_allclose * fix formula format
Configuration menu - View commit details
-
Copy full SHA for d5c6386 - Browse repository at this point
Copy the full SHA d5c6386View commit details -
Configuration menu - View commit details
-
Copy full SHA for 99d8ba4 - Browse repository at this point
Copy the full SHA 99d8ba4View commit details -
[Sparse] Fix indices (#47190) (#47226)
当前无法从Tensor中获取到SparseTensor的sparse_dim,无法准确推断出indices的shape,所以目前先以3D点云模型为主,输入的SparseTensor的维度是5D的,其中非零元素是一维向量,所以indices是[4, -1]。
zhangkaihuo authoredOct 25, 2022 Configuration menu - View commit details
-
Copy full SHA for 942ab42 - Browse repository at this point
Copy the full SHA 942ab42View commit details
Commits on Oct 26, 2022
-
[Cherry-pick][Release/2.4]Refine the memory usage of fused_attention …
…and fused_feedforward ops (#47235) * fix fused_attention fused_feedforward * fix ci * fix ci * fix ci PADDLE_GET_CONST * fix ci ut
Configuration menu - View commit details
-
Copy full SHA for 9a6dd8f - Browse repository at this point
Copy the full SHA 9a6dd8fView commit details -
Added workaround for elementwise oneDNN kernel (#47080) (#47342)
* return proper state * fix for dims * fix Co-authored-by: jakpiase <jakpia21@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for 7c6550a - Browse repository at this point
Copy the full SHA 7c6550aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 12e6dfc - Browse repository at this point
Copy the full SHA 12e6dfcView commit details -
[Cherry-pick][audio] fix tess split fold (#47350)
* fix tess split fold * format
Configuration menu - View commit details
-
Copy full SHA for 85094bc - Browse repository at this point
Copy the full SHA 85094bcView commit details -
Configuration menu - View commit details
-
Copy full SHA for a16ef9f - Browse repository at this point
Copy the full SHA a16ef9fView commit details -
Fix inference performance problem caused by selecting cudnn kernel of…
Configuration menu - View commit details
-
Copy full SHA for 0369cd0 - Browse repository at this point
Copy the full SHA 0369cd0View commit details
Commits on Oct 27, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 99cec1a - Browse repository at this point
Copy the full SHA 99cec1aView commit details -
[cherry-pick] add batch_norm_kernel (#47394)
* cherry-pick #46359 and resolve conflict
zhangkaihuo authoredOct 27, 2022 Configuration menu - View commit details
-
Copy full SHA for b143e00 - Browse repository at this point
Copy the full SHA b143e00View commit details -
[Cherry-pick Release/2.4] Fix multi_tensor adam and momentum bug when…
… the parameter is list of dict (#47372) * reformat file by black * fix multi_tensor adam/momentum bug
Configuration menu - View commit details
-
Copy full SHA for 94240e2 - Browse repository at this point
Copy the full SHA 94240e2View commit details
Commits on Oct 28, 2022
-
Configuration menu - View commit details
-
Copy full SHA for eec93bd - Browse repository at this point
Copy the full SHA eec93bdView commit details -
[cherry-pick]add sync_batch_norm_bn and deliver indices_dict (#47407)
add sync_batch_norm_bn and deliver indices_dict
zhangkaihuo authoredOct 28, 2022 Configuration menu - View commit details
-
Copy full SHA for 0fa8309 - Browse repository at this point
Copy the full SHA 0fa8309View commit details -
[Cherry-pick][JIT] Add Predictor for JITLayer (#47379) (#47419)
* [JIT] Add Predictor for JITLayer (#47379) * add predictor_engine * add predictor_engine * fix zero shape * fix lodTensor * fix unittest * fix code style * update CmakeList * fix new executor
Configuration menu - View commit details
-
Copy full SHA for c42929c - Browse repository at this point
Copy the full SHA c42929cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7618cbd - Browse repository at this point
Copy the full SHA 7618cbdView commit details
Commits on Oct 29, 2022
-
[JITLayer]Enable OneDNN on CPU and Fix zero shape (#47428) (#47436)
* [JITLayer]Enable OneDNN on CPU and Fix zero shape
Configuration menu - View commit details
-
Copy full SHA for f478844 - Browse repository at this point
Copy the full SHA f478844View commit details -
Configuration menu - View commit details
-
Copy full SHA for 26465cd - Browse repository at this point
Copy the full SHA 26465cdView commit details -
[Cherry-pick][Release/2.4]Add fused_allreduce_gradients_with_group fo…
…r PPFleetX (#47458) * reformat hybrid_parallel_util.py by black * add fused_allreduce_gradients_with_group * add scale * fix ci
Configuration menu - View commit details
-
Copy full SHA for df64e79 - Browse repository at this point
Copy the full SHA df64e79View commit details
Commits on Oct 31, 2022
-
[cherry-pick] update dygraph PTQ export_model api (#47415)
* update dygraph PTQ export_model api * remove postprocess
Configuration menu - View commit details
-
Copy full SHA for 12b9b03 - Browse repository at this point
Copy the full SHA 12b9b03View commit details -
[Cherry-pick][audio] rm kaiser window in audio get_window function &&…
… rm audio utils(#47469) (#47479) * [audio] rm kaiser window in audio get_window function && rm audio utils (#47469) * rm kaiser window in audio window function * rm paddle audio utils which is redundant * rm kaiser in test_audio_functions.py Conflicts: python/paddle/audio/utils/error.py python/paddle/tests/test_audio_functions.py * format
Configuration menu - View commit details
-
Copy full SHA for f93e9a5 - Browse repository at this point
Copy the full SHA f93e9a5View commit details -
* update codestyle * [AutoParallel] fix fp16 for subblock (#47189) * [AutoParallel] fix fp16 for subblock * fix engine * fix comment * [AutoParallel] fix engine _build and cost method (#47263) * fix engine build method * fix import * update engine cost * update raise error * update cmakelist * revert optimizer * revert optimizer * fix unittest * fix unittest Co-authored-by: caozhou <caozhou@radi.ac.cn> Co-authored-by: caozhou <caozhou@radi.ac.cn>
Configuration menu - View commit details
-
Copy full SHA for 4b3589f - Browse repository at this point
Copy the full SHA 4b3589fView commit details
Commits on Nov 1, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 0201ccc - Browse repository at this point
Copy the full SHA 0201cccView commit details -
Configuration menu - View commit details
-
Copy full SHA for 5ffd4af - Browse repository at this point
Copy the full SHA 5ffd4afView commit details -
Configuration menu - View commit details
-
Copy full SHA for 23c05f2 - Browse repository at this point
Copy the full SHA 23c05f2View commit details -
[cherry-pick][code-gen] Support code-gen for opmaker of sparse op (#4…
…6993) (#47417) * support generating code of opmaker for backward op invoke forward op (#46912) * [code-gen] Support code-gen for opmaker of sparse op (#46993) * support generating code of opmaker for backward op invoke forward op * gsupport code-gen of opmaker for sparse op * refind logic of choose phi kernrel * fix complie budg * fix code_gen bug * fix bug * fix kernel signature code-gen * fix complie bug of VarType * fix complie bug of VarType * fix test_sparse_conv_op * fix test_sparse_norm_op * [Phi] Refactor logic of judging whether having a phi kernrel (#46920) * refind logic of choose phi kernrel * fix complie budg * update cmake
Configuration menu - View commit details
-
Copy full SHA for 601626a - Browse repository at this point
Copy the full SHA 601626aView commit details -
[cherry-pick]Fix english documents of sparse api (#47496)
Fix english documents of sparse api
zhangkaihuo authoredNov 1, 2022 Configuration menu - View commit details
-
Copy full SHA for 61953b9 - Browse repository at this point
Copy the full SHA 61953b9View commit details
Commits on Nov 2, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 7a1cf27 - Browse repository at this point
Copy the full SHA 7a1cf27View commit details
Commits on Nov 3, 2022
-
[cherry pick] fix memory copy in prepare_data of FusedMultiTransforme…
…r pass (#47308) * fix memory copy in prepare_data. test=develop * add cache_kv fp16 support. test=develop * fit for simplify_with_basic_ops_pass. test=develop
Configuration menu - View commit details
-
Copy full SHA for ba4fbe7 - Browse repository at this point
Copy the full SHA ba4fbe7View commit details -
support unbalanced data for pipeline (#47199) (#47569)
* add unbalanced data * fix utest
Configuration menu - View commit details
-
Copy full SHA for d4bf8b1 - Browse repository at this point
Copy the full SHA d4bf8b1View commit details -
[Sparse] Unified api args name (#47529) (#47627)
Unified api args name
zhangkaihuo authoredNov 3, 2022 Configuration menu - View commit details
-
Copy full SHA for 75088bb - Browse repository at this point
Copy the full SHA 75088bbView commit details -
Fix ComputePropagateScalesMkldnnPass of MKLDNN (#47574) (#47639)
* add constant_folding_pass pass for mkldnn int8 * update UpdateScaleOpInOutScales
Configuration menu - View commit details
-
Copy full SHA for 559b975 - Browse repository at this point
Copy the full SHA 559b975View commit details -
Configuration menu - View commit details
-
Copy full SHA for 99c872f - Browse repository at this point
Copy the full SHA 99c872fView commit details
Commits on Nov 4, 2022
-
[cherry-pick2.4]for CodeStyle (#47608)
* only run pre-commit * only run pre-commit
Configuration menu - View commit details
-
Copy full SHA for cfee9c1 - Browse repository at this point
Copy the full SHA cfee9c1View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1ccd3ea - Browse repository at this point
Copy the full SHA 1ccd3eaView commit details