Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Cherry-pick]Move sum op to PHI && Fix MetaTensor's bug when run infermeta #49341

Closed
wants to merge 243 commits into from
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
243 commits
Select commit Hold shift + click to select a range
29c44eb
[cherry-pick] Allow manaully set py_reader name in standalone executo…
From00 Sep 13, 2022
0903020
cherry pick softmax infer kernel (#45957)
JZZ-NOTE Sep 13, 2022
e223cf7
delete new executor log (#45917)
pangyoki Sep 14, 2022
9d5003d
Fix compile (#45996) (#46027)
b3602sss Sep 14, 2022
20d168d
fix transformer bug, test=kunlun (#45983)
ykkk2333 Sep 14, 2022
925e84b
[chery-pick] Fix namespace error (#45925) (#46029)
engineer1109 Sep 14, 2022
2fac8ab
set device id before op run (#45994)
zhiqiu Sep 14, 2022
2ca6590
cherry pick delay tensorrt log (#45958)
JZZ-NOTE Sep 14, 2022
5130b0a
merge python lib (#46013)
JZZ-NOTE Sep 14, 2022
61a3e30
fix trt multiclass_nms3 (#45166) (#46034)
zhangjun Sep 15, 2022
91677eb
Fix arm fp16 compile error (#45991) (#46048)
chenwhql Sep 15, 2022
526e032
Modify callstacklevel flag for c++ (#46059)
JZZ-NOTE Sep 15, 2022
9012e8b
fix distributed bug caused by fill_any_like (#45978) (#46041)
Charles-hit Sep 15, 2022
2680a71
Delete eigen header in data_type.h (#46036) (#46066)
zyfncg Sep 15, 2022
0793311
General Plugin Mechanism (#45355) (#46070)
weishengying Sep 15, 2022
903c87b
Support 0 shapes input Tensor for MKL slice (#45930) (#46072)
0x45f Sep 15, 2022
0048695
[ Dy2Static ] Fix bugs when select inputs meeting different shape or …
2742195759 Sep 15, 2022
8caaf85
[Cherry-pick] Normalize yaml name and label (#46052)
chenwhql Sep 16, 2022
e25e947
[cherry-pick][jit] Jit skip forward (#45926)
zh794390558 Sep 16, 2022
4e09e40
(cherry-pick)Fix split infershape in static mode and add convert rule…
Charles-hit Sep 16, 2022
2c2cd79
modify windows build script (#46099)
JZZ-NOTE Sep 16, 2022
a76fa41
V2.4 - cherry-pick (#46126)
ziyoujiyi Sep 17, 2022
f6dd201
fix compilation errors on mac arm64 (#46135)
yuanlehome Sep 17, 2022
4261ae3
Unify core avx and core_noavx to libpaddle (#46095) (#46113)
chenwhql Sep 19, 2022
a5d4f57
[Cherry-pick] fix bug for TransformedDistribution (#46157)
MayYouBeProsperous Sep 19, 2022
7a6db0a
rename fleetx, develop=document_fix (#46141)
ForFishes Sep 19, 2022
dabb8f2
Revert "Simplify size op impl (#45808)" (#46168)
chenwhql Sep 19, 2022
f4ec156
convfusion_cache (#46054)
xiaoxiaohehe001 Sep 19, 2022
e468e93
[Eager] Optimize log (#45783) (#46133)
JiabinYang Sep 19, 2022
860f607
fix broadcast kernel (#46158)
sneaxiy Sep 19, 2022
c5cc427
[Cherry-pick][Auto Parallel] Improve the APIs (#46164)
aoyulong Sep 19, 2022
e5dc9d6
refactor mp. (#45803) (#46121)
wuhuachaocoding Sep 19, 2022
db368d5
Add INT8 support for fused_multi_transformer_op (#45284) (#46169)
minghaoBD Sep 19, 2022
6b59a07
fix_recover_remove_padding kernel (#46050) (#46198)
Wangzheee Sep 19, 2022
855fdde
fix (#46125)
zhaocaibei123 Sep 19, 2022
1c7e95c
[Cherry-pick] Support bmm and bmm_grad in xpu (#45887) (#46132)
JiabinYang Sep 19, 2022
7f0c1f0
Remove redundant code in pe engine (#46110) (#46145)
0x45f Sep 19, 2022
adab3c5
(cherry-pick)support some op backward refuse forward (#46201)
Charles-hit Sep 19, 2022
707d838
cherry-pick 46152 (#46183)
jiweibo Sep 19, 2022
a056601
Add symbolic shape deduction function for general Plugin mechanism (#…
weishengying Sep 19, 2022
d1ce974
[cherry-pick] [dy2static] support user to use decorator in their prog…
feifei-111 Sep 19, 2022
45a3c65
[JitLayer]Save property meta file to correct path (#46131) (#46195)
0x45f Sep 19, 2022
ad8beaa
[cherry-pick] add abs,mean,sum,ge,gt,pow,etc higher-order differentia…
cxxly Sep 19, 2022
be84cac
[vision.ops.nms] Fix return order error and duplicate results with sp…
RichardWooSJTU Sep 19, 2022
4bced24
Recompute unify incubate (#46073) (#46210)
wuhuachaocoding Sep 19, 2022
ac8cce2
Fix wrong eigen header include (#46082) (#46202)
zyfncg Sep 20, 2022
7712ce1
[Cherry-pick] Fix static check (#46253)
sljlp Sep 20, 2022
c0324e8
[Cherry-pick] Update layoutautotune for inplace (#45826) (#46226)
AnnaTrainingG Sep 20, 2022
4f28a4c
Run_program_op add scope cache & reuse (#45813) (#46223)
zhangbo9674 Sep 20, 2022
c384b00
[Inference] fix preln_residual_bias_fuse_pass bug in TNT_small model …
zhoutianzi666 Sep 20, 2022
bc92d5f
(cherry-pick)Support some op refuse forward and fix some bugs (#46211)
Charles-hit Sep 20, 2022
38c0fd0
[Eager] Fix linspace error in amp (#46088) (#46206)
JiabinYang Sep 20, 2022
fd8ec4a
[Cherry-pick] Sparse add InferMeta (#46235)
Sep 20, 2022
5034030
Fix TransDataBackend Error when call unsqueeze using MKL Tensor (#460…
0x45f Sep 20, 2022
1418a71
[cherry-pick] Refine thread pool config of interpretercore (#46219)
zhiqiu Sep 20, 2022
654807c
[Paddle-TRT] Support matmul_v2 in Paddle-TensorRT (#46177)
zhoutianzi666 Sep 20, 2022
759736d
[Release/2.4][Cherry-pick] Fix bug of reduce_sum op (#46160)
GhostScreaming Sep 20, 2022
da173c4
[Cherry-pick] Fix amp error cp (#46272)
JiabinYang Sep 20, 2022
c43ebfc
[Cherry-Pick][AutoParallel] change import way and fix strategy (#46270)
zhaoyinglia Sep 20, 2022
42e56f6
[PolishComments] Polish some code comments (#46032) (#46261)
jiahy0825 Sep 20, 2022
adb2f5e
[Paddle-TRT] Full support for ops with persistable input (#45545) (#4…
zhoutianzi666 Sep 20, 2022
a43f960
[cherry-pick][xpu] update xdnn activations (#46282)
houj04 Sep 20, 2022
230b9a8
[Paddle-TRT][Cherry-Pick]Fix cast bug (#46293)
zhoutianzi666 Sep 20, 2022
bab1109
Revert "Optimiza params sync between CPU and GPU. (#45805)" (#46274)
jiweibo Sep 20, 2022
461099c
fix miss return error. (#46298)
jiweibo Sep 20, 2022
3e8b322
cherry-pick V2.4 (#46294)
ziyoujiyi Sep 20, 2022
b027652
remove tmp fp32 var for gaussian_random (#46285)
GuoxiaWang Sep 21, 2022
311eb87
update for py3.6 bug (#46325)
wuhuachaocoding Sep 21, 2022
f3d5c85
fix CI sccache install dir (#46267) (#46327)
sljlp Sep 21, 2022
b74c092
[JitLayer]Erase out vars in scope to avoid data rewritinig (#46249) (…
0x45f Sep 21, 2022
cc3e7cd
[Cherry-pick][BugFix]Fix pooling output_size bug if encounter list[Te…
Aurelius84 Sep 21, 2022
083853c
[Auto Parallel] fix lazyinit (#46355) (#46382)
zhaoyinglia Sep 22, 2022
372505b
[Dygraph] Fix bugs of mp in eager mode (#46303) (#46396)
haohongxiang Sep 22, 2022
7eb046c
logger manager (#45909) (#46087)
sljlp Sep 22, 2022
484377c
[Cherry-Pick][BugFix]Fix reduce_mean/min/sum/prod, cumsum grad_op inf…
Aurelius84 Sep 23, 2022
6a50833
fix compile problem (#46354), test=kunlun (#46383)
zyfncg Sep 23, 2022
980292c
fix bug in convert call: tranform the static func. (#46278) (#46365)
2742195759 Sep 23, 2022
cbf3f4b
[BugFix] fixed a bug that deco_name can't be parsed corrected (#46297…
feifei-111 Sep 23, 2022
55f73ba
[OpAttr]Fix dropout2d/3d static API (#46434)
Aurelius84 Sep 23, 2022
0ab2672
Remove audio ParameterError (#46316) (#46468)
SmileGoat Sep 24, 2022
6990edf
[cherrypick] Fix elementwise_sub sign reverse for mkldnn (#46107)
zh794390558 Sep 26, 2022
4a8aa6d
fix conflict (#46388)
feifei-111 Sep 26, 2022
536d9d8
cherry-pick V2.4 (#46358)
ziyoujiyi Sep 26, 2022
8089a1f
change use_calc_stream to sync_op (#46182) (#46493)
LiYuRio Sep 27, 2022
dc3a3f1
tirl bool for jit (#46513)
zh794390558 Sep 27, 2022
0cc2251
[cherry-pick] clear extra attrs of some ops in OpMaker (#45845, #4598…
zyfncg Sep 27, 2022
5711bbe
fix shard_index kernel (#46491) (#46511)
zhaoyinglia Sep 27, 2022
5dab0b0
[AutoParallel] fix amp o1 (#46391) (#46481)
zhaoyinglia Sep 27, 2022
a77a6f6
remove trt_reshape2_matmul_fuse_pass (#46363)
zhoutianzi666 Sep 28, 2022
b2e4211
[cherry-pick] Clear extra attrs of some ops in OpMaker (#46150, #4632…
zyfncg Sep 28, 2022
1c22ed7
Fix libpaddle soname mismatch error (#46344) (#46576)
chenwhql Sep 28, 2022
3f35e63
refine dy2st glog (#46415) (#46438)
zhangbo9674 Sep 28, 2022
f5956be
[CherryPick][Fix] Remove std::trunc() in FloorDivideFunctor and Inver…
Bobholamovic Sep 29, 2022
d90db9b
Fix the half precision problem of general plugin (#46580)
weishengying Sep 29, 2022
d67da3d
[cherry-pick] Open the clip_extra flag in save_inference_model (#46577)
zyfncg Sep 29, 2022
a58663f
[cherry-pick] Add FP16 support for uniform in dygraph mode on Nvidia …
Courtesy-Xs Sep 29, 2022
4282af6
[Dy2Static] refactor the return transformer (#45900) (#46205)
2742195759 Oct 9, 2022
3cc3f60
[PHI] Shape op migration (#46051) (#46724)
Silv3S Oct 10, 2022
44ecae6
[PHI] migrate softmax_grad kernel (#46257) (#46725)
Silv3S Oct 10, 2022
51a91fe
[PHI] Migrate slice, slice_grad, split, pad and pad3d oneDNN kernels …
Silv3S Oct 10, 2022
25d61cd
[cherry-pick] [PHI] Migrate sgd and stack oneDNN kernels (#46374) (#4…
Silv3S Oct 10, 2022
fdd0d6d
[cherry-pick] [PHI] Migrate concat+grad, expand+grad, fill_constant ……
Silv3S Oct 10, 2022
d8daf64
[Dy2St]Fix Regex DeprecationWarning in PY3 (#46829)
Aurelius84 Oct 10, 2022
a0e0341
Fix gather op convert for Paddle-TensorRT (#46779) (#46825)
Oct 10, 2022
d091d1b
optimize Paddle-TRT performance (#46684)
yuanlehome Oct 11, 2022
6a6c749
[BugFix]Fix concat bugs when call onednn kernel (#46518) (#46845)
YuanRisheng Oct 11, 2022
9cc3f69
Cherry pick for dygraph pp (#46876)
FeixLiu Oct 11, 2022
f556549
speedup ChannelClipAndQuantDequantKernelQuantAxis1 kernel (#46471) (#…
ceci3 Oct 11, 2022
7b3837e
Revert pool+grad oneDNN kernel conversion (#45989) (#46860)
Silv3S Oct 11, 2022
2bcbf8b
[cherry-pick] [PHI] relu6_grad kernel (#46501) (#46862)
Silv3S Oct 11, 2022
2c6bd4a
hard_swish grad (#46857)
Silv3S Oct 11, 2022
2190da2
add seed check (#46858)
Silv3S Oct 11, 2022
b051455
set_value_op: add support for complex types (#46885)
Oct 11, 2022
08d233f
cherry pick pr46536 (#46901)
ronny1996 Oct 12, 2022
61273c0
[Cherry-pick]Update layout autotune for module with no modified (#465…
AnnaTrainingG Oct 12, 2022
0280c0b
[cherry-pick] [PHI] transpose2_grad op migration (#46139) (#46873)
Silv3S Oct 13, 2022
100a075
[Cherry-pick] Add fp16 dtype support for set_value op (#46906)
Courtesy-Xs Oct 13, 2022
d90aaa6
interpretercore thread not always spin (#46687) (#46952)
zhangbo9674 Oct 13, 2022
b8677c0
[Paddle-TRT] support new quant format from slim (#46022) (#46979)
zhoutianzi666 Oct 14, 2022
535d757
[cherry-pick 2.4][inference] fix reshape2 opteller (#46871)
zhangjun Oct 14, 2022
4b47265
[BUG]Fix expand_as_v2 bug while X and Y with different dtype (#46950)…
Aurelius84 Oct 14, 2022
2744432
[Dy2St]Remove usless cast operation to speed up FP16 training (#46851…
Aurelius84 Oct 14, 2022
8f1ac7c
Add bmm convert (#47011)
xiaoxiaohehe001 Oct 14, 2022
84333cf
update quantization new format (#46529)
yghstill Oct 14, 2022
82db499
cherry-pick 46942 (#47015)
jiweibo Oct 14, 2022
5c2bea1
Fix nvcc lazy (#46879)
xiaoxiaohehe001 Oct 14, 2022
8bfd45a
[Cherry-Pick]Move valid check from python to kernel (#46980)
ZzSean Oct 17, 2022
a1cdbad
rm fp16 dtype_check (#46739) (#46866)
gglin001 Oct 17, 2022
bd89be1
[IPU] paddle-inference support custom-ops (#45235) (#46868)
gglin001 Oct 17, 2022
d913bc9
fix ut timeout 2 (#45233) (#46867)
gglin001 Oct 17, 2022
8c6c79a
update to sdk3.0 (#46865) (#46892)
gglin001 Oct 17, 2022
7eef05c
[cherry-pick] Fix the bug of exporting model in dygraph QAT (#47028)
yghstill Oct 17, 2022
976af0d
Optimize performance of depthwise_conv (#46896)
ZzSean Oct 17, 2022
10225d2
[cherry-pick]Sparse static graph (#46838)
Oct 17, 2022
5fba2a9
[Cherry-pick] Collective communication APIs (#46922)
HermitSun Oct 17, 2022
5f6b9f1
[Cherry pick] trt pool2d adaptive ifx (#47069)
wwbitejotunn Oct 18, 2022
b84edd9
[cherry-pick] Fix perf issues of mp/pp/fuse in eager mode (#47071)
haohongxiang Oct 18, 2022
5b64214
Cherry pick for sharding (#47061)
FeixLiu Oct 18, 2022
5a44c12
support shape tensor is the input of trt-subgraph (#47066)
zhoutianzi666 Oct 18, 2022
5fef043
[cherry-pick 2.4] add sparse api transpose/reshape/is_same_shape (#47…
zhwesky2010 Oct 18, 2022
2cc8797
Add symbolic shape deduction function for unfold, scatter_nd_add, p_n…
weishengying Oct 18, 2022
de6f15b
reconstruct code for convert_fp16 (#46428) (#47087)
jiweibo Oct 18, 2022
fcb9c0b
[ cherrypick] Construct exec and ctx only once in cond op to speed up…
zh794390558 Oct 19, 2022
f08c104
update audio api examples (#46938) (#47126)
SmileGoat Oct 19, 2022
69515e9
[Dy2St]Fix recurrent op eager deletion pass error in dy2st (#47105) (…
0x45f Oct 19, 2022
1d015f1
Add enable_partial_send_recv switch in pipeline_configs (#46992) (#47…
GhostScreaming Oct 19, 2022
247ef47
[CherryPick] Support TypeHint for function decorated by @to_static (#…
2742195759 Oct 19, 2022
5a9befe
[Dy2Stat]Polish @to_static temporary file directory to speed up trans…
Aurelius84 Oct 19, 2022
66dccd7
Add unsigned int8 scale propagation (#46378) (#47156)
yeliang2258 Oct 19, 2022
23f2a4e
[cherry-pick] strided_slice grad add fp16 support (#47159)
Oct 19, 2022
90b3179
[Cherry-Pick][AutoParallel] auto_parallel cherry-pick to release2.4 (…
zhaoyinglia Oct 19, 2022
c894d91
Fix cannot import `paddle.distributed` in python 3.6 on release/2.4 (…
HermitSun Oct 20, 2022
da7d2f2
[Cherry-pick][Release/2.4] support pure bfloat16 for more ops
sneaxiy Oct 20, 2022
c74bf01
[Cherry-pick][Release/2.4] Fix some operators when the tensor.numel()…
sneaxiy Oct 20, 2022
4c92524
add _get_phi_kernel_name interface (#47033)
JZZ-NOTE Oct 20, 2022
9ed1454
[Cherry-pick] layernorm shift partation enhance (#47086)
wwbitejotunn Oct 20, 2022
09b1923
add get ops scripts (#47049)
JZZ-NOTE Oct 20, 2022
68c4ac3
[Paddle-TRT][Cherry-Pick]Rewrite strided_slice converter using shap…
zhoutianzi666 Oct 20, 2022
c2d344d
[cherry-pick] Fix quantize model deploy bug in MKLDNN (#47119)
yeliang2258 Oct 20, 2022
3d647b1
fix problem of persistable var saving in QAT (#47203)
yghstill Oct 20, 2022
6712e26
Add value check & error message for gather_tree (#47051) (#47221)
FrostML Oct 20, 2022
396427a
[cherry pick] Add FusedMultiTransformer fuse pass for GPT3 (#47150)
heavengate Oct 20, 2022
50d4fa5
[cherry-pick 2.4] remove incubate of all paddle sparse api (#47183)
zhwesky2010 Oct 20, 2022
c0ed872
[Cherry-pick] Simplify conv codes and fix cache and autotune bugs. (#…
Xreki Oct 20, 2022
29c9f02
Add paddle audio dataset & backend (#45939) (#47230)
SmileGoat Oct 20, 2022
8739497
Add infer prune function (#47047)
JZZ-NOTE Oct 21, 2022
001c8a6
add pdsa-2022-001, test=document_fix (#47228)
VigiZhang Oct 21, 2022
d1fedc5
[CustomDevice] turn on WITH_CUSTOM_DEVICE when WITH_PYTHON=ON (#47165)
ronny1996 Oct 21, 2022
281891c
support qat in sharding stage2 (#47169) (#47240)
haohongxiang Oct 21, 2022
d42a1dc
[Paddle-TRT] Fix conv2d (#47034)
zhoutianzi666 Oct 21, 2022
82f1e1b
fix send for old dygraph mode by passing use_calc_stream to the send …
sljlp Oct 24, 2022
5c85f1a
Support BF16 training for sharding (#46846) (#47246)
GhostScreaming Oct 24, 2022
9780eb7
Fix virtualpp with mp/recompute bugs (#47242) (#47249)
FeixLiu Oct 24, 2022
caf2751
fix import in python3.6 (#47275)
zhaoyinglia Oct 24, 2022
e8d6339
Fix hAPI bug of not compatible with LayerHook (#47001) (#47283)
parap1uie-s Oct 24, 2022
6454133
[CodeStyle] add black config to release2.4 (#47146)
SigureMo Oct 24, 2022
d5c6386
[cherry-pick] add prior_box and box_coder for paddle.vision.ops (#46786)
nemonameless Oct 25, 2022
99d8ba4
[geometric] fix english doc (#46485) (#47317)
DesmonDay Oct 25, 2022
942ab42
[Sparse] Fix indices (#47190) (#47226)
Oct 25, 2022
9a6dd8f
[Cherry-pick][Release/2.4]Refine the memory usage of fused_attention …
sneaxiy Oct 26, 2022
7c6550a
Added workaround for elementwise oneDNN kernel (#47080) (#47342)
yeliang2258 Oct 26, 2022
12e6dfc
[Cherry-Pick][Dy2Stat]Fix module loading OSError in multiprocess (#47…
Aurelius84 Oct 26, 2022
85094bc
[Cherry-pick][audio] fix tess split fold (#47350)
SmileGoat Oct 26, 2022
a16ef9f
fix a bug that print log twice (#47336) (#47343)
sljlp Oct 26, 2022
0369cd0
Fix inference performance problem caused by selecting cudnn kernel of…
zyfncg Oct 26, 2022
99cec1a
fix slice bug (#47349) (#47376)
wanghuancoder Oct 27, 2022
b143e00
[cherry-pick] add batch_norm_kernel (#47394)
Oct 27, 2022
94240e2
[Cherry-pick Release/2.4] Fix multi_tensor adam and momentum bug when…
sneaxiy Oct 27, 2022
eec93bd
support multiclass_nms in int8 (#47337)
zhoutianzi666 Oct 28, 2022
0fa8309
[cherry-pick]add sync_batch_norm_bn and deliver indices_dict (#47407)
Oct 28, 2022
c42929c
[Cherry-pick][JIT] Add Predictor for JITLayer (#47379) (#47419)
Aurelius84 Oct 28, 2022
7618cbd
[Dy2St]Fix abnormal growth of memory in train mode and no_grad for Dy…
0x45f Oct 28, 2022
f478844
[JITLayer]Enable OneDNN on CPU and Fix zero shape (#47428) (#47436)
Aurelius84 Oct 29, 2022
26465cd
fix paddle.audio.function.get_window security error (#47453)
SmileGoat Oct 29, 2022
df64e79
[Cherry-pick][Release/2.4]Add fused_allreduce_gradients_with_group fo…
sneaxiy Oct 29, 2022
12b9b03
[cherry-pick] update dygraph PTQ export_model api (#47415)
yghstill Oct 31, 2022
f93e9a5
[Cherry-pick][audio] rm kaiser window in audio get_window function &&…
SmileGoat Oct 31, 2022
4b3589f
2.4/fix engine build (#47462)
zhaoyinglia Oct 31, 2022
0201ccc
fix p2p comm memory release logic (#47497) (#47517)
FeixLiu Nov 1, 2022
5ffd4af
add missing scale parameter (#47522)
sneaxiy Nov 1, 2022
23c05f2
add pdsa-2022-002 (#47486) (#47526)
VigiZhang Nov 1, 2022
601626a
[cherry-pick][code-gen] Support code-gen for opmaker of sparse op (#4…
zyfncg Nov 1, 2022
61953b9
[cherry-pick]Fix english documents of sparse api (#47496)
Nov 1, 2022
7a1cf27
[geometric] Optimize graph sample speed (#47531) (#47548)
DesmonDay Nov 2, 2022
ba4fbe7
[cherry pick] fix memory copy in prepare_data of FusedMultiTransforme…
heavengate Nov 3, 2022
d4bf8b1
support unbalanced data for pipeline (#47199) (#47569)
ForFishes Nov 3, 2022
75088bb
[Sparse] Unified api args name (#47529) (#47627)
Nov 3, 2022
559b975
Fix ComputePropagateScalesMkldnnPass of MKLDNN (#47574) (#47639)
yeliang2258 Nov 3, 2022
99c872f
FC/matmul(v2) + scale fuse pass (#47420)
Silv3S Nov 3, 2022
cfee9c1
[cherry-pick2.4]for CodeStyle (#47608)
Ligoml Nov 4, 2022
72e1eb6
[CherryPick] Cherry pick #45916 #46031 #47299 (#47610)
2742195759 Nov 4, 2022
7145db6
Revert "SparseConv support duplicate coordinates (#44976)" (#45202) (…
Nov 7, 2022
58c47e8
[Cherry-pick][BugFix]Fix set_attr modify underly type (#47500) (#47566)
Aurelius84 Nov 7, 2022
764cea0
【Cherry-pick PR47666】add cudnn error if compiled cudnn version is inc…
pangyoki Nov 7, 2022
d580983
[cherry-pick2.4]fix numpy issue in codeblock examples (#47664)
Ligoml Nov 7, 2022
3a01478
update the split logic for uniform (#47670) (#47705)
FeixLiu Nov 7, 2022
cf668ab
[cherry-pick2.4]docs fix (#47669)
Ligoml Nov 7, 2022
d0e19af
[CHERRY-PICK] Added caching to oneDNN FC and op+unsqueeze2 and op+res…
jakpiase Nov 8, 2022
b3e4571
Support cuda 11 on jetson (#47742)
chalsliu Nov 8, 2022
34f67a8
add fuse_multi_transformer passes to fp16. test=develop (#47733)
heavengate Nov 8, 2022
ea5f44b
[cherry-pick] Squeeze2 and transpose2 fuse using oneDNN(#47712)
zh794390558 Nov 9, 2022
51248f8
[Cherry-pick] remove functions not belong to public-api from __all__ …
zoooo0820 Nov 9, 2022
76b883c
【Cherry-pick PR47743】change cudnn error to cuda error if compiled cud…
pangyoki Nov 10, 2022
ff642c6
[Cherry-pick] Fix python link error (#47811)
chenwhql Nov 10, 2022
2e9e65d
【cherry-pick】update Recompute doc (#47784)
wuhuachaocoding Nov 10, 2022
3a6cc57
Fuse multi transformer layer pass (#47541) (#47830)
RichardWooSJTU Nov 10, 2022
4465ba2
rename fw_bw func name of interleave pp (#47571) (#47862)
haohongxiang Nov 11, 2022
5033b6c
Fix slice bugs in MKLDNN when input dims are zeros (#46671) (#47887)
yeliang2258 Nov 11, 2022
3fa7a73
Fix mac link python (#48017)
wanghuancoder Nov 16, 2022
1490aaa
[cherry-pick2.4]en-docs warning&error fix (#48332)
ustiniankw Nov 24, 2022
b9b7f00
update (#48350)
jiweibo Nov 25, 2022
a2f61fe
Fix wrong eigen header include in data_type.h (#48157) (#48260)
zyfncg Nov 25, 2022
7a0b862
Cherrypick NV fixes to release/2.4 (#48263)
zlsh80826 Nov 28, 2022
9e2ba9b
[cherry-pick] updating mul and matmul with set_mem_desc and fix squee…
yeliang2258 Nov 29, 2022
ddcd1b6
[cherry-pick][Inference] support mixed precision inference (#49077)
yuanlehome Dec 19, 2022
cdab3a4
Fix nullptr to TestFuseGemmEpilogueReluBWDFP* (#48997) (#49090)
ForFishes Dec 20, 2022
fb19648
cherry-pick #75b734 (#49201)
Dec 21, 2022
7c36b88
fix unittests (#49203) (#49210)
Aganlengzi Dec 21, 2022
612bdb1
[Docs]update readme; test=document_fix (#49246)
Ligoml Dec 22, 2022
11c7f57
Fix mixed precision bug (#49239)
yuanlehome Dec 22, 2022
5d29a5b
fix unittest in post training quantization (#49257)
yghstill Dec 22, 2022
b187c95
cherry-pick 45860
YuanRisheng Sep 19, 2022
dfa1d69
[BUG FIX]Fix MetaTensor's bug when run infermeta (#46265)
YuanRisheng Sep 22, 2022
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
3 changes: 3 additions & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -53,6 +53,7 @@ model_test

Testing
tools/__pycache__
tools/nvcc_lazy

# This file is automatically generated.
# TODO(zhiqiang) Move this file to build directory.
Expand All @@ -70,7 +71,9 @@ paddle/fluid/pybind/eager_op_function.cc

# these files (directories) are generated before build system generation
paddle/fluid/operators/generated_op.cc
paddle/fluid/operators/generated_sparse_op.cc
paddle/phi/ops/compat/generated_sig.cc
paddle/phi/ops/compat/generated_sparse_sig.cc
paddle/phi/api/yaml/parsed_apis/
python/paddle/utils/code_gen/
paddle/fluid/pybind/tmp_eager_op_function_impl.h
Expand Down
8 changes: 4 additions & 4 deletions .pre-commit-config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,10 +4,10 @@ repos:
hooks:
- id: remove-crlf
files: (?!.*third_party)^.*$ | (?!.*book)^.*$
- repo: https://github.com/google/yapf
rev: v0.32.0
- repo: https://github.com/psf/black.git
rev: 22.8.0
hooks:
- id: yapf
- id: black
files: (.*\.(py|bzl)|BUILD|.*\.BUILD|WORKSPACE)$
exclude: |
(?x)^(
Expand All @@ -24,7 +24,7 @@ repos:
files: (?!.*third_party)^.*$ | (?!.*book)^.*$
- id: end-of-file-fixer
- id: sort-simple-yaml
files: (api|backward|api_[a-z_]+)\.yaml$
files: (op|backward|op_[a-z_]+)\.yaml$
- repo: local
hooks:
- id: clang-format
Expand Down
3 changes: 0 additions & 3 deletions .style.yapf

This file was deleted.

4 changes: 2 additions & 2 deletions CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -249,7 +249,7 @@ option(WITH_SYSTEM_BLAS "Use system blas library" OFF)
option(WITH_DISTRIBUTE "Compile with distributed support" OFF)
option(WITH_BRPC_RDMA "Use brpc rdma as the rpc protocal" OFF)
option(ON_INFER "Turn on inference optimization and inference-lib generation"
OFF)
ON)
################################ Internal Configurations #######################################
option(WITH_NV_JETSON "Compile PaddlePaddle with NV JETSON" OFF)
option(WITH_PROFILER "Compile PaddlePaddle with GPU profiler and gperftools"
Expand Down Expand Up @@ -336,7 +336,7 @@ endif()

if(LINUX
AND NOT WITH_CUSTOM_DEVICE
AND NOT ON_INFER)
AND WITH_PYTHON)
set(WITH_CUSTOM_DEVICE
ON
CACHE BOOL "Enable Custom Device when compiling for Linux" FORCE)
Expand Down
8 changes: 4 additions & 4 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -15,12 +15,12 @@ English | [简体中文](./README_cn.md)
Welcome to the PaddlePaddle GitHub.

PaddlePaddle, as the first independent R&D deep learning platform in China, has been officially open-sourced to professional communities since 2016. It is an industrial platform with advanced technologies and rich features that cover core deep learning frameworks, basic model libraries, end-to-end development kits, tools & components as well as service platforms.
PaddlePaddle is originated from industrial practices with dedication and commitments to industrialization. It has been widely adopted by a wide range of sectors including manufacturing, agriculture, enterprise service, and so on while serving more than 4.7 million developers, 180,000 companies and generating 560,000 models. With such advantages, PaddlePaddle has helped an increasing number of partners commercialize AI.
PaddlePaddle is originated from industrial practices with dedication and commitments to industrialization. It has been widely adopted by a wide range of sectors including manufacturing, agriculture, enterprise service, and so on while serving more than 5.35 million developers, 200,000 companies and generating 670,000 models. With such advantages, PaddlePaddle has helped an increasing number of partners commercialize AI.


## Installation

### Latest PaddlePaddle Release: [v2.3](https://github.com/PaddlePaddle/Paddle/tree/release/2.3)
### Latest PaddlePaddle Release: [v2.4](https://github.com/PaddlePaddle/Paddle/tree/release/2.4)

Our vision is to enable deep learning for everyone via PaddlePaddle.
Please refer to our [release announcement](https://github.com/PaddlePaddle/Paddle/releases) to track the latest features of PaddlePaddle.
Expand Down Expand Up @@ -89,8 +89,8 @@ We provide [English](https://www.paddlepaddle.org.cn/documentation/docs/en/guide

## Courses

- [Server Deployments](https://aistudio.baidu.com/aistudio/course/introduce/19084): Courses intorducing high performance server deployments via local and remote services.
- [Edge Deployments](https://aistudio.baidu.com/aistudio/course/introduce/22690): Courses intorducing edge deployments from mobile, IoT to web and applets.
- [Server Deployments](https://aistudio.baidu.com/aistudio/course/introduce/19084): Courses introducing high performance server deployments via local and remote services.
- [Edge Deployments](https://aistudio.baidu.com/aistudio/course/introduce/22690): Courses introducing edge deployments from mobile, IoT to web and applets.

## Copyright and License
PaddlePaddle is provided under the [Apache-2.0 license](LICENSE).
25 changes: 9 additions & 16 deletions README_cn.md
Original file line number Diff line number Diff line change
Expand Up @@ -15,11 +15,11 @@

欢迎来到 PaddlePaddle GitHub

飞桨(PaddlePaddle)以百度多年的深度学习技术研究和业务应用为基础,是中国首个自主研发、功能完备、 开源开放的产业级深度学习平台,集深度学习核心训练和推理框架、基础模型库、端到端开发套件和丰富的工具组件于一体。目前,飞桨累计开发者477万,服务企业18万家,基于飞桨开源深度学习平台产生了56万个模型。飞桨助力开发者快速实现AI想法,快速上线AI业务。帮助越来越多的行业完成AI赋能,实现产业智能化升级。
飞桨(PaddlePaddle)以百度多年的深度学习技术研究和业务应用为基础,是中国首个自主研发、功能完备、 开源开放的产业级深度学习平台,集深度学习核心训练和推理框架、基础模型库、端到端开发套件和丰富的工具组件于一体。目前,飞桨累计开发者535万,服务企业20万家,基于飞桨开源深度学习平台产生了67万个模型。飞桨助力开发者快速实现AI想法,快速上线AI业务。帮助越来越多的行业完成AI赋能,实现产业智能化升级。

## 安装

### PaddlePaddle最新版本: [v2.3](https://github.com/PaddlePaddle/Paddle/tree/release/2.3)
### PaddlePaddle最新版本: [v2.4](https://github.com/PaddlePaddle/Paddle/tree/release/2.4)

跟进PaddlePaddle最新特性请参考我们的[版本说明](https://github.com/PaddlePaddle/Paddle/releases)

Expand Down Expand Up @@ -63,32 +63,25 @@ PaddlePaddle用户可领取**免费Tesla V100在线算力资源**,训练模型
我们提供 [英文](https://www.paddlepaddle.org.cn/documentation/docs/en/guides/index_en.html) 和
[中文](https://www.paddlepaddle.org.cn/documentation/docs/zh/guides/index_cn.html) 文档

- [使用指南](https://www.paddlepaddle.org.cn/documentation/docs/zh/guides/index_cn.html)
- [使用指南](https://www.paddlepaddle.org.cn/documentation/docs/zh/guides/index_cn.html):或许您想从深度学习基础开始学习飞桨

或许您想从深度学习基础开始学习飞桨

- [应用实践](https://www.paddlepaddle.org.cn/documentation/docs/zh/tutorial/index_cn.html)
- [应用实践](https://www.paddlepaddle.org.cn/documentation/docs/zh/tutorial/index_cn.html):使用飞桨搭建您的模型,更高效的完成深度学习任务


- [API Reference](https://www.paddlepaddle.org.cn/documentation/docs/zh/api/index_cn.html)
- [API 文档](https://www.paddlepaddle.org.cn/documentation/docs/zh/api/index_cn.html):新的 API 支持代码更少更简洁的程序

新的API支持代码更少更简洁的程序


- [贡献方式](https://www.paddlepaddle.org.cn/documentation/docs/zh/guides/08_contribution/index_cn.html)

欢迎您的贡献!
- [贡献方式](https://www.paddlepaddle.org.cn/documentation/docs/zh/guides/08_contribution/index_cn.html):欢迎您的贡献!

## 交流与反馈

- 欢迎您通过[Github Issues](https://github.com/PaddlePaddle/Paddle/issues)来提交问题、报告与建议
- QQ群: 441226485 (PaddlePaddle)
- [论坛](https://aistudio.baidu.com/paddle/forum): 欢迎大家在PaddlePaddle论坛分享在使用PaddlePaddle中遇到的问题和经验, 营造良好的论坛氛围
- [论坛](https://aistudio.baidu.com/paddle/forum): 欢迎大家在PaddlePaddle论坛分享在使用PaddlePaddle中遇到的问题和经验营造良好的论坛氛围

## 课程

- [服务器部署](https://aistudio.baidu.com/aistudio/course/introduce/19084): 详细介绍高性能服务器端部署实操,包含本地端及服务化Serving部署等
- [端侧部署](https://aistudio.baidu.com/aistudio/course/introduce/22690): 详细介绍端侧多场景部署实操,从移端端设备、IoT、网页到小程序部署
- [端侧部署](https://aistudio.baidu.com/aistudio/course/introduce/22690): 详细介绍端侧多场景部署实操,从移动端设备、IoT、网页到小程序部署

## 版权和许可证
PaddlePaddle由[Apache-2.0 license](LICENSE)提供
4 changes: 2 additions & 2 deletions SECURITY.md
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ We will indicate the bug fix in the release of PaddlePaddle, and publish the vul

### What is a vulnerability?

In the process of computation graphs in PaddlePaddle, models can perform arbitrary computations , including reading and writing files, communicating with the network, etc. It may cause memory exhaustion, deadlock, etc., which will lead to unexpected behavior of PaddlePaddle. We consider these behavior to be security vulnerabilities only if they are out of the intention of the operation involved.
In the process of computation graphs in PaddlePaddle, models can perform arbitrary computations , including reading and writing files, communicating with the network, etc. It may cause memory exhaustion, deadlock, etc., which will lead to unexpected behavior of PaddlePaddle. We consider these behavior to be security vulnerabilities only if they are out of the intention of the operation involved.



Expand All @@ -60,4 +60,4 @@ If malicious input can trigger memory corruption or non-clean exit, such bug is



[security advisories](https://github.com/PaddlePaddle/Paddle/blob/develop/security/README.md)
[security advisories](./security/README.md)
2 changes: 1 addition & 1 deletion SECURITY_cn.md
Original file line number Diff line number Diff line change
Expand Up @@ -46,4 +46,4 @@

如果输入非预期的参数后,对飞桨代码造成了内存破坏,或者非干净退出,这类行为被认定为存在安全问题。

### [安全公告](https://github.com/PaddlePaddle/Paddle/blob/develop/security/README_cn.md)
### [安全公告](./security/README_cn.md)
1 change: 1 addition & 0 deletions cmake/cuda.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ if(WITH_NV_JETSON)
add_definitions(-DWITH_NV_JETSON)
set(paddle_known_gpu_archs "53 62 72")
set(paddle_known_gpu_archs10 "53 62 72")
set(paddle_known_gpu_archs11 "53 62 72")
elseif(NEW_RELEASE_ALL)
message("Using New Release Strategy - All Arches Packge")
add_definitions(-DNEW_RELEASE_ALL)
Expand Down
17 changes: 11 additions & 6 deletions cmake/experiments/cuda_module_loading_lazy.cmake
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
# Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
Expand All @@ -16,16 +16,15 @@
# cuda moduel lazy loading is supported by CUDA 11.7+
# this experiment option makes Paddle supports lazy loading before CUDA 11.7.

option(EXP_CUDA_MODULE_LOADING_LAZY "enable lazy cuda module loading" OFF)
if(${EXP_CUDA_MODULE_LOADING_LAZY})
if(NOT ${ON_INFER} OR NOT ${LINUX})
if(LINUX)
if(NOT ON_INFER)
message(
"EXP_CUDA_MODULE_LOADING_LAZY only works with ON_INFER=ON on Linux platforms"
)
return()
endif()
if(NOT ${CUDA_FOUND})
message("EXP_CUDA_MODULE_LOADING_LAZY only works with CUDA")
if(NOT WITH_GPU)
message("EXP_CUDA_MODULE_LOADING_LAZY only works with GPU")
return()
endif()
if(${CUDA_VERSION} VERSION_GREATER_EQUAL "11.7")
Expand All @@ -41,6 +40,12 @@ if(${EXP_CUDA_MODULE_LOADING_LAZY})
CACHE BOOL "" FORCE)
set(CMAKE_CUDA_FLAGS "--cudart shared")
enable_language(CUDA)
execute_process(
COMMAND "rm" "-rf" "${CMAKE_SOURCE_DIR}/tools/nvcc_lazy"
COMMAND "chmod" "755" "${CMAKE_SOURCE_DIR}/tools/nvcc_lazy.sh"
COMMAND "bash" "${CMAKE_SOURCE_DIR}/tools/nvcc_lazy.sh"
"${CMAKE_SOURCE_DIR}/tools/nvcc_lazy" "${CUDA_TOOLKIT_ROOT_DIR}")
execute_process(COMMAND "chmod" "755" "${CMAKE_SOURCE_DIR}/tools/nvcc_lazy")
set(CUDA_NVCC_EXECUTABLE
"${CMAKE_SOURCE_DIR}/tools/nvcc_lazy"
CACHE FILEPATH "" FORCE)
Expand Down
4 changes: 2 additions & 2 deletions cmake/external/gloo.cmake
100644 → 100755
Original file line number Diff line number Diff line change
Expand Up @@ -25,8 +25,8 @@ set(GLOO_LIBRARY_DIR
"${GLOO_INSTALL_DIR}/lib"
CACHE PATH "gloo library directory." FORCE)
# As we add extra features for gloo, we use the non-official repo
set(GLOO_REPOSITORY ${GIT_URL}/sandyhouse/gloo.git)
set(GLOO_TAG v0.0.2)
set(GLOO_REPOSITORY ${GIT_URL}/ziyoujiyi/gloo.git)
set(GLOO_TAG v0.0.3)
set(GLOO_LIBRARIES
"${GLOO_INSTALL_DIR}/lib/libgloo.a"
CACHE FILEPATH "gloo library." FORCE)
Expand Down
2 changes: 0 additions & 2 deletions cmake/external/protobuf.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -268,8 +268,6 @@ function(build_protobuf TARGET_NAME BUILD_FOR_HOST)
DOWNLOAD_DIR ${PROTOBUF_SOURCE_DIR}
DOWNLOAD_COMMAND rm -rf arm_protobuf.tar.gz && wget --no-check-certificate
${ARM_PROTOBUF_URL} && tar zxvf arm_protobuf.tar.gz
#DOWNLOAD_COMMAND cp /home/wangbin44/Paddle/build/arm_protobuf.tar.gz .
# && tar zxvf arm_protobuf.tar.gz
UPDATE_COMMAND ""
CMAKE_ARGS -DCMAKE_INSTALL_PREFIX:PATH=${PROTOBUF_INSTALL_DIR}
-DCMAKE_BUILD_TYPE:STRING=${THIRD_PARTY_BUILD_TYPE}
Expand Down
1 change: 1 addition & 0 deletions cmake/external/rocksdb.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -35,6 +35,7 @@ ExternalProject_Add(
CMAKE_ARGS -DCMAKE_CXX_COMPILER=${CMAKE_CXX_COMPILER}
-DCMAKE_C_COMPILER=${CMAKE_C_COMPILER}
-DWITH_BZ2=OFF
-DPORTABLE=1
-DWITH_GFLAGS=OFF
-DCMAKE_CXX_FLAGS=${ROCKSDB_CMAKE_CXX_FLAGS}
-DCMAKE_C_FLAGS=${CMAKE_C_FLAGS}
Expand Down
6 changes: 3 additions & 3 deletions cmake/external/xpu.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,8 @@ set(XPU_RT_LIB_NAME "libxpurt.so")

if(NOT DEFINED XPU_BASE_URL)
set(XPU_BASE_URL_WITHOUT_DATE
"https://baidu-kunlun-product.cdn.bcebos.com/KL-SDK/klsdk-dev")
set(XPU_BASE_URL "${XPU_BASE_URL_WITHOUT_DATE}/20220907")
"https://baidu-kunlun-product.su.bcebos.com/KL-SDK/klsdk-dev")
set(XPU_BASE_URL "${XPU_BASE_URL_WITHOUT_DATE}/20220919")
else()
set(XPU_BASE_URL "${XPU_BASE_URL}")
endif()
Expand All @@ -19,7 +19,7 @@ endif()
if(NOT DEFINED XPU_XDNN_BASE_URL)
set(XPU_XDNN_BASE_URL_WITHOUT_DATE
"https://klx-sdk-release-public.su.bcebos.com/xdnn/dev")
set(XPU_XDNN_BASE_URL "${XPU_XDNN_BASE_URL_WITHOUT_DATE}/20220907")
set(XPU_XDNN_BASE_URL "${XPU_XDNN_BASE_URL_WITHOUT_DATE}/20220919")
else()
set(XPU_XDNN_BASE_URL "${XPU_XDNN_BASE_URL}")
endif()
Expand Down
1 change: 0 additions & 1 deletion cmake/external/xxhash.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -97,5 +97,4 @@ endif()

add_library(xxhash STATIC IMPORTED GLOBAL)
set_property(TARGET xxhash PROPERTY IMPORTED_LOCATION ${XXHASH_LIBRARIES})
include_directories(${XXHASH_INCLUDE_DIR})
add_dependencies(xxhash extern_xxhash)
57 changes: 56 additions & 1 deletion cmake/operators.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -510,7 +510,7 @@ function(op_library TARGET)
if(WITH_MKLDNN AND ${mkldnn_cc_srcs_len} GREATER 0)
# Append first implemented MKLDNN activation operator
if(${MKLDNN_FILE} STREQUAL "activation_mkldnn_op")
file(APPEND ${pybind_file} "USE_OP_DEVICE_KERNEL(gelu, MKLDNN);\n")
file(APPEND ${pybind_file} "USE_OP_DEVICE_KERNEL(softplus, MKLDNN);\n")
elseif(${MKLDNN_FILE} STREQUAL "conv_mkldnn_op")
file(APPEND ${pybind_file}
"USE_OP_DEVICE_KERNEL_WITH_CUSTOM_TYPE(conv2d, MKLDNN, FP32);\n")
Expand Down Expand Up @@ -611,3 +611,58 @@ function(register_operators)
endif()
endif()
endfunction()

function(prune_pybind_h)
set(op_list ${OP_LIST})

list(APPEND op_list "load_combine")
list(APPEND op_list "tensorrt_engine")

# add fused_op in op_list
list(APPEND op_list "fc")
list(APPEND op_list "conv2d_fusion")
list(APPEND op_list "fusion_seqconv_eltadd_relu")
list(APPEND op_list "fusion_seqpool_cvm_concat")
list(APPEND op_list "fusion_gru")
list(APPEND op_list "fusion_seqexpand_concat_fc")
list(APPEND op_list "fusion_repeated_fc_relu")
list(APPEND op_list "fusion_squared_mat_sub")

# add plugin_op in op_list
list(APPEND op_list "anchor_generator")

file(STRINGS ${pybind_file} op_registry_list)

file(WRITE ${pybind_file_prune} "")
file(
APPEND ${pybind_file_prune}
"// Generated by the paddle/fluid/operators/CMakeLists.txt. DO NOT EDIT!\n"
)

# add USE_OP_ITSELF for all op in op_list
foreach(op_name IN LISTS op_list)
file(APPEND ${pybind_file_prune} "USE_OP_ITSELF(${op_name});\n")
endforeach()

foreach(op_registry IN LISTS op_registry_list)
if(NOT ${op_registry} EQUAL "")
foreach(op_name IN LISTS op_list)
string(FIND ${op_registry} "(${op_name})" index1)
string(FIND ${op_registry} "(${op_name}," index2)
string(FIND ${op_registry} "USE_OP_ITSELF" index3)
if(((NOT ${index1} EQUAL "-1") OR (NOT ${index2} EQUAL "-1"))
AND (${index3} EQUAL "-1"))
file(APPEND ${pybind_file_prune} "${op_registry}\n")
endif()
endforeach()
endif()
endforeach()

file(WRITE ${pybind_file} "")
file(STRINGS ${pybind_file_prune} op_registry_list_tmp)
foreach(op_name IN LISTS op_registry_list_tmp)
if(NOT ${op_name} EQUAL "")
file(APPEND ${pybind_file} "${op_name}\n")
endif()
endforeach()
endfunction()
Loading