-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Gpugraph.0621 #144
base: develop
Are you sure you want to change the base?
Gpugraph.0621 #144
Commits on Apr 27, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 3faa1d8 - Browse repository at this point
Copy the full SHA 3faa1d8View commit details
Commits on Apr 28, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 3a70c42 - Browse repository at this point
Copy the full SHA 3a70c42View commit details -
Configuration menu - View commit details
-
Copy full SHA for 09455a6 - Browse repository at this point
Copy the full SHA 09455a6View commit details -
Configuration menu - View commit details
-
Copy full SHA for acb8ac0 - Browse repository at this point
Copy the full SHA acb8ac0View commit details
Commits on Apr 29, 2022
-
Configuration menu - View commit details
-
Copy full SHA for ff5fa32 - Browse repository at this point
Copy the full SHA ff5fa32View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7950086 - Browse repository at this point
Copy the full SHA 7950086View commit details -
Configuration menu - View commit details
-
Copy full SHA for 24cb259 - Browse repository at this point
Copy the full SHA 24cb259View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7798771 - Browse repository at this point
Copy the full SHA 7798771View commit details -
Configuration menu - View commit details
-
Copy full SHA for 641fcac - Browse repository at this point
Copy the full SHA 641fcacView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7762561 - Browse repository at this point
Copy the full SHA 7762561View commit details
Commits on May 1, 2022
-
Merge pull request xuewujiao#26 from DesmonDay/gpu_graph_engine2
Add actual neighbor sample result
Configuration menu - View commit details
-
Copy full SHA for 7cfe661 - Browse repository at this point
Copy the full SHA 7cfe661View commit details -
Configuration menu - View commit details
-
Copy full SHA for 918bb56 - Browse repository at this point
Copy the full SHA 918bb56View commit details -
Merge branch 'gpu_graph_engine2' of https://github.com/seemingwang/Pa…
…ddle into gpu_graph_engine2
Configuration menu - View commit details
-
Copy full SHA for 2e44d09 - Browse repository at this point
Copy the full SHA 2e44d09View commit details -
Configuration menu - View commit details
-
Copy full SHA for b63e2e2 - Browse repository at this point
Copy the full SHA b63e2e2View commit details -
Configuration menu - View commit details
-
Copy full SHA for cfe8d4a - Browse repository at this point
Copy the full SHA cfe8d4aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 0d18f57 - Browse repository at this point
Copy the full SHA 0d18f57View commit details -
Configuration menu - View commit details
-
Copy full SHA for 0ad3e46 - Browse repository at this point
Copy the full SHA 0ad3e46View commit details
Commits on May 4, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 2606895 - Browse repository at this point
Copy the full SHA 2606895View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3cfac05 - Browse repository at this point
Copy the full SHA 3cfac05View commit details -
Configuration menu - View commit details
-
Copy full SHA for 38f7b15 - Browse repository at this point
Copy the full SHA 38f7b15View commit details
Commits on May 10, 2022
-
Configuration menu - View commit details
-
Copy full SHA for d1a74f2 - Browse repository at this point
Copy the full SHA d1a74f2View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2e2dd2a - Browse repository at this point
Copy the full SHA 2e2dd2aView commit details
Commits on May 11, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 3c33403 - Browse repository at this point
Copy the full SHA 3c33403View commit details -
Configuration menu - View commit details
-
Copy full SHA for 080e5c9 - Browse repository at this point
Copy the full SHA 080e5c9View commit details
Commits on May 12, 2022
-
Configuration menu - View commit details
-
Copy full SHA for b00a116 - Browse repository at this point
Copy the full SHA b00a116View commit details -
Configuration menu - View commit details
-
Copy full SHA for 97c3f0c - Browse repository at this point
Copy the full SHA 97c3f0cView commit details -
Configuration menu - View commit details
-
Copy full SHA for aaa137e - Browse repository at this point
Copy the full SHA aaa137eView commit details -
Configuration menu - View commit details
-
Copy full SHA for 08a301f - Browse repository at this point
Copy the full SHA 08a301fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 12168b0 - Browse repository at this point
Copy the full SHA 12168b0View commit details
Commits on May 14, 2022
-
Configuration menu - View commit details
-
Copy full SHA for ec89107 - Browse repository at this point
Copy the full SHA ec89107View commit details
Commits on May 15, 2022
-
Configuration menu - View commit details
-
Copy full SHA for c221167 - Browse repository at this point
Copy the full SHA c221167View commit details -
Configuration menu - View commit details
-
Copy full SHA for ab005da - Browse repository at this point
Copy the full SHA ab005daView commit details
Commits on May 16, 2022
-
Configuration menu - View commit details
-
Copy full SHA for e5c8a41 - Browse repository at this point
Copy the full SHA e5c8a41View commit details -
Configuration menu - View commit details
-
Copy full SHA for 715ed06 - Browse repository at this point
Copy the full SHA 715ed06View commit details
Commits on May 17, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 79a6cd5 - Browse repository at this point
Copy the full SHA 79a6cd5View commit details
Commits on May 18, 2022
-
Configuration menu - View commit details
-
Copy full SHA for c805cc0 - Browse repository at this point
Copy the full SHA c805cc0View commit details -
Configuration menu - View commit details
-
Copy full SHA for 32a4341 - Browse repository at this point
Copy the full SHA 32a4341View commit details
Commits on May 23, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 3f3185e - Browse repository at this point
Copy the full SHA 3f3185eView commit details -
Merge pull request #1 from Thunderbrook/gpugraph_0523
[GpuGraph] GraphInsGenerator
Configuration menu - View commit details
-
Copy full SHA for 676a92c - Browse repository at this point
Copy the full SHA 676a92cView commit details
Commits on May 24, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 8b0b194 - Browse repository at this point
Copy the full SHA 8b0b194View commit details -
Configuration menu - View commit details
-
Copy full SHA for 91f5c32 - Browse repository at this point
Copy the full SHA 91f5c32View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6672b82 - Browse repository at this point
Copy the full SHA 6672b82View commit details -
Configuration menu - View commit details
-
Copy full SHA for 112cf61 - Browse repository at this point
Copy the full SHA 112cf61View commit details -
Configuration menu - View commit details
-
Copy full SHA for 88f13ec - Browse repository at this point
Copy the full SHA 88f13ecView commit details -
root committed
May 24, 2022 Configuration menu - View commit details
-
Copy full SHA for ed569b4 - Browse repository at this point
Copy the full SHA ed569b4View commit details -
Merge pull request xuewujiao#5 from Thunderbrook/gpugraph_deepwalk
[GpuGraph] deepwalk
Configuration menu - View commit details
-
Copy full SHA for 63e501d - Browse repository at this point
Copy the full SHA 63e501dView commit details
Commits on May 25, 2022
-
Configuration menu - View commit details
-
Copy full SHA for dc8dbf2 - Browse repository at this point
Copy the full SHA dc8dbf2View commit details -
Configuration menu - View commit details
-
Copy full SHA for f43d085 - Browse repository at this point
Copy the full SHA f43d085View commit details -
Configuration menu - View commit details
-
Copy full SHA for a9b5445 - Browse repository at this point
Copy the full SHA a9b5445View commit details -
Merge pull request xuewujiao#3 from seemingwang/gpu_graph_engine2
split graph table
Configuration menu - View commit details
-
Copy full SHA for fc28b23 - Browse repository at this point
Copy the full SHA fc28b23View commit details
Commits on May 26, 2022
-
root committed
May 26, 2022 Configuration menu - View commit details
-
Copy full SHA for f24db08 - Browse repository at this point
Copy the full SHA f24db08View commit details -
root committed
May 26, 2022 Configuration menu - View commit details
-
Copy full SHA for 8daf1a2 - Browse repository at this point
Copy the full SHA 8daf1a2View commit details -
root committed
May 26, 2022 Configuration menu - View commit details
-
Copy full SHA for 308a394 - Browse repository at this point
Copy the full SHA 308a394View commit details -
Merge pull request xuewujiao#6 from Thunderbrook/gpugraph_deepwalk
[GpuGraph] remove useless variables and adjust log level
Configuration menu - View commit details
-
Copy full SHA for 47b82ac - Browse repository at this point
Copy the full SHA 47b82acView commit details -
root committed
May 26, 2022 Configuration menu - View commit details
-
Copy full SHA for ee7f0e8 - Browse repository at this point
Copy the full SHA ee7f0e8View commit details -
Merge pull request xuewujiao#7 from huwei02/gpugraph
fix compile error
Configuration menu - View commit details
-
Copy full SHA for 319a201 - Browse repository at this point
Copy the full SHA 319a201View commit details
Commits on May 27, 2022
-
support slot_feature, storage and loading
root committedMay 27, 2022 Configuration menu - View commit details
-
Copy full SHA for c313938 - Browse repository at this point
Copy the full SHA c313938View commit details
Commits on May 28, 2022
-
root committed
May 28, 2022 Configuration menu - View commit details
-
Copy full SHA for 0fba9f2 - Browse repository at this point
Copy the full SHA 0fba9f2View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1d95522 - Browse repository at this point
Copy the full SHA 1d95522View commit details -
Merge pull request xuewujiao#8 from huwei02/gpugraph
support slot_feature, storage and loading
Configuration menu - View commit details
-
Copy full SHA for 9f8d4c2 - Browse repository at this point
Copy the full SHA 9f8d4c2View commit details
Commits on May 29, 2022
-
root committed
May 29, 2022 Configuration menu - View commit details
-
Copy full SHA for b6a4a0e - Browse repository at this point
Copy the full SHA b6a4a0eView commit details -
merge dymf branch (PaddlePaddle#42714)
merge dymf branch
Configuration menu - View commit details
-
Copy full SHA for 94fd1e3 - Browse repository at this point
Copy the full SHA 94fd1e3View commit details -
Configuration menu - View commit details
-
Copy full SHA for 716f908 - Browse repository at this point
Copy the full SHA 716f908View commit details -
add dymf support of gpups
Configuration menu - View commit details
-
Copy full SHA for 81644ce - Browse repository at this point
Copy the full SHA 81644ceView commit details -
Configuration menu - View commit details
-
Copy full SHA for 484ca4c - Browse repository at this point
Copy the full SHA 484ca4cView commit details -
Configuration menu - View commit details
-
Copy full SHA for e14d1f0 - Browse repository at this point
Copy the full SHA e14d1f0View commit details
Commits on May 31, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 453ddf5 - Browse repository at this point
Copy the full SHA 453ddf5View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6225444 - Browse repository at this point
Copy the full SHA 6225444View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1c2d8d4 - Browse repository at this point
Copy the full SHA 1c2d8d4View commit details -
Configuration menu - View commit details
-
Copy full SHA for c51530a - Browse repository at this point
Copy the full SHA c51530aView commit details -
root committed
May 31, 2022 Configuration menu - View commit details
-
Copy full SHA for 61a9cfa - Browse repository at this point
Copy the full SHA 61a9cfaView commit details -
Merge pull request xuewujiao#9 from Thunderbrook/gpugraph_deepwalk
[Gpugraph] change graph_sample interface
Configuration menu - View commit details
-
Copy full SHA for 2a16ca7 - Browse repository at this point
Copy the full SHA 2a16ca7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 8253f4e - Browse repository at this point
Copy the full SHA 8253f4eView commit details -
Merge pull request xuewujiao#13 from lxsbupt/gpugraph
Add gpu_graph_utils.h
Configuration menu - View commit details
-
Copy full SHA for f24c3e9 - Browse repository at this point
Copy the full SHA f24c3e9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 78a0ed3 - Browse repository at this point
Copy the full SHA 78a0ed3View commit details -
root committed
May 31, 2022 Configuration menu - View commit details
-
Copy full SHA for 283a430 - Browse repository at this point
Copy the full SHA 283a430View commit details -
Configuration menu - View commit details
-
Copy full SHA for e6cfbde - Browse repository at this point
Copy the full SHA e6cfbdeView commit details
Commits on Jun 1, 2022
-
Merge pull request xuewujiao#28 from DesmonDay/gpugraph
Delete old sample interface
Configuration menu - View commit details
-
Copy full SHA for 1d313a8 - Browse repository at this point
Copy the full SHA 1d313a8View commit details -
Configuration menu - View commit details
-
Copy full SHA for 0c4104d - Browse repository at this point
Copy the full SHA 0c4104dView commit details -
Configuration menu - View commit details
-
Copy full SHA for a9e5973 - Browse repository at this point
Copy the full SHA a9e5973View commit details -
Merge pull request xuewujiao#10 from danleifeng/mul_dims
support dynamic mf dims
Configuration menu - View commit details
-
Copy full SHA for b01e178 - Browse repository at this point
Copy the full SHA b01e178View commit details -
Configuration menu - View commit details
-
Copy full SHA for eaa58a3 - Browse repository at this point
Copy the full SHA eaa58a3View commit details -
Configuration menu - View commit details
-
Copy full SHA for eae32ae - Browse repository at this point
Copy the full SHA eae32aeView commit details -
Merge pull request xuewujiao#12 from seemingwang/gpu_graph_engine3
changing int64 key to uint64 for graph engine
Configuration menu - View commit details
-
Copy full SHA for 2457680 - Browse repository at this point
Copy the full SHA 2457680View commit details -
root committed
Jun 1, 2022 Configuration menu - View commit details
-
Copy full SHA for 440da15 - Browse repository at this point
Copy the full SHA 440da15View commit details
Commits on Jun 2, 2022
-
root committed
Jun 2, 2022 Configuration menu - View commit details
-
Copy full SHA for a11d1e3 - Browse repository at this point
Copy the full SHA a11d1e3View commit details -
root committed
Jun 2, 2022 Configuration menu - View commit details
-
Copy full SHA for b624aec - Browse repository at this point
Copy the full SHA b624aecView commit details -
Merge pull request xuewujiao#15 from Thunderbrook/gpugraph_deepwalk
[GpuGraph] metapath
Configuration menu - View commit details
-
Copy full SHA for d01a280 - Browse repository at this point
Copy the full SHA d01a280View commit details -
Configuration menu - View commit details
-
Copy full SHA for 4fb5c13 - Browse repository at this point
Copy the full SHA 4fb5c13View commit details -
Merge pull request xuewujiao#17 from Thunderbrook/gpugraph_0602
adapt uint64 and iterate edge table
Configuration menu - View commit details
-
Copy full SHA for 20a1aa3 - Browse repository at this point
Copy the full SHA 20a1aa3View commit details
Commits on Jun 6, 2022
-
Optimize memory overhead for gpugraph. (xuewujiao#18)
* Optimize memory overhead for gpugraph. * Optimize memory overhead for gpugraph.
Configuration menu - View commit details
-
Copy full SHA for d7eae03 - Browse repository at this point
Copy the full SHA d7eae03View commit details
Commits on Jun 8, 2022
-
[GpuGraph] same first_node_type share start (xuewujiao#19)
* same first nodetype share start * strategy
Configuration menu - View commit details
-
Copy full SHA for 750e343 - Browse repository at this point
Copy the full SHA 750e343View commit details -
search and fill slot_feature (xuewujiao#20)
* search and fill slot_feature * search and fill slot_feature, fix compile error * search and fill slot_feature, rename 8 as slot_num_ Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 1816fc2 - Browse repository at this point
Copy the full SHA 1816fc2View commit details -
remove debug code (xuewujiao#21)
* search and fill slot_feature * search and fill slot_feature, fix compile error * search and fill slot_feature, rename 8 as slot_num_ * remove debug code Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 8c65693 - Browse repository at this point
Copy the full SHA 8c65693View commit details -
remove debug code (xuewujiao#22)
* search and fill slot_feature * search and fill slot_feature, fix compile error * search and fill slot_feature, rename 8 as slot_num_ * remove debug code * remove debug code Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for ccccb5f - Browse repository at this point
Copy the full SHA ccccb5fView commit details
Commits on Jun 10, 2022
-
* load edge parallel * load node and edge parallel * add node and edge parallel * comment slot fea parse * clear some log * add some log Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 3bde5e4 - Browse repository at this point
Copy the full SHA 3bde5e4View commit details
Commits on Jun 11, 2022
-
dedup feature, opt parse feature (xuewujiao#25)
* dedup feature, opt parse feature * remove debug code Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 68c2cc4 - Browse repository at this point
Copy the full SHA 68c2cc4View commit details
Commits on Jun 12, 2022
-
Configuration menu - View commit details
-
Copy full SHA for ba5d709 - Browse repository at this point
Copy the full SHA ba5d709View commit details
Commits on Jun 13, 2022
-
[GpuGraph] add enforce in alloc and free (xuewujiao#27)
* solve int overflow * enforce for alloc and free
Configuration menu - View commit details
-
Copy full SHA for 77b007e - Browse repository at this point
Copy the full SHA 77b007eView commit details -
* Optimize memory overhead for gpugraph. * Optimize memory overhead for gpugraph. * Add debug codes for HBM
Configuration menu - View commit details
-
Copy full SHA for a7ce0cf - Browse repository at this point
Copy the full SHA a7ce0cfView commit details -
add whether load node and edge parallel flag (xuewujiao#28)
* add whether load node and edge parallel flag * add whether load node and edge parallel flag * add whether load node and edge parallel flag Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for d6f1800 - Browse repository at this point
Copy the full SHA d6f1800View commit details -
opt slot_feature (xuewujiao#29)
Co-authored-by: huwei02@baidu.com <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 8e712f0 - Browse repository at this point
Copy the full SHA 8e712f0View commit details
Commits on Jun 14, 2022
-
move feature_buf into if-branch (xuewujiao#31)
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 0c33297 - Browse repository at this point
Copy the full SHA 0c33297View commit details
Commits on Jun 15, 2022
-
add get all neighbor (xuewujiao#35)
* add whether load node and edge parallel flag * add whether load node and edge parallel flag * add whether load node and edge parallel flag * add get all neighbor * add get all neighbor * add get all neighbor * add get all neighbor Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 0969831 - Browse repository at this point
Copy the full SHA 0969831View commit details
Commits on Jun 16, 2022
-
[GpuGraph]direct access (xuewujiao#36)
* direct access * format * log level
Configuration menu - View commit details
-
Copy full SHA for 8ce3cf9 - Browse repository at this point
Copy the full SHA 8ce3cf9View commit details -
support graph inference (xuewujiao#34)
* gpu_graph_infer * simplify infer * fix * remove logs * remove logs * change logs
Configuration menu - View commit details
-
Copy full SHA for f19ca37 - Browse repository at this point
Copy the full SHA f19ca37View commit details -
add adam/sharedadam optimzier for gpups (xuewujiao#32)
* add adam/sharedadam optimzier for gpups;edit optimizer struct;test=develop * remove useless code;test=develop * remove useless code;test=develop * remove useless code;test=develop * remove useless code;test=develop * remove useless code;test=develop
Configuration menu - View commit details
-
Copy full SHA for e5b7978 - Browse repository at this point
Copy the full SHA e5b7978View commit details
Commits on Jun 21, 2022
-
Optimizing the zero key problem in the push phase
root committedJun 21, 2022 Configuration menu - View commit details
-
Copy full SHA for 39b8659 - Browse repository at this point
Copy the full SHA 39b8659View commit details -
get_features: int64_t -> uint64_t (xuewujiao#38)
* get_features: int64_t -> uint64_t * remove useless var Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 33ba59a - Browse repository at this point
Copy the full SHA 33ba59aView commit details -
Optimizing the zero key problem in the push phase (xuewujiao#40)
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0008.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 4179390 - Browse repository at this point
Copy the full SHA 4179390View commit details -
fix compile error (xuewujiao#43)
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 75630e9 - Browse repository at this point
Copy the full SHA 75630e9View commit details -
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJun 21, 2022 Configuration menu - View commit details
-
Copy full SHA for c6193e1 - Browse repository at this point
Copy the full SHA c6193e1View commit details
Commits on Jun 22, 2022
-
* fix config does not take effect; test=develop * remove useless code;test=develop
Configuration menu - View commit details
-
Copy full SHA for d5011c7 - Browse repository at this point
Copy the full SHA d5011c7View commit details -
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJun 22, 2022 Configuration menu - View commit details
-
Copy full SHA for c5b6b59 - Browse repository at this point
Copy the full SHA c5b6b59View commit details -
Optimize CUDA thread parallelism in MergeGrad phase
root committedJun 22, 2022 Configuration menu - View commit details
-
Copy full SHA for bc32d9d - Browse repository at this point
Copy the full SHA bc32d9dView commit details -
Optimize CUDA thread parallelism in MergeGrad phase
root committedJun 22, 2022 Configuration menu - View commit details
-
Copy full SHA for 960469b - Browse repository at this point
Copy the full SHA 960469bView commit details
Commits on Jun 23, 2022
-
load parallel between nodes and edges (xuewujiao#37)
* load parallel between nodes and edges
Configuration menu - View commit details
-
Copy full SHA for c176c00 - Browse repository at this point
Copy the full SHA c176c00View commit details -
speed up dumpfile (xuewujiao#42)
* speed up gpugraph infer,using sprintf
Configuration menu - View commit details
-
Copy full SHA for 35a704f - Browse repository at this point
Copy the full SHA 35a704fView commit details -
reconstruct table,reduce hbm 20.8G (xuewujiao#46)
* reconstruct table,reduce hbm 20.8G
Configuration menu - View commit details
-
Copy full SHA for fdf59b6 - Browse repository at this point
Copy the full SHA fdf59b6View commit details
Commits on Jun 24, 2022
-
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJun 24, 2022 Configuration menu - View commit details
-
Copy full SHA for 5d29e89 - Browse repository at this point
Copy the full SHA 5d29e89View commit details -
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJun 24, 2022 Configuration menu - View commit details
-
Copy full SHA for 2279fc1 - Browse repository at this point
Copy the full SHA 2279fc1View commit details -
Merge branch 'gpugraph.0621' of https://github.com/lxsbupt/Paddle int…
…o gpugraph.0621
root committedJun 24, 2022 Configuration menu - View commit details
-
Copy full SHA for cf65277 - Browse repository at this point
Copy the full SHA cf65277View commit details -
* Optimizing the zero key problem in the push phase * Optimize CUDA thread parallelism in MergeGrad phase * Optimize CUDA thread parallelism in MergeGrad phase Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0008.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 79675fd - Browse repository at this point
Copy the full SHA 79675fdView commit details
Commits on Jun 26, 2022
-
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJun 26, 2022 Configuration menu - View commit details
-
Copy full SHA for 26a0779 - Browse repository at this point
Copy the full SHA 26a0779View commit details
Commits on Jun 27, 2022
-
fix slot_feature in infer (xuewujiao#47)
* fix slot_feature in infer
Configuration menu - View commit details
-
Copy full SHA for 851d336 - Browse repository at this point
Copy the full SHA 851d336View commit details
Commits on Jun 28, 2022
-
opt cpu to gpu load (xuewujiao#50)
* opt cpu to gpu load:1. remove vector copy;2. parallel * remove useless code
Configuration menu - View commit details
-
Copy full SHA for 87d49ec - Browse repository at this point
Copy the full SHA 87d49ecView commit details -
fix adam with multi dim (xuewujiao#39)
* fix adam with multi dim; test=develop
Configuration menu - View commit details
-
Copy full SHA for 7263442 - Browse repository at this point
Copy the full SHA 7263442View commit details
Commits on Jun 29, 2022
-
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJun 29, 2022 Configuration menu - View commit details
-
Copy full SHA for 1dcffaf - Browse repository at this point
Copy the full SHA 1dcffafView commit details -
Performance optimization, segment gradient merging
root committedJun 29, 2022 Configuration menu - View commit details
-
Copy full SHA for b19d610 - Browse repository at this point
Copy the full SHA b19d610View commit details
Commits on Jun 30, 2022
-
Configuration menu - View commit details
-
Copy full SHA for e7f3193 - Browse repository at this point
Copy the full SHA e7f3193View commit details -
Performance optimization, segment gradient merging
root committedJun 30, 2022 Configuration menu - View commit details
-
Copy full SHA for b277cfc - Browse repository at this point
Copy the full SHA b277cfcView commit details
Commits on Jul 4, 2022
-
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJul 4, 2022 Configuration menu - View commit details
-
Copy full SHA for f5302c8 - Browse repository at this point
Copy the full SHA f5302c8View commit details -
Supports segmented gradient merging.
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJul 4, 2022 Configuration menu - View commit details
-
Copy full SHA for bb64ead - Browse repository at this point
Copy the full SHA bb64eadView commit details -
Supports segmented gradient merging
root committedJul 4, 2022 Configuration menu - View commit details
-
Copy full SHA for 20e5aeb - Browse repository at this point
Copy the full SHA 20e5aebView commit details -
Optimize CUDA thread parallelism in MergeGrad phase,6.5h-3.25h
Configuration menu - View commit details
-
Copy full SHA for 1ce2252 - Browse repository at this point
Copy the full SHA 1ce2252View commit details -
fix inference batch_size configuration (xuewujiao#57)
* gpu_graph_infer * simplify infer * fix * remove logs * remove logs * change logs * speed up dumpfile * reduce inference time * replace to_string * fix inference * change log * fix infer batch_size * remove log * check tensor initialization
Configuration menu - View commit details
-
Copy full SHA for 967cb54 - Browse repository at this point
Copy the full SHA 967cb54View commit details
Commits on Jul 6, 2022
-
Optimize graph loading and sample loading performance (xuewujiao#60)
* Optimize graph loading and sample loading performance,gpu load graph from 172s to 124s, cpu load graph from 92s to 45s
Configuration menu - View commit details
-
Copy full SHA for 33f3c9b - Browse repository at this point
Copy the full SHA 33f3c9bView commit details
Commits on Jul 11, 2022
-
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJul 11, 2022 Configuration menu - View commit details
-
Copy full SHA for 7aad6ff - Browse repository at this point
Copy the full SHA 7aad6ffView commit details -
Optimize pullsparse and increase keys aggregation
root committedJul 11, 2022 Configuration menu - View commit details
-
Copy full SHA for b7ba6a9 - Browse repository at this point
Copy the full SHA b7ba6a9View commit details -
Optimize pullsparse and increase keys aggregation (xuewujiao#61)
dedup keys when pullsparse
Configuration menu - View commit details
-
Copy full SHA for 84d4b98 - Browse repository at this point
Copy the full SHA 84d4b98View commit details -
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJul 11, 2022 Configuration menu - View commit details
-
Copy full SHA for 08d3499 - Browse repository at this point
Copy the full SHA 08d3499View commit details
Commits on Jul 12, 2022
-
pull push dedup, use pull feature value clip size (xuewujiao#62)
pull push dedup, clip pull feature value
Configuration menu - View commit details
-
Copy full SHA for 15aeb84 - Browse repository at this point
Copy the full SHA 15aeb84View commit details -
Configuration menu - View commit details
-
Copy full SHA for cf46d7b - Browse repository at this point
Copy the full SHA cf46d7bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 988aa30 - Browse repository at this point
Copy the full SHA 988aa30View commit details
Commits on Jul 13, 2022
-
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJul 13, 2022 Configuration menu - View commit details
-
Copy full SHA for 0667391 - Browse repository at this point
Copy the full SHA 0667391View commit details -
add result check and add parameter hhm cost stat (xuewujiao#66)
add result check and add parameter hhm cost stat
Configuration menu - View commit details
-
Copy full SHA for 22e1835 - Browse repository at this point
Copy the full SHA 22e1835View commit details
Commits on Jul 15, 2022
-
Configuration menu - View commit details
-
Copy full SHA for ce40d68 - Browse repository at this point
Copy the full SHA ce40d68View commit details
Commits on Jul 20, 2022
-
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJul 20, 2022 Configuration menu - View commit details
-
Copy full SHA for 24437bf - Browse repository at this point
Copy the full SHA 24437bfView commit details
Commits on Jul 21, 2022
-
apply shard_num (xuewujiao#55)
* save performance;test=develop * increase shard_num
Configuration menu - View commit details
-
Copy full SHA for e540156 - Browse repository at this point
Copy the full SHA e540156View commit details
Commits on Jul 22, 2022
-
feature learning rate (xuewujiao#69)
splite node and slot feature learning rate
Configuration menu - View commit details
-
Copy full SHA for d05b5b5 - Browse repository at this point
Copy the full SHA d05b5b5View commit details -
Configuration menu - View commit details
-
Copy full SHA for 933d14f - Browse repository at this point
Copy the full SHA 933d14fView commit details
Commits on Jul 25, 2022
-
[gpups]refine adam aceessor (xuewujiao#58)
* [gpups]refine adam aceessor;test=develop * template;test=develop * fix adam accessor:template;test=develop * add feature_value.cu;test=develop * Merge pull request * format; test=develop * format; test=develop * format; test=develop * add ut; test=develop * fix format * fix format * format
Configuration menu - View commit details
-
Copy full SHA for 3c79186 - Browse repository at this point
Copy the full SHA 3c79186View commit details
Commits on Jul 26, 2022
-
fix bug in slot_feature (xuewujiao#71)
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for e7ca266 - Browse repository at this point
Copy the full SHA e7ca266View commit details -
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedJul 26, 2022 Configuration menu - View commit details
-
Copy full SHA for a6e6232 - Browse repository at this point
Copy the full SHA a6e6232View commit details
Commits on Jul 27, 2022
-
add FLAGS_gpugraph_load_node_list_into_hbm to decide wether load node…
…_list into hbm (xuewujiao#72) Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 416c558 - Browse repository at this point
Copy the full SHA 416c558View commit details -
Configuration menu - View commit details
-
Copy full SHA for bba158c - Browse repository at this point
Copy the full SHA bba158cView commit details
Commits on Jul 28, 2022
-
format general accessor (xuewujiao#74)
* format general accessor;test=develop * format general accessor;test=develop
Configuration menu - View commit details
-
Copy full SHA for eed5f6b - Browse repository at this point
Copy the full SHA eed5f6bView commit details
Commits on Jul 29, 2022
-
[GPUGraph] enable exit trainer when an worker has no instance (xuewuj…
…iao#75) * add barrier * train mode
Configuration menu - View commit details
-
Copy full SHA for a0a6e81 - Browse repository at this point
Copy the full SHA a0a6e81View commit details -
Configuration menu - View commit details
-
Copy full SHA for fa55719 - Browse repository at this point
Copy the full SHA fa55719View commit details
Commits on Aug 4, 2022
-
remove isolate node_id from embdding table (xuewujiao#79)
* remove isolate node_id from embdding table Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for a7017fc - Browse repository at this point
Copy the full SHA a7017fcView commit details -
merge develop to gpugraph (xuewujiao#77)
* fix python3.10 compile bug on window (PaddlePaddle#44330) * Fix random seed for several unit tests (PaddlePaddle#44135) * Fix test_functional_conv2d_transpose random seed * Fix random seed and use np.testing * Fix random seed for test_lu_unpack_op * Fix test_autograd_functional_dynamic random seed * Remove boost library (PaddlePaddle#44092) * add fused token prune op and plugin (PaddlePaddle#44281) * add fused token prune op and plugin * Fix run inference bug for standalone executor (PaddlePaddle#44340) * xpu-paddlepaddle-33 [任务] matmul单测 timeout (PaddlePaddle#44333) test=kunlun * [IPU] add custom-op UTs 0/N (PaddlePaddle#44328) * add custom-op UTs 0 * add authors Co-authored-by: Allen Guo <alleng@graphcore.ai> Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai> Co-authored-by: Zhaorui Chen <zhaoruic@graphcore.ai> Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai> Co-authored-by: Zhaorui Chen <zhaoruic@graphcore.ai> * [IPU] add custom-op UTs 1/N (PaddlePaddle#44329) * add custom-op UTs 1 * add authors Co-authored-by: Allen Guo <alleng@graphcore.ai> Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai> Co-authored-by: Zhaorui Chen <zhaoruic@graphcore.ai> * update url Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai> Co-authored-by: Zhaorui Chen <zhaoruic@graphcore.ai> * support KL2 multi-card training, *test=kunlun (PaddlePaddle#43889) * update xccl lib * use separate streams for compute/comm on XPU * add broadcast op to xpu2_op_list * Remove auto to_pascal_case for args in op generator (PaddlePaddle#44350) * remove auto to_pascal_case for args in op generator * fix yaml config * Standard sparse conv name (PaddlePaddle#44353) * [Eager] eager variable back sync (PaddlePaddle#44343) * eager variable back sync * [ Phi Kernel ] Transfer as_real to phi. (PaddlePaddle#44263) * transfer as_real to phi * fix erros * blocking: True -> False * [Eager]Fix assert statement (PaddlePaddle#43492) * Not rename pb file to avoid re-compile (PaddlePaddle#44370) * [Phi] Migrate solve kernel to phi (PaddlePaddle#44363) * draft version * draft version * draft version * migrate solve kernel to phi * polish * polish * re useless header file, fix a bug in grad_kernel_impl * add header file in need * [auto parallel] remove comm init control (PaddlePaddle#44385) * [CustomDevice] remove unused file (PaddlePaddle#44358) * [Paddle-TRT] reshape fill_constant (PaddlePaddle#44314) * reshape fill_constant * commit * commit * set seed for uts (PaddlePaddle#44372) * [Paddle-TRT] remove useless code in fc (PaddlePaddle#44382) * remove useless code in fc * [Paddle-TRT] Fix cast (PaddlePaddle#44312) * fix_cast * fix_cast * commit * Polish jit layer cmakelists to hide some message (PaddlePaddle#44351) * Enable inference multi stream ci test (PaddlePaddle#44275) * test * update * fix bug of old pp (PaddlePaddle#44361) * add xpu resnet_unit (PaddlePaddle#44297) * add xpu resnet_unit *test=kunlun * tmp *test=kunlun * add blacklist in prim2orig interface (PaddlePaddle#44383) * [Plugin] Fix Custom device in eager mode, test=develop (PaddlePaddle#43952) * [Plugin] Fix Custom device in eager mode, test=develop * update test case, test=develop * update ut for coverage, test=develop * add ipu support for standalone executor. (PaddlePaddle#44342) * fix typos in template for codegen of operators (PaddlePaddle#44364) * fix duplicate slice logic in _grad (PaddlePaddle#44396) * [MLU] fix mlu ctest final. (PaddlePaddle#44404) * fix data transform bug of interpolate op (PaddlePaddle#44401) * [Sparse] Add sparse matmul kernel(coo*dense->dense) (PaddlePaddle#44346) * fix new autodiff api docs (PaddlePaddle#44341) * fix build error in low arch (PaddlePaddle#44391) * [new api] add new api paddle.vision.ops.distribute_fpn_proposals (PaddlePaddle#43736) * add distribute_fpn_proposals * change to new dygraph * fix doc and example code * change fluid impl to current version * update (PaddlePaddle#44418) * [Paddle-TRT] Shape sum fix scale (PaddlePaddle#44394) * shape sum * add shape, sum trt layer * [Phi] Migrate infermeta and add yaml for solve op (PaddlePaddle#44379) * migrate solve kernel to phi * re useless header file, fix a bug in grad_kernel_impl * add header file in need * add yaml for solve op * fix solve_sig.cc ArgumentMapping and update tests case * disable legacy dygraph check in op_test * rm solve_op.cc / solve_sig.cc and migrate yaml config * Update op_test.py disable legacy dygraph check when check_eager is True * add labels for infer ut (PaddlePaddle#44279) * add labels for infer ut * add RUN_TYPE=INFER for cpp ut * fix formaterror * update * Add mfence for XPU2 KP (PaddlePaddle#44258) * remove include of all.h in resnet_basic_block_op_xpu.cc, test=kunlun (PaddlePaddle#44423) * Rename BOOST_GET macros (PaddlePaddle#44368) * Rename BOOST_GET macros * Fix conflicts * [new API] add paddle.vision.ops.generate_proposals (PaddlePaddle#43611) * add generate_proposals into paddle.vision * remove class api * im_info -> img_size * change fluid impl to current version * Accelerate inference period in op Cache method (PaddlePaddle#43857) * Added pad3d and pad2d FP32 FWD oneDNN kernels (PaddlePaddle#43990) * Piotrek's changes for pad3d * my changes * first version of pad3d, single copy, unnecessary reads * optimized pad3d kernel * test upadte * removed magic numbers * added support for pad2d * reverted two files * reverted one old change * added support for Paddings tensor * CI fix * CI fix * fixed timeout of tests * fixed typo * changes to GetKernelTypeForVar * Revert "changes to GetKernelTypeForVar" This reverts commit 4691061. * added AsExtra() to pad2d Co-authored-by: Piotr Paturej <piotr.paturej@intel.com> * add save_cache/patch (PaddlePaddle#44420) * add save_cache/patch * add pybind * remove pybind * remove const_cast * add fleet * Standard name of sparse pool (PaddlePaddle#44344) * move eig operator from fluid to phi (PaddlePaddle#44398) * move eig operator from fluid to phi * add eig_grad unitest, upgrade IsComplexType() from fluid to phi * [Phi]Move angle op to phi (PaddlePaddle#44393) * Move angle op to phi * Replace mutable_data using Alloc * Remove some include * Try to fix windows ci error * include math.h to fix windows ci error * Fix kernel name * Move angle_grad infershape * [Eager]release gil when run backward (PaddlePaddle#44433) * release gil when run backward * compile phi/backends into one static library (PaddlePaddle#44373) * compile into one static library * fix xpu compile * fix xpu compile * fix inference compile * fix inference compile * add custom test * revert one file * [IPU] Add more Ops (PaddlePaddle#44414) * [IPU] Add more Ops * update boost API * Clean CI_SKIP_CPP_TEST (PaddlePaddle#44412) * Add dependency for read op in standalone executor (PaddlePaddle#44362) * Add dependency for read op in standalone executor * Fix CI errors * Add UT * add_dependency -> dependency_utils * Fix CI errors * Add distro in ci docker (PaddlePaddle#44332) * add distro zstd * test * test * add pip3.8 * [Phi] migrate as_complex kernel to phi (PaddlePaddle#44438) * migrate as_complex kernel to phi * support as_complex and as_real in phi * rm GetExpectedKernelType for AsRealOp * [GPUPS]FleetWrapper initialize (PaddlePaddle#44441) * fix FleetWrapper initialize * [XPU][NPU] (1) add device_guard. (2) add support for LoDTensorArray of sum op. (PaddlePaddle#44367) * device_guard support xpu. test=kunlun * sum op of xpu support LoDTensorArray. add test for while op of xpu. test=kunlun. * [IPU] add Op uts (PaddlePaddle#44415) * transfer block_id to CreateVarNode in multi_devices_graph_pass (PaddlePaddle#44366) * fix CreateVarNode in multi_devices_graph_pass * Revert "Fix var duplication bug for graph_to_program_pass (PaddlePaddle#44278)" This reverts commit a2c4c86. * 【GPUPS】Adam accessor (PaddlePaddle#43919) * add adam/sharedadam optimzier for gpups;edit optimizer struct;test=develop * [Phi] migrate sync_batch_norm to phi (PaddlePaddle#44369) * [GPUPS]Fix psgpuwrapper initialization (PaddlePaddle#44468) * Update ps_gpu_wrapper.h * Update ps_gpu_wrapper.h * Update ps_gpu_wrapper.cc * [Phi] migrate exponential kernel to phi (PaddlePaddle#44376) * [Phi] migrate exponential kernel to phi * fix comment * fix CI * [PHI] move diag_embed op to phi. (PaddlePaddle#44408) * move diag_embed to phi. * [MLU] set_value performance optimizing (PaddlePaddle#44390) * Update api changing approve members (PaddlePaddle#44463) * update api approve members, test=document_fix * add qingqnig into list, test=document_fix * fix bug,test=document_fix (PaddlePaddle#44478) * [Phi] migrate clip_by_norm to phi (PaddlePaddle#44458) * add eigen3 dependency for phi_backends (PaddlePaddle#44479) * remove fleet_13 ut in parallel_UT_rule.py; test=develop (PaddlePaddle#44477) * [PHI]Seperate xshape kernel from normal kernel (PaddlePaddle#44315) * seperate xshape kernel from normal kernel * fix bugs in infermeta * fix compile bugs * fix compile bugs * [AutoParallel] fix unittest with paddle.distributed.launch (PaddlePaddle#44439) * fix unittest * fix log_dir * _enable_legacy_dygraph * [Phi] add temporal_shift yaml (PaddlePaddle#44409) * add temporal_shift yaml and unittest * [Paddle inference] Add conv_fusion_fp16 (PaddlePaddle#44435) * convfusionfp16 * convfusionfp16 * convfusionfp16 * fix some convert error found in tipc. (PaddlePaddle#44457) * fix some error found in tipc. * update * [BugFix]Fix randint_like bugs when save program that don't need use tensor's value (PaddlePaddle#44446) * fix bugs of random * fix unittest error * fix unittest bugs * add adaptive pool and softmax with cross entropy supports different axis, * test = kunlun (PaddlePaddle#44428) * add xpu pnorm op and fix pool op, *test=kunlun * add adaptive pool, and softmax with cross entropy supports different axis, *test=kunlun * add slot attr for push sparse op (PaddlePaddle#44422) * add slot attr for push sparse op * add pybind * remove fleet * add unittest * fix * [Dy2Sta]Fix Segment Fault while training multi-card if params have no grad (PaddlePaddle#44485) * [Dy2Sta]Fix Segment Fault while training multi-card if params have no grad * fix unittest * fix tensor stream error in custom op (PaddlePaddle#44500) * Replace with dygraph op calling method. (PaddlePaddle#44331) * Replace with dygraph op calling method. * [JitLayer]Pybind PEFunction and call phi api in layer_test (PaddlePaddle#44465) * Support predictor function in JitLayer * Pybind PEFunction * Pybind PEFunction and call phi api in layer_test * Call sqrt phi API * Polish flags * Fix comments * [Sparse] Add sparse addmm kernel (dense+coo*dense->dense,dense+csr*dense->dense) (PaddlePaddle#44451) * [Eager] bilinear_tensor_product yaml (PaddlePaddle#44459) * bilinear_tensor_product yaml * [ Phi ] svd transfer (PaddlePaddle#44392) * svd cpu forward * svd gpu forward * transfer the backward of svd * remove cusolver in svd_grad * svd kernel bug fix * fix bugs * fix bugs. * fix bug * [Paddle-TRT] fix_fill_constant (PaddlePaddle#44481) * fix_fill_constant * fix_fill_constant * fix_ernie * [MLU] transpose avg_pool2d to NHWC for better performance. (PaddlePaddle#44475) * [jit] jit support property.proto (PaddlePaddle#44337) * add property.proto, can compiled * property get and deserilize * support get float * format code * format code * add unittest * add more set method * fix grammar error * Update paddle/fluid/jit/property.h Co-authored-by: Aurelius84 <zhangliujie@baidu.com> * Update paddle/fluid/jit/property.cc Co-authored-by: Aurelius84 <zhangliujie@baidu.com> * Update paddle/fluid/jit/property.cc Co-authored-by: Aurelius84 <zhangliujie@baidu.com> * Update paddle/fluid/jit/property.cc Co-authored-by: Aurelius84 <zhangliujie@baidu.com> * fix comment * fix error throw * fix property save unit test * fix error info * fix copyright and header import * reorder jit property tensor datatype Co-authored-by: Aurelius84 <zhangliujie@baidu.com> * [ Dy2static ] infer_program may be incorrect in amp mode. (PaddlePaddle#44487) * fix the outputs of net is x,x * add unittest for duplicate output * fix * fix _infer_program use the original program not the amp program. * get _***program_id back and avoid duplicate cache ing * fix * Fc fp16 (PaddlePaddle#44505) * fc support fp16 * add a ‘,’ on paddle_pass_builder.cc * fc support fp16 on non-cuda. * add batch stream (PaddlePaddle#44524) * shufflechannelfix (PaddlePaddle#44516) * fix arg_max to select first index (PaddlePaddle#44521) * [MLU] add floor kernel and grid_sampler kernel (PaddlePaddle#44498) * commit (PaddlePaddle#44534) * [CustomDevice] register Copy for custom device (PaddlePaddle#44200) * [CustomDevice] register Copy for custom device * [CustomDevice] register Copy for custom device * [CustomDevice] register Copy for custom device * merge and add uts * merge and add uts * fix for blocking and unittests coverage * (modified) fc support fp16 (PaddlePaddle#44540) * Add code of occupancy computing on DCU and avoid threadID bug for DCU profiler (PaddlePaddle#44520) * add xpu lars_momentum/pow2_decay (PaddlePaddle#44448) *test=kunlun * [phi] move inverse op from fluid to phi (PaddlePaddle#44471) * move inverse from fluid to phi with unitest bug * fix bug, add eager op yaml * support send_partial, recv_partial and allgather_partial in ProcessGroupNCCL (PaddlePaddle#44444) * [Sparse]add sparse unary api(expm1/deg2rad/rad2deg/relu6/leaky_relu) (PaddlePaddle#44432) * Fc fp16 (PaddlePaddle#44558) * (modified) fc support fp16 * __CUDA_ARCH__ version * delete half * delete half * Fix bug of amp code-gen (PaddlePaddle#44570) * fix bug of amp code_gen * fix bug * [JitLayer]Fix jit.save error when save params combined (PaddlePaddle#44504) * Fix jit.save error when save params combined * Change dict_value to list * [Phi] Migrate squared_l2_norm_op to phi (PaddlePaddle#44492) * add swish using TensorRT layer (PaddlePaddle#44561) * update * empty commit * update * update * update * Phi gird sampler migration (PaddlePaddle#44562) * add_ymal_utest for phi grid_sampler op * skip dist test cases if mlu card number only one, test=develop (PaddlePaddle#44549) * [dy2st]Add ProgramHelper to polish build program logic in autoparallel.Engine (PaddlePaddle#44513) * [dy2st]Add ProgramHelper to polish build program logic in autoparallel.Engine * refine code * 【Hackathon No.21】为 Paddle 新增 SoftMarginLoss (PaddlePaddle#42364) * 2022-04-28 * 2022-04-28_V2 * 2022-04-30 * 2022-04-30_V2 * 2022-05-01 * 2022-05-02 * 2022-05-02_V2 * 2022-05-05_V1 * 2022-05-06_V1 * 2022-05-07_V1 * Update loss.py * 2022-05-07_V2 * 2022-05-13_V1 * Update test_soft_margin_loss.py * Update loss.py * Update loss.py * 2022-05-16_V1 * 2022-05-19_V1 * 2022-05-20_V1 * Update test_soft_margin_loss.py * 2022-06-01_V1 * 2022-06-05 * 2022-06-07 * 2022-06-07 * 2022-06-08 * 2022-06-08_V2 * 2022-06-17-code_style * Modify python * 2022-06-20 * for * for CI;test=document_fix Co-authored-by: Ligoml <39876205+Ligoml@users.noreply.github.com> * [MLU]transpose convbpf output to HWCN for better performance (PaddlePaddle#44552) * Fc fp16 (PaddlePaddle#44578) * (modified) fc support fp16 * __CUDA_ARCH__ version * delete half * delete half * add half support * add half support * add half support * [Auto Parallel] Add dist op cost (PaddlePaddle#44146) * update comp cost * add dist default op cost * add dist fill constant batch size like op cost * add elewise op cost * add fill_constant_batch_size_like op cost unittest * add unittest and remove fill_constant_batch_size_like grad op cost * add to cmakelist * fix unittest bug * Improve CI unittest parallel execution strategy (PaddlePaddle#44334) * paralle_test_daily * test=paralle_test_daily * test=paralle_test_daily * test=parallel_test_daily * test=paralle_test_daily * test=paralle_test_daily * test=paralle_test_daily * test=paralle_test_daily * test=paralle_test_daily * test=paralle_test_daily * test=paralle_test_daily * test=paralle_test_daily * test pre_test_bak * test cfs * test_cfs,test=paralle_test_daily * test_cfs,test=paralle_test_daily * fix nightly test name,test=paralle_test_daily * fix nightly test name,test=paralle_test_daily * test ci parallel speed * refine parallel rule,test=paralle_test_daily * Move bmm OP from fluid to phi (PaddlePaddle#44496) * [PHI]Move slogdeterminant op to phi (PaddlePaddle#44547) * Move slogdeterminant op to phi * Add yaml and unit test for slogdeterminant * Rename pybind_boost_header.h (PaddlePaddle#44592) * unify data type and property enum value (PaddlePaddle#44585) * inference multi stream support handle lazy init. (PaddlePaddle#44563) * multi stream support handle lazy init. * support eigen lazy init * update * fix ci problem * Remove ControlDepVar in GraphToBlock (PaddlePaddle#44591) * transfer the svd infer into phi infermeta (PaddlePaddle#44528) * transfer the svd infer into phi infermeta * remove the svd.h * modify svd api * fix svd error by insert optional * Einsum grad complex (PaddlePaddle#44598) * add complex for einsum grad kernel * pass the ci * add reverse yaml (PaddlePaddle#44518) * add reverse yaml * Set more attrs in ReplaceScaleLossGradOp (PaddlePaddle#44576) * Set more attrs in ReplaceScaleLossGradOp * Fix typos * Fix CI errors * Add UT * [Phi] Migrate box coder to phi. (PaddlePaddle#44550) * fix behavior of device_id=None in Tensor.cuda (PaddlePaddle#44515) * fix behavior of device_id=None in Tensor.cuda * fix CI * fix windows cuda11.7 bug (PaddlePaddle#44601) * add horizontal federation learning ps feature (PaddlePaddle#44327) * back fl * delete ssl cert * . * make warning * . * unittest paral degree * solve unittest * heter & multi cloud commm ready * . * . * fl-ps v1.0 * . * support N + N mode * . * . * . * . * delete print * . * . * . * . * fix bug * . * . * fl-ps with coordinator ready * merge dev * update message parse only * update fl client scheduler * fix bug * update multithreads sync * fix ci errors * update role_maker.py * update role_maker.py * fix ci error: windows py import error * fix ci error: windows py import error * fix windows ci pylib import error * add dump fields & params * try to fix windows import fleet error * fix ps FLAGS error * [MLU] rollback cntoolkit vetsion to 2.8.5 (PaddlePaddle#44595) * [CustomDevice] add blas_axpby api for gradient_accumulator (PaddlePaddle#44584) * add sin,cos,exp primitive operators (PaddlePaddle#44345) * Optimize sparse convolution (PaddlePaddle#43576) * Merge kProgramDescs in GraphToProgram (PaddlePaddle#44526) * [Eager] Add warpctc yaml (PaddlePaddle#44617) * Add a feed op before each input parameter var. (PaddlePaddle#44499) * Add a feed op before each input parameter var. * Fix some issues about the unit test build_cinn_pass_test. * fix record event for operator type in new dygraph (PaddlePaddle#44582) * fix new dygraph record event for op * update unit test * fix bug of elementwise_add_grad, *test=kunlun (PaddlePaddle#44545) * fix bug of elementwise_add_grad, *test=kunlun * fix bug, *test=kunlun * rm pooling_t, *test=kunlun * fix bug of ew_add_grad when inplace, *test=kunlun * [IPU] small bug fix (PaddlePaddle#44473) * sync misc changes * add authors Co-authored-by: Zhaorui Chen <zhaoruic@graphcore.ai> * up x * Revert "up x" This reverts commit f3fde45. * add guarg for ipu Co-authored-by: Zhaorui Chen <zhaoruic@graphcore.ai> * support auto fallback to cpu kernel for cusom device (PaddlePaddle#44639) * fix dygraph bugs in broadcast_to api. (PaddlePaddle#44612) * add set_dtype for inverse_op (PaddlePaddle#44618) * refine overalls.cmake (PaddlePaddle#44623) * [PHI]Add yaml and unittest for bmm op (PaddlePaddle#44625) Add yaml and unittest for bmm op * Phi average accumulates migration (PaddlePaddle#44554) * move average_accumulates op to phi kernel * new exe not support pg (PaddlePaddle#44628) * [CustomDevice]fix phi kernel header (PaddlePaddle#44637) * [CustomDevice] add process_group_xccl ut (PaddlePaddle#44632) * [CustomDevice] add process_group_xccl ut * update * Fix conv api name (PaddlePaddle#44636) * [DCU] Fix NAN problem when training BERT on DUC platform (PaddlePaddle#44643) * [JitLayer]Remove include fluid head files in JitLayer (PaddlePaddle#44597) * Remove include fluid head files in JitLayer * Format code * Remove const to fix ci error * Fix param error * Polish jit layer include and cp some headers to python/include * Fix comment * [jit] jit.save support property serialization (PaddlePaddle#44581) * jit.save support peropty serilization * extract set property function * fix property test file name * fix typing error * fix typing error * fix test coverage * Replaced add_custom_command with add_custom_target in xpu_kp_cmake (PaddlePaddle#44619) * Replaced add_custom_command with add_custom_target in xpu_kp_cmake * add adagrad and rmsprop yaml (PaddlePaddle#44631) * [phi] move crop_tensor kernel from fluid to phi (PaddlePaddle#44574) * move crop_tensor from fluid to phi * delete fluid header files * fix crop_tensor_op dygraph_mode bug * modify header files, add out tensor check * fix RemoveIntermediateOut in fuse_elewise_add_act_pass while converting graph to program (PaddlePaddle#44593) * fix RemoveNode in fuse_elewise_add_act_pass * fix * change pointer to share_ptr * fix * fix * fix format * fix * fix graph_safe_remove_nodes * fix UTs on physical ipu (PaddlePaddle#44647) * [IPU] add more loss ops (PaddlePaddle#44646) * add more loss ops * add authors Co-authored-by: Zhaorui Chen <zhaoruic@graphcore.ai> Co-authored-by: Zhaorui Chen <zhaoruic@graphcore.ai> * add g_ipuplace_pytype (PaddlePaddle#44648) * Strided slice fp16 (PaddlePaddle#44653) * [MLU]fix sync_batch_norm and concat_grad op (PaddlePaddle#44586) * retain dist op returns (PaddlePaddle#44634) * xpu unittest grad compute supports more types, *test=kunlun (PaddlePaddle#44606) * [Eager] Add hierarchical_sigmoid yaml (PaddlePaddle#44638) * add matrix_nms in python/paddle/vision/ops.py (PaddlePaddle#44357) * [auto parallel] bug fix for op has sub_block attr created with copy_from (PaddlePaddle#44664) * Change the way to set attributes for grad op maker (PaddlePaddle#44514) * fix typos in template for codegen of operators * change the way to set attributes for grad op maker * [XPU] add top_k op (PaddlePaddle#44656) * [XPU] add top_k op. test=kunlun * [XPU] add top_k op. test=kunlun * use PADDLE_ENFORCE_XDNN_NOT_NULL to check pointer. test=kunlun * Support broadcast tensor in phi system (PaddlePaddle#44590) * [PHI] Move spectral_norm to phi (PaddlePaddle#44577) * Add kernel declarations * Copy kernel implementation code * Transfer implementation code * Fix: Move out_grad to first * Register new kernels * Remove old kernels * Move out_grad to last * Fix bugs * Transfer infermeta * Add yaml files * Add blank line * Fix code style * Optimize directory structure Co-authored-by: Bobholamovic <linmanhui@baidu.com> * Complete the dtypes for all_gather, add all_gather_object api (PaddlePaddle#44417) * [Eager] refactor general_grad and fix some bugs (PaddlePaddle#44611) * refactor general_grad and fix some bugs * add TODO: support prune logic deeper * support log_grad op, *test=kunlun (PaddlePaddle#44662) * [LAUNCH] add distributed launch check tools (PaddlePaddle#44495) * add launch test * launch test for cpu * bs 1 * Move api(lgamma) from legacy_api.yaml to api.yaml (PaddlePaddle#44355) * Move api(lgamma) from legacy_api.yaml to api.yaml * Move api(lgamma) from legacy_api.yaml to api.yaml * Move api(lgamma) from legacy_api.yaml to api.yaml * modify code style * add x to X mapping * add definition of lgamma * delete redundant lgamma definitions * Modify code comments * Modify ops.py code format * add lgamma single test and lgamma api in fluid * Optimized lgamma unittest * Move frame kernel to phi (PaddlePaddle#44615) * Move frame OP to phi、add frame OP yaml config and supplement single test * add Header file of in_dygraph_mode * Modify variable name and FrameGradInferMeta multiplex UnchangedInferMeta * move seq2col to phi * delete elementwise pow in xpu_kp_list (PaddlePaddle#44661) * [MLU] fix log_softmax mode selection. (PaddlePaddle#44669) * adapt for resnet (PaddlePaddle#44685) * Fix some problem of kernel fallback in C++ API (PaddlePaddle#44681) * support auto fallback to cpu kernel for cusom device * fix some problem of kernel fallback * fix bugs of lstsq (PaddlePaddle#44689) * migrate dirichlet kernel to phi (PaddlePaddle#44434) * migrate dirichlet op kernel to phi * fix dirichlet sample memory leak * [phi]move softsign from fluid to phi (PaddlePaddle#44616) * test_activation_op unitest error, yaml & activation.py in_dygraph_mode incomplete * fix test_activation_op unitest error, add yaml and dygraph test * fix code style with pre-commit * try to fix namespace error of abs in activation_functor.h * fix namespace error of abs * [Paddle Inference] Support depthwise_conv2d fp16. (PaddlePaddle#44642) * depthwise_fp16 * depthwise_fp16 * depthwise_fp16 * depthwise_fp16 * fix logging debug level (PaddlePaddle#44684) * back fl * delete ssl cert * . * make warning * . * unittest paral degree * solve unittest * heter & multi cloud commm ready * . * . * fl-ps v1.0 * . * support N + N mode * . * . * . * . * delete print * . * . * . * . * fix bug * . * . * fl-ps with coordinator ready * merge dev * update message parse only * update fl client scheduler * fix bug * update multithreads sync * fix ci errors * update role_maker.py * update role_maker.py * fix ci error: windows py import error * fix ci error: windows py import error * fix windows ci pylib import error * add dump fields & params * try to fix windows import fleet error * fix ps FLAGS error * fix logging risk * fix logging possible risk * Skip CUDA Graph case for standalone executor (PaddlePaddle#44693) * [Eager] fix lerp grad kernel logic (PaddlePaddle#44705) * clone ort_predictor reuse session (PaddlePaddle#44703) * [XPU] add sampling_id op, add top_k op, update xdnn api. test=kunlun (PaddlePaddle#44704) * fused_fc_elementwise_layernorm_op support fp16 (PaddlePaddle#44710) * fused_fc_elementwise_layernorm support fp16 * fused_fc_elementwise_layernorm support double * [Phi] Add yaml for assign_value (PaddlePaddle#44596) * [Phi] Add yaml for assign_value * [Phi] Fix the bug of the assign api and modify the unittest * [Phi] Fix the bug when the tensor does not have the backend info * [Phi] Replace the functional-style cast init by the brace-init * [Phi] Cast the data explicitly * [PHI] Move lu to phi (PaddlePaddle#44605) * Add kernel declarations * Copy kernel implementation code * Transfer implementation code * Register new kernels * Remove old kernels * Fix code style * Fix bugs * mutable_data->HostAlloc * Transfer infermeta * Add yaml and update python api * Add PADDLE_WITH_HIP check * Update unittests * Fix bugs * Fix bugs * Optimize directory structure * Add output checks * lu_impl.h->lu_kernel_impl.h Co-authored-by: Bobholamovic <linmanhui@baidu.com> * [MLU] add pytest for mlu strided_slice kernel (PaddlePaddle#44523) * Support backward final hook (PaddlePaddle#44686) * update to sdk2.6.0 (PaddlePaddle#44673) * move CUDAStream to phi (PaddlePaddle#44529) * init * move CUDAStream to phi * fix compilation * merge develop * add stream_owned_ member * split cuda_stream.h * fix cpu compile * fix constructor * fix bug * fix windows compile * fix inference test_levit * fix windows tests * [Auto parallel] Optimization Tuning (PaddlePaddle#43782) * fixed bug for pass & engine * fixed bug for benchmark GPT-3 * add tuner & profiler * add algorithms & config * skip cast trt convert when input dtype is bool (PaddlePaddle#44716) * skip cast trt convert when input dtype is bool * [LAUNCH] fix set args bug (PaddlePaddle#44717) * Phi softplus migration (PaddlePaddle#44542) * add yaml and utests of phi softplus add yaml of softplus fix softplus bug in phi * update utests * bug fix * bug fix for test_layers * layer api match * match def and doc in ops.py * doc polish * fix unwanted modified of thresholded_relu * style imporve * 【PaddlePaddle Hackathon 3 No.15】为 Paddle 新增 count_nonzero (PaddlePaddle#44169) * add count_nonzero api * remove grad test * [WIP] Matmul v1 & v2 unification -- part 1 (PaddlePaddle#44640) * - Unit tests to be debugged - fix - refactor - diagnostic - more diagnostic - fix - Fix number two - fix - fix - fix - alpha added - more fixes - compilation fix - removed diagnostic code - cosmetic fixes * lint * add FLAGS_enable_api_kernel_fallback (PaddlePaddle#44706) * add FLAGS_enable_api_kernel_fallback * deal with more cases * add ut for coverage * phi_multiclass_nms3 (PaddlePaddle#44613) * add some fp16 op for kunlun resnet50 model (PaddlePaddle#44672) * add some fp16 op for kunlun resnet50 model *test=kunlun * tmp *test=kunlun * add dist op costs (PaddlePaddle#44701) * [API/OP] Migrate Lstsq op into phi (PaddlePaddle#44318) * migrate lstsq op * update * fix bugs for CIs * update * fix bugs * add uts * update * update * update * fix bugs of jip * fix bugs of hip * update * update according to review * update * update * update * update * Add sparse SyncBatchNorm (PaddlePaddle#43520) * add sparse SyncBatchNorm * unify fluid::CUDADeviceContext and phi::GpuContext (PaddlePaddle#44723) * remove cudaDeviceContext * remove more template * fix rocm compile * 【PaddlePaddle Hackathon 3 No.12】为 Paddle 新增 pairwise_distance (PaddlePaddle#44161) * add paddle.nn.functional.pairwise_distance (cattidea#273) * remove the test case for undefined behavior Co-authored-by: SigureMo <sigure.qaq@gmail.com> * Phi prior box (PaddlePaddle#44431) * phi_prior_box * add float[] support * phi_prior_box_optest * update * ort backend support output mutable data (PaddlePaddle#44724) * [PHI] Move lu_unpack to phi (PaddlePaddle#44674) * Add kernel declarations * Copy kernel implementation code * Transfer implementation code * Register new kernels * Remove old kernels * Fix code style * Fix bugs * mutable_data->HostAlloc * Transfer infermeta * Add yaml and update python api * Add PADDLE_WITH_HIP check * Update unittests * Add kernel declarations * Copy kernel implementation code * Transfer kernel implementation code * Register new kernels * Remove old kernels * Add lu_unpack_sig * Fix bugs * Fix bugs * Fix bugs * Optimize directory structure * Add output checks * Update include files * lu_impl.h->lu_kernel_impl.h * Transfer infermeta * Add yaml and update python api * Add check_eager Co-authored-by: Bobholamovic <linmanhui@baidu.com> * update document of quantile and nanquantile; test=document_fix (PaddlePaddle#42413) * migrate reduce_amin,reduce_amax kernel to phi (PaddlePaddle#44698) * [Paddle Inference] add varlen_token_prune plugin, pass, convert (PaddlePaddle#44733) * add varlen_token_prune plugin, pass, convert * support build with Ninja on Linux (PaddlePaddle#44210) * support ninja * fix mkldnn on windows * fix mkldnn on windows up1 * up2 * up3 * fix gflags * BUILD_BYPRODUCTS_OPTION -> BUILD_BYPRODUCTS_ARGS * use CMAKE_COMMAND * up x * migrate overlap_add and overlap_add_grad op (PaddlePaddle#44739) * update code format * add ymal and test * update for comments * Fix to CI (PaddlePaddle#44744) * - fix * - another fix * lint * infer context fix place error. (PaddlePaddle#44726) * infer context fix place error. * update * update * [operator migration] Migrate unstack_op and nms_op (PaddlePaddle#44424) * update unstack_op * update unstack_op * update unstack_op * fix unstack test * update unstack * update with remote * fix unstack_test.py * temp_save_change_nms_op * add nms test * update nms fix * update unstack_op * temp save change * finish fix nms_op * pass nms test * fix CI * fix ops test * save change * fix code style * fix code style * fix ci and codestyle * fix ci Co-authored-by: ShiningZhang <zhang_liang1991@126.com> * Update linalg.py (PaddlePaddle#44347) * Fix test and doc (PaddlePaddle#44735) * fix test and doc * fix all_gather_object with various length, test=allcases (PaddlePaddle#44718) * update manipulation.py paddle.moveaxis (PaddlePaddle#44191) * [CI] CI for Distributed (PaddlePaddle#44085) * generate_unify_header supports excludes (PaddlePaddle#44761) * [JitLayer]Polish PEFuntion to speed up JitLayer and fix memory leak (PaddlePaddle#44738) * Polish PEFuntion to speed up JitLayer * Polish PEFunction code * Fix comments * paddle2onnx update version to 1.0.0rc2 (PaddlePaddle#44759) * set parallel_job according to CUDA memory in Windows CI unittest (PaddlePaddle#44695) * set parallel_job according to CUDA memory * fix bug: add whitespace between conten and [] or condition wont work * [Sparse] optimize sparse attention (PaddlePaddle#44743) * GPUGraph merge to develop (PaddlePaddle#44594) Co-authored-by: seemingwang <zsasuke@qq.com> Co-authored-by: DesmonDay <908660116@qq.com> Co-authored-by: seemingwang <seemingwang@users.noreply.github.com> Co-authored-by: Thunderbrook <a754913769@163.com> Co-authored-by: xuewujiao <105861147+xuewujiao@users.noreply.github.com> Co-authored-by: root <root@yq01-sys-hic-k8s-v100-box-a225-0693.yq01.baidu.com> Co-authored-by: Thunderbrook <52529258+Thunderbrook@users.noreply.github.com> Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com> Co-authored-by: huwei02 <53012141+huwei02@users.noreply.github.com> Co-authored-by: yaoxuefeng <yaoxuefeng@baidu.com> Co-authored-by: lxsbupt <luoxsbupt@163.com> Co-authored-by: miaoli06 <106585574+miaoli06@users.noreply.github.com> Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0008.yq01.baidu.com> Co-authored-by: chao9527 <33347532+chao9527@users.noreply.github.com> Co-authored-by: qingshui <qshuihu@gmail.com> Co-authored-by: yangjunchao <yangjunchao@baidu.com> * Revert for cmake static library errors on XPU KP PaddlePaddle#44762 * unify gpu context (PaddlePaddle#44740) * remove cudaDeviceContext * remove more template * fix rocm compile * remove alias name CUDADeviceContext * fix compile * fix tests * revert changes * API doc(en) Bugs fix in 第四期体验评估 (PaddlePaddle#44749) * fix docs(en) bugs;test=document_fix * update paddle.add docs;test=document_fix * update paddle.where docs;test=document_fix * for ci;test=document_fix * Update manipulation.py * update paddle.where;test=document_fix Co-authored-by: Ligoml <39876205+Ligoml@users.noreply.github.com> * Modify the output result annotation under the lerp function (PaddlePaddle#44035) * Refactor build_op_downstream_map for standalone executor (PaddlePaddle#44729) * Refactor build_op_downstream_map for standalone executor * Add some comments * update xpu.cmake to 20220731, test=kunlun (PaddlePaddle#44767) * fix ut new_group_api (PaddlePaddle#44764) * support beam_search operator on xpu. test=kunlun (PaddlePaddle#44720) * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * [phi] add yolov3_loss yaml and unittest (PaddlePaddle#44476) * add yaml and unittest * update yaml * update backward yaml and unittest * update yaml * add Yolov3LossGradInferMeta * update yolov3_loss_op.cc * fix bug * code format * Update manipulation.py for rot90() (PaddlePaddle#44038) * fix compile error;test=develop * fix compile error;test=develop * fix compile;test=develop Co-authored-by: Sing_chan <51314274+betterpig@users.noreply.github.com> Co-authored-by: zlsh80826 <rewang@nvidia.com> Co-authored-by: Ruibiao Chen <chenruibiao@baidu.com> Co-authored-by: RichardWooSJTU <37864677+RichardWooSJTU@users.noreply.github.com> Co-authored-by: taixiurong <taixiurong@126.com> Co-authored-by: Allen Guo <alleng@graphcore.ai> Co-authored-by: Zhixin Yao <zhixiny@graphcore.ai> Co-authored-by: Zhaorui Chen <zhaoruic@graphcore.ai> Co-authored-by: zhangxiaoci <zhangxiaoci@baidu.com> Co-authored-by: zyfncg <zhangyunfei07@baidu.com> Co-authored-by: zhangkaihuo <zhangkaihuo@baidu.com> Co-authored-by: wanghuancoder <wanghuan29@baidu.com> Co-authored-by: xiongkun <xiongkun03@baidu.com> Co-authored-by: Aurelius84 <zhangliujie@baidu.com> Co-authored-by: Leo Chen <chenqiuliang@baidu.com> Co-authored-by: Weilong Wu <veyron_wu@163.com> Co-authored-by: caozhou <48191911+Caozhou1995@users.noreply.github.com> Co-authored-by: ronnywang <ronny1996@163.com> Co-authored-by: zhoutianzi666 <39978853+zhoutianzi666@users.noreply.github.com> Co-authored-by: Haohongxiang <86215757+haohongxiang@users.noreply.github.com> Co-authored-by: WangZhen <23097963+0x45f@users.noreply.github.com> Co-authored-by: Wilber <jiweibo@baidu.com> Co-authored-by: ShenLiang <1422485404@qq.com> Co-authored-by: QingshuChen <chenqingshu@baidu.com> Co-authored-by: levi131 <83750468+levi131@users.noreply.github.com> Co-authored-by: Qi Li <qili93@qq.com> Co-authored-by: 王明冬 <78149749+winter-wang@users.noreply.github.com> Co-authored-by: Feiyu Chan <chenfeiyu@baidu.com> Co-authored-by: Xiaoxu Chen <chenxx_id@163.com> Co-authored-by: Chenxiao Niu <ncxinhanzhong@gmail.com> Co-authored-by: Zhou Wei <1183042833@qq.com> Co-authored-by: JYChen <zoooo0820@qq.com> Co-authored-by: YUNSHEN XIE <1084314248@qq.com> Co-authored-by: niuliling123 <51102941+niuliling123@users.noreply.github.com> Co-authored-by: zhangyikun02 <48021248+zhangyk0314@users.noreply.github.com> Co-authored-by: huzhiqiang <912790387@qq.com> Co-authored-by: jakpiase <jakpia21@gmail.com> Co-authored-by: Piotr Paturej <piotr.paturej@intel.com> Co-authored-by: zhaocaibei123 <48509226+zhaocaibei123@users.noreply.github.com> Co-authored-by: freeliuzc <lzc842650834@gmail.com> Co-authored-by: tianshuo78520a <707759223@qq.com> Co-authored-by: zmxdream <zhangminxu01@baidu.com> Co-authored-by: houj04 <35131887+houj04@users.noreply.github.com> Co-authored-by: pangyoki <pangyoki@126.com> Co-authored-by: lyq <30404405+affectionlu@users.noreply.github.com> Co-authored-by: Zhong Hui <zhonghui.net@gmail.com> Co-authored-by: fuyou765 <64373205+fuyou765@users.noreply.github.com> Co-authored-by: Chen Weihang <chenweihang@baidu.com> Co-authored-by: YuanRisheng <yuanrisheng@baidu.com> Co-authored-by: zhaoyingli <86812880+zhaoyinglia@users.noreply.github.com> Co-authored-by: ccrrong <101700995+ccrrong@users.noreply.github.com> Co-authored-by: xiaoxiaohehe001 <49090790+xiaoxiaohehe001@users.noreply.github.com> Co-authored-by: ykkk2333 <77383312+ykkk2333@users.noreply.github.com> Co-authored-by: Li Min <11663212+limin2021@users.noreply.github.com> Co-authored-by: Hui Zhang <zhtclz@foxmail.com> Co-authored-by: ming1753 <61511741+ming1753@users.noreply.github.com> Co-authored-by: cifar10 <41565156+cifar10@users.noreply.github.com> Co-authored-by: fwenguang <95677191+fwenguang@users.noreply.github.com> Co-authored-by: Aganlengzi <aganlengzi@gmail.com> Co-authored-by: yuguo <948529990@qq.com> Co-authored-by: Zhang Jun <ewalker@live.cn> Co-authored-by: Wang Bojun <105858416+wwbitejotunn@users.noreply.github.com> Co-authored-by: yangguohao <70266361+yangguohao@users.noreply.github.com> Co-authored-by: Ligoml <39876205+Ligoml@users.noreply.github.com> Co-authored-by: Lux et Veritas <1004239791@qq.com> Co-authored-by: zhangbo9674 <82555433+zhangbo9674@users.noreply.github.com> Co-authored-by: BiynXu <62832681+BiynXu@users.noreply.github.com> Co-authored-by: ziyoujiyi <73728031+ziyoujiyi@users.noreply.github.com> Co-authored-by: Zhen Wang <wangzhen31@baidu.com> Co-authored-by: chenjian <chenjian26@baidu.com> Co-authored-by: helen88 <z8hanghuan@126.com> Co-authored-by: Yuang Liu <liuyuang@baidu.com> Co-authored-by: qipengh <huangqipeng@cambricon.com> Co-authored-by: shangliang Xu <ghostxsl@users.noreply.github.com> Co-authored-by: Jiabin Yang <360788950@qq.com> Co-authored-by: Lin Manhui <mhlin425@whu.edu.cn> Co-authored-by: Bobholamovic <linmanhui@baidu.com> Co-authored-by: LiYuRio <63526175+LiYuRio@users.noreply.github.com> Co-authored-by: kuizhiqing <kuizhiqing@baidu.com> Co-authored-by: Charles-hit <56987902+Charles-hit@users.noreply.github.com> Co-authored-by: HongyuJia <jiahongyu@baidu.com> Co-authored-by: heliqi <1101791222@qq.com> Co-authored-by: Yulong Ao <aoyulong@baidu.com> Co-authored-by: JZ-LIANG <jianzhongliang10@gmail.com> Co-authored-by: thunder95 <290844930@qq.com> Co-authored-by: Jacek Czaja <jacek.czaja@intel.com> Co-authored-by: zhiboniu <31800336+zhiboniu@users.noreply.github.com> Co-authored-by: Ainavo <57820731+Ainavo@users.noreply.github.com> Co-authored-by: SigureMo <sigure.qaq@gmail.com> Co-authored-by: Asthestarsfalll <72954905+Asthestarsfalll@users.noreply.github.com> Co-authored-by: Wangzheee <634486483@qq.com> Co-authored-by: Thomas Young <35565423+HexToString@users.noreply.github.com> Co-authored-by: ShiningZhang <zhang_liang1991@126.com> Co-authored-by: OccupyMars2025 <31559413+OccupyMars2025@users.noreply.github.com> Co-authored-by: mrcangye <mrcangye@email.cn> Co-authored-by: Roc <30228238+sljlp@users.noreply.github.com> Co-authored-by: seemingwang <zsasuke@qq.com> Co-authored-by: DesmonDay <908660116@qq.com> Co-authored-by: seemingwang <seemingwang@users.noreply.github.com> Co-authored-by: Thunderbrook <a754913769@163.com> Co-authored-by: xuewujiao <105861147+xuewujiao@users.noreply.github.com> Co-authored-by: root <root@yq01-sys-hic-k8s-v100-box-a225-0693.yq01.baidu.com> Co-authored-by: Thunderbrook <52529258+Thunderbrook@users.noreply.github.com> Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com> Co-authored-by: huwei02 <53012141+huwei02@users.noreply.github.com> Co-authored-by: yaoxuefeng <yaoxuefeng@baidu.com> Co-authored-by: lxsbupt <luoxsbupt@163.com> Co-authored-by: miaoli06 <106585574+miaoli06@users.noreply.github.com> Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0008.yq01.baidu.com> Co-authored-by: chao9527 <33347532+chao9527@users.noreply.github.com> Co-authored-by: qingshui <qshuihu@gmail.com> Co-authored-by: yangjunchao <yangjunchao@baidu.com> Co-authored-by: yang131313 <lisy928472889@163.com> Co-authored-by: mengqingchun02 <103740521+mengqingchun02@users.noreply.github.com> Co-authored-by: 熊峻峰 <xiongjunfeng@sina.com>
Configuration menu - View commit details
-
Copy full SHA for 0610216 - Browse repository at this point
Copy the full SHA 0610216View commit details -
【GPUPS】fix load model in pscore mode (xuewujiao#78)
* fix load model;test=develop * fix load model;test=develop * fix load model;test=develop
Configuration menu - View commit details
-
Copy full SHA for a59566c - Browse repository at this point
Copy the full SHA a59566cView commit details
Commits on Aug 8, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 4782a8f - Browse repository at this point
Copy the full SHA 4782a8fView commit details
Commits on Aug 16, 2022
-
change load node and edge from local to cpu (xuewujiao#83)
* change load node and edge * remove useless code Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for c946cf3 - Browse repository at this point
Copy the full SHA c946cf3View commit details
Commits on Aug 18, 2022
-
extract pull sparse as single stage(xuewujiao#85)
Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 2bb8eb0 - Browse repository at this point
Copy the full SHA 2bb8eb0View commit details -
sync gpugraph to gpugraph_v2 (xuewujiao#86)
* change load node and edge from local to cpu (xuewujiao#83) * change load node and edge * remove useless code Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com> * extract pull sparse as single stage(xuewujiao#85) Co-authored-by: yangjunchao <yangjunchao@baidu.com> Co-authored-by: miaoli06 <106585574+miaoli06@users.noreply.github.com> Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com> Co-authored-by: chao9527 <33347532+chao9527@users.noreply.github.com> Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for a26eaa5 - Browse repository at this point
Copy the full SHA a26eaa5View commit details -
Configuration menu - View commit details
-
Copy full SHA for f78b197 - Browse repository at this point
Copy the full SHA f78b197View commit details
Commits on Aug 19, 2022
-
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedAug 19, 2022 Configuration menu - View commit details
-
Copy full SHA for 375410c - Browse repository at this point
Copy the full SHA 375410cView commit details
Commits on Aug 22, 2022
-
[GPUGraph] graph sample v2 (xuewujiao#87)
* change load node and edge from local to cpu (xuewujiao#83) * change load node and edge * remove useless code Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com> * extract pull sparse as single stage(xuewujiao#85) Co-authored-by: yangjunchao <yangjunchao@baidu.com> * support ssdsparsetable;test=develop (xuewujiao#81) * graph sample v2 * remove log Co-authored-by: miaoli06 <106585574+miaoli06@users.noreply.github.com> Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com> Co-authored-by: chao9527 <33347532+chao9527@users.noreply.github.com> Co-authored-by: yangjunchao <yangjunchao@baidu.com> Co-authored-by: danleifeng <52735331+danleifeng@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 5529e61 - Browse repository at this point
Copy the full SHA 5529e61View commit details
Commits on Aug 23, 2022
-
root committed
Aug 23, 2022 Configuration menu - View commit details
-
Copy full SHA for 6ef6d45 - Browse repository at this point
Copy the full SHA 6ef6d45View commit details -
Configuration menu - View commit details
-
Copy full SHA for a087a6d - Browse repository at this point
Copy the full SHA a087a6dView commit details
Commits on Aug 25, 2022
-
compatible whole HBM mode (xuewujiao#91)
Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 1372dc6 - Browse repository at this point
Copy the full SHA 1372dc6View commit details
Commits on Aug 29, 2022
-
Add graphsage data process (xuewujiao#94)
* Add graphsage withoud self-loop * change next_num_nodes line * temp commit * move id tensor position * change vlog num * merge gpugraph, add graphsage optimization, refactor reindex, add graphsage infer * add review * add graphsage optimize, add geometric * update fill_dvalue * add geometric * v2 version * v2 version * delete unused code * update fill_dvalue
Configuration menu - View commit details
-
Copy full SHA for d56e8b7 - Browse repository at this point
Copy the full SHA d56e8b7View commit details -
* compatible whole HBM mode * unify flag for graph emd storage mode and graph struct storage mode * format Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 1ccd5e0 - Browse repository at this point
Copy the full SHA 1ccd5e0View commit details -
split generate batch into multi stage (xuewujiao#92)
* split generate batch into multi stage * fix conflict Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 0e04d15 - Browse repository at this point
Copy the full SHA 0e04d15View commit details -
[GpuGraph] Uniq feature (xuewujiao#95)
* uniq feature * uniq feature * uniq feature
Configuration menu - View commit details
-
Copy full SHA for 6a8ea82 - Browse repository at this point
Copy the full SHA 6a8ea82View commit details
Commits on Aug 31, 2022
-
[GpuGraph] global startid (xuewujiao#98)
* uniq feature * uniq feature * uniq feature * global startid
Configuration menu - View commit details
-
Copy full SHA for 591978f - Browse repository at this point
Copy the full SHA 591978fView commit details
Commits on Sep 2, 2022
-
load node edge seperately and release graph (xuewujiao#99)
* load node edge seperately and release graph * load node edge seperately and release graph Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 58a3ae2 - Browse repository at this point
Copy the full SHA 58a3ae2View commit details
Commits on Sep 5, 2022
-
Optimize graphsage data process (xuewujiao#96)
* add sample update, update reindex * temp commit * fix offset calc * optimize reindex * delete unnecessary code * add for loop speed up * add speed optimize * change common_graph VLOG type * add kernel3 update for review * temp commit * delete unused code
Configuration menu - View commit details
-
Copy full SHA for fe1653c - Browse repository at this point
Copy the full SHA fe1653cView commit details -
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedSep 5, 2022 Configuration menu - View commit details
-
Copy full SHA for 5f6460a - Browse repository at this point
Copy the full SHA 5f6460aView commit details
Commits on Sep 6, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 854959e - Browse repository at this point
Copy the full SHA 854959eView commit details
Commits on Sep 8, 2022
-
optimize begin pass and end pass (xuewujiao#106)
Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 2bc6bf7 - Browse repository at this point
Copy the full SHA 2bc6bf7View commit details
Commits on Sep 9, 2022
-
Configuration menu - View commit details
-
Copy full SHA for b8b0d26 - Browse repository at this point
Copy the full SHA b8b0d26View commit details -
[GPUGraph] fix FillOneStep args (xuewujiao#107)
* fix ins no * fix FillOnestep args
Configuration menu - View commit details
-
Copy full SHA for 582d236 - Browse repository at this point
Copy the full SHA 582d236View commit details -
Optimize graphsage data process by merging neighbor sampling
* add sample update, update reindex * temp commit * fix offset calc * optimize reindex * delete unnecessary code * add for loop speed up * add speed optimize * change common_graph VLOG type * add kernel3 update for review * temp commit * temp commit * fix illegal memory access * fix illegal memory access * add success run merge sample kernel * delete unused code * delete unused code * delete unused code * add merge sample optimize * add neighbor_sample_kernel3 * fix move_result bug * add comment, delete unused code * reset code * fix single-direction bug
Configuration menu - View commit details
-
Copy full SHA for 9441c97 - Browse repository at this point
Copy the full SHA 9441c97View commit details
Commits on Sep 13, 2022
-
fix bug for whole hbm mode (xuewujiao#110)
Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for cc71f56 - Browse repository at this point
Copy the full SHA cc71f56View commit details
Commits on Sep 14, 2022
-
[GPUGraph] fix infer && add infer_table_cap (xuewujiao#108)
* fix ins no * fix FillOnestep args * fix infer && add infer table cap * fix infer
Configuration menu - View commit details
-
Copy full SHA for 1462c54 - Browse repository at this point
Copy the full SHA 1462c54View commit details -
Configuration menu - View commit details
-
Copy full SHA for 55739cf - Browse repository at this point
Copy the full SHA 55739cfView commit details
Commits on Sep 15, 2022
-
【PSCORE】perform ssd sparse table (xuewujiao#111)
* perform ssd sparsetable;test=develop Conflicts: paddle/fluid/framework/fleet/ps_gpu_wrapper.cc * perform ssd sparsetable;test=develop * remove debug code; * remove debug code; * add jemalloc cmake;test=develop * fix wrapper;test=develop
Configuration menu - View commit details
-
Copy full SHA for aed3925 - Browse repository at this point
Copy the full SHA aed3925View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2d61dde - Browse repository at this point
Copy the full SHA 2d61ddeView commit details
Commits on Sep 16, 2022
-
[GpuGraph] optimize shuffle batch (xuewujiao#115)
* fix sample core * optimize shuffle batch
Configuration menu - View commit details
-
Copy full SHA for eef65cd - Browse repository at this point
Copy the full SHA eef65cdView commit details -
release gpu mem when sample end (xuewujiao#116)
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 057885a - Browse repository at this point
Copy the full SHA 057885aView commit details -
fix class not found err (xuewujiao#118)
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 7d65319 - Browse repository at this point
Copy the full SHA 7d65319View commit details -
optimize sample (xuewujiao#117)
* optimize sample * optimize sample Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 04110a5 - Browse repository at this point
Copy the full SHA 04110a5View commit details
Commits on Sep 17, 2022
-
fix clear gpu mem (xuewujiao#119)
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 357a451 - Browse repository at this point
Copy the full SHA 357a451View commit details
Commits on Sep 20, 2022
-
fix sample core (xuewujiao#121)
Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 9f9c0ab - Browse repository at this point
Copy the full SHA 9f9c0abView commit details -
Configuration menu - View commit details
-
Copy full SHA for c59c59e - Browse repository at this point
Copy the full SHA c59c59eView commit details -
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedSep 20, 2022 Configuration menu - View commit details
-
Copy full SHA for 54a04ac - Browse repository at this point
Copy the full SHA 54a04acView commit details
Commits on Sep 23, 2022
-
Update graphsage speed(thrust -> cub), fix sample async bug (xuewujia…
…o#120) * fix deepwalk sample kernel, fix inplace kernel sync bug * update v2 sample * change kernel name for readability * delete unused kernel
Configuration menu - View commit details
-
Copy full SHA for 950067b - Browse repository at this point
Copy the full SHA 950067bView commit details -
support slot_feature with different length (xuewujiao#124)
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for bb5a184 - Browse repository at this point
Copy the full SHA bb5a184View commit details
Commits on Sep 26, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 99d3a96 - Browse repository at this point
Copy the full SHA 99d3a96View commit details -
【graphsage】don't alloc a new d_feature_buf if the old one is enough (x…
…uewujiao#128) * add graphsage slot feature * update graphsage slot feature * update graphsage slot feature
Configuration menu - View commit details
-
Copy full SHA for e49641a - Browse repository at this point
Copy the full SHA e49641aView commit details
Commits on Sep 27, 2022
-
* add ssd cache;test=develop * add ssd cache;test=develop * add ssd cache;test=develop
Configuration menu - View commit details
-
Copy full SHA for bdd64ef - Browse repository at this point
Copy the full SHA bdd64efView commit details
Commits on Sep 28, 2022
-
add multi epoch train & fix train table change ins & save infer embed…
…ing (xuewujiao#129) * add multi epoch train & fix train table change ins & save infer embedding * change epoch finish judge * change epoch finish change Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for c6a07b2 - Browse repository at this point
Copy the full SHA c6a07b2View commit details
Commits on Oct 9, 2022
-
* Add debug log * Add debug log Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0008.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 4380355 - Browse repository at this point
Copy the full SHA 4380355View commit details
Commits on Oct 10, 2022
-
Configuration menu - View commit details
-
Copy full SHA for baf678f - Browse repository at this point
Copy the full SHA baf678fView commit details -
fix bug in slot_feature (xuewujiao#134)
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-ab2-0009.yq01.baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 7899bd6 - Browse repository at this point
Copy the full SHA 7899bd6View commit details
Commits on Oct 18, 2022
-
[GpuGraph] cherry pick var slot feature && fix load multi path node (x…
…uewujiao#136) * optimize mem in uniq slot feature * cherry-pick var slot_feature Co-authored-by: huwei02 <53012141+huwei02@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 4e8b290 - Browse repository at this point
Copy the full SHA 4e8b290View commit details -
[GpuGraph] fix kernel overflow (xuewujiao#138)
* optimize mem in uniq slot feature * cherry-pick var slot_feature * fix kernel overflow && add max feature num flag Co-authored-by: huwei02 <53012141+huwei02@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for dde1573 - Browse repository at this point
Copy the full SHA dde1573View commit details
Commits on Oct 19, 2022
-
Configuration menu - View commit details
-
Copy full SHA for 3fd553a - Browse repository at this point
Copy the full SHA 3fd553aView commit details
Commits on Oct 25, 2022
-
slot feature secondary storage (xuewujiao#140)
* slot feature secondary storage * slot feature secondary storage Co-authored-by: yangjunchao <yangjunchao@baidu.com>
Configuration menu - View commit details
-
Copy full SHA for 9c96863 - Browse repository at this point
Copy the full SHA 9c96863View commit details
Commits on Nov 5, 2022
-
Merge remote-tracking branch 'upstream/gpugraph' into gpugraph.0621
root committedNov 5, 2022 Configuration menu - View commit details
-
Copy full SHA for b6556e9 - Browse repository at this point
Copy the full SHA b6556e9View commit details
Commits on Nov 7, 2022
-
root committed
Nov 7, 2022 Configuration menu - View commit details
-
Copy full SHA for 5cc8f0c - Browse repository at this point
Copy the full SHA 5cc8f0cView commit details
Commits on Dec 7, 2022
-
Configuration menu - View commit details
-
Copy full SHA for ecec983 - Browse repository at this point
Copy the full SHA ecec983View commit details -
Merge branch 'gpugraph.0621' of https://github.com/lxsbupt/Paddle int…
…o gpugraph.0621
Configuration menu - View commit details
-
Copy full SHA for 70eb479 - Browse repository at this point
Copy the full SHA 70eb479View commit details