-
Notifications
You must be signed in to change notification settings - Fork 244
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
adding unique indices test for batch write #1014
Merged
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
zhexuany
force-pushed
the
adding_more_test
branch
3 times, most recently
from
August 13, 2019 10:21
7c8c72a
to
bfb3e6c
Compare
/run-all-tests mode=full tidb=v3.0.2 |
zhexuany
force-pushed
the
adding_more_test
branch
4 times, most recently
from
August 16, 2019 03:25
dd17633
to
2e9094e
Compare
/run-all-tests mode=full tidb=v3.0.2 |
marsishandsome
approved these changes
Aug 16, 2019
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
others LGTM
core/src/test/scala/com/pingcap/tispark/datasource/BaseDataSourceTest.scala
Outdated
Show resolved
Hide resolved
core/src/test/scala/com/pingcap/tispark/datasource/BaseDataSourceTest.scala
Outdated
Show resolved
Hide resolved
core/src/test/scala/org/apache/spark/sql/test/generator/Schema.scala
Outdated
Show resolved
Hide resolved
zhexuany
force-pushed
the
adding_more_test
branch
from
August 16, 2019 07:40
2e9094e
to
7f95d2c
Compare
zhexuany
force-pushed
the
adding_more_test
branch
from
August 16, 2019 09:40
a754d7f
to
6d2ec0d
Compare
/run-all-tests mode=full tidb=v3.0.2 |
marsishandsome
added a commit
that referenced
this pull request
Aug 29, 2019
* 2.2.0 -> 2.3.0 (#947) * Add tests for primary key (#948) * add changelog (#955) * add multi-column tests (#954) * fix range partition throw UnsupportedSyntaxException error (#960) * fix view parsing problem (#953) * make tispark can read from a hash partition table (#966) * increase ci worker number (#965) * update readme for tispark-2.1.2 release (#968) * update document for pyspark (#975) * fix one jar bug (#972) * adding common port number used by spark cluster (#973) * fix cost model in table scan (#977) * create an UninitializedType for TypeDecimal (#979) * update sparkr doc (#976) * use spark-2.4.3 to run ut (#978) * use spark-2.4.3 to run ut * fix ci * a better design for get auto table id (#980) * fix bug: ci SpecialTiDBTypeTestSuite failed with tidb-3.0.1 (#984) * improve TiConfiguration getPdAddrsString function (#963) * bump grpc to 1.17 (#982) * Add multiple-column PK tests (#970) * add retry for batchGet (#986) * use tispark self-made m2 cahce file (#990) * add spark sql document for batch write (#991) * add auto mode for test.data.load (#994) * fix typo (#996) * fix index scan bug (#995) * refine doc (#1003) * add tidb-3.0 compatibility document (#998) * add tidb-3.0 compatibility document * address code review * address code review * add log4j config document (#1008) * refactor batch write region pre-split (#999) * add ci simple mode (#1012) * clean up redundant code (#997) * prohibit agg or groupby pushdown on double read (#1004) * remove split region code (#1015) * add supported scala version (#1013) * Fix scala compiler version (#1010) * fix reflection bug for hdp release (#1017) (#1018) (cherry picked from commit 118b12e) * check by grammarly (#1022) * add benchmark result for batch write (#1025) * release tispark 2.1.3 (#1026) (#1035) (cherry picked from commit 107eb2b) * support setting random seed in daily regression test (#1032) * Remove create in tisession (#1021) * set tikv region size from 96M to 1M (#1031) * adding unique indices test for batch write (#1014) * use one unique seed (#1043) * remove unused code (#1030) * adding batch write pk insertion test (#1044) * fix table not found bug in TiSession because of synchronization (#1041) * fix test failure (#1051) * fix reflection bug: pass in different arguments for different version of same function (#1037) (#1052) (cherry picked from commit a5462c2) * Adding pk and unique index test for batch write (#1049) * fix distinct without alias bug: disable pushdown aggregate with alias (#1054) * improve the doc (#1053) * Refactor RegionStoreClient logic (#989) * using stream rather removeIf (#1057) * Remove redundant pre-write/commit logic in LockResolverTest (#1062) * adding recreate flag when create tisession (#1064) * fix issue 1047 (#1066) * cleanup code in TiBatchWrite (#1067) * release tispark-2.1.4 (#1068) (#1069) (cherry picked from commit fd8068a) * update document for tispark-2.1.4 release (#1070)
wfxxh
pushed a commit
to wanfangdata/tispark
that referenced
this pull request
Jun 30, 2023
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
What problem does this PR solve?
Improve the coverage of batch-write on unique indices.
What is changed and how it works?
In this PR, we are trying to improve the coverage of batch-write on unique indices.
In order to test the correctness of batch-write, we have to create different schemas with different numbers of the index on the schema. Additionally, we have to prepare data and use batch-write api to perform the insertion.
After discussion with @birdstorm, we decided we should generate the schema, so does the data.
To generate different schemas, we create an array of
dataTypes
. Our schema generator enumerates over thedataTypes
and chooses some of them to become unique indices.Once we have schema, we generate data randomly.
With schema and data, it is ready to test our batch-write api.
Check List
Tests