-
Notifications
You must be signed in to change notification settings - Fork 237
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] failures parquet_test test:read_round_trip #3195
Comments
tgravescs
added
bug
Something isn't working
? - Needs Triage
Need team to review and classify
labels
Aug 11, 2021
pxLi
changed the title
[BUG] Dataproc build failures parquet_test test:read_round_trip
[BUG] build failures parquet_test test:read_round_trip
Aug 12, 2021
this failed in nightly tests w/ spark standalone 3.1.1 and 3.1.2
|
pxLi
changed the title
[BUG] build failures parquet_test test:read_round_trip
[BUG] failures parquet_test test:read_round_trip
Aug 12, 2021
12 test failures
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
A bunch of these failures:
08:01:21 FAILED integration_tests/src/main/python/parquet_test.py::test_read_round_trip[parquet-reader_confs2-read_parquet_sql-[Byte, Short, Integer, Long, Float, Double, String, Boolean, Date, Timestamp, Array(Byte), Array(Long), Array(String), Array(Date), Array(Timestamp), Array(Decimal(18,0)), Array(Array(Byte)), Struct(['child0', Array(Byte)],['child1', Byte],['child2', Float],['child3', Decimal(18,0)]), Array(Struct(['child0', String],['child1', Double],['child2', Integer])), Map(Boolean(not_null),Boolean), Map(Byte(not_null),Byte), Map(Short(not_null),Short), Map(Integer(not_null),Integer), Map(Long(not_null),Long), Map(Float(not_null),Float), Map(Double(not_null),Double), Map(Date(not_null),Date), Map(Timestamp(not_null),Timestamp), Map(String(not_null),String), Map(String(not_null),Array(String)), Map(RepeatSeq(not_null)(Integer(not_null)),Long), Map(String(not_null),Map(String(not_null),String)), Decimal(18,0), Decimal(7,3), Decimal(10,10), Decimal(9,0), Decimal(18,15)]]
08:01:21 �[1m�[31mE Caused by: org.apache.spark.SparkUpgradeException: You may get a different result due to the upgrading of Spark 3.0: writing dates before 1582-10-15 or timestamps before 1900-01-01T00:00:00Z into Parquet INT96 files can be dangerous, as the files may be read by Spark 2.x or legacy versions of Hive later, which uses a legacy hybrid calendar that is different from Spark 3.0+'s Proleptic Gregorian calendar. See more details in SPARK-31404. You can set spark.sql.legacy.parquet.int96RebaseModeInWrite to 'LEGACY' to rebase the datetime values w.r.t. the calendar difference during writing, to get maximum interoperability. Or set spark.sql.legacy.parquet.int96RebaseModeInWrite to 'CORRECTED' to write the datetime values as it is, if you are 100% sure that the written files will only be read by Spark 3.0+ or other systems that use Proleptic Gregorian calendar.�[0m
The text was updated successfully, but these errors were encountered: