|
17 | 17 |
|
18 | 18 | package org.apache.spark.sql.execution |
19 | 19 |
|
| 20 | +import org.apache.spark.sql.catalyst.util.DateTimeTestUtils._ |
20 | 21 | import org.apache.spark.sql.connector.InMemoryTableCatalog |
| 22 | +import org.apache.spark.sql.internal.SQLConf |
21 | 23 | import org.apache.spark.sql.test.{ExamplePoint, ExamplePointUDT, SharedSparkSession} |
22 | 24 |
|
23 | 25 | class HiveResultSuite extends SharedSparkSession { |
24 | 26 | import testImplicits._ |
25 | 27 |
|
| 28 | + private def withOutstandingZoneIds(f: => Unit): Unit = { |
| 29 | + for { |
| 30 | + jvmZoneId <- outstandingZoneIds |
| 31 | + sessionZoneId <- outstandingZoneIds |
| 32 | + } { |
| 33 | + withDefaultTimeZone(jvmZoneId) { |
| 34 | + withSQLConf(SQLConf.SESSION_LOCAL_TIMEZONE.key -> sessionZoneId.getId) { |
| 35 | + f |
| 36 | + } |
| 37 | + } |
| 38 | + } |
| 39 | + } |
| 40 | + |
26 | 41 | test("date formatting in hive result") { |
27 | | - val dates = Seq("2018-12-28", "1582-10-03", "1582-10-04", "1582-10-15") |
28 | | - val df = dates.toDF("a").selectExpr("cast(a as date) as b") |
29 | | - val result = HiveResult.hiveResultString(df) |
30 | | - assert(result == dates) |
31 | | - val df2 = df.selectExpr("array(b)") |
32 | | - val result2 = HiveResult.hiveResultString(df2) |
33 | | - assert(result2 == dates.map(x => s"[$x]")) |
| 42 | + withOutstandingZoneIds { |
| 43 | + val dates = Seq("2018-12-28", "1582-10-03", "1582-10-04", "1582-10-15") |
| 44 | + val df = dates.toDF("a").selectExpr("cast(a as date) as b") |
| 45 | + val result = HiveResult.hiveResultString(df) |
| 46 | + assert(result == dates) |
| 47 | + val df2 = df.selectExpr("array(b)") |
| 48 | + val result2 = HiveResult.hiveResultString(df2) |
| 49 | + assert(result2 == dates.map(x => s"[$x]")) |
| 50 | + } |
34 | 51 | } |
35 | 52 |
|
36 | 53 | test("timestamp formatting in hive result") { |
37 | | - val timestamps = Seq( |
38 | | - "2018-12-28 01:02:03", |
39 | | - "1582-10-03 01:02:03", |
40 | | - "1582-10-04 01:02:03", |
41 | | - "1582-10-15 01:02:03") |
42 | | - val df = timestamps.toDF("a").selectExpr("cast(a as timestamp) as b") |
43 | | - val result = HiveResult.hiveResultString(df) |
44 | | - assert(result == timestamps) |
45 | | - val df2 = df.selectExpr("array(b)") |
46 | | - val result2 = HiveResult.hiveResultString(df2) |
47 | | - assert(result2 == timestamps.map(x => s"[$x]")) |
| 54 | + withOutstandingZoneIds { |
| 55 | + val timestamps = Seq( |
| 56 | + "2018-12-28 01:02:03", |
| 57 | + "1582-10-03 01:02:03", |
| 58 | + "1582-10-04 01:02:03", |
| 59 | + "1582-10-15 01:02:03") |
| 60 | + val df = timestamps.toDF("a").selectExpr("cast(a as timestamp) as b") |
| 61 | + val result = HiveResult.hiveResultString(df) |
| 62 | + assert(result == timestamps) |
| 63 | + val df2 = df.selectExpr("array(b)") |
| 64 | + val result2 = HiveResult.hiveResultString(df2) |
| 65 | + assert(result2 == timestamps.map(x => s"[$x]")) |
| 66 | + } |
48 | 67 | } |
49 | 68 |
|
50 | 69 | test("toHiveString correctly handles UDTs") { |
|
0 commit comments