@@ -91,7 +91,8 @@ class OrcQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterAll {
9191 val tempDir = getTempFilePath(" orcTest" ).getCanonicalPath
9292 val range = (0 to 255 )
9393 val data = sparkContext.parallelize(range)
94- .map(x => AllDataTypes (s " $x" , x, x.toLong, x.toFloat, x.toDouble, x.toShort, x.toByte, x % 2 == 0 ))
94+ .map(x =>
95+ AllDataTypes (s " $x" , x, x.toLong, x.toFloat,x.toDouble, x.toShort, x.toByte, x % 2 == 0 ))
9596 data.toDF().saveAsOrcFile(tempDir)
9697 checkAnswer(
9798 TestHive .orcFile(tempDir),
@@ -101,7 +102,8 @@ class OrcQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterAll {
101102
102103 test(" read/write binary data" ) {
103104 val tempDir = getTempFilePath(" orcTest" ).getCanonicalPath
104- sparkContext.parallelize(BinaryData (" test" .getBytes(" utf8" )) :: Nil ).toDF().saveAsOrcFile(tempDir)
105+ sparkContext.parallelize(BinaryData (" test" .getBytes(" utf8" )) :: Nil )
106+ .toDF().saveAsOrcFile(tempDir)
105107 TestHive .orcFile(tempDir)
106108 .map(r => new String (r(0 ).asInstanceOf [Array [Byte ]], " utf8" ))
107109 .collect().toSeq == Seq (" test" )
@@ -136,7 +138,8 @@ class OrcQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterAll {
136138 rdd.foreach {
137139 // '===' does not like string comparison?
138140 row : Row => {
139- assert(row.getString(1 ).equals(s " val_ $counter" ), s " row $counter value ${row.getString(1 )} does not match val_ $counter" )
141+ assert(row.getString(1 ).equals(s " val_ $counter" ),
142+ s " row $counter value ${row.getString(1 )} does not match val_ $counter" )
140143 counter = counter + 1
141144 }
142145 }
@@ -173,7 +176,7 @@ class OrcQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterAll {
173176
174177 // We only support zlib in hive0.12.0 now
175178 test(" Default Compression options for writing to an Orcfile" ) {
176- // TODO: support other compress codec
179+ // TODO: support other compress codec
177180 var tempDir = getTempFilePath(" orcTest" ).getCanonicalPath
178181 val rdd = sparkContext.parallelize((1 to 100 ))
179182 .map(i => TestRDDEntry (i, s " val_ $i" ))
@@ -184,7 +187,7 @@ class OrcQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterAll {
184187 }
185188
186189 // Following codec is supported in hive-0.13.1, ignore it now
187- ignore(" Other Compression options for writing to an Orcfile only supported in hive 0.13.1 and above" ) {
190+ ignore(" Other Compression options for writing to an Orcfile - 0.13.1 and above" ) {
188191 TestHive .sparkContext.hadoopConfiguration.set(orcDefaultCompressVar, " SNAPPY" )
189192 var tempDir = getTempFilePath(" orcTest" ).getCanonicalPath
190193 val rdd = sparkContext.parallelize((1 to 100 ))
0 commit comments