Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,9 @@ public enum OriginalType {
DECIMAL,
DATE,
TIME_MILLIS,
TIME_MICROS,
TIMESTAMP_MILLIS,
TIMESTAMP_MICROS,
UINT_8,
UINT_16,
UINT_32,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -424,7 +424,9 @@ protected PrimitiveType build(String name) {
Preconditions.checkState(primitiveType == PrimitiveTypeName.INT32,
originalType.toString() + " can only annotate INT32");
break;
case TIME_MICROS:
case TIMESTAMP_MILLIS:
case TIMESTAMP_MICROS:
case UINT_64:
case INT_64:
Preconditions.checkState(primitiveType == PrimitiveTypeName.INT64,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -509,7 +509,7 @@ public Type call() throws Exception {
@Test
public void testInt64Annotations() {
OriginalType[] types = new OriginalType[] {
TIMESTAMP_MILLIS, UINT_64, INT_64};
TIME_MICROS, TIMESTAMP_MILLIS, TIMESTAMP_MICROS, UINT_64, INT_64};
for (OriginalType logicalType : types) {
PrimitiveType expected = new PrimitiveType(REQUIRED, INT64, "col", logicalType);
PrimitiveType date = Types.required(INT64).as(logicalType).named("col");
Expand All @@ -520,7 +520,7 @@ public void testInt64Annotations() {
@Test
public void testInt64AnnotationsRejectNonInt64() {
OriginalType[] types = new OriginalType[] {
TIMESTAMP_MILLIS, UINT_64, INT_64};
TIME_MICROS, TIMESTAMP_MILLIS, TIMESTAMP_MICROS, UINT_64, INT_64};
for (final OriginalType logicalType : types) {
PrimitiveTypeName[] nonInt64 = new PrimitiveTypeName[]{
BOOLEAN, INT32, INT96, DOUBLE, FLOAT, BINARY
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,6 @@
import java.util.Set;
import java.util.concurrent.ConcurrentHashMap;

import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.parquet.CorruptStatistics;
import org.apache.parquet.Log;
import org.apache.parquet.hadoop.metadata.ColumnPath;
Expand Down Expand Up @@ -336,8 +335,12 @@ OriginalType getOriginalType(ConvertedType type) {
return OriginalType.DATE;
case TIME_MILLIS:
return OriginalType.TIME_MILLIS;
case TIME_MICROS:
return OriginalType.TIME_MICROS;
case TIMESTAMP_MILLIS:
return OriginalType.TIMESTAMP_MILLIS;
case TIMESTAMP_MICROS:
return OriginalType.TIMESTAMP_MICROS;
case INTERVAL:
return OriginalType.INTERVAL;
case INT_8:
Expand Down Expand Up @@ -384,8 +387,12 @@ ConvertedType getConvertedType(OriginalType type) {
return ConvertedType.DATE;
case TIME_MILLIS:
return ConvertedType.TIME_MILLIS;
case TIME_MICROS:
return ConvertedType.TIME_MICROS;
case TIMESTAMP_MILLIS:
return ConvertedType.TIMESTAMP_MILLIS;
case TIMESTAMP_MICROS:
return ConvertedType.TIMESTAMP_MICROS;
case INTERVAL:
return ConvertedType.INTERVAL;
case INT_8:
Expand Down
2 changes: 1 addition & 1 deletion pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@
<shade.prefix>shaded.parquet</shade.prefix>
<hadoop.version>1.1.0</hadoop.version>
<cascading.version>2.5.3</cascading.version>
<parquet.format.version>2.3.0-incubating</parquet.format.version>
<parquet.format.version>2.3.1</parquet.format.version>
<previous.version>1.7.0</previous.version>
<thrift.executable>thrift</thrift.executable>
<scala.version>2.10.4</scala.version>
Expand Down