|
87 | 87 | import org.apache.parquet.column.statistics.LongStatistics;
|
88 | 88 | import org.apache.parquet.column.statistics.SizeStatistics;
|
89 | 89 | import org.apache.parquet.column.statistics.Statistics;
|
| 90 | +import org.apache.parquet.crypto.DecryptionPropertiesFactory; |
| 91 | +import org.apache.parquet.crypto.EncryptionPropertiesFactory; |
| 92 | +import org.apache.parquet.crypto.FileDecryptionProperties; |
| 93 | +import org.apache.parquet.crypto.InternalFileDecryptor; |
90 | 94 | import org.apache.parquet.example.Paper;
|
91 | 95 | import org.apache.parquet.example.data.Group;
|
92 | 96 | import org.apache.parquet.example.data.simple.SimpleGroup;
|
@@ -635,18 +639,49 @@ public void randomTestFilterMetaData() {
|
635 | 639 | }
|
636 | 640 |
|
637 | 641 | @Test
|
638 |
| - public void testNullFieldMetadataDebugLogging() { |
| 642 | + public void testFieldMetadataDebugLogging() { |
639 | 643 | MessageType schema = parseMessageType("message test { optional binary some_null_field; }");
|
640 | 644 | org.apache.parquet.hadoop.metadata.FileMetaData fileMetaData =
|
641 |
| - new org.apache.parquet.hadoop.metadata.FileMetaData(schema, new HashMap<String, String>(), null); |
642 |
| - List<BlockMetaData> blockMetaDataList = new ArrayList<BlockMetaData>(); |
| 645 | + new org.apache.parquet.hadoop.metadata.FileMetaData( |
| 646 | + schema, |
| 647 | + new HashMap<>(), |
| 648 | + null, |
| 649 | + org.apache.parquet.hadoop.metadata.FileMetaData.EncryptionType.UNENCRYPTED, |
| 650 | + null); |
| 651 | + List<BlockMetaData> blockMetaDataList = new ArrayList<>(); |
643 | 652 | BlockMetaData blockMetaData = new BlockMetaData();
|
644 | 653 | blockMetaData.addColumn(createColumnChunkMetaData());
|
645 | 654 | blockMetaDataList.add(blockMetaData);
|
646 | 655 | ParquetMetadata metadata = new ParquetMetadata(fileMetaData, blockMetaDataList);
|
647 | 656 | ParquetMetadata.toJSON(metadata);
|
648 | 657 | }
|
649 | 658 |
|
| 659 | + @Test |
| 660 | + public void testEncryptedFieldMetadataDebugLogging() { |
| 661 | + Configuration conf = new Configuration(); |
| 662 | + conf.set( |
| 663 | + EncryptionPropertiesFactory.CRYPTO_FACTORY_CLASS_PROPERTY_NAME, |
| 664 | + "org.apache.parquet.crypto.SampleDecryptionPropertiesFactory"); |
| 665 | + DecryptionPropertiesFactory decryptionPropertiesFactory = DecryptionPropertiesFactory.loadFactory(conf); |
| 666 | + FileDecryptionProperties decryptionProperties = |
| 667 | + decryptionPropertiesFactory.getFileDecryptionProperties(conf, null); |
| 668 | + |
| 669 | + MessageType schema = parseMessageType("message test { optional binary some_null_field; }"); |
| 670 | + |
| 671 | + org.apache.parquet.hadoop.metadata.FileMetaData fileMetaData = |
| 672 | + new org.apache.parquet.hadoop.metadata.FileMetaData( |
| 673 | + schema, |
| 674 | + new HashMap<>(), |
| 675 | + null, |
| 676 | + org.apache.parquet.hadoop.metadata.FileMetaData.EncryptionType.ENCRYPTED_FOOTER, |
| 677 | + new InternalFileDecryptor(decryptionProperties)); |
| 678 | + |
| 679 | + List<BlockMetaData> blockMetaDataList = new ArrayList<>(); |
| 680 | + ParquetMetadata metadata = new ParquetMetadata(fileMetaData, blockMetaDataList); |
| 681 | + ParquetMetadata.toJSON(metadata); |
| 682 | + System.out.println(ParquetMetadata.toPrettyJSON(metadata)); |
| 683 | + } |
| 684 | + |
650 | 685 | @Test
|
651 | 686 | public void testMetadataToJson() {
|
652 | 687 | ParquetMetadata metadata = new ParquetMetadata(null, null);
|
|
0 commit comments