Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 7 additions & 0 deletions pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -77,6 +77,7 @@
<module>presto-blackhole</module>
<module>presto-memory</module>
<module>presto-orc</module>
<module>presto-parquet</module>
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you retitle this commit "Move Parquet reader to new presto-parquet module"?

<module>presto-rcfile</module>
<module>presto-hive</module>
<module>presto-hive-hadoop2</module>
Expand Down Expand Up @@ -187,6 +188,12 @@
<version>${project.version}</version>
</dependency>

<dependency>
<groupId>com.facebook.presto</groupId>
<artifactId>presto-parquet</artifactId>
<version>${project.version}</version>
</dependency>

<dependency>
<groupId>com.facebook.presto</groupId>
<artifactId>presto-rcfile</artifactId>
Expand Down
5 changes: 5 additions & 0 deletions presto-hive/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -32,6 +32,11 @@
<artifactId>presto-orc</artifactId>
</dependency>

<dependency>
<groupId>com.facebook.presto</groupId>
<artifactId>presto-parquet</artifactId>
</dependency>

<dependency>
<groupId>com.facebook.presto</groupId>
<artifactId>presto-memory-context</artifactId>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -101,8 +101,6 @@ public class HiveClientConfig
private DataSize textMaxLineLength = new DataSize(100, MEGABYTE);

private boolean useParquetColumnNames;
private boolean parquetOptimizedReaderEnabled = true;
private boolean parquetPredicatePushdownEnabled = true;

private boolean assumeCanonicalPartitionKeys;

Expand Down Expand Up @@ -670,34 +668,6 @@ public HiveClientConfig setVerifyChecksum(boolean verifyChecksum)
return this;
}

@Deprecated
public boolean isParquetPredicatePushdownEnabled()
{
return parquetPredicatePushdownEnabled;
}

@Deprecated
@Config("hive.parquet-predicate-pushdown.enabled")
public HiveClientConfig setParquetPredicatePushdownEnabled(boolean parquetPredicatePushdownEnabled)
{
this.parquetPredicatePushdownEnabled = parquetPredicatePushdownEnabled;
return this;
}

@Deprecated
public boolean isParquetOptimizedReaderEnabled()
{
return parquetOptimizedReaderEnabled;
}

@Deprecated
@Config("hive.parquet-optimized-reader.enabled")
public HiveClientConfig setParquetOptimizedReaderEnabled(boolean parquetOptimizedReaderEnabled)
{
this.parquetOptimizedReaderEnabled = parquetOptimizedReaderEnabled;
return this;
}

public boolean isUseOrcColumnNames()
{
return useOrcColumnNames;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,6 @@
import com.facebook.presto.hive.orc.DwrfPageSourceFactory;
import com.facebook.presto.hive.orc.OrcPageSourceFactory;
import com.facebook.presto.hive.parquet.ParquetPageSourceFactory;
import com.facebook.presto.hive.parquet.ParquetRecordCursorProvider;
import com.facebook.presto.hive.rcfile.RcFilePageSourceFactory;
import com.facebook.presto.spi.connector.ConnectorNodePartitioningProvider;
import com.facebook.presto.spi.connector.ConnectorPageSinkProvider;
Expand Down Expand Up @@ -76,7 +75,6 @@ public void configure(Binder binder)
newExporter(binder).export(NamenodeStats.class).as(generatedNameOf(NamenodeStats.class, connectorId));

Multibinder<HiveRecordCursorProvider> recordCursorProviderBinder = newSetBinder(binder, HiveRecordCursorProvider.class);
recordCursorProviderBinder.addBinding().to(ParquetRecordCursorProvider.class).in(Scopes.SINGLETON);
recordCursorProviderBinder.addBinding().to(GenericHiveRecordCursorProvider.class).in(Scopes.SINGLETON);

binder.bind(HiveWriterStats.class).in(Scopes.SINGLETON);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -61,8 +61,6 @@ public final class HiveSessionProperties
private static final String ORC_OPTIMIZED_WRITER_MAX_DICTIONARY_MEMORY = "orc_optimized_writer_max_dictionary_memory";
private static final String HIVE_STORAGE_FORMAT = "hive_storage_format";
private static final String RESPECT_TABLE_FORMAT = "respect_table_format";
private static final String PARQUET_PREDICATE_PUSHDOWN_ENABLED = "parquet_predicate_pushdown_enabled";
private static final String PARQUET_OPTIMIZED_READER_ENABLED = "parquet_optimized_reader_enabled";
private static final String PARQUET_USE_COLUMN_NAME = "parquet_use_column_names";
private static final String PARQUET_WRITER_BLOCK_SIZE = "parquet_writer_block_size";
private static final String PARQUET_WRITER_PAGE_SIZE = "parquet_writer_page_size";
Expand Down Expand Up @@ -221,16 +219,6 @@ public HiveSessionProperties(HiveClientConfig hiveClientConfig, OrcFileWriterCon
"Write new partitions using table format rather than default storage format",
hiveClientConfig.isRespectTableFormat(),
false),
booleanProperty(
PARQUET_OPTIMIZED_READER_ENABLED,
"Experimental: Parquet: Enable optimized reader",
hiveClientConfig.isParquetOptimizedReaderEnabled(),
false),
booleanProperty(
PARQUET_PREDICATE_PUSHDOWN_ENABLED,
"Experimental: Parquet: Enable predicate pushdown for Parquet",
hiveClientConfig.isParquetPredicatePushdownEnabled(),
false),
booleanProperty(
PARQUET_USE_COLUMN_NAME,
"Experimental: Parquet: Access Parquet columns using names from the file",
Expand Down Expand Up @@ -313,11 +301,6 @@ public static InsertExistingPartitionsBehavior getInsertExistingPartitionsBehavi
return session.getProperty(INSERT_EXISTING_PARTITIONS_BEHAVIOR, InsertExistingPartitionsBehavior.class);
}

public static boolean isParquetOptimizedReaderEnabled(ConnectorSession session)
{
return session.getProperty(PARQUET_OPTIMIZED_READER_ENABLED, Boolean.class);
}

public static boolean isOrcBloomFiltersEnabled(ConnectorSession session)
{
return session.getProperty(ORC_BLOOM_FILTERS_ENABLED, Boolean.class);
Expand Down Expand Up @@ -415,11 +398,6 @@ public static boolean isRespectTableFormat(ConnectorSession session)
return session.getProperty(RESPECT_TABLE_FORMAT, Boolean.class);
}

public static boolean isParquetPredicatePushdownEnabled(ConnectorSession session)
{
return session.getProperty(PARQUET_PREDICATE_PUSHDOWN_ENABLED, Boolean.class);
}

public static boolean isUseParquetColumnNames(ConnectorSession session)
{
return session.getProperty(PARQUET_USE_COLUMN_NAME, Boolean.class);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@
package com.facebook.presto.hive.parquet;

import com.facebook.presto.hive.FileFormatDataSourceStats;
import com.facebook.presto.parquet.ParquetDataSource;
import com.facebook.presto.spi.PrestoException;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FileSystem;
Expand Down
Loading