public class ParquetRecordReader extends CommonParquetRecordReader
CommonParquetRecordReader.Metricfooter, fragmentContext, NUM_RECORDS_TO_READ_NOT_SPECIFIED, operatorContext, parquetReaderStatsDEFAULT_TEXT_COLS_TO_READALLOCATOR_INITIAL_RESERVATION, ALLOCATOR_MAX_RESERVATION| Constructor and Description |
|---|
ParquetRecordReader(FragmentContext fragmentContext,
long numRecordsToRead,
org.apache.hadoop.fs.Path path,
int rowGroupIndex,
org.apache.hadoop.fs.FileSystem fs,
org.apache.parquet.hadoop.CodecFactory codecFactory,
org.apache.parquet.hadoop.metadata.ParquetMetadata footer,
List<SchemaPath> columns,
ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) |
ParquetRecordReader(FragmentContext fragmentContext,
org.apache.hadoop.fs.Path path,
int rowGroupIndex,
org.apache.hadoop.fs.FileSystem fs,
org.apache.parquet.hadoop.CodecFactory codecFactory,
org.apache.parquet.hadoop.metadata.ParquetMetadata footer,
List<SchemaPath> columns,
ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) |
ParquetRecordReader(FragmentContext fragmentContext,
org.apache.hadoop.fs.Path path,
int rowGroupIndex,
long numRecordsToRead,
org.apache.hadoop.fs.FileSystem fs,
org.apache.parquet.hadoop.CodecFactory codecFactory,
org.apache.parquet.hadoop.metadata.ParquetMetadata footer,
List<SchemaPath> columns,
ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) |
| Modifier and Type | Method and Description |
|---|---|
void |
allocate(Map<String,ValueVector> vectorMap) |
void |
close() |
RecordBatchSizerManager |
getBatchSizesMgr() |
org.apache.parquet.hadoop.CodecFactory |
getCodecFactory() |
ParquetReaderUtility.DateCorruptionStatus |
getDateCorruptionStatus()
Flag indicating if the old non-standard data format appears
in this file, see DRILL-4203.
|
protected List<SchemaPath> |
getDefaultColumnsToRead() |
org.apache.hadoop.fs.FileSystem |
getFileSystem() |
FragmentContext |
getFragmentContext() |
org.apache.hadoop.fs.Path |
getHadoopPath() |
OperatorContext |
getOperatorContext() |
ReadState |
getReadState() |
int |
getRowGroupIndex() |
int |
next()
Read the next record batch from the file using the reader and read state
created previously.
|
void |
setup(OperatorContext operatorContext,
OutputMutator output)
Prepare the Parquet reader.
|
String |
toString() |
boolean |
useBulkReader() |
closeStats, handleAndRaise, initNumRecordsToRead, updateRowGroupsStatsgetColumns, hasNext, isSkipQuery, isStarQuery, setColumns, transformColumnspublic ParquetRecordReader(FragmentContext fragmentContext, org.apache.hadoop.fs.Path path, int rowGroupIndex, long numRecordsToRead, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.hadoop.CodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus)
public ParquetRecordReader(FragmentContext fragmentContext, org.apache.hadoop.fs.Path path, int rowGroupIndex, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.hadoop.CodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus)
public ParquetRecordReader(FragmentContext fragmentContext, long numRecordsToRead, org.apache.hadoop.fs.Path path, int rowGroupIndex, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.hadoop.CodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus)
public ParquetReaderUtility.DateCorruptionStatus getDateCorruptionStatus()
public org.apache.parquet.hadoop.CodecFactory getCodecFactory()
public org.apache.hadoop.fs.Path getHadoopPath()
public org.apache.hadoop.fs.FileSystem getFileSystem()
public int getRowGroupIndex()
public RecordBatchSizerManager getBatchSizesMgr()
public OperatorContext getOperatorContext()
public FragmentContext getFragmentContext()
public boolean useBulkReader()
public ReadState getReadState()
public void setup(OperatorContext operatorContext, OutputMutator output) throws ExecutionSetupException
operatorContext - operator context for the readeroutput - The place where output for a particular scan should be written. The record reader is responsible for
mutating the set of schema values for that particular record.ExecutionSetupExceptionpublic void allocate(Map<String,ValueVector> vectorMap) throws OutOfMemoryException
allocate in interface RecordReaderallocate in class AbstractRecordReaderOutOfMemoryExceptionpublic int next()
public void close()
protected List<SchemaPath> getDefaultColumnsToRead()
getDefaultColumnsToRead in class AbstractRecordReaderpublic String toString()
toString in class AbstractRecordReaderCopyright © 2021 The Apache Software Foundation. All rights reserved.