Java Code Examples for org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch#getMaxSize()
The following examples show how to use
org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch#getMaxSize() .
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example 1
Source File: AvroToOrcRecordConverter.java From datacollector with Apache License 2.0 | 6 votes |
public static void addAvroRecord( VectorizedRowBatch batch, GenericRecord record, TypeDescription orcSchema, int orcBatchSize, Writer writer ) throws IOException { for (int c = 0; c < batch.numCols; c++) { ColumnVector colVector = batch.cols[c]; final String thisField = orcSchema.getFieldNames().get(c); final TypeDescription type = orcSchema.getChildren().get(c); Object fieldValue = record.get(thisField); Schema.Field avroField = record.getSchema().getField(thisField); addToVector(type, colVector, avroField.schema(), fieldValue, batch.size); } batch.size++; if (batch.size % orcBatchSize == 0 || batch.size == batch.getMaxSize()) { writer.addRowBatch(batch); batch.reset(); batch.size = 0; } }
Example 2
Source File: MDSHiveDirectVectorizedReader.java From multiple-dimension-spread with Apache License 2.0 | 5 votes |
@Override public boolean next( final NullWritable key, final VectorizedRowBatch outputBatch ) throws IOException { outputBatch.reset(); setting.setPartitionValues( outputBatch ); if( indexSize <= currentIndex ){ if( ! currentReader.hasNext() ){ updateCounter( currentReader.getReadStats() ); outputBatch.endOfFile = true; isEnd = true; return false; } while( ! setSpread() ){ if( ! currentReader.hasNext() ){ updateCounter( currentReader.getReadStats() ); outputBatch.endOfFile = true; isEnd = true; return false; } } } int maxSize = outputBatch.getMaxSize(); if( indexSize < currentIndex + maxSize ){ maxSize = indexSize - currentIndex; } for( int colIndex : needColumnIds ){ assignors[colIndex].setColumnVector( outputBatch.cols[colIndex] , currentIndexList , currentIndex , maxSize ); } outputBatch.size = maxSize; currentIndex += maxSize; if( indexSize <= currentIndex && ! currentReader.hasNext() ){ outputBatch.endOfFile = true; } return outputBatch.size > 0; }
Example 3
Source File: SqlInterpreterTest.java From zeppelin with Apache License 2.0 | 5 votes |
public File createORCFile(int[] values) throws IOException { File file = File.createTempFile("zeppelin-flink-input", ".orc"); file.delete(); Path path = new Path(file.getAbsolutePath()); Configuration conf = new Configuration(); conf.set("orc.compress", "snappy"); TypeDescription schema = TypeDescription.fromString("struct<msg:int>"); Writer writer = OrcFile.createWriter(path, OrcFile.writerOptions(conf) .setSchema(schema)); VectorizedRowBatch batch = schema.createRowBatch(); LongColumnVector x = (LongColumnVector) batch.cols[0]; for (int i = 0; i < values.length; ++i) { int row = batch.size++; x.vector[row] = values[i]; // If the batch is full, write it out and start over. if (batch.size == batch.getMaxSize()) { writer.addRowBatch(batch); batch.reset(); } } if (batch.size != 0) { writer.addRowBatch(batch); batch.reset(); } writer.close(); return file; }