org.apache.hadoop.io.BooleanWritable Java Examples
The following examples show how to use
org.apache.hadoop.io.BooleanWritable.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example #1
Source File: CommonHadoopShim.java From pentaho-hadoop-shims with Apache License 2.0 | 6 votes |
@Override public Class<? extends Writable> getHadoopWritableCompatibleClass( ValueMetaInterface kettleType ) { if ( kettleType == null ) { return NullWritable.class; } switch ( kettleType.getType() ) { case ValueMetaInterface.TYPE_STRING: case ValueMetaInterface.TYPE_BIGNUMBER: case ValueMetaInterface.TYPE_DATE: return Text.class; case ValueMetaInterface.TYPE_INTEGER: return LongWritable.class; case ValueMetaInterface.TYPE_NUMBER: return DoubleWritable.class; case ValueMetaInterface.TYPE_BOOLEAN: return BooleanWritable.class; case ValueMetaInterface.TYPE_BINARY: return BytesWritable.class; default: return Text.class; } }
Example #2
Source File: TestJoinTupleWritable.java From big-c with Apache License 2.0 | 6 votes |
public void testNestedIterable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable sTuple = makeTuple(writs); assertTrue("Bad count", writs.length == verifIter(writs, sTuple, 0)); }
Example #3
Source File: TestJoinTupleWritable.java From big-c with Apache License 2.0 | 6 votes |
public void testIterable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable t = new TupleWritable(writs); for (int i = 0; i < 6; ++i) { t.setWritten(i); } verifIter(writs, t, 0); }
Example #4
Source File: TestJoinTupleWritable.java From big-c with Apache License 2.0 | 6 votes |
private Writable[] makeRandomWritables() { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; return writs; }
Example #5
Source File: TypeConverterFactory.java From pentaho-hadoop-shims with Apache License 2.0 | 6 votes |
/** * Determine the Hadoop writable type to pass Kettle type back to Hadoop as. * * @param kettleType * @return Java type to convert {@code kettleType} to when sending data back to Hadoop. */ public static Class<? extends Writable> getWritableForKettleType( ValueMetaInterface kettleType ) { if ( kettleType == null ) { return NullWritable.class; } switch ( kettleType.getType() ) { case ValueMetaInterface.TYPE_STRING: case ValueMetaInterface.TYPE_BIGNUMBER: case ValueMetaInterface.TYPE_DATE: return Text.class; case ValueMetaInterface.TYPE_INTEGER: return LongWritable.class; case ValueMetaInterface.TYPE_NUMBER: return DoubleWritable.class; case ValueMetaInterface.TYPE_BOOLEAN: return BooleanWritable.class; case ValueMetaInterface.TYPE_BINARY: return BytesWritable.class; default: return Text.class; } }
Example #6
Source File: AdmmReducerContextWritable.java From laser with Apache License 2.0 | 6 votes |
public void write(DataOutput out) throws IOException { new BooleanWritable(null != context.getXUpdated()).write(out); if (null != context.getXUpdated()) { writer.set(context.getXUpdated()); writer.write(out); } new BooleanWritable(null != context.getUInitial()).write(out); if (null != context.getUInitial()) { writer.set(context.getUInitial()); writer.write(out); } new BooleanWritable(null != context.getZUpdated()).write(out); if (null != context.getZUpdated()) { writer.set(context.getZUpdated()); writer.write(out); } new DoubleWritable(context.getRho()).write(out); new DoubleWritable(context.getLambdaValue()).write(out); new DoubleWritable(context.getPrimalObjectiveValue()).write(out); new LongWritable(context.getCount()).write(out); }
Example #7
Source File: TestTupleWritable.java From hadoop-gpu with Apache License 2.0 | 6 votes |
public void testIterable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable t = new TupleWritable(writs); for (int i = 0; i < 6; ++i) { t.setWritten(i); } verifIter(writs, t, 0); }
Example #8
Source File: TestJoinTupleWritable.java From hadoop with Apache License 2.0 | 6 votes |
public void testWritable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable sTuple = makeTuple(writs); ByteArrayOutputStream out = new ByteArrayOutputStream(); sTuple.write(new DataOutputStream(out)); ByteArrayInputStream in = new ByteArrayInputStream(out.toByteArray()); TupleWritable dTuple = new TupleWritable(); dTuple.readFields(new DataInputStream(in)); assertTrue("Failed to write/read tuple", sTuple.equals(dTuple)); }
Example #9
Source File: TestJoinTupleWritable.java From hadoop with Apache License 2.0 | 6 votes |
public void testNestedIterable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable sTuple = makeTuple(writs); assertTrue("Bad count", writs.length == verifIter(writs, sTuple, 0)); }
Example #10
Source File: TestJoinTupleWritable.java From hadoop with Apache License 2.0 | 6 votes |
public void testIterable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable t = new TupleWritable(writs); for (int i = 0; i < 6; ++i) { t.setWritten(i); } verifIter(writs, t, 0); }
Example #11
Source File: TestJoinTupleWritable.java From hadoop with Apache License 2.0 | 6 votes |
private Writable[] makeRandomWritables() { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; return writs; }
Example #12
Source File: TestTupleWritable.java From big-c with Apache License 2.0 | 6 votes |
public void testWritable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable sTuple = makeTuple(writs); ByteArrayOutputStream out = new ByteArrayOutputStream(); sTuple.write(new DataOutputStream(out)); ByteArrayInputStream in = new ByteArrayInputStream(out.toByteArray()); TupleWritable dTuple = new TupleWritable(); dTuple.readFields(new DataInputStream(in)); assertTrue("Failed to write/read tuple", sTuple.equals(dTuple)); }
Example #13
Source File: HadoopCloverConvert.java From CloverETL-Engine with GNU Lesser General Public License v2.1 | 6 votes |
@SuppressWarnings("rawtypes") public static Class cloverType2Hadoop(DataFieldMetadata field) throws IOException{ switch (field.getDataType()){ case BOOLEAN: return BooleanWritable.class; case BYTE: case CBYTE: return BytesWritable.class; case DATE: return LongWritable.class; case INTEGER: return IntWritable.class; case LONG: return LongWritable.class; case NUMBER: return DoubleWritable.class; case STRING: return Text.class; default: throw new IOException(String.format("Unsupported CloverDX data type \"%s\" of field \"%s\" in conversion to Hadoop.",field.getDataType().getName(),field.getName())); } }
Example #14
Source File: DataWritableWriter.java From parquet-mr with Apache License 2.0 | 6 votes |
private void writePrimitive(final Writable value) { if (value == null) { return; } if (value instanceof DoubleWritable) { recordConsumer.addDouble(((DoubleWritable) value).get()); } else if (value instanceof BooleanWritable) { recordConsumer.addBoolean(((BooleanWritable) value).get()); } else if (value instanceof FloatWritable) { recordConsumer.addFloat(((FloatWritable) value).get()); } else if (value instanceof IntWritable) { recordConsumer.addInteger(((IntWritable) value).get()); } else if (value instanceof LongWritable) { recordConsumer.addLong(((LongWritable) value).get()); } else if (value instanceof ShortWritable) { recordConsumer.addInteger(((ShortWritable) value).get()); } else if (value instanceof ByteWritable) { recordConsumer.addInteger(((ByteWritable) value).get()); } else if (value instanceof BigDecimalWritable) { throw new UnsupportedOperationException("BigDecimal writing not implemented"); } else if (value instanceof BinaryWritable) { recordConsumer.addBinary(((BinaryWritable) value).getBinary()); } else { throw new IllegalArgumentException("Unknown value type: " + value + " " + value.getClass()); } }
Example #15
Source File: TestTupleWritable.java From hadoop with Apache License 2.0 | 6 votes |
private Writable[] makeRandomWritables() { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; return writs; }
Example #16
Source File: TestTupleWritable.java From RDFS with Apache License 2.0 | 6 votes |
public void testIterable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable t = new TupleWritable(writs); for (int i = 0; i < 6; ++i) { t.setWritten(i); } verifIter(writs, t, 0); }
Example #17
Source File: TestTupleWritable.java From hadoop-gpu with Apache License 2.0 | 6 votes |
public void testWritable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable sTuple = makeTuple(writs); ByteArrayOutputStream out = new ByteArrayOutputStream(); sTuple.write(new DataOutputStream(out)); ByteArrayInputStream in = new ByteArrayInputStream(out.toByteArray()); TupleWritable dTuple = new TupleWritable(); dTuple.readFields(new DataInputStream(in)); assertTrue("Failed to write/read tuple", sTuple.equals(dTuple)); }
Example #18
Source File: UDAFToOrderedList.java From incubator-hivemall with Apache License 2.0 | 6 votes |
@Override public Object terminatePartial(@SuppressWarnings("deprecation") AggregationBuffer agg) throws HiveException { QueueAggregationBuffer myagg = (QueueAggregationBuffer) agg; Pair<List<Object>, List<Object>> tuples = myagg.drainQueue(); if (tuples == null) { return null; } List<Object> keyList = tuples.getKey(); List<Object> valueList = tuples.getValue(); Object[] partialResult = new Object[outKV || outVK ? 5 : 4]; partialResult[0] = valueList; partialResult[1] = keyList; partialResult[2] = new IntWritable(myagg.size); partialResult[3] = new BooleanWritable(myagg.reverseOrder); if (myagg.outKV) { partialResult[4] = new BooleanWritable(true); } else if (myagg.outVK) { partialResult[4] = new BooleanWritable(true); } return partialResult; }
Example #19
Source File: LookupBuilderReducer.java From incubator-retired-blur with Apache License 2.0 | 6 votes |
@Override protected void reduce(Text rowId, Iterable<NullWritable> nothing, Reducer<Text, NullWritable, Text, BooleanWritable>.Context context) throws IOException, InterruptedException { if (_matcher == null) { _matcher = getMergeSortRowIdMatcher(rowId, context); } if (_writer == null) { _writer = getRowIdWriter(rowId, context); } _writer.append(rowId, NullWritable.get()); _rowIds.increment(1); if (_action == null) { _action = new Action() { @Override public void found(Text rowId) throws IOException { _rowIdsToUpdate.increment(1); try { context.write(rowId, new BooleanWritable(true)); } catch (InterruptedException e) { throw new IOException(e); } } }; } _matcher.lookup(rowId, _action); }
Example #20
Source File: SequenceFileLoader.java From spork with Apache License 2.0 | 6 votes |
protected Object translateWritableToPigDataType(Writable w, byte dataType) { switch(dataType) { case DataType.CHARARRAY: return ((Text) w).toString(); case DataType.BYTEARRAY: BytesWritable bw = (BytesWritable) w; // Make a copy return new DataByteArray(bw.getBytes(), 0, bw.getLength()); case DataType.BOOLEAN: return ((BooleanWritable) w).get(); case DataType.INTEGER: return ((IntWritable) w).get(); case DataType.LONG: return ((LongWritable) w).get(); case DataType.FLOAT: return ((FloatWritable) w).get(); case DataType.DOUBLE: return ((DoubleWritable) w).get(); case DataType.BYTE: return ((ByteWritable) w).get(); case DataType.DATETIME: return ((DateTimeWritable) w).get(); } return null; }
Example #21
Source File: ConfigurationManagerIntegrationTest.java From metron with Apache License 2.0 | 6 votes |
@Test public void testPushAll() throws Exception { // push all configs; parser, enrichment, indexing, etc pushAllConfigs(); // validate final Set<String> sensorsInZookeeper = new HashSet<>(); final BooleanWritable foundGlobal = new BooleanWritable(false); ConfigurationsUtils.visitConfigs(client, new ConfigurationsUtils.ConfigurationVisitor() { @Override public void visit(ConfigurationType configurationType, String name, String data) { assertTrue(data.length() > 0); validateConfig(name, configurationType, data); if(configurationType == GLOBAL) { validateConfig(name, configurationType, data); foundGlobal.set(true); } else { sensorsInZookeeper.add(name); } } }); assertTrue(foundGlobal.get()); assertEquals(sensorsInZookeeper, sensors); }
Example #22
Source File: RecordkeyAdapterTest.java From pxf with Apache License 2.0 | 6 votes |
/** * Test convertKeyValue for several calls of the same type */ @Test public void convertKeyValueManyCalls() { Boolean key = true; mockLog(); initRecordkeyAdapter(); runConvertKeyValue(key, new BooleanWritable(key)); verifyLog("converter initialized for type " + key.getClass() + " (key value: " + key + ")"); for (int i = 0; i < 5; ++i) { key = (i % 2) == 0; runConvertKeyValue(key, new BooleanWritable(key)); } verifyLogOnlyOnce(); }
Example #23
Source File: CompactWritablesDeserializer.java From Cubert with Apache License 2.0 | 6 votes |
private static final WritableComparable<?> createWritable(DataType type) { switch (type) { case BOOLEAN: return new BooleanWritable(); case BYTE: return new ByteWritable(); case INT: return new IntWritable(); case LONG: return new LongWritable(); case FLOAT: return new FloatWritable(); case DOUBLE: return new DoubleWritable(); case STRING: return new Text(); default: return null; } }
Example #24
Source File: TestTupleWritable.java From RDFS with Apache License 2.0 | 6 votes |
public void testNestedIterable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable sTuple = makeTuple(writs); assertTrue("Bad count", writs.length == verifIter(writs, sTuple, 0)); }
Example #25
Source File: DefaultStratosphereTypeConverter.java From stratosphere with Apache License 2.0 | 6 votes |
@SuppressWarnings("unchecked") private<T> T convert(Record stratosphereType, int pos, Class<T> hadoopType) { if(hadoopType == LongWritable.class ) { return (T) new LongWritable((stratosphereType.getField(pos, LongValue.class)).getValue()); } if(hadoopType == org.apache.hadoop.io.Text.class) { return (T) new Text((stratosphereType.getField(pos, StringValue.class)).getValue()); } if(hadoopType == org.apache.hadoop.io.IntWritable.class) { return (T) new IntWritable((stratosphereType.getField(pos, IntValue.class)).getValue()); } if(hadoopType == org.apache.hadoop.io.FloatWritable.class) { return (T) new FloatWritable((stratosphereType.getField(pos, FloatValue.class)).getValue()); } if(hadoopType == org.apache.hadoop.io.DoubleWritable.class) { return (T) new DoubleWritable((stratosphereType.getField(pos, DoubleValue.class)).getValue()); } if(hadoopType == org.apache.hadoop.io.BooleanWritable.class) { return (T) new BooleanWritable((stratosphereType.getField(pos, BooleanValue.class)).getValue()); } if(hadoopType == org.apache.hadoop.io.ByteWritable.class) { return (T) new ByteWritable((stratosphereType.getField(pos, ByteValue.class)).getValue()); } throw new RuntimeException("Unable to convert Stratosphere type ("+stratosphereType.getClass().getCanonicalName()+") to Hadoop."); }
Example #26
Source File: TestTupleWritable.java From RDFS with Apache License 2.0 | 6 votes |
public void testWritable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable sTuple = makeTuple(writs); ByteArrayOutputStream out = new ByteArrayOutputStream(); sTuple.write(new DataOutputStream(out)); ByteArrayInputStream in = new ByteArrayInputStream(out.toByteArray()); TupleWritable dTuple = new TupleWritable(); dTuple.readFields(new DataInputStream(in)); assertTrue("Failed to write/read tuple", sTuple.equals(dTuple)); }
Example #27
Source File: HadoopCloverConvert.java From CloverETL-Engine with GNU Lesser General Public License v2.1 | 6 votes |
@SuppressWarnings("rawtypes") public static DataFieldType hadoopType2Clover( Class data) throws IOException{ if (data == BooleanWritable.class){ return DataFieldType.BOOLEAN; }else if (data == BytesWritable.class){ return DataFieldType.BYTE; }else if (data == LongWritable.class){ return DataFieldType.LONG; }else if (data == IntWritable.class){ return DataFieldType.INTEGER; }else if (data == DoubleWritable.class){ return DataFieldType.NUMBER; }else if (data == Text.class){ return DataFieldType.STRING; }else{ throw new IOException(String.format("Unsupported Hadoop data/Class type \"%s\" in conversion from Hadoop to Clover.",data.getName())); } }
Example #28
Source File: InternalUtilities.java From marklogic-contentpump with Apache License 2.0 | 5 votes |
/** * Create new XdmValue from value type and Writables. * */ public static XdmValue newValue(ValueType valueType, Object value) { if (value instanceof Text) { return ValueFactory.newValue(valueType, ((Text)value).toString()); } else if (value instanceof BytesWritable) { return ValueFactory.newValue(valueType, ((BytesWritable)value).getBytes()); } else if (value instanceof IntWritable) { return ValueFactory.newValue(valueType, ((IntWritable)value).get()); } else if (value instanceof LongWritable) { return ValueFactory.newValue(valueType, ((LongWritable)value).get()); } else if (value instanceof VIntWritable) { return ValueFactory.newValue(valueType, ((VIntWritable)value).get()); } else if (value instanceof VLongWritable) { return ValueFactory.newValue(valueType, ((VLongWritable)value).get()); } else if (value instanceof BooleanWritable) { return ValueFactory.newValue(valueType, ((BooleanWritable)value).get()); } else if (value instanceof FloatWritable) { return ValueFactory.newValue(valueType, ((FloatWritable)value).get()); } else if (value instanceof DoubleWritable) { return ValueFactory.newValue(valueType, ((DoubleWritable)value).get()); } else if (value instanceof MarkLogicNode) { return ValueFactory.newValue(valueType, ((MarkLogicNode)value).get()); } else { throw new UnsupportedOperationException("Value " + value.getClass().getName() + " is unsupported."); } }
Example #29
Source File: TypedBytesWritableOutput.java From big-c with Apache License 2.0 | 5 votes |
public void write(Writable w) throws IOException { if (w instanceof TypedBytesWritable) { writeTypedBytes((TypedBytesWritable) w); } else if (w instanceof BytesWritable) { writeBytes((BytesWritable) w); } else if (w instanceof ByteWritable) { writeByte((ByteWritable) w); } else if (w instanceof BooleanWritable) { writeBoolean((BooleanWritable) w); } else if (w instanceof IntWritable) { writeInt((IntWritable) w); } else if (w instanceof VIntWritable) { writeVInt((VIntWritable) w); } else if (w instanceof LongWritable) { writeLong((LongWritable) w); } else if (w instanceof VLongWritable) { writeVLong((VLongWritable) w); } else if (w instanceof FloatWritable) { writeFloat((FloatWritable) w); } else if (w instanceof DoubleWritable) { writeDouble((DoubleWritable) w); } else if (w instanceof Text) { writeText((Text) w); } else if (w instanceof ArrayWritable) { writeArray((ArrayWritable) w); } else if (w instanceof MapWritable) { writeMap((MapWritable) w); } else if (w instanceof SortedMapWritable) { writeSortedMap((SortedMapWritable) w); } else if (w instanceof Record) { writeRecord((Record) w); } else { writeWritable(w); // last resort } }
Example #30
Source File: ListDefaulterMapValueValidator.java From jumbune with GNU Lesser General Public License v3.0 | 5 votes |
/*** * validates whether the given input value is valid * @return boolean isValid */ @SuppressWarnings("rawtypes") @Override public boolean isPatternValid( WritableComparable value ) { boolean isValid = false; BooleanWritable booleanWritable=(BooleanWritable) value; boolean isDefaulter = booleanWritable.get(); // if is defaulter is true then only valid input isValid = isDefaulter; return isValid; }