Java Code Examples for org.apache.hadoop.io.DefaultStringifier#load()
The following examples show how to use
org.apache.hadoop.io.DefaultStringifier#load() .
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example 1
Source File: ContentOutputFormat.java From marklogic-contentpump with Apache License 2.0 | 6 votes |
protected LinkedMapWritable getForestStatusMap(Configuration conf) throws IOException { String forestHost = conf.get(OUTPUT_FOREST_HOST); if (forestHost != null) { //Restores the object from the configuration. LinkedMapWritable fhmap = DefaultStringifier.load(conf, OUTPUT_FOREST_HOST, LinkedMapWritable.class); // must be in fast load mode, otherwise won't reach here String s = conf.get(ASSIGNMENT_POLICY); //EXECUTION_MODE must have a value in mlcp; //default is "distributed" in hadoop connector String mode = conf.get(EXECUTION_MODE, MODE_DISTRIBUTED); if (MODE_DISTRIBUTED.equals(mode)) { AssignmentPolicy.Kind policy = AssignmentPolicy.Kind.forName(s); am.initialize(policy, fhmap, conf.getInt(BATCH_SIZE, 10)); } return fhmap; } else { throw new IOException("Forest host map not found"); } }
Example 2
Source File: GenericRecordExportMapper.java From aliyun-maxcompute-data-collectors with Apache License 2.0 | 5 votes |
@Override protected void setup(Context context) throws IOException, InterruptedException { super.setup(context); Configuration conf = context.getConfiguration(); // Instantiate a copy of the user's class to hold and parse the record. String recordClassName = conf.get( ExportJobBase.SQOOP_EXPORT_TABLE_CLASS_KEY); if (null == recordClassName) { throw new IOException("Export table class name (" + ExportJobBase.SQOOP_EXPORT_TABLE_CLASS_KEY + ") is not set!"); } try { Class cls = Class.forName(recordClassName, true, Thread.currentThread().getContextClassLoader()); recordImpl = (SqoopRecord) ReflectionUtils.newInstance(cls, conf); } catch (ClassNotFoundException cnfe) { throw new IOException(cnfe); } if (null == recordImpl) { throw new IOException("Could not instantiate object of type " + recordClassName); } columnTypes = DefaultStringifier.load(conf, AVRO_COLUMN_TYPES_MAP, MapWritable.class); }
Example 3
Source File: ParquetExportMapper.java From aliyun-maxcompute-data-collectors with Apache License 2.0 | 5 votes |
@Override protected void setup(Context context) throws IOException, InterruptedException { super.setup(context); Configuration conf = context.getConfiguration(); // Instantiate a copy of the user's class to hold and parse the record. String recordClassName = conf.get( ExportJobBase.SQOOP_EXPORT_TABLE_CLASS_KEY); if (null == recordClassName) { throw new IOException("Export table class name (" + ExportJobBase.SQOOP_EXPORT_TABLE_CLASS_KEY + ") is not set!"); } try { Class cls = Class.forName(recordClassName, true, Thread.currentThread().getContextClassLoader()); recordImpl = (SqoopRecord) ReflectionUtils.newInstance(cls, conf); } catch (ClassNotFoundException cnfe) { throw new IOException(cnfe); } if (null == recordImpl) { throw new IOException("Could not instantiate object of type " + recordClassName); } columnTypes = DefaultStringifier.load(conf, AVRO_COLUMN_TYPES_MAP, MapWritable.class); }
Example 4
Source File: RDFInputFormat.java From marklogic-contentpump with Apache License 2.0 | 5 votes |
protected LinkedMapWritable getRoleMap(TaskAttemptContext context) throws IOException{ //Restores the object from the configuration. Configuration conf = context.getConfiguration(); LinkedMapWritable fhmap = null; if(conf.get(ConfigConstants.CONF_ROLE_MAP)!=null) { fhmap = DefaultStringifier.load(conf, ConfigConstants.CONF_ROLE_MAP, LinkedMapWritable.class); } return fhmap; }
Example 5
Source File: RDFInputFormat.java From marklogic-contentpump with Apache License 2.0 | 5 votes |
protected String getServerVersion(TaskAttemptContext context) throws IOException{ //Restores the object from the configuration. Configuration conf = context.getConfiguration(); Text version = DefaultStringifier.load(conf, ConfigConstants.CONF_ML_VERSION, Text.class); return version.toString(); }
Example 6
Source File: MarkLogicOutputFormat.java From marklogic-contentpump with Apache License 2.0 | 5 votes |
protected TextArrayWritable getHosts(Configuration conf) throws IOException { String forestHost = conf.get(OUTPUT_FOREST_HOST); if (forestHost != null) { // Restores the object from the configuration. TextArrayWritable hosts = DefaultStringifier.load(conf, OUTPUT_FOREST_HOST, TextArrayWritable.class); return hosts; } else { throw new IOException("Forest host map not found"); } }
Example 7
Source File: SqoopHCatExportHelper.java From aliyun-maxcompute-data-collectors with Apache License 2.0 | 4 votes |
public SqoopHCatExportHelper(Configuration conf, boolean isOdps) throws IOException, InterruptedException { this.isOdps = isOdps; if (!isOdps) { colTypesJava = DefaultStringifier.load(conf, SqoopHCatUtilities.HCAT_DB_OUTPUT_COLTYPES_JAVA, MapWritable.class); colTypesSql = DefaultStringifier.load(conf, SqoopHCatUtilities.HCAT_DB_OUTPUT_COLTYPES_SQL, MapWritable.class); } // Instantiate a copy of the user's class to hold and parse the record. String recordClassName = conf.get( ExportJobBase.SQOOP_EXPORT_TABLE_CLASS_KEY); if (null == recordClassName) { throw new IOException("Export table class name (" + ExportJobBase.SQOOP_EXPORT_TABLE_CLASS_KEY + ") is not set!"); } bigDecimalFormatString = conf.getBoolean( ImportJobBase.PROPERTY_BIGDECIMAL_FORMAT, ImportJobBase.PROPERTY_BIGDECIMAL_FORMAT_DEFAULT); debugHCatExportMapper = conf.getBoolean( SqoopHCatUtilities.DEBUG_HCAT_EXPORT_MAPPER_PROP, false); try { Class<?> cls = Class.forName(recordClassName, true, Thread.currentThread().getContextClassLoader()); sqoopRecord = (SqoopRecord) ReflectionUtils.newInstance(cls, conf); } catch (ClassNotFoundException cnfe) { throw new IOException(cnfe); } if (null == sqoopRecord) { throw new IOException("Could not instantiate object of type " + recordClassName); } String inputJobInfoStr = conf.get(HCatConstants.HCAT_KEY_JOB_INFO); jobInfo = (InputJobInfo) HCatUtil.deserialize(inputJobInfoStr); HCatSchema tableSchema = jobInfo.getTableInfo().getDataColumns(); HCatSchema partitionSchema = jobInfo.getTableInfo().getPartitionColumns(); hCatFullTableSchema = new HCatSchema(tableSchema.getFields()); for (HCatFieldSchema hfs : partitionSchema.getFields()) { hCatFullTableSchema.append(hfs); } }
Example 8
Source File: TransformOutputFormat.java From marklogic-contentpump with Apache License 2.0 | 4 votes |
/** * initialize mimetype map if not initialized, return the map * * @return * @throws IOException */ private LinkedMapWritable getMimetypesMap() throws IOException { if (mimetypeMap != null) return mimetypeMap; String mtmap = conf.get(ConfigConstants.CONF_MIMETYPES); if (mtmap != null) { mimetypeMap = DefaultStringifier.load(conf, ConfigConstants.CONF_MIMETYPES, LinkedMapWritable.class); return mimetypeMap; } String[] hosts = conf.getStrings(OUTPUT_HOST); Session session = null; ResultSequence result = null; for (int i = 0; i < hosts.length; i++) { try { String host = hosts[i]; ContentSource cs = InternalUtilities.getOutputContentSource(conf, host); session = cs.newSession(); AdhocQuery query = session.newAdhocQuery(MIMETYPES_QUERY); RequestOptions options = new RequestOptions(); options.setDefaultXQueryVersion("1.0-ml"); query.setOptions(options); result = session.submitRequest(query); if (!result.hasNext()) throw new IOException( "Server-side transform requires MarkLogic 7 or later"); mimetypeMap = new LinkedMapWritable(); while (result.hasNext()) { String suffs = result.next().asString(); Text format = new Text(result.next().asString()); // some extensions are in a space separated string for (String s : suffs.split(" ")) { Text suff = new Text(s); mimetypeMap.put(suff, format); } } return mimetypeMap; } catch (Exception e) { if (e.getCause() instanceof ServerConnectionException) { LOG.warn("Unable to connect to " + hosts[i] + " to query destination information"); continue; } LOG.error(e.getMessage(), e); throw new IOException(e); } finally { if (result != null) { result.close(); } if (session != null) { session.close(); } } } return null; }