Java Code Examples for org.elasticsearch.common.xcontent.support.XContentMapValues#nodeIntegerValue()
The following examples show how to use
org.elasticsearch.common.xcontent.support.XContentMapValues#nodeIntegerValue() .
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example 1
Source File: GitHubRiver.java From elasticsearch-river-github with Apache License 2.0 | 5 votes |
@SuppressWarnings({"unchecked"}) @Inject public GitHubRiver(RiverName riverName, RiverSettings settings, Client client) { super(riverName, settings); this.client = client; if (!settings.settings().containsKey("github")) { throw new IllegalArgumentException("Need river settings - owner and repository."); } // get settings Map<String, Object> githubSettings = (Map<String, Object>) settings.settings().get("github"); owner = XContentMapValues.nodeStringValue(githubSettings.get("owner"), null); repository = XContentMapValues.nodeStringValue(githubSettings.get("repository"), null); index = String.format("%s&%s", owner, repository); userRequestedInterval = XContentMapValues.nodeIntegerValue(githubSettings.get("interval"), 60); // auth (optional) username = null; password = null; if (githubSettings.containsKey("authentication")) { Map<String, Object> auth = (Map<String, Object>) githubSettings.get("authentication"); username = XContentMapValues.nodeStringValue(auth.get("username"), null); password = XContentMapValues.nodeStringValue(auth.get("password"), null); } // endpoint (optional - default to github.com) endpoint = XContentMapValues.nodeStringValue(githubSettings.get("endpoint"), "https://api.github.com"); logger.info("Created GitHub river."); }
Example 2
Source File: Neo4jDriver.java From elasticsearch-river-neo4j with Apache License 2.0 | 5 votes |
@Inject public Neo4jDriver(RiverName riverName, RiverSettings settings, @RiverIndexName final String riverIndexName, final Client client) { super(riverName, settings); this.client = client; uri = XContentMapValues.nodeStringValue(XContentMapValues.extractValue("neo4j.uri", settings.settings()), DEFAULT_NEO_URI); List<Object> neo4jLabels = XContentMapValues.extractRawValues("neo4j.labels", settings.settings()); String label; if(XContentMapValues.isArray(neo4jLabels)) { for (Object neo4jLabel : neo4jLabels) { label = XContentMapValues.nodeStringValue(neo4jLabel, null); labels.add(DynamicLabel.label(label)); } } timestampField = XContentMapValues.nodeStringValue(XContentMapValues.extractValue("neo4j.timestampField", settings.settings()), DEFAULT_NEO_TIMESTAMP_FIELD); interval = XContentMapValues.nodeIntegerValue(XContentMapValues.extractValue("neo4j.interval", settings.settings()), DEFAULT_NEO_INTERVAL); index = XContentMapValues.nodeStringValue(XContentMapValues.extractValue("index.name", settings.settings()), DEFAULT_NEO_INDEX); type = XContentMapValues.nodeStringValue(XContentMapValues.extractValue("index.type", settings.settings()), DEFAULT_NEO_TYPE); indexFromLabel = XContentMapValues.nodeStringValue(XContentMapValues.extractValue("index.name.label", settings.settings()), null); typeFromLabel = XContentMapValues.nodeStringValue(XContentMapValues.extractValue("index.type.label", settings.settings()), null); logger.debug("Neo4j settings [uri={}]", new Object[]{uri}); logger.debug("River settings [indexName={}, type={}, interval={}, timestampField={}, indexLabel={}, " + "typelabel={}]", new Object[]{index, type, interval, timestampField, indexFromLabel, typeFromLabel} ); }
Example 3
Source File: IsPrimeSearchScriptFactory.java From elasticsearch-native-script-example with Apache License 2.0 | 5 votes |
/** * This method is called for every search on every shard. * * @param params list of script parameters passed with the query * @return new native script */ @Override public ExecutableScript newScript(@Nullable Map<String, Object> params) { // Example of a mandatory string parameter // The XContentMapValues helper class can be used to simplify parameter parsing String fieldName = params == null ? null : XContentMapValues.nodeStringValue(params.get("field"), defaultFieldName); if (!Strings.hasLength(fieldName)) { throw new IllegalArgumentException("Missing the field parameter"); } // Example of an optional integer parameter int certainty = params == null ? 10 : XContentMapValues.nodeIntegerValue(params.get("certainty"), 10); return new IsPrimeSearchScript(fieldName, certainty); }
Example 4
Source File: StringFieldMapper.java From Elasticsearch with Apache License 2.0 | 4 votes |
@Override public Mapper.Builder parse(String name, Map<String, Object> node, ParserContext parserContext) throws MapperParsingException { StringFieldMapper.Builder builder = stringField(name); parseTextField(builder, name, node, parserContext); for (Iterator<Map.Entry<String, Object>> iterator = node.entrySet().iterator(); iterator.hasNext();) { Map.Entry<String, Object> entry = iterator.next(); String propName = Strings.toUnderscoreCase(entry.getKey()); Object propNode = entry.getValue(); if (propName.equals("null_value")) { if (propNode == null) { throw new MapperParsingException("Property [null_value] cannot be null."); } builder.nullValue(propNode.toString()); iterator.remove(); } else if (propName.equals("search_quote_analyzer")) { NamedAnalyzer analyzer = parserContext.analysisService().analyzer(propNode.toString()); if (analyzer == null) { throw new MapperParsingException("Analyzer [" + propNode.toString() + "] not found for field [" + name + "]"); } builder.searchQuotedAnalyzer(analyzer); iterator.remove(); } else if (propName.equals("position_increment_gap") || parserContext.indexVersionCreated().before(Version.V_2_0_0) && propName.equals("position_offset_gap")) { int newPositionIncrementGap = XContentMapValues.nodeIntegerValue(propNode, -1); if (newPositionIncrementGap < 0) { throw new MapperParsingException("positions_increment_gap less than 0 aren't allowed."); } builder.positionIncrementGap(newPositionIncrementGap); // we need to update to actual analyzers if they are not set in this case... // so we can inject the position increment gap... if (builder.fieldType().indexAnalyzer() == null) { builder.fieldType().setIndexAnalyzer(parserContext.analysisService().defaultIndexAnalyzer()); } if (builder.fieldType().searchAnalyzer() == null) { builder.fieldType().setSearchAnalyzer(parserContext.analysisService().defaultSearchAnalyzer()); } if (builder.fieldType().searchQuoteAnalyzer() == null) { builder.fieldType().setSearchQuoteAnalyzer(parserContext.analysisService().defaultSearchQuoteAnalyzer()); } iterator.remove(); } else if (propName.equals("ignore_above")) { builder.ignoreAbove(XContentMapValues.nodeIntegerValue(propNode, -1)); iterator.remove(); } else if (parseMultiField(builder, name, parserContext, propName, propNode)) { iterator.remove(); } } return builder; }
Example 5
Source File: KafkaRiverConfig.java From elasticsearch-river-kafka with Apache License 2.0 | 4 votes |
public KafkaRiverConfig(RiverSettings settings) { if (settings.settings().containsKey("kafka")) { Map<String, Object> kafkaSettings = (Map<String, Object>) settings.settings().get("kafka"); topic = (String)kafkaSettings.get("topic"); zookeeper = XContentMapValues.nodeStringValue(kafkaSettings.get("zookeeper"), "localhost"); factoryClass = XContentMapValues.nodeStringValue(kafkaSettings.get("message_handler_factory_class"), "org.elasticsearch.river.kafka.JsonMessageHandlerFactory"); brokerHost = XContentMapValues.nodeStringValue(kafkaSettings.get("broker_host"), "localhost"); brokerPort = XContentMapValues.nodeIntegerValue(kafkaSettings.get("broker_port"), 9092); partition = XContentMapValues.nodeIntegerValue(kafkaSettings.get("partition"), 0); } else { zookeeper = "localhost"; brokerHost = "localhost"; brokerPort = 9092; topic = "default_topic"; partition = 0; factoryClass = "org.elasticsearch.river.kafka.JsonMessageHandlerFactory"; } if (settings.settings().containsKey("index")) { Map<String, Object> indexSettings = (Map<String, Object>) settings.settings().get("index"); bulkSize = XContentMapValues.nodeIntegerValue(indexSettings.get("bulk_size_bytes"), 10*1024*1024); if (indexSettings.containsKey("bulk_timeout")) { bulkTimeout = TimeValue.parseTimeValue(XContentMapValues.nodeStringValue(indexSettings.get("bulk_timeout"), "10ms"), TimeValue.timeValueMillis(10000)); } else { bulkTimeout = TimeValue.timeValueMillis(10); } } else { bulkSize = 10*1024*1024; bulkTimeout = TimeValue.timeValueMillis(10000); } if (settings.settings().containsKey("statsd")) { Map<String, Object> statsdSettings = (Map<String, Object>) settings.settings().get("statsd"); statsdHost = (String)statsdSettings.get("host"); statsdPort = XContentMapValues.nodeIntegerValue(statsdSettings.get("port"), 8125); statsdPrefix = XContentMapValues.nodeStringValue(statsdSettings.get("prefix"), "es-kafka-river"); } else { statsdHost = null; statsdPort = -1; statsdPrefix = null; } }
Example 6
Source File: S3River.java From es-amazon-s3-river with Apache License 2.0 | 4 votes |
@Inject @SuppressWarnings({ "unchecked" }) protected S3River(RiverName riverName, RiverSettings settings, Client client, ThreadPool threadPool) throws Exception{ super(riverName, settings); this.client = client; this.threadPool = threadPool; this.riverStatus = RiverStatus.UNKNOWN; // Deal with connector settings. if (settings.settings().containsKey("amazon-s3")){ Map<String, Object> feed = (Map<String, Object>)settings.settings().get("amazon-s3"); // Retrieve feed settings. String feedname = XContentMapValues.nodeStringValue(feed.get("name"), null); String bucket = XContentMapValues.nodeStringValue(feed.get("bucket"), null); String pathPrefix = XContentMapValues.nodeStringValue(feed.get("pathPrefix"), null); String downloadHost = XContentMapValues.nodeStringValue(feed.get("download_host"), null); int updateRate = XContentMapValues.nodeIntegerValue(feed.get("update_rate"), 15 * 60 * 1000); boolean jsonSupport = XContentMapValues.nodeBooleanValue(feed.get("json_support"), false); double indexedCharsRatio = XContentMapValues.nodeDoubleValue(feed.get("indexed_chars_ratio"), 0.0); String[] includes = S3RiverUtil.buildArrayFromSettings(settings.settings(), "amazon-s3.includes"); String[] excludes = S3RiverUtil.buildArrayFromSettings(settings.settings(), "amazon-s3.excludes"); // Retrieve connection settings. String accessKey = XContentMapValues.nodeStringValue(feed.get("accessKey"), null); String secretKey = XContentMapValues.nodeStringValue(feed.get("secretKey"), null); boolean useIAMRoleForEC2 = XContentMapValues.nodeBooleanValue(feed.get("use_EC2_IAM"), false); feedDefinition = new S3RiverFeedDefinition(feedname, bucket, pathPrefix, downloadHost, updateRate, Arrays.asList(includes), Arrays.asList(excludes), accessKey, secretKey, useIAMRoleForEC2, jsonSupport, indexedCharsRatio); } else { logger.error("You didn't define the amazon-s3 settings. Exiting... See https://github.com/lbroudoux/es-amazon-s3-river"); indexName = null; typeName = null; bulkSize = 100; feedDefinition = null; s3 = null; return; } // Deal with index settings if provided. if (settings.settings().containsKey("index")) { Map<String, Object> indexSettings = (Map<String, Object>)settings.settings().get("index"); indexName = XContentMapValues.nodeStringValue(indexSettings.get("index"), riverName.name()); typeName = XContentMapValues.nodeStringValue(indexSettings.get("type"), S3RiverUtil.INDEX_TYPE_DOC); bulkSize = XContentMapValues.nodeIntegerValue(indexSettings.get("bulk_size"), 100); } else { indexName = riverName.name(); typeName = S3RiverUtil.INDEX_TYPE_DOC; bulkSize = 100; } // We need to connect to Amazon S3 after ensure mandatory settings are here. if (feedDefinition.getBucket() == null){ logger.error("Amazon S3 bucket should not be null. Please fix this."); throw new IllegalArgumentException("Amazon S3 bucket should not be null."); } // Connect using the appropriate authentication process. if (feedDefinition.getAccessKey() == null && feedDefinition.getSecretKey() == null) { s3 = new S3Connector(feedDefinition.isUseIAMRoleForEC2()); } else { s3 = new S3Connector(feedDefinition.getAccessKey(), feedDefinition.getSecretKey()); } try { s3.connectUserBucket(feedDefinition.getBucket(), feedDefinition.getPathPrefix()); } catch (AmazonS3Exception ase){ logger.error("Exception while connecting Amazon S3 user bucket. " + "Either access key, secret key, IAM Role or bucket name are incorrect"); throw ase; } this.riverStatus = RiverStatus.INITIALIZED; }
Example 7
Source File: RiverConfig.java From elasticsearch-river-kafka with Apache License 2.0 | 4 votes |
public RiverConfig(RiverName riverName, RiverSettings riverSettings) { // Extract kafka related configuration if (riverSettings.settings().containsKey("kafka")) { Map<String, Object> kafkaSettings = (Map<String, Object>) riverSettings.settings().get("kafka"); topic = (String) kafkaSettings.get(TOPIC); zookeeperConnect = XContentMapValues.nodeStringValue(kafkaSettings.get(ZOOKEEPER_CONNECT), "localhost"); zookeeperConnectionTimeout = XContentMapValues.nodeIntegerValue(kafkaSettings.get(ZOOKEEPER_CONNECTION_TIMEOUT), 10000); messageType = MessageType.fromValue(XContentMapValues.nodeStringValue(kafkaSettings.get(MESSAGE_TYPE), MessageType.JSON.toValue())); } else { zookeeperConnect = "localhost"; zookeeperConnectionTimeout = 10000; topic = "elasticsearch-river-kafka"; messageType = MessageType.JSON; } // Extract ElasticSearch related configuration if (riverSettings.settings().containsKey("index")) { Map<String, Object> indexSettings = (Map<String, Object>) riverSettings.settings().get("index"); indexName = XContentMapValues.nodeStringValue(indexSettings.get(INDEX_NAME), riverName.name()); typeName = XContentMapValues.nodeStringValue(indexSettings.get(MAPPING_TYPE), "status"); bulkSize = XContentMapValues.nodeIntegerValue(indexSettings.get(BULK_SIZE), 100); concurrentRequests = XContentMapValues.nodeIntegerValue(indexSettings.get(CONCURRENT_REQUESTS), 1); actionType = ActionType.fromValue(XContentMapValues.nodeStringValue(indexSettings.get(ACTION_TYPE), ActionType.INDEX.toValue())); flushInterval = TimeValue.parseTimeValue(XContentMapValues.nodeStringValue(indexSettings.get(FLUSH_INTERVAL), "12h"), FLUSH_12H); } else { indexName = riverName.name(); typeName = "status"; bulkSize = 100; concurrentRequests = 1; actionType = ActionType.INDEX; flushInterval = FLUSH_12H; } // Extract StatsD related configuration if (riverSettings.settings().containsKey("statsd")) { Map<String, Object> statsdSettings = (Map<String, Object>) riverSettings.settings().get("statsd"); statsdHost = XContentMapValues.nodeStringValue(statsdSettings.get(STATSD_HOST), "localhost"); statsdPrefix = XContentMapValues.nodeStringValue(statsdSettings.get(STATSD_PREFIX), "kafka_river"); statsdPort = XContentMapValues.nodeIntegerValue(statsdSettings.get(STATSD_PORT), 8125); statsdIntervalInSeconds = XContentMapValues.nodeIntegerValue(statsdSettings.get(STATSD_INTERVAL_IN_SECONDS), 10); } }
Example 8
Source File: TextFieldMapper.java From crate with Apache License 2.0 | 4 votes |
@Override public Mapper.Builder parse(String fieldName, Map<String, Object> node, ParserContext parserContext) throws MapperParsingException { TextFieldMapper.Builder builder = new TextFieldMapper.Builder(fieldName); builder.fieldType().setIndexAnalyzer(parserContext.getIndexAnalyzers().getDefaultIndexAnalyzer()); builder.fieldType().setSearchAnalyzer(parserContext.getIndexAnalyzers().getDefaultSearchAnalyzer()); builder.fieldType().setSearchQuoteAnalyzer(parserContext.getIndexAnalyzers().getDefaultSearchQuoteAnalyzer()); parseTextField(builder, fieldName, node, parserContext); for (Iterator<Map.Entry<String, Object>> iterator = node.entrySet().iterator(); iterator.hasNext();) { Map.Entry<String, Object> entry = iterator.next(); String propName = entry.getKey(); Object propNode = entry.getValue(); if (propName.equals("position_increment_gap")) { int newPositionIncrementGap = XContentMapValues.nodeIntegerValue(propNode, -1); builder.positionIncrementGap(newPositionIncrementGap); iterator.remove(); } else if (propName.equals("fielddata")) { builder.fielddata(XContentMapValues.nodeBooleanValue(propNode, "fielddata")); iterator.remove(); } else if (propName.equals("eager_global_ordinals")) { builder.eagerGlobalOrdinals(XContentMapValues.nodeBooleanValue(propNode, "eager_global_ordinals")); iterator.remove(); } else if (propName.equals("fielddata_frequency_filter")) { Map<?,?> frequencyFilter = (Map<?, ?>) propNode; double minFrequency = XContentMapValues.nodeDoubleValue(frequencyFilter.remove("min"), 0); double maxFrequency = XContentMapValues.nodeDoubleValue(frequencyFilter.remove("max"), Integer.MAX_VALUE); int minSegmentSize = XContentMapValues.nodeIntegerValue(frequencyFilter.remove("min_segment_size"), 0); builder.fielddataFrequencyFilter(minFrequency, maxFrequency, minSegmentSize); DocumentMapperParser.checkNoRemainingFields(propName, frequencyFilter, parserContext.indexVersionCreated()); iterator.remove(); } else if (propName.equals("index_prefixes")) { Map<?, ?> indexPrefix = (Map<?, ?>) propNode; int minChars = XContentMapValues.nodeIntegerValue(indexPrefix.remove("min_chars"), Defaults.INDEX_PREFIX_MIN_CHARS); int maxChars = XContentMapValues.nodeIntegerValue(indexPrefix.remove("max_chars"), Defaults.INDEX_PREFIX_MAX_CHARS); builder.indexPrefixes(minChars, maxChars); DocumentMapperParser.checkNoRemainingFields(propName, indexPrefix, parserContext.indexVersionCreated()); iterator.remove(); } else if (propName.equals("index_phrases")) { builder.indexPhrases(XContentMapValues.nodeBooleanValue(propNode, "index_phrases")); iterator.remove(); } } return builder; }