Java Code Examples for backtype.storm.Config#setDebug()
The following examples show how to use
backtype.storm.Config#setDebug() .
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example 1
Source File: Runner.java From storm-camel-example with Apache License 2.0 | 6 votes |
public static final void main(final String[] args) throws Exception { final ApplicationContext applicationContext = new ClassPathXmlApplicationContext("applicationContext.xml"); final JmsProvider jmsQueueProvider = new SpringJmsProvider(applicationContext, "jmsConnectionFactory", "notificationQueue"); final TopologyBuilder topologyBuilder = new TopologyBuilder(); final JmsBolt jmsBolt = new JmsBolt(); jmsBolt.setJmsProvider(jmsQueueProvider); jmsBolt.setJmsMessageProducer((session, input) -> { final String json = "{\"word\":\"" + input.getString(0) + "\", \"count\":" + String.valueOf(input.getInteger(1)) + "}"; return session.createTextMessage(json); }); topologyBuilder.setSpout("wordGenerator", new RandomWordFeeder()); topologyBuilder.setBolt("counter", new WordCounterBolt()).shuffleGrouping("wordGenerator"); topologyBuilder.setBolt("jmsBolt", jmsBolt).shuffleGrouping("counter"); final Config config = new Config(); config.setDebug(false); final LocalCluster cluster = new LocalCluster(); cluster.submitTopology("word-count", config, topologyBuilder.createTopology()); }
Example 2
Source File: ExampleRunner.java From flowmix with Apache License 2.0 | 6 votes |
public void run() { StormTopology topology = new FlowmixBuilder() .setFlowLoader(new SimpleFlowLoaderSpout(provider.getFlows(), 60000)) .setEventsLoader(new MockEventGeneratorSpout(getMockEvents(), 10)) .setOutputBolt(new PrinterBolt()) .setParallelismHint(6) .create() .createTopology(); Config conf = new Config(); conf.setNumWorkers(20); conf.setMaxSpoutPending(5000); conf.setDebug(false); conf.registerSerialization(BaseEvent.class, EventSerializer.class); conf.setSkipMissingKryoRegistrations(false); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("example-topology", conf, topology); }
Example 3
Source File: KafkaThroughput.java From flink-perf with Apache License 2.0 | 5 votes |
public static void main(String[] args) throws AlreadyAliveException, InvalidTopologyException, UnknownHostException, InterruptedException { final ParameterTool pt = ParameterTool.fromArgs(args); TopologyBuilder builder = new TopologyBuilder(); BrokerHosts hosts = new ZkHosts(pt.getRequired("zookeeper")); SpoutConfig spoutConfig = new SpoutConfig(hosts, pt.getRequired("topic"), "/" + pt.getRequired("topic"), UUID.randomUUID().toString()); spoutConfig.scheme = new SchemeAsMultiScheme(new StringScheme()); KafkaSpout kafkaSpout = new KafkaSpout(spoutConfig); builder.setSpout("source", kafkaSpout, pt.getInt("sourceParallelism")); builder.setBolt("sink", new Throughput.Sink(pt), pt.getInt("sinkParallelism")).noneGrouping("source"); Config conf = new Config(); conf.setDebug(false); if (!pt.has("local")) { conf.setNumWorkers(pt.getInt("par", 2)); StormSubmitter.submitTopologyWithProgressBar("kafka-spout-"+pt.get("name", "no_name"), conf, builder.createTopology()); } else { conf.setMaxTaskParallelism(pt.getInt("par", 2)); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("kafka-spout", conf, builder.createTopology()); Thread.sleep(300000); cluster.shutdown(); } }
Example 4
Source File: CounterTopology.java From storm-kafka-examples with Apache License 2.0 | 5 votes |
/** * @param args * http://www.programcreek.com/java-api-examples/index.php?api=storm.kafka.KafkaSpout */ public static void main(String[] args) { try{ //设置喷发节点并分配并发数,该并发数将会控制该对象在集群中的线程数(6个) String zkhost = "wxb-1:2181,wxb-2:2181,wxb-3:2181"; String topic = "order"; String groupId = "id"; int spoutNum = 3; int boltNum = 1; ZkHosts zkHosts = new ZkHosts(zkhost);//kafaka所在的zookeeper SpoutConfig spoutConfig = new SpoutConfig(zkHosts, topic, "/order", groupId); // create /order /id spoutConfig.scheme = new SchemeAsMultiScheme(new StringScheme()); KafkaSpout kafkaSpout = new KafkaSpout(spoutConfig); TopologyBuilder builder = new TopologyBuilder(); builder.setSpout("spout", kafkaSpout, spoutNum); builder.setBolt("check", new CheckOrderBolt(), boltNum).shuffleGrouping("spout"); builder.setBolt("counter", new CounterBolt(),boltNum).shuffleGrouping("check"); Config config = new Config(); config.setDebug(true); if(args!=null && args.length > 0) { config.setNumWorkers(2); StormSubmitter.submitTopology(args[0], config, builder.createTopology()); } else { config.setMaxTaskParallelism(2); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("Wordcount-Topology", config, builder.createTopology()); Thread.sleep(500000); cluster.shutdown(); } }catch (Exception e) { e.printStackTrace(); } }
Example 5
Source File: LocalTopologyRunner.java From C2-Github-commit-count with MIT License | 5 votes |
public static void main(String[] args) { TopologyBuilder builder = new TopologyBuilder(); builder.setSpout("commit-feed-listener", new CommitFeedListener()); builder .setBolt("email-extractor", new EmailExtractor()) .shuffleGrouping("commit-feed-listener"); builder .setBolt("email-counter", new EmailCounter()) .fieldsGrouping("email-extractor", new Fields("email")); Config config = new Config(); config.setDebug(true); StormTopology topology = builder.createTopology(); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("github-commit-count-topology", config, topology); Utils.sleep(TEN_MINUTES); cluster.killTopology("github-commit-count-topology"); cluster.shutdown(); }
Example 6
Source File: FirstStoryDetection.java From first-stories-twitter with MIT License | 5 votes |
private static Config createTopologyConfiguration(Properties prop, boolean localMode) { Config conf = new Config(); List<String> dprcServers = new ArrayList<String>(); dprcServers.add("localhost"); conf.put(Config.DRPC_SERVERS, dprcServers); conf.put(Config.DRPC_PORT, 3772); if (!localMode) conf.put(Config.STORM_CLUSTER_MODE, new String("distributed")); conf.put("UNIQUE_WORDS_EXPECTED", prop.getProperty("UNIQUE_WORDS_EXPECTED")); conf.put("PATH_TO_OOV_FILE", prop.getProperty("PATH_TO_OOV_FILE")); conf.put("L", prop.getProperty("L")); conf.put("BucketsParallelism", prop.getProperty("BucketsParallelism")); conf.put("k", prop.getProperty("k")); conf.put("QUEUE_SIZE", prop.getProperty("QUEUE_SIZE")); List<String> countAggKeepFields = new ArrayList<String>(); countAggKeepFields.add("tweet_obj"); countAggKeepFields.add("coltweet_obj"); conf.put("countAggKeepFields", countAggKeepFields); conf.put("THRESHOLD", prop.getProperty("THRESHOLD")); conf.put("RECENT_TWEETS_TO_COMPARE_WITH", prop.getProperty("RECENT_TWEETS_TO_COMPARE_WITH")); conf.setDebug(false); conf.setNumWorkers(Integer.valueOf((String) prop .get("NUMBER_OF_WORKERS"))); conf.setMaxSpoutPending(50000000); return conf; }
Example 7
Source File: StormAbstractCloudLiveTest.java From brooklyn-library with Apache License 2.0 | 5 votes |
public boolean submitTopology(StormTopology stormTopology, String topologyName, int numOfWorkers, boolean debug, long timeoutMs) { if (log.isDebugEnabled()) log.debug("Connecting to NimbusClient: {}", nimbus.getConfig(Storm.NIMBUS_HOSTNAME)); Config conf = new Config(); conf.setDebug(debug); conf.setNumWorkers(numOfWorkers); // TODO - confirm this creats the JAR correctly String jar = createJar( new File(Os.mergePaths(ResourceUtils.create(this).getClassLoaderDir(), "org/apache/brooklyn/entity/messaging/storm/topologies")), "org/apache/brooklyn/entity/messaging/storm/"); System.setProperty("storm.jar", jar); long startMs = System.currentTimeMillis(); long endMs = (timeoutMs == -1) ? Long.MAX_VALUE : (startMs + timeoutMs); long currentTime = startMs; Throwable lastError = null; int attempt = 0; while (currentTime <= endMs) { currentTime = System.currentTimeMillis(); if (attempt != 0) Time.sleep(Duration.ONE_SECOND); if (log.isTraceEnabled()) log.trace("trying connection to {} at time {}", nimbus.getConfig(Storm.NIMBUS_HOSTNAME), currentTime); try { StormSubmitter.submitTopology(topologyName, conf, stormTopology); return true; } catch (Exception e) { if (shouldRetryOn(e)) { if (log.isDebugEnabled()) log.debug("Attempt {} failed connecting to {} ({})", new Object[] {attempt + 1, nimbus.getConfig(Storm.NIMBUS_HOSTNAME), e.getMessage()}); lastError = e; } else { throw Throwables.propagate(e); } } attempt++; } log.warn("unable to connect to Nimbus client: ", lastError); Assert.fail(); return false; }
Example 8
Source File: TransactionalWordsTest.java From jstorm with Apache License 2.0 | 5 votes |
@Test public void test_transaction_word() { try { MemoryTransactionalSpout spout = new MemoryTransactionalSpout(DATA, new Fields("word"), PARTITION_TAKE_PER_BATCH); TransactionalTopologyBuilder builder = new TransactionalTopologyBuilder("top-n-words", "spout", spout, 2); builder.setBolt("count", new KeyedCountUpdater(), 5).fieldsGrouping("spout", new Fields("word")); builder.setBolt("bucketize", new Bucketize()).shuffleGrouping("count"); builder.setBolt("buckets", new BucketCountUpdater(), 5).fieldsGrouping("bucketize", new Fields("bucket")); LocalCluster cluster = new LocalCluster(); Config config = new Config(); config.setDebug(true); config.setMaxSpoutPending(3); cluster.submitTopology("top-n-topology", config, builder.buildTopology()); JStormUtils.sleepMs(60 * 1000); assertEquals(false, outOfOrder.get() ); assertNotSame(0, receiveCounter1.get()); assertNotSame(0, receiveCounter2.get()); //cluster.killTopology("top-n-topology"); //cluster.shutdown(); } catch (Exception e) { e.printStackTrace(); Assert.fail("Failed to run simple transaction"); } }
Example 9
Source File: SpeedViolationTopology.java From ignite-book-code-samples with GNU General Public License v3.0 | 5 votes |
public static void main(String[] args) throws Exception { if (getProperties() == null || getProperties().isEmpty()) { System.out.println("Property file <ignite-storm.property> is not found or empty"); return; } // Ignite Stream Ibolt final StormStreamer<String, String> stormStreamer = new StormStreamer<>(); stormStreamer.setAutoFlushFrequency(10L); stormStreamer.setAllowOverwrite(true); stormStreamer.setCacheName(getProperties().getProperty("cache.name")); stormStreamer.setIgniteTupleField(getProperties().getProperty("tuple.name")); stormStreamer.setIgniteConfigFile(getProperties().getProperty("ignite.spring.xml")); TopologyBuilder builder = new TopologyBuilder(); builder.setSpout("spout", new FileSourceSpout(), 1); builder.setBolt("limit", new SpeedLimitBolt(), 1).fieldsGrouping("spout", new Fields("trafficLog")); // set ignite bolt builder.setBolt("ignite-bolt", stormStreamer, STORM_EXECUTORS).shuffleGrouping("limit"); Config conf = new Config(); conf.setDebug(false); conf.setMaxTaskParallelism(1); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("speed-violation", conf, builder.createTopology()); Thread.sleep(10000); cluster.shutdown(); }
Example 10
Source File: Throughput.java From flink-perf with Apache License 2.0 | 4 votes |
public static void main(String[] args) throws Exception { ParameterTool pt = ParameterTool.fromArgs(args); int par = pt.getInt("para"); TopologyBuilder builder = new TopologyBuilder(); builder.setSpout("source0", new Generator(pt), pt.getInt("sourceParallelism")); int i = 0; for(; i < pt.getInt("repartitions", 1) - 1;i++) { System.out.println("adding source"+i+" --> source"+(i+1)); builder.setBolt("source"+(i+1), new RepartPassThroughBolt(pt), pt.getInt("sinkParallelism")).fieldsGrouping("source" + i, new Fields("id")); } System.out.println("adding final source"+i+" --> sink"); builder.setBolt("sink", new Sink(pt), pt.getInt("sinkParallelism")).fieldsGrouping("source"+i, new Fields("id")); Config conf = new Config(); conf.setDebug(false); //System.exit(1); if(pt.has("ft") || pt.has("maxPending")) { conf.setMaxSpoutPending(pt.getInt("maxPending", 1000)); } if (!pt.has("local")) { conf.setNumWorkers(par); StormSubmitter.submitTopologyWithProgressBar("throughput-"+pt.get("name", "no_name"), conf, builder.createTopology()); } else { conf.setMaxTaskParallelism(par); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("throughput", conf, builder.createTopology()); Thread.sleep(300000); cluster.shutdown(); } }
Example 11
Source File: StormTopologySubmitter.java From samoa with Apache License 2.0 | 4 votes |
public static void main(String[] args) throws IOException{ Properties props = StormSamoaUtils.getProperties(); String uploadedJarLocation = props.getProperty(StormJarSubmitter.UPLOADED_JAR_LOCATION_KEY); if(uploadedJarLocation == null){ logger.error("Invalid properties file. It must have key {}", StormJarSubmitter.UPLOADED_JAR_LOCATION_KEY); return; } List<String> tmpArgs = new ArrayList<String>(Arrays.asList(args)); int numWorkers = StormSamoaUtils.numWorkers(tmpArgs); args = tmpArgs.toArray(new String[0]); StormTopology stormTopo = StormSamoaUtils.argsToTopology(args); Config conf = new Config(); conf.putAll(Utils.readStormConfig()); conf.putAll(Utils.readCommandLineOpts()); conf.setDebug(false); conf.setNumWorkers(numWorkers); String profilerOption = props.getProperty(StormTopologySubmitter.YJP_OPTIONS_KEY); if(profilerOption != null){ String topoWorkerChildOpts = (String) conf.get(Config.TOPOLOGY_WORKER_CHILDOPTS); StringBuilder optionBuilder = new StringBuilder(); if(topoWorkerChildOpts != null){ optionBuilder.append(topoWorkerChildOpts); optionBuilder.append(' '); } optionBuilder.append(profilerOption); conf.put(Config.TOPOLOGY_WORKER_CHILDOPTS, optionBuilder.toString()); } Map<String, Object> myConfigMap = new HashMap<String, Object>(conf); StringWriter out = new StringWriter(); try { JSONValue.writeJSONString(myConfigMap, out); } catch (IOException e) { System.out.println("Error in writing JSONString"); e.printStackTrace(); return; } Config config = new Config(); config.putAll(Utils.readStormConfig()); String nimbusHost = (String) config.get(Config.NIMBUS_HOST); NimbusClient nc = new NimbusClient(nimbusHost); String topologyName = stormTopo.getTopologyName(); try { System.out.println("Submitting topology with name: " + topologyName); nc.getClient().submitTopology(topologyName, uploadedJarLocation, out.toString(), stormTopo.getStormBuilder().createTopology()); System.out.println(topologyName + " is successfully submitted"); } catch (AlreadyAliveException aae) { System.out.println("Fail to submit " + topologyName + "\nError message: " + aae.get_msg()); } catch (InvalidTopologyException ite) { System.out.println("Invalid topology for " + topologyName); ite.printStackTrace(); } catch (TException te) { System.out.println("Texception for " + topologyName); te.printStackTrace(); } }
Example 12
Source File: StormDoTask.java From samoa with Apache License 2.0 | 4 votes |
/** * The main method. * * @param args the arguments */ public static void main(String[] args) { List<String> tmpArgs = new ArrayList<String>(Arrays.asList(args)); boolean isLocal = isLocal(tmpArgs); int numWorker = StormSamoaUtils.numWorkers(tmpArgs); args = tmpArgs.toArray(new String[0]); //convert the arguments into Storm topology StormTopology stormTopo = StormSamoaUtils.argsToTopology(args); String topologyName = stormTopo.getTopologyName(); Config conf = new Config(); conf.putAll(Utils.readStormConfig()); conf.setDebug(false); if(isLocal){ //local mode conf.setMaxTaskParallelism(numWorker); backtype.storm.LocalCluster cluster = new backtype.storm.LocalCluster(); cluster.submitTopology(topologyName , conf, stormTopo.getStormBuilder().createTopology()); backtype.storm.utils.Utils.sleep(600*1000); cluster.killTopology(topologyName); cluster.shutdown(); }else{ //cluster mode conf.setNumWorkers(numWorker); try { backtype.storm.StormSubmitter.submitTopology(topologyName, conf, stormTopo.getStormBuilder().createTopology()); } catch (backtype.storm.generated.AlreadyAliveException ale) { ale.printStackTrace(); } catch (backtype.storm.generated.InvalidTopologyException ite) { ite.printStackTrace(); } } }
Example 13
Source File: SampleTopology.java From aws-big-data-blog with Apache License 2.0 | 4 votes |
public static void main(String[] args) throws IllegalArgumentException, KeeperException, InterruptedException, AlreadyAliveException, InvalidTopologyException, IOException { String propertiesFile = null; String mode = null; if (args.length != 2) { printUsageAndExit(); } else { propertiesFile = args[0]; mode = args[1]; } configure(propertiesFile); final KinesisSpoutConfig config = new KinesisSpoutConfig(streamName, zookeeperEndpoint).withZookeeperPrefix(zookeeperPrefix) .withInitialPositionInStream(initialPositionInStream) .withRegion(Regions.fromName(regionName)); final KinesisSpout spout = new KinesisSpout(config, new CustomCredentialsProviderChain(), new ClientConfiguration()); TopologyBuilder builder = new TopologyBuilder(); LOG.info("Using Kinesis stream: " + config.getStreamName()); // Using number of shards as the parallelism hint for the spout. builder.setSpout("Kinesis", spout, 2); builder.setBolt("Parse", new ParseReferrerBolt(), 6).shuffleGrouping("Kinesis"); builder.setBolt("Count", new RollingCountBolt(5, 2,elasticCacheRedisEndpoint), 6).fieldsGrouping("Parse", new Fields("referrer")); //builder.setBolt("Count", new CountReferrerBolt(), 12).fieldsGrouping("Parse", new Fields("referrer")); Config topoConf = new Config(); topoConf.setFallBackOnJavaSerialization(true); topoConf.setDebug(false); if (mode.equals("LocalMode")) { LOG.info("Starting sample storm topology in LocalMode ..."); new LocalCluster().submitTopology("test_spout", topoConf, builder.createTopology()); } else if (mode.equals("RemoteMode")) { topoConf.setNumWorkers(1); topoConf.setMaxSpoutPending(5000); LOG.info("Submitting sample topology " + topologyName + " to remote cluster."); StormSubmitter.submitTopology(topologyName, topoConf, builder.createTopology()); } else { printUsageAndExit(); } }
Example 14
Source File: SkewedRollingTopWords.java From jstorm with Apache License 2.0 | 4 votes |
private static Config createTopologyConfiguration() { Config conf = new Config(); conf.setDebug(true); return conf; }
Example 15
Source File: StormDoTask.java From incubator-samoa with Apache License 2.0 | 4 votes |
/** * The main method. * * @param args * the arguments */ public static void main(String[] args) { List<String> tmpArgs = new ArrayList<String>(Arrays.asList(args)); boolean isLocal = isLocal(tmpArgs); int numWorker = StormSamoaUtils.numWorkers(tmpArgs); args = tmpArgs.toArray(new String[0]); // convert the arguments into Storm topology StormTopology stormTopo = StormSamoaUtils.argsToTopology(args); String topologyName = stormTopo.getTopologyName(); Config conf = new Config(); conf.putAll(Utils.readStormConfig()); conf.setDebug(false); if (isLocal) { // local mode conf.setMaxTaskParallelism(numWorker); backtype.storm.LocalCluster cluster = new backtype.storm.LocalCluster(); cluster.submitTopology(topologyName, conf, stormTopo.getStormBuilder().createTopology()); backtype.storm.utils.Utils.sleep(600 * 1000); cluster.killTopology(topologyName); cluster.shutdown(); } else { // cluster mode conf.setNumWorkers(numWorker); try { backtype.storm.StormSubmitter.submitTopology(topologyName, conf, stormTopo.getStormBuilder().createTopology()); } catch (backtype.storm.generated.AlreadyAliveException ale) { ale.printStackTrace(); } catch (backtype.storm.generated.InvalidTopologyException ite) { ite.printStackTrace(); } } }
Example 16
Source File: HdfsTopology.java From storm-kafka-examples with Apache License 2.0 | 4 votes |
public static void main(String[] args) { try{ String zkhost = "wxb-1:2181,wxb-2:2181,wxb-3:2181"; String topic = "order"; String groupId = "id"; int spoutNum = 3; int boltNum = 1; ZkHosts zkHosts = new ZkHosts(zkhost);//kafaka所在的zookeeper SpoutConfig spoutConfig = new SpoutConfig(zkHosts, topic, "/order", groupId); // create /order /id spoutConfig.scheme = new SchemeAsMultiScheme(new StringScheme()); KafkaSpout kafkaSpout = new KafkaSpout(spoutConfig); // HDFS bolt // use "|" instead of "," for field delimiter RecordFormat format = new DelimitedRecordFormat() .withFieldDelimiter("|"); // sync the filesystem after every 1k tuples SyncPolicy syncPolicy = new CountSyncPolicy(1000); // rotate files when they reach 5MB FileRotationPolicy rotationPolicy = new FileSizeRotationPolicy(5.0f, FileSizeRotationPolicy.Units.MB); // FileRotationPolicy rotationPolicy = new TimedRotationPolicy(1.0f, TimedRotationPolicy.TimeUnit.MINUTES); FileNameFormat fileNameFormat = new DefaultFileNameFormat() .withPath("/tmp/").withPrefix("order_").withExtension(".log"); HdfsBolt hdfsBolt = new HdfsBolt() .withFsUrl("hdfs://wxb-1:8020") .withFileNameFormat(fileNameFormat) .withRecordFormat(format) .withRotationPolicy(rotationPolicy) .withSyncPolicy(syncPolicy); TopologyBuilder builder = new TopologyBuilder(); builder.setSpout("spout", kafkaSpout, spoutNum); builder.setBolt("check", new CheckOrderBolt(), boltNum).shuffleGrouping("spout"); builder.setBolt("counter", new CounterBolt(),boltNum).shuffleGrouping("check"); builder.setBolt("hdfs", hdfsBolt,boltNum).shuffleGrouping("counter"); Config config = new Config(); config.setDebug(true); if(args!=null && args.length > 0) { config.setNumWorkers(2); StormSubmitter.submitTopology(args[0], config, builder.createTopology()); } else { config.setMaxTaskParallelism(2); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("Wordcount-Topology", config, builder.createTopology()); Thread.sleep(500000); cluster.shutdown(); } }catch (Exception e) { e.printStackTrace(); } }
Example 17
Source File: WordCountTopology.java From ignite-book-code-samples with GNU General Public License v3.0 | 4 votes |
public static void main(String[] args) throws Exception { // Ignite Stream Ibolt final StormStreamer<String, String> stormStreamer = new StormStreamer<>(); stormStreamer.setAutoFlushFrequency(10L); stormStreamer.setAllowOverwrite(true); stormStreamer.setCacheName("testCache"); stormStreamer.setIgniteTupleField("ignite"); stormStreamer.setIgniteConfigFile("/Users/shamim/Development/workshop/assembla/ignite-book/chapters/chapter-cep-storm/src/main/resources/example-ignite.xml"); //Used to build the topology TopologyBuilder builder = new TopologyBuilder(); //Add the spout, with a name of 'spout' //and parallelism hint of 5 executors builder.setSpout("spout", new RandomSentenceSpout(), 5); //Add the SplitSentence bolt, with a name of 'split' //and parallelism hint of 8 executors //shufflegrouping subscribes to the spout, and equally distributes //tuples (sentences) across instances of the SplitSentence bolt builder.setBolt("split", new SplitSentence(), 8).shuffleGrouping("spout"); //Add the counter, with a name of 'count' //and parallelism hint of 12 executors //fieldsgrouping subscribes to the split bolt, and //ensures that the same word is sent to the same instance (group by field 'word') builder.setBolt("count", new WordCount(), 12).fieldsGrouping("split", new Fields("word")); // set ignite bolt builder.setBolt("ignite-bolt", stormStreamer,STORM_EXECUTORS).shuffleGrouping("count"); //new configuration Config conf = new Config(); //Set to false to disable debug information // when running in production mode. conf.setDebug(false); //If there are arguments, we are running on a cluster if (args != null && args.length > 0) { //parallelism hint to set the number of workers conf.setNumWorkers(3); //submit the topology StormSubmitter.submitTopology(args[0], conf, builder.createTopology()); } //Otherwise, we are running locally else { //Cap the maximum number of executors that can be spawned //for a component to 3 conf.setMaxTaskParallelism(3); //LocalCluster is used to run locally LocalCluster cluster = new LocalCluster(); //submit the topology cluster.submitTopology("word-count", conf, builder.createTopology()); //sleep Thread.sleep(10000); //shut down the cluster cluster.shutdown(); } }
Example 18
Source File: WordCountTopologyNode.java From flink-perf with Apache License 2.0 | 3 votes |
public static void main(String[] args) throws Exception { TopologyBuilder builder = new TopologyBuilder(); builder.setSpout("spout", new RandomSentence(), 5); builder.setBolt("split", new SplitSentence(), 8).shuffleGrouping("spout"); builder.setBolt("count", new WordCount(), 12).fieldsGrouping("split", new Fields("word")); Config conf = new Config(); conf.setDebug(true); if (args != null && args.length > 0) { conf.setNumWorkers(3); StormSubmitter.submitTopologyWithProgressBar(args[0], conf, builder.createTopology()); } else { conf.setMaxTaskParallelism(3); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("word-count", conf, builder.createTopology()); Thread.sleep(10000); cluster.shutdown(); } }
Example 19
Source File: WordCountTopology.java From flink-perf with Apache License 2.0 | 3 votes |
public static void main(String[] args) throws Exception { TopologyBuilder builder = new TopologyBuilder(); builder.setSpout("spout", new RandomSentenceSpout(), 5); builder.setBolt("split", new SplitSentence(), 8).shuffleGrouping("spout"); builder.setBolt("count", new WordCount(), 12).fieldsGrouping("split", new Fields("word")); Config conf = new Config(); conf.setDebug(true); if (args != null && args.length > 0) { conf.setNumWorkers(3); StormSubmitter.submitTopologyWithProgressBar(args[0], conf, builder.createTopology()); } else { conf.setMaxTaskParallelism(3); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("word-count", conf, builder.createTopology()); Thread.sleep(10000); cluster.shutdown(); } }
Example 20
Source File: TridentForwardThroughput.java From flink-perf with Apache License 2.0 | 3 votes |
public static void main(String[] args) throws Exception { ParameterTool pt = ParameterTool.fromArgs(args); int par = pt.getInt("para"); TridentTopology topology = new TridentTopology(); Stream sourceStream = topology.newStream("source", new Generator(pt)).parallelismHint(pt.getInt("sourceParallelism")); sourceStream.localOrShuffle().each(FIELDS, new Sink(pt), new Fields("dontcare")); Config conf = new Config(); conf.setDebug(false); // conf.setMaxSpoutPending(pt.getInt("maxPending", 1000)); //System.exit(1); if (!pt.has("local")) { conf.setNumWorkers(par); StormSubmitter.submitTopologyWithProgressBar("forward-throughput-"+pt.get("name", "no_name"), conf, topology.build()); } else { conf.setMaxTaskParallelism(par); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("forward-throughput", conf, topology.build()); Thread.sleep(300000); cluster.shutdown(); } }