org.joda.time.Duration Java Examples
The following examples show how to use
org.joda.time.Duration.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example #1
Source File: CacheFactory.java From freehealth-connector with GNU Affero General Public License v3.0 | 6 votes |
public static <K, V> Cache<K, V> newInstance(CacheFactory.CacheType cacheType, String cacheName, CacheInformation.ExpiryType expiryType, Duration expiryDuration) { try { ConfigurableFactoryHelper helper; switch(cacheType) { case MEMORY: helper = new ConfigurableFactoryHelper("be.ehealth.technicalconnector.cache.memory.impl", DEFAULT_CACHE_MEMORY_IMPL); break; case PERSISTENT: helper = new ConfigurableFactoryHelper("be.ehealth.technicalconnector.cache.persistent.impl", DEFAULT_CACHE_PERSISTENT_IMPL); break; default: throw new IllegalArgumentException("Unsupported cache type [" + cacheType + "]"); } Map<String, Object> options = new HashMap(); options.put("cacheName", cacheName); if (expiryType != null) { options.put("cacheExpiryType", CacheFactory.ExpiryType.valueOf(expiryType.name())); } options.put("cacheExpiryDuration", expiryDuration); return (Cache)helper.getImplementation(options); } catch (TechnicalConnectorException var6) { throw new ConfigurationException(var6); } }
Example #2
Source File: SampleTest.java From beam with Apache License 2.0 | 6 votes |
@Test @Category(NeedsRunner.class) public void testSampleAnyZero() { PCollection<Integer> input = pipeline.apply( Create.timestamped(ImmutableList.of(tv(0), tv(1), tv(2), tv(3), tv(4), tv(5))) .withCoder(BigEndianIntegerCoder.of())); PCollection<Integer> output = input .apply(Window.into(FixedWindows.of(Duration.standardSeconds(3)))) .apply(Sample.any(0)); PAssert.that(output) .inWindow(new IntervalWindow(new Instant(0), Duration.standardSeconds(3))) .satisfies(new VerifyCorrectSample<>(0, EMPTY)); PAssert.that(output) .inWindow(new IntervalWindow(new Instant(3000), Duration.standardSeconds(3))) .satisfies(new VerifyCorrectSample<>(0, EMPTY)); pipeline.run(); }
Example #3
Source File: PeriodicStaleCheckingCacheStrategy.java From alchemy with MIT License | 6 votes |
private void invalidateAllIfStale(CachingContext context) { if (!lock.compareAndSet(false, true)) { return; } try { final Duration elapsed = new Duration(lastSync, DateTime.now()); if (!elapsed.isLongerThan(period)) { return; } lastSync = DateTime.now(); } finally { lock.set(false); } if (context.checkIfAnyStale()) { context.invalidateAll(true); } }
Example #4
Source File: DistinctTest.java From beam with Apache License 2.0 | 6 votes |
@Test public void testBuild_Windowing() { final PCollection<String> dataset = TestUtils.createMockDataset(TypeDescriptors.strings()); final PCollection<String> uniq = Distinct.of(dataset) .windowBy(FixedWindows.of(org.joda.time.Duration.standardHours(1))) .triggeredBy(DefaultTrigger.of()) .accumulationMode(AccumulationMode.DISCARDING_FIRED_PANES) .output(); final Distinct distinct = (Distinct) TestUtils.getProducer(uniq); assertTrue(distinct.getWindow().isPresent()); @SuppressWarnings("unchecked") final WindowDesc<?> windowDesc = WindowDesc.of((Window) distinct.getWindow().get()); assertEquals( FixedWindows.of(org.joda.time.Duration.standardHours(1)), windowDesc.getWindowFn()); assertEquals(DefaultTrigger.of(), windowDesc.getTrigger()); }
Example #5
Source File: VirtualRoutingResource.java From cloudstack with Apache License 2.0 | 6 votes |
public boolean configure(final String name, final Map<String, Object> params) throws ConfigurationException { _name = name; _params = params; String value = (String)params.get("ssh.sleep"); _sleep = NumbersUtil.parseInt(value, 10) * 1000; value = (String)params.get("ssh.retry"); _retry = NumbersUtil.parseInt(value, 36); value = (String)params.get("ssh.port"); _port = NumbersUtil.parseInt(value, 3922); value = (String)params.get("router.aggregation.command.each.timeout"); _eachTimeout = Duration.standardSeconds(NumbersUtil.parseInt(value, (int)VRScripts.VR_SCRIPT_EXEC_TIMEOUT.getStandardSeconds())); if (s_logger.isDebugEnabled()){ s_logger.debug("The router.aggregation.command.each.timeout in seconds is set to " + _eachTimeout.getStandardSeconds()); } if (_vrDeployer == null) { throw new ConfigurationException("Unable to find the resource for VirtualRouterDeployer!"); } _vrAggregateCommandsSet = new HashMap<>(); return true; }
Example #6
Source File: TestPubsubSignal.java From beam with Apache License 2.0 | 6 votes |
/** Wait for a success signal for {@code duration}. */ public void waitForSuccess(Duration duration) throws IOException { SubscriptionPath resultSubscriptionPath = PubsubClient.subscriptionPathFromName( pipelineOptions.getProject(), "result-subscription-" + String.valueOf(ThreadLocalRandom.current().nextLong())); pubsub.createSubscription( resultTopicPath, resultSubscriptionPath, (int) duration.getStandardSeconds()); String result = pollForResultForDuration(resultSubscriptionPath, duration); if (!RESULT_SUCCESS_MESSAGE.equals(result)) { throw new AssertionError(result); } }
Example #7
Source File: TestPubsub.java From beam with Apache License 2.0 | 6 votes |
/** * Check if topics exist. * * @param project GCP project identifier. * @param timeoutDuration Joda duration that sets a period of time before checking times out. */ public void checkIfAnySubscriptionExists(String project, Duration timeoutDuration) throws InterruptedException, IllegalArgumentException, IOException, TimeoutException { if (timeoutDuration.getMillis() <= 0) { throw new IllegalArgumentException(String.format("timeoutDuration should be greater than 0")); } DateTime startTime = new DateTime(); int sizeOfSubscriptionList = 0; while (sizeOfSubscriptionList == 0 && Seconds.secondsBetween(new DateTime(), startTime).getSeconds() < timeoutDuration.toStandardSeconds().getSeconds()) { // Sleep 1 sec Thread.sleep(1000); sizeOfSubscriptionList = listSubscriptions(projectPathFromPath(String.format("projects/%s", project)), topicPath()) .size(); } if (sizeOfSubscriptionList > 0) { return; } else { throw new TimeoutException("Timed out when checking if topics exist for " + topicPath()); } }
Example #8
Source File: WindowingTest.java From beam with Apache License 2.0 | 6 votes |
@Test @Category(ValidatesRunner.class) public void testNonPartitioningWindowing() { PCollection<String> input = p.apply( Create.timestamped( TimestampedValue.of("a", new Instant(1)), TimestampedValue.of("a", new Instant(7)), TimestampedValue.of("b", new Instant(8)))); PCollection<String> output = input.apply(new WindowedCount(SlidingWindows.of(new Duration(10)).every(new Duration(5)))); PAssert.that(output) .containsInAnyOrder( output("a", 1, 1, -5, 5), output("a", 2, 5, 0, 10), output("a", 1, 10, 5, 15), output("b", 1, 8, 0, 10), output("b", 1, 10, 5, 15)); p.run(); }
Example #9
Source File: CachingResolverTest.java From monsoon with BSD 3-Clause "New" or "Revised" License | 6 votes |
@Test public void getTuplesTest() throws Exception { final CompletableFuture<Collection<ResolverTuple>> fut = new CompletableFuture<>(); final List<ResolverTuple> RESULT = new ArrayList<>(); when(mockedAsyncResolver.getTuples()).thenReturn((CompletableFuture) fut); final Collection<ResolverTuple> entries; try (CachingResolver cr = new CachingResolver(mockedAsyncResolver, Duration.standardMinutes(10), Duration.standardMinutes(60))) { fut.complete(RESULT); Thread.sleep(5000); // Racy test, but this should work in most cases to allow callbacks to propagate. entries = cr.getTuples(); } assertSame(RESULT, entries); }
Example #10
Source File: CountByKeyTest.java From beam with Apache License 2.0 | 6 votes |
@Test public void testBuild() { final PCollection<String> dataset = TestUtils.createMockDataset(TypeDescriptors.strings()); final FixedWindows windowing = FixedWindows.of(org.joda.time.Duration.standardHours(1)); final DefaultTrigger trigger = DefaultTrigger.of(); final PCollection<KV<String, Long>> counted = CountByKey.named("CountByKey1") .of(dataset) .keyBy(s -> s) .windowBy(windowing) .triggeredBy(trigger) .discardingFiredPanes() .withAllowedLateness(Duration.millis(1000)) .output(); final CountByKey count = (CountByKey) TestUtils.getProducer(counted); assertTrue(count.getName().isPresent()); assertEquals("CountByKey1", count.getName().get()); assertNotNull(count.getKeyExtractor()); assertTrue(count.getWindow().isPresent()); final WindowDesc<?> desc = WindowDesc.of((Window<?>) count.getWindow().get()); assertEquals(windowing, desc.getWindowFn()); assertEquals(trigger, desc.getTrigger()); assertEquals(AccumulationMode.DISCARDING_FIRED_PANES, desc.getAccumulationMode()); assertEquals(Duration.millis(1000), desc.getAllowedLateness()); }
Example #11
Source File: WindowedOperatorTest.java From attic-apex-malhar with Apache License 2.0 | 6 votes |
@Test public void testKeyedAccumulation() { KeyedWindowedOperatorImpl<String, Long, MutableLong, Long> windowedOperator = createDefaultKeyedWindowedOperator(false); windowedOperator.setWindowOption(new WindowOption.TimeWindows(Duration.millis(1000))); windowedOperator.setup(testMeta.operatorContext); windowedOperator.beginWindow(1); windowedOperator.processTuple(new Tuple.TimestampedTuple<>(BASE + 100L, new KeyValPair<>("a", 2L))); windowedOperator.processTuple(new Tuple.TimestampedTuple<>(BASE + 200L, new KeyValPair<>("a", 3L))); windowedOperator.processTuple(new Tuple.TimestampedTuple<>(BASE + 300L, new KeyValPair<>("b", 4L))); windowedOperator.processTuple(new Tuple.TimestampedTuple<>(BASE + 150L, new KeyValPair<>("b", 5L))); windowedOperator.endWindow(); Assert.assertEquals(1, keyedDataStorage.size()); Assert.assertEquals(5L, keyedDataStorage.get(new Window.TimeWindow(BASE, 1000), "a").longValue()); Assert.assertEquals(9L, keyedDataStorage.get(new Window.TimeWindow(BASE, 1000), "b").longValue()); windowedOperator.teardown(); }
Example #12
Source File: PeriodicImpulse.java From beam with Apache License 2.0 | 6 votes |
@Override public PCollection<Instant> expand(PBegin input) { PCollection<Instant> result = input .apply( Create.<PeriodicSequence.SequenceDefinition>of( new PeriodicSequence.SequenceDefinition( startTimestamp, stopTimestamp, fireInterval))) .apply(PeriodicSequence.create()); if (this.applyWindowing) { result = result.apply( Window.<Instant>into(FixedWindows.of(Duration.millis(fireInterval.getMillis())))); } return result; }
Example #13
Source File: SamzaPipelineResult.java From beam with Apache License 2.0 | 6 votes |
@Override public State waitUntilFinish(@Nullable Duration duration) { try { if (duration == null) { runner.waitForFinish(); } else { runner.waitForFinish(java.time.Duration.ofMillis(duration.getMillis())); } } catch (Exception e) { throw new Pipeline.PipelineExecutionException(e); } final StateInfo stateInfo = getStateInfo(); if (listener != null && (stateInfo.state == State.DONE || stateInfo.state == State.FAILED)) { listener.onFinish(); } if (stateInfo.state == State.FAILED) { throw stateInfo.error; } LOG.info("Pipeline finished. Final state: {}", stateInfo.state); return stateInfo.state; }
Example #14
Source File: FluentBackoff.java From beam with Apache License 2.0 | 5 votes |
private FluentBackoff( double exponent, Duration initialBackoff, Duration maxBackoff, Duration maxCumulativeBackoff, int maxRetries) { this.exponent = exponent; this.initialBackoff = initialBackoff; this.maxBackoff = maxBackoff; this.maxRetries = maxRetries; this.maxCumulativeBackoff = maxCumulativeBackoff; }
Example #15
Source File: DeduplicateTest.java From beam with Apache License 2.0 | 5 votes |
@Test @Category({NeedsRunner.class, UsesTestStream.class}) public void testInDifferentWindows() { Instant base = new Instant(0); TestStream<String> values = TestStream.create(StringUtf8Coder.of()) .advanceWatermarkTo(base) .addElements( TimestampedValue.of("k1", base), TimestampedValue.of("k2", base.plus(Duration.standardSeconds(10))), TimestampedValue.of("k3", base.plus(Duration.standardSeconds(20))), TimestampedValue.of("k1", base.plus(Duration.standardSeconds(30))), TimestampedValue.of("k2", base.plus(Duration.standardSeconds(40))), TimestampedValue.of("k3", base.plus(Duration.standardSeconds(50))), TimestampedValue.of("k4", base.plus(Duration.standardSeconds(60))), TimestampedValue.of("k5", base.plus(Duration.standardSeconds(70))), TimestampedValue.of("k6", base.plus(Duration.standardSeconds(80)))) .advanceWatermarkToInfinity(); PCollection<String> distinctValues = p.apply(values) .apply(Window.into(FixedWindows.of(Duration.standardSeconds(30)))) .apply(Deduplicate.values()); PAssert.that(distinctValues) .inWindow(new IntervalWindow(base, base.plus(Duration.standardSeconds(30)))) .containsInAnyOrder("k1", "k2", "k3"); PAssert.that(distinctValues) .inWindow( new IntervalWindow( base.plus(Duration.standardSeconds(30)), base.plus(Duration.standardSeconds(60)))) .containsInAnyOrder("k1", "k2", "k3"); PAssert.that(distinctValues) .inWindow( new IntervalWindow( base.plus(Duration.standardSeconds(60)), base.plus(Duration.standardSeconds(90)))) .containsInAnyOrder("k4", "k5", "k6"); p.run(); }
Example #16
Source File: StatefulTeamScoreTest.java From deployment-examples with MIT License | 5 votes |
/** * Tests that {@link UpdateTeamScoreFn} {@link org.apache.beam.sdk.transforms.DoFn} outputs * correctly for multiple teams. */ @Test public void testScoreUpdatesPerTeam() { TestStream<KV<String, GameActionInfo>> createEvents = TestStream.create(KvCoder.of(StringUtf8Coder.of(), AvroCoder.of(GameActionInfo.class))) .advanceWatermarkTo(baseTime) .addElements( event(TestUser.RED_ONE, 50, Duration.standardSeconds(10)), event(TestUser.RED_TWO, 50, Duration.standardSeconds(20)), event(TestUser.BLUE_ONE, 70, Duration.standardSeconds(30)), event(TestUser.BLUE_TWO, 80, Duration.standardSeconds(40)), event(TestUser.BLUE_TWO, 50, Duration.standardSeconds(50))) .advanceWatermarkToInfinity(); PCollection<KV<String, Integer>> teamScores = p.apply(createEvents).apply(ParDo.of(new UpdateTeamScoreFn(100))); String redTeam = TestUser.RED_ONE.getTeam(); String blueTeam = TestUser.BLUE_ONE.getTeam(); PAssert.that(teamScores) .inWindow(GlobalWindow.INSTANCE) .containsInAnyOrder(KV.of(redTeam, 100), KV.of(blueTeam, 150), KV.of(blueTeam, 200)); p.run().waitUntilFinish(); }
Example #17
Source File: OrFinallyStateMachineTest.java From beam with Apache License 2.0 | 5 votes |
/** * Tests that if the first trigger rewinds to be non-finished in the merged window, then it * becomes the currently active trigger again, with real triggers. */ @Test public void testShouldFireAfterMerge() throws Exception { tester = TriggerStateMachineTester.forTrigger( AfterEachStateMachine.inOrder( AfterPaneStateMachine.elementCountAtLeast(5) .orFinally(AfterWatermarkStateMachine.pastEndOfWindow()), RepeatedlyStateMachine.forever(AfterPaneStateMachine.elementCountAtLeast(1))), Sessions.withGapDuration(Duration.millis(10))); // Finished the orFinally in the first window tester.injectElements(1); IntervalWindow firstWindow = new IntervalWindow(new Instant(1), new Instant(11)); assertFalse(tester.shouldFire(firstWindow)); tester.advanceInputWatermark(new Instant(11)); assertTrue(tester.shouldFire(firstWindow)); tester.fireIfShouldFire(firstWindow); // Set up second window where it is not done tester.injectElements(5); IntervalWindow secondWindow = new IntervalWindow(new Instant(5), new Instant(15)); assertFalse(tester.shouldFire(secondWindow)); // Merge them, if the merged window were on the second trigger, it would be ready tester.mergeWindows(); IntervalWindow mergedWindow = new IntervalWindow(new Instant(1), new Instant(15)); assertFalse(tester.shouldFire(mergedWindow)); // Now adding 3 more makes the main trigger ready to fire tester.injectElements(1, 2, 3, 4, 5); tester.mergeWindows(); assertTrue(tester.shouldFire(mergedWindow)); }
Example #18
Source File: MovingBoundaryTimeBucketAssignerTest.java From attic-apex-malhar with Apache License 2.0 | 5 votes |
@Test public void testNumBuckets() { testMeta.timeBucketAssigner.setExpireBefore(Duration.standardHours(1)); testMeta.timeBucketAssigner.setBucketSpan(Duration.standardMinutes(30)); testMeta.timeBucketAssigner.setup(testMeta.mockManagedStateContext); Assert.assertEquals("num buckets", 2, testMeta.timeBucketAssigner.getNumBuckets()); testMeta.timeBucketAssigner.teardown(); }
Example #19
Source File: FixedWindowsTest.java From beam with Apache License 2.0 | 5 votes |
@Test public void testFixedOffsetWindow() throws Exception { Map<IntervalWindow, Set<String>> expected = new HashMap<>(); expected.put(new IntervalWindow(new Instant(-5), new Instant(5)), set(1, 2)); expected.put(new IntervalWindow(new Instant(5), new Instant(15)), set(5, 9, 10, 11)); expected.put(new IntervalWindow(new Instant(95), new Instant(105)), set(100)); assertEquals( expected, runWindowFn( FixedWindows.of(new Duration(10)).withOffset(new Duration(5)), Arrays.asList(1L, 2L, 5L, 9L, 10L, 11L, 100L))); }
Example #20
Source File: BigQuerySinkTest.java From feast with Apache License 2.0 | 5 votes |
@Test public void uniqueJobIdPerWindow() { TestStream<FeatureRow> featureRowTestStream = TestStream.create(ProtoCoder.of(FeatureRow.class)) .advanceWatermarkTo(Instant.now()) .addElements(generateRow("myproject/fs")) .addElements(generateRow("myproject/fs")) .advanceWatermarkTo(Instant.now().plus(Duration.standardSeconds(10))) .addElements(generateRow("myproject/fs")) .addElements(generateRow("myproject/fs")) .advanceWatermarkToInfinity(); FeatureSink sink = makeSink( ValueProvider.StaticValueProvider.of(bigQuery), p.apply( "StaticSpecs", Create.of( ImmutableMap.of( FeatureSetReference.of(spec.getProject(), spec.getName(), 1), spec)))); p.apply(featureRowTestStream).apply(sink.writer()); p.run(); assertThat(jobService.getAllJobs().size(), is(2)); assertThat( jobService.getAllJobs().stream() .map(j -> j.getJobReference().getJobId()) .distinct() .count(), is(2L)); }
Example #21
Source File: DnsMessageTransportTest.java From nomulus with Apache License 2.0 | 5 votes |
@Test public void testTimeoutReceivingResponse() throws Exception { InputStream mockInputStream = mock(InputStream.class); when(mockInputStream.read()).thenThrow(new SocketTimeoutException("testing")); when(mockSocket.getInputStream()).thenReturn(mockInputStream); when(mockSocket.getOutputStream()).thenReturn(new ByteArrayOutputStream()); Duration testTimeout = Duration.standardSeconds(1); DnsMessageTransport resolver = new DnsMessageTransport(mockFactory, UPDATE_HOST, testTimeout); Message expectedQuery = new Message(); assertThrows(SocketTimeoutException.class, () -> resolver.send(expectedQuery)); verify(mockSocket).setSoTimeout((int) testTimeout.getMillis()); }
Example #22
Source File: VirtualRoutingResource.java From cloudstack with Apache License 2.0 | 5 votes |
public boolean configureHostParams(final Map<String, String> params) { if (_params.get("router.aggregation.command.each.timeout") != null) { String value = (String)params.get("router.aggregation.command.each.timeout"); _eachTimeout = Duration.standardSeconds(NumbersUtil.parseLong(value, 600)); if (s_logger.isDebugEnabled()){ s_logger.debug("The router.aggregation.command.each.timeout in seconds is set to " + _eachTimeout.getStandardSeconds()); } } return true; }
Example #23
Source File: WatermarkPolicyTest.java From beam with Apache License 2.0 | 5 votes |
@Test public void shouldAdvanceWatermarkToNowWithProcessingTimePolicy() { WatermarkPolicy policy = WatermarkPolicyFactory.withProcessingTimePolicy().createWatermarkPolicy(); mockStatic(Instant.class); Instant time1 = NOW.minus(Duration.standardSeconds(5)); Instant time2 = NOW.minus(Duration.standardSeconds(4)); when(Instant.now()).thenReturn(time1).thenReturn(time2); assertThat(policy.getWatermark()).isEqualTo(time1); assertThat(policy.getWatermark()).isEqualTo(time2); }
Example #24
Source File: SetupOteCommandTest.java From nomulus with Apache License 2.0 | 5 votes |
/** Verify TLD creation. */ private void verifyTldCreation( String tldName, String roidSuffix, TldState tldState, boolean isEarlyAccess) { Registry registry = Registry.get(tldName); assertThat(registry).isNotNull(); assertThat(registry.getRoidSuffix()).isEqualTo(roidSuffix); assertThat(registry.getTldState(DateTime.now(UTC))).isEqualTo(tldState); assertThat(registry.getDnsWriters()).containsExactly("VoidDnsWriter"); assertThat(registry.getPremiumList()).isNotNull(); assertThat(registry.getPremiumList().getName()).isEqualTo("default_sandbox_list"); assertThat(registry.getAddGracePeriodLength()).isEqualTo(Duration.standardMinutes(60)); assertThat(registry.getRedemptionGracePeriodLength()).isEqualTo(Duration.standardMinutes(10)); assertThat(registry.getPendingDeleteLength()).isEqualTo(Duration.standardMinutes(5)); ImmutableSortedMap<DateTime, Money> eapFeeSchedule = registry.getEapFeeScheduleAsMap(); if (!isEarlyAccess) { assertThat(eapFeeSchedule) .isEqualTo(ImmutableSortedMap.of(new DateTime(0), Money.of(CurrencyUnit.USD, 0))); } else { assertThat(eapFeeSchedule) .isEqualTo( ImmutableSortedMap.of( new DateTime(0), Money.of(CurrencyUnit.USD, 0), DateTime.parse("2018-03-01T00:00:00Z"), Money.of(CurrencyUnit.USD, 100), DateTime.parse("2030-03-01T00:00:00Z"), Money.of(CurrencyUnit.USD, 0))); } }
Example #25
Source File: DataCatalogBigQueryIT.java From beam with Apache License 2.0 | 5 votes |
@Test public void testRead() throws Exception { TableReference bqTable = bigQuery.tableReference(); // Streaming inserts do not work with DIRECT_READ mode, there is a several hour lag. PCollection<Row> data = writePipeline.apply(Create.of(row(1, "name1"), row(2, "name2"), row(3, "name3"))); data.apply( BigQueryIO.<Row>write() .withSchema(BigQueryUtils.toTableSchema(ID_NAME_SCHEMA)) .withFormatFunction(BigQueryUtils.toTableRow()) .withMethod(Method.FILE_LOADS) .to(bqTable)); writePipeline.run().waitUntilFinish(Duration.standardMinutes(2)); String tableId = String.format( "bigquery.`table`.`%s`.`%s`.`%s`", bqTable.getProjectId(), bqTable.getDatasetId(), bqTable.getTableId()); readPipeline .getOptions() .as(BeamSqlPipelineOptions.class) .setPlannerName(queryPlanner.getCanonicalName()); try (DataCatalogTableProvider tableProvider = DataCatalogTableProvider.create( readPipeline.getOptions().as(DataCatalogPipelineOptions.class))) { PCollection<Row> result = readPipeline.apply( "query", SqlTransform.query("SELECT id, name FROM " + tableId) .withDefaultTableProvider("datacatalog", tableProvider)); PAssert.that(result).containsInAnyOrder(row(1, "name1"), row(2, "name2"), row(3, "name3")); readPipeline.run().waitUntilFinish(Duration.standardMinutes(2)); } }
Example #26
Source File: EppResourceUtilsTest.java From nomulus with Apache License 2.0 | 5 votes |
@Test public void testLoadAtPointInTime_beforeCreated_returnsNull() { clock.advanceOneMilli(); // Don't save a commit log, we shouldn't need one. HostResource host = persistResource( newHostResource("ns1.cat.tld").asBuilder() .setCreationTimeForTest(clock.nowUtc()) .build()); assertThat(loadAtPointInTime(host, clock.nowUtc().minus(Duration.millis(1))).now()).isNull(); }
Example #27
Source File: DynamoDBFibonacciRetryerTest.java From emr-dynamodb-connector with Apache License 2.0 | 5 votes |
@Test(expected = RuntimeException.class) public void testRetryableASEException2() throws Exception { AmazonServiceException ase = new AmazonServiceException("Test"); ase.setErrorCode("ArbitRetryableException"); ase.setStatusCode(503); when(call.call()).thenThrow(ase); DynamoDBFibonacciRetryer retryer = new DynamoDBFibonacciRetryer(Duration.standardSeconds(10)); try { retryer.runWithRetry(call, null, null); } finally { verify(call, atLeast(2)).call(); verify(call, atMost(15)).call(); } }
Example #28
Source File: CollectHistoryServer.java From monsoon with BSD 3-Clause "New" or "Revised" License | 5 votes |
@Override public Stream<Collection<CollectHistory.NamedEvaluation>> evaluate(Map<String, ? extends TimeSeriesMetricExpression> query, DateTime begin, Duration stepSize) { LOG.log(Level.FINE, "request received({0}, {1})", new Object[]{begin, stepSize}); Stream<Collection<CollectHistory.NamedEvaluation>> result = history.evaluate(query, begin, stepSize); LOG.log(Level.FINE, "returning({0}, {1}) => {2}", new Object[]{begin, stepSize, result}); return result; }
Example #29
Source File: ReduceFnRunnerTest.java From beam with Apache License 2.0 | 5 votes |
@Test public void noEmptyPanesFinalIfNonEmpty() throws Exception { ReduceFnTester<Integer, Iterable<Integer>, IntervalWindow> tester = ReduceFnTester.nonCombining( WindowingStrategy.of(FixedWindows.of(Duration.millis(10))) .withTrigger( Repeatedly.forever( AfterFirst.of( AfterPane.elementCountAtLeast(2), AfterWatermark.pastEndOfWindow()))) .withMode(AccumulationMode.ACCUMULATING_FIRED_PANES) .withAllowedLateness(Duration.millis(100)) .withTimestampCombiner(TimestampCombiner.EARLIEST) .withClosingBehavior(ClosingBehavior.FIRE_IF_NON_EMPTY)); tester.advanceInputWatermark(new Instant(0)); tester.injectElements( TimestampedValue.of(1, new Instant(1)), TimestampedValue.of(2, new Instant(2))); tester.advanceInputWatermark(new Instant(20)); tester.advanceInputWatermark(new Instant(250)); List<WindowedValue<Iterable<Integer>>> output = tester.extractOutput(); assertThat( output, contains( // Trigger with 2 elements isSingleWindowedValue(containsInAnyOrder(1, 2), 1, 0, 10), // Trigger for the empty on time pane isSingleWindowedValue(containsInAnyOrder(1, 2), 9, 0, 10))); }
Example #30
Source File: AwsSessionCredentialProvider.java From cloudbreak with Apache License 2.0 | 5 votes |
private AwsSessionCredentials checkExpirationTime(AwsSessionCredentials sessionCredentials) { final Date expirationTime = sessionCredentials.getExpiration(); if (expirationTimeCheck && expirationTime != null) { DateTime expirationDateTime = new DateTime(expirationTime); DateTime nowDateTime = new DateTime(new Date()); if (expirationDateTime.isAfter(nowDateTime.toDate().getTime()) || new Duration(expirationDateTime, nowDateTime).getStandardMinutes() < forceTokenRefreshTimeInMin) { LOGGER.debug("Force retrieving session credentials because of expiration time is too close."); sessionCredentials = awsSessionCredentialClient.retrieveSessionCredentials(awsCredentialView); } } return sessionCredentials; }