org.apache.flink.runtime.taskmanager.NoOpTaskActions Java Examples
The following examples show how to use
org.apache.flink.runtime.taskmanager.NoOpTaskActions.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example #1
Source File: StreamNetworkBenchmarkEnvironment.java From Flink-CEPplus with Apache License 2.0 | 6 votes |
protected ResultPartitionWriter createResultPartition( JobID jobId, ResultPartitionID partitionId, NetworkEnvironment environment, int channels) throws Exception { ResultPartition resultPartition = new ResultPartition( "sender task", new NoOpTaskActions(), jobId, partitionId, ResultPartitionType.PIPELINED_BOUNDED, channels, 1, environment.getResultPartitionManager(), new NoOpResultPartitionConsumableNotifier(), ioManager, false); environment.setupPartition(resultPartition); return resultPartition; }
Example #2
Source File: ResultPartitionTest.java From flink with Apache License 2.0 | 6 votes |
/** * Tests {@link ResultPartition#addBufferConsumer(BufferConsumer, int)} on a working partition. * * @param partitionType the result partition type to set up */ private void testAddOnPartition(final ResultPartitionType partitionType) throws Exception { ResultPartitionConsumableNotifier notifier = mock(ResultPartitionConsumableNotifier.class); JobID jobId = new JobID(); TaskActions taskActions = new NoOpTaskActions(); ResultPartitionWriter consumableNotifyingPartitionWriter = createConsumableNotifyingResultPartitionWriter( partitionType, taskActions, jobId, notifier); BufferConsumer bufferConsumer = createFilledBufferConsumer(BufferBuilderTestUtils.BUFFER_SIZE); try { // partition.add() adds the bufferConsumer without recycling it (if not spilling) consumableNotifyingPartitionWriter.addBufferConsumer(bufferConsumer, 0); assertFalse("bufferConsumer should not be recycled (still in the queue)", bufferConsumer.isRecycled()); } finally { if (!bufferConsumer.isRecycled()) { bufferConsumer.close(); } // should have been notified for pipelined partitions if (partitionType.isPipelined()) { verify(notifier, times(1)) .notifyPartitionConsumable(eq(jobId), eq(consumableNotifyingPartitionWriter.getPartitionId()), eq(taskActions)); } } }
Example #3
Source File: StreamNetworkBenchmarkEnvironment.java From flink with Apache License 2.0 | 6 votes |
protected ResultPartitionWriter createResultPartition( JobID jobId, ResultPartitionID partitionId, NettyShuffleEnvironment environment, int channels) throws Exception { ResultPartitionWriter resultPartitionWriter = new ResultPartitionBuilder() .setResultPartitionId(partitionId) .setResultPartitionType(ResultPartitionType.PIPELINED_BOUNDED) .setNumberOfSubpartitions(channels) .setResultPartitionManager(environment.getResultPartitionManager()) .setupBufferPoolFactoryFromNettyShuffleEnvironment(environment) .build(); ResultPartitionWriter consumableNotifyingPartitionWriter = new ConsumableNotifyingResultPartitionWriterDecorator( new NoOpTaskActions(), jobId, resultPartitionWriter, new NoOpResultPartitionConsumableNotifier()); consumableNotifyingPartitionWriter.setup(); return consumableNotifyingPartitionWriter; }
Example #4
Source File: ResultPartitionTest.java From flink with Apache License 2.0 | 6 votes |
/** * Tests {@link ResultPartition#addBufferConsumer(BufferConsumer, int)} on a working partition. * * @param partitionType the result partition type to set up */ private void testAddOnPartition(final ResultPartitionType partitionType) throws Exception { ResultPartitionConsumableNotifier notifier = mock(ResultPartitionConsumableNotifier.class); JobID jobId = new JobID(); TaskActions taskActions = new NoOpTaskActions(); ResultPartitionWriter consumableNotifyingPartitionWriter = createConsumableNotifyingResultPartitionWriter( partitionType, taskActions, jobId, notifier); BufferConsumer bufferConsumer = createFilledFinishedBufferConsumer(BufferBuilderTestUtils.BUFFER_SIZE); try { // partition.add() adds the bufferConsumer without recycling it (if not spilling) consumableNotifyingPartitionWriter.addBufferConsumer(bufferConsumer, 0); assertFalse("bufferConsumer should not be recycled (still in the queue)", bufferConsumer.isRecycled()); } finally { if (!bufferConsumer.isRecycled()) { bufferConsumer.close(); } // should have been notified for pipelined partitions if (partitionType.isPipelined()) { verify(notifier, times(1)) .notifyPartitionConsumable(eq(jobId), eq(consumableNotifyingPartitionWriter.getPartitionId()), eq(taskActions)); } } }
Example #5
Source File: StreamNetworkBenchmarkEnvironment.java From flink with Apache License 2.0 | 6 votes |
public ResultPartitionWriter createResultPartitionWriter(int partitionIndex) throws Exception { ResultPartitionWriter resultPartitionWriter = new ResultPartitionBuilder() .setResultPartitionId(partitionIds[partitionIndex]) .setResultPartitionType(ResultPartitionType.PIPELINED_BOUNDED) .setNumberOfSubpartitions(channels) .setResultPartitionManager(senderEnv.getResultPartitionManager()) .setupBufferPoolFactoryFromNettyShuffleEnvironment(senderEnv) .build(); ResultPartitionWriter consumableNotifyingPartitionWriter = new ConsumableNotifyingResultPartitionWriterDecorator( new NoOpTaskActions(), jobId, resultPartitionWriter, new NoOpResultPartitionConsumableNotifier()); consumableNotifyingPartitionWriter.setup(); return consumableNotifyingPartitionWriter; }
Example #6
Source File: ResultPartitionTest.java From flink with Apache License 2.0 | 5 votes |
/** * Tests {@link ResultPartition#addBufferConsumer} on a partition which has already finished. * * @param partitionType the result partition type to set up */ private void testAddOnFinishedPartition(final ResultPartitionType partitionType) throws Exception { BufferConsumer bufferConsumer = createFilledBufferConsumer(BufferBuilderTestUtils.BUFFER_SIZE); ResultPartitionConsumableNotifier notifier = mock(ResultPartitionConsumableNotifier.class); JobID jobId = new JobID(); TaskActions taskActions = new NoOpTaskActions(); ResultPartitionWriter consumableNotifyingPartitionWriter = createConsumableNotifyingResultPartitionWriter( partitionType, taskActions, jobId, notifier); try { consumableNotifyingPartitionWriter.finish(); reset(notifier); // partition.add() should fail consumableNotifyingPartitionWriter.addBufferConsumer(bufferConsumer, 0); Assert.fail("exception expected"); } catch (IllegalStateException e) { // expected => ignored } finally { if (!bufferConsumer.isRecycled()) { bufferConsumer.close(); Assert.fail("bufferConsumer not recycled"); } // should not have notified either verify(notifier, never()).notifyPartitionConsumable( eq(jobId), eq(consumableNotifyingPartitionWriter.getPartitionId()), eq(taskActions)); } }
Example #7
Source File: ResultPartitionTest.java From flink with Apache License 2.0 | 5 votes |
/** * Tests {@link ResultPartition#addBufferConsumer} on a partition which has already been released. * * @param partitionType the result partition type to set up */ private void testAddOnReleasedPartition(final ResultPartitionType partitionType) throws Exception { BufferConsumer bufferConsumer = createFilledBufferConsumer(BufferBuilderTestUtils.BUFFER_SIZE); ResultPartitionConsumableNotifier notifier = mock(ResultPartitionConsumableNotifier.class); JobID jobId = new JobID(); TaskActions taskActions = new NoOpTaskActions(); ResultPartition partition = partitionType == ResultPartitionType.BLOCKING ? createPartition(partitionType, fileChannelManager) : createPartition(partitionType); ResultPartitionWriter consumableNotifyingPartitionWriter = ConsumableNotifyingResultPartitionWriterDecorator.decorate( Collections.singleton(PartitionTestUtils.createPartitionDeploymentDescriptor(partitionType)), new ResultPartitionWriter[] {partition}, taskActions, jobId, notifier)[0]; try { partition.release(); // partition.add() silently drops the bufferConsumer but recycles it consumableNotifyingPartitionWriter.addBufferConsumer(bufferConsumer, 0); assertTrue(partition.isReleased()); } finally { if (!bufferConsumer.isRecycled()) { bufferConsumer.close(); Assert.fail("bufferConsumer not recycled"); } // should not have notified either verify(notifier, never()).notifyPartitionConsumable(eq(jobId), eq(partition.getPartitionId()), eq(taskActions)); } }
Example #8
Source File: RecordWriterTest.java From flink with Apache License 2.0 | 5 votes |
/** * Tests that the RecordWriter is available iif the respective LocalBufferPool has at-least one available buffer. */ @Test public void testIsAvailableOrNot() throws Exception { // setup final NetworkBufferPool globalPool = new NetworkBufferPool(10, 128, 2); final BufferPool localPool = globalPool.createBufferPool(1, 1, null, 1, Integer.MAX_VALUE); final ResultPartitionWriter resultPartition = new ResultPartitionBuilder() .setBufferPoolFactory(p -> localPool) .build(); resultPartition.setup(); final ResultPartitionWriter partitionWrapper = new ConsumableNotifyingResultPartitionWriterDecorator( new NoOpTaskActions(), new JobID(), resultPartition, new NoOpResultPartitionConsumableNotifier()); final RecordWriter recordWriter = createRecordWriter(partitionWrapper); try { // record writer is available because of initial available global pool assertTrue(recordWriter.getAvailableFuture().isDone()); // request one buffer from the local pool to make it unavailable afterwards final BufferBuilder bufferBuilder = resultPartition.getBufferBuilder(0); assertNotNull(bufferBuilder); assertFalse(recordWriter.getAvailableFuture().isDone()); // recycle the buffer to make the local pool available again final Buffer buffer = BufferBuilderTestUtils.buildSingleBuffer(bufferBuilder); buffer.recycleBuffer(); assertTrue(recordWriter.getAvailableFuture().isDone()); assertEquals(recordWriter.AVAILABLE, recordWriter.getAvailableFuture()); } finally { localPool.lazyDestroy(); globalPool.destroy(); } }
Example #9
Source File: ResultPartitionTest.java From flink with Apache License 2.0 | 5 votes |
/** * Tests {@link ResultPartition#addBufferConsumer} on a partition which has already finished. * * @param partitionType the result partition type to set up */ private void testAddOnFinishedPartition(final ResultPartitionType partitionType) throws Exception { BufferConsumer bufferConsumer = createFilledFinishedBufferConsumer(BufferBuilderTestUtils.BUFFER_SIZE); ResultPartitionConsumableNotifier notifier = mock(ResultPartitionConsumableNotifier.class); JobID jobId = new JobID(); TaskActions taskActions = new NoOpTaskActions(); ResultPartitionWriter consumableNotifyingPartitionWriter = createConsumableNotifyingResultPartitionWriter( partitionType, taskActions, jobId, notifier); try { consumableNotifyingPartitionWriter.finish(); reset(notifier); // partition.add() should fail consumableNotifyingPartitionWriter.addBufferConsumer(bufferConsumer, 0); Assert.fail("exception expected"); } catch (IllegalStateException e) { // expected => ignored } finally { if (!bufferConsumer.isRecycled()) { bufferConsumer.close(); Assert.fail("bufferConsumer not recycled"); } // should not have notified either verify(notifier, never()).notifyPartitionConsumable( eq(jobId), eq(consumableNotifyingPartitionWriter.getPartitionId()), eq(taskActions)); } }
Example #10
Source File: ResultPartitionTest.java From flink with Apache License 2.0 | 5 votes |
/** * Tests {@link ResultPartition#addBufferConsumer} on a partition which has already been released. * * @param partitionType the result partition type to set up */ private void testAddOnReleasedPartition(final ResultPartitionType partitionType) throws Exception { BufferConsumer bufferConsumer = createFilledFinishedBufferConsumer(BufferBuilderTestUtils.BUFFER_SIZE); ResultPartitionConsumableNotifier notifier = mock(ResultPartitionConsumableNotifier.class); JobID jobId = new JobID(); TaskActions taskActions = new NoOpTaskActions(); ResultPartition partition = partitionType == ResultPartitionType.BLOCKING ? createPartition(partitionType, fileChannelManager) : createPartition(partitionType); ResultPartitionWriter consumableNotifyingPartitionWriter = ConsumableNotifyingResultPartitionWriterDecorator.decorate( Collections.singleton(PartitionTestUtils.createPartitionDeploymentDescriptor(partitionType)), new ResultPartitionWriter[] {partition}, taskActions, jobId, notifier)[0]; try { partition.release(); // partition.add() silently drops the bufferConsumer but recycles it consumableNotifyingPartitionWriter.addBufferConsumer(bufferConsumer, 0); assertTrue(partition.isReleased()); } finally { if (!bufferConsumer.isRecycled()) { bufferConsumer.close(); Assert.fail("bufferConsumer not recycled"); } // should not have notified either verify(notifier, never()).notifyPartitionConsumable(eq(jobId), eq(partition.getPartitionId()), eq(taskActions)); } }
Example #11
Source File: SpillableSubpartitionTest.java From Flink-CEPplus with Apache License 2.0 | 4 votes |
/** * Tests a fix for FLINK-12544. * * @see <a href="https://issues.apache.org/jira/browse/FLINK-12544">FLINK-12544</a> */ @Test public void testConcurrentRequestAndReleaseMemory() throws Exception { final ExecutorService executor = Executors.newFixedThreadPool(2); final NetworkBufferPool networkBufferPool = new NetworkBufferPool(10, 32); try { final CountDownLatch blockLatch = new CountDownLatch(1); final CountDownLatch doneLatch = new CountDownLatch(1); final IOManager ioManager = new IOManagerAsyncWithCountDownLatch(blockLatch, doneLatch); final ResultPartitionWithCountDownLatch partition = new ResultPartitionWithCountDownLatch( "Test", new NoOpTaskActions(), new JobID(), new ResultPartitionID(), ResultPartitionType.BLOCKING, 1, 1, new ResultPartitionManager(), new NoOpResultPartitionConsumableNotifier(), ioManager, true, doneLatch, blockLatch); final BufferPool bufferPool = networkBufferPool.createBufferPool(1, 1, Optional.of(partition)); partition.registerBufferPool(bufferPool); final BufferBuilder firstBuffer = bufferPool.requestBufferBuilderBlocking(); partition.addBufferConsumer(firstBuffer.createBufferConsumer(), 0); // Finishes the buffer consumer which could be recycled during SpillableSubpartition#releaseMemory firstBuffer.finish(); Future<Void> future = executor.submit(new Callable<Void>() { @Override public Void call() throws Exception { //Occupies the lock in SpillableSubpartition#releaseMemory, trying to get the lock in LocalBufferPool#recycle partition.releaseMemory(1); return null; } }); final CompletableFuture<?> firstCallFuture = partition.getFirstCallFuture(); firstCallFuture.thenRunAsync(() -> { try { // There are no available buffers in pool, so trigger release memory in SpillableSubpartition. // Occupies the lock in LocalBufferPool, and trying to get the lock in SpillableSubpartition. BufferBuilder secondBuffer = bufferPool.requestBufferBuilderBlocking(); assertThat(firstBuffer, is(equalTo(secondBuffer))); } catch (IOException | InterruptedException ex) { fail("Should not throw any exceptions!"); } }, executor); future.get(); } finally { networkBufferPool.destroyAllBufferPools(); networkBufferPool.destroy(); executor.shutdown(); } }
Example #12
Source File: StreamNetworkBenchmarkEnvironment.java From Flink-CEPplus with Apache License 2.0 | 4 votes |
private InputGate createInputGate( JobID jobId, IntermediateDataSetID dataSetID, ExecutionAttemptID executionAttemptID, final TaskManagerLocation senderLocation, NetworkEnvironment environment, final int channels) throws IOException { InputGate[] gates = new InputGate[channels]; for (int channel = 0; channel < channels; ++channel) { int finalChannel = channel; InputChannelDeploymentDescriptor[] channelDescriptors = Arrays.stream(partitionIds) .map(partitionId -> new InputChannelDeploymentDescriptor( partitionId, localMode ? ResultPartitionLocation.createLocal() : ResultPartitionLocation.createRemote(new ConnectionID(senderLocation, finalChannel)))) .toArray(InputChannelDeploymentDescriptor[]::new); final InputGateDeploymentDescriptor gateDescriptor = new InputGateDeploymentDescriptor( dataSetID, ResultPartitionType.PIPELINED_BOUNDED, channel, channelDescriptors); SingleInputGate gate = SingleInputGate.create( "receiving task[" + channel + "]", jobId, executionAttemptID, gateDescriptor, environment, new NoOpTaskActions(), UnregisteredMetricGroups.createUnregisteredTaskMetricGroup().getIOMetricGroup()); environment.setupInputGate(gate); gates[channel] = gate; } if (channels > 1) { return new UnionInputGate(gates); } else { return gates[0]; } }
Example #13
Source File: RecordWriterTest.java From flink with Apache License 2.0 | 4 votes |
@Test public void testIdleTime() throws IOException, InterruptedException { // setup final NetworkBufferPool globalPool = new NetworkBufferPool(10, 128, 2); final BufferPool localPool = globalPool.createBufferPool(1, 1, null, 1, Integer.MAX_VALUE); final ResultPartitionWriter resultPartition = new ResultPartitionBuilder() .setBufferPoolFactory(p -> localPool) .build(); resultPartition.setup(); final ResultPartitionWriter partitionWrapper = new ConsumableNotifyingResultPartitionWriterDecorator( new NoOpTaskActions(), new JobID(), resultPartition, new NoOpResultPartitionConsumableNotifier()); final RecordWriter recordWriter = createRecordWriter(partitionWrapper); BufferBuilder builder = recordWriter.requestNewBufferBuilder(0); BufferBuilderTestUtils.fillBufferBuilder(builder, 1).finish(); ResultSubpartitionView readView = resultPartition.getSubpartition(0).createReadView(new NoOpBufferAvailablityListener()); Buffer buffer = readView.getNextBuffer().buffer(); // idle time is zero when there is buffer available. assertEquals(0, recordWriter.getIdleTimeMsPerSecond().getCount()); CountDownLatch syncLock = new CountDownLatch(1); AtomicReference<BufferBuilder> asyncRequestResult = new AtomicReference<>(); final Thread requestThread = new Thread(new Runnable() { @Override public void run() { try { // notify that the request thread start to run. syncLock.countDown(); // wait for buffer. asyncRequestResult.set(recordWriter.requestNewBufferBuilder(0)); } catch (Exception e) { } } }); requestThread.start(); // wait until request thread start to run. syncLock.await(); Thread.sleep(10); //recycle the buffer buffer.recycleBuffer(); requestThread.join(); assertThat(recordWriter.getIdleTimeMsPerSecond().getCount(), Matchers.greaterThan(0L)); assertNotNull(asyncRequestResult.get()); }