Java Code Examples for org.springframework.kafka.test.utils.KafkaTestUtils#getSingleRecord()
The following examples show how to use
org.springframework.kafka.test.utils.KafkaTestUtils#getSingleRecord() .
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example 1
Source File: DeserializtionErrorHandlerByBinderTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 7 votes |
@Test @Ignore public void test() { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("foos"); template.sendDefault(1, 7, "hello"); Map<String, Object> consumerProps = KafkaTestUtils.consumerProps("foobar", "false", embeddedKafka); consumerProps.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); DefaultKafkaConsumerFactory<String, String> cf = new DefaultKafkaConsumerFactory<>( consumerProps); Consumer<String, String> consumer1 = cf.createConsumer(); embeddedKafka.consumeFromAnEmbeddedTopic(consumer1, "error.foos.foobar-group"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer1, "error.foos.foobar-group"); assertThat(cr.value()).isEqualTo("hello"); assertThat(cr.partition()).isEqualTo(0); // Ensuring that the deserialization was indeed done by the binder verify(conversionDelegate).deserializeOnInbound(any(Class.class), any(KStream.class)); }
Example 2
Source File: MultipleFunctionsInSameAppTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 6 votes |
private void receiveAndValidate(String in, String... out) throws InterruptedException { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>(senderProps); try { KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic(in); template.sendDefault("coffee"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, out[0]); assertThat(cr.value().contains("coffee")).isTrue(); template.sendDefault("electronics"); cr = KafkaTestUtils.getSingleRecord(consumer, out[1]); assertThat(cr.value().contains("electronics")).isTrue(); Assert.isTrue(countDownLatch.await(5, TimeUnit.SECONDS), "Analyze (BiConsumer) method didn't receive all the expected records"); } finally { pf.destroy(); } }
Example 3
Source File: KafkaStreamsBinderWordCountBranchesFunctionTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 6 votes |
private void receiveAndValidate(ConfigurableApplicationContext context) throws Exception { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>(senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("words"); template.sendDefault("english"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, "counts"); assertThat(cr.value().contains("\"word\":\"english\",\"count\":1")).isTrue(); template.sendDefault("french"); template.sendDefault("french"); cr = KafkaTestUtils.getSingleRecord(consumer, "foo"); assertThat(cr.value().contains("\"word\":\"french\",\"count\":2")).isTrue(); template.sendDefault("spanish"); template.sendDefault("spanish"); template.sendDefault("spanish"); cr = KafkaTestUtils.getSingleRecord(consumer, "bar"); assertThat(cr.value().contains("\"word\":\"spanish\",\"count\":3")).isTrue(); }
Example 4
Source File: KafkaStreamsBinderWordCountIntegrationTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 6 votes |
private void receiveAndValidate(String in, String out) { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); try { KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic(in); template.sendDefault("foobar"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, out); assertThat(cr.value().contains("\"word\":\"foobar\",\"count\":1")).isTrue(); } finally { pf.destroy(); } }
Example 5
Source File: KafkaStreamsNativeEncodingDecodingTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 6 votes |
@Test public void test() { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("decode-words"); template.sendDefault("foobar"); StopWatch stopWatch = new StopWatch(); stopWatch.start(); System.out.println("Starting: "); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, "decode-counts"); stopWatch.stop(); System.out.println("Total time: " + stopWatch.getTotalTimeSeconds()); assertThat(cr.value().equals("Count for foobar : 1")).isTrue(); verify(conversionDelegate).serializeOnOutbound(any(KStream.class)); verify(conversionDelegate).deserializeOnInbound(any(Class.class), any(KStream.class)); }
Example 6
Source File: KafkaStreamsNativeEncodingDecodingTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 6 votes |
@Test public void test() throws Exception { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("decode-words-1"); template.sendDefault("foobar"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, "decode-counts-1"); assertThat(cr.value().equals("Count for foobar : 1")).isTrue(); verify(conversionDelegate, never()).serializeOnOutbound(any(KStream.class)); verify(conversionDelegate, never()).deserializeOnInbound(any(Class.class), any(KStream.class)); }
Example 7
Source File: KafkaStreamsBranchingSampleTests.java From spring-cloud-stream-samples with Apache License 2.0 | 6 votes |
@Test public void testKafkaStreamsWordCountProcessor() throws InterruptedException { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>(senderProps); try { KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("words"); template.sendDefault("english"); template.sendDefault("french"); template.sendDefault("spanish"); Thread.sleep(2000); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, "english-counts", 5000); assertThat(cr.value().contains("english")).isTrue(); cr = KafkaTestUtils.getSingleRecord(consumer, "french-counts", 5000); assertThat(cr.value().contains("french")).isTrue(); cr = KafkaTestUtils.getSingleRecord(consumer, "spanish-counts", 5000); assertThat(cr.value().contains("spanish")).isTrue(); } finally { pf.destroy(); } }
Example 8
Source File: OutboundValueNullSkippedConversionTest.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
@Test public void testOutboundNullValueIsHandledGracefully() throws Exception { SpringApplication app = new SpringApplication( OutboundNullApplication.class); app.setWebApplicationType(WebApplicationType.NONE); try (ConfigurableApplicationContext context = app.run("--server.port=0", "--spring.jmx.enabled=false", "--spring.cloud.stream.bindings.input.destination=words", "--spring.cloud.stream.bindings.output.destination=counts", "--spring.cloud.stream.bindings.output.producer.useNativeEncoding=false", "--spring.cloud.stream.kafka.streams.default.consumer.application-id=testOutboundNullValueIsHandledGracefully", "--spring.cloud.stream.kafka.streams.binder.configuration.commit.interval.ms=1000", "--spring.cloud.stream.kafka.streams.binder.configuration.default.key.serde" + "=org.apache.kafka.common.serialization.Serdes$StringSerde", "--spring.cloud.stream.kafka.streams.binder.configuration.default.value.serde" + "=org.apache.kafka.common.serialization.Serdes$StringSerde", "--spring.cloud.stream.kafka.streams.timeWindow.length=5000", "--spring.cloud.stream.kafka.streams.timeWindow.advanceBy=0", "--spring.cloud.stream.kafka.binder.brokers=" + embeddedKafka.getBrokersAsString())) { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); try { KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("words"); template.sendDefault("foobar"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, "counts"); assertThat(cr.value() == null).isTrue(); } finally { pf.destroy(); } } }
Example 9
Source File: ProducerOnlyTransactionTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
@Test public void testProducerTx() { this.sender.DoInTransaction(this.output); assertThat(this.sender.isInTx()).isTrue(); Map<String, Object> props = KafkaTestUtils.consumerProps("consumeTx", "false", embeddedKafka.getEmbeddedKafka()); props.put(ConsumerConfig.KEY_DESERIALIZER_CLASS_CONFIG, ByteArrayDeserializer.class); props.put(ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG, ByteArrayDeserializer.class); props.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); props.put(ConsumerConfig.ISOLATION_LEVEL_CONFIG, IsolationLevel.READ_COMMITTED.name().toLowerCase()); Consumer<?, ?> consumer = new KafkaConsumer<>(props); embeddedKafka.getEmbeddedKafka().consumeFromAllEmbeddedTopics(consumer); ConsumerRecord<?, ?> record = KafkaTestUtils.getSingleRecord(consumer, "output"); assertThat(record.value()).isEqualTo("foo".getBytes()); }
Example 10
Source File: KafkaStreamsInteractiveQueryIntegrationTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
private void receiveAndValidateFoo(ConfigurableApplicationContext context) { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("foos"); template.sendDefault("{\"id\":\"123\"}"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, "counts-id"); assertThat(cr.value().contains("Count for product with ID 123: 1")).isTrue(); ProductCountApplication.Foo foo = context .getBean(ProductCountApplication.Foo.class); assertThat(foo.getProductStock(123).equals(1L)); // perform assertions on HostInfo related methods in InteractiveQueryService InteractiveQueryService interactiveQueryService = context .getBean(InteractiveQueryService.class); HostInfo currentHostInfo = interactiveQueryService.getCurrentHostInfo(); assertThat(currentHostInfo.host() + ":" + currentHostInfo.port()) .isEqualTo(embeddedKafka.getBrokersAsString()); HostInfo hostInfo = interactiveQueryService.getHostInfo("prod-id-count-store", 123, new IntegerSerializer()); assertThat(hostInfo.host() + ":" + hostInfo.port()) .isEqualTo(embeddedKafka.getBrokersAsString()); HostInfo hostInfoFoo = interactiveQueryService .getHostInfo("prod-id-count-store-foo", 123, new IntegerSerializer()); assertThat(hostInfoFoo).isNull(); final List<HostInfo> hostInfos = interactiveQueryService.getAllHostsInfo("prod-id-count-store"); assertThat(hostInfos.size()).isEqualTo(1); final HostInfo hostInfo1 = hostInfos.get(0); assertThat(hostInfo1.host() + ":" + hostInfo1.port()) .isEqualTo(embeddedKafka.getBrokersAsString()); }
Example 11
Source File: KafkaStreamsBinderWordCountFunctionTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
private void receiveAndValidate(String in, String out) { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>(senderProps); try { KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic(in); template.sendDefault("foobar"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, out); assertThat(cr.value().contains("\"word\":\"foobar\",\"count\":1")).isTrue(); } finally { pf.destroy(); } }
Example 12
Source File: KafkaStreamsBinderWordCountFunctionTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
@Test public void testKstreamWordCountFunctionWithCustomProducerStreamPartitioner() throws Exception { SpringApplication app = new SpringApplication(WordCountProcessorApplication.class); app.setWebApplicationType(WebApplicationType.NONE); try (ConfigurableApplicationContext context = app.run( "--server.port=0", "--spring.jmx.enabled=false", "--spring.cloud.stream.bindings.process-in-0.destination=words-2", "--spring.cloud.stream.bindings.process-out-0.destination=counts-2", "--spring.cloud.stream.bindings.process-out-0.producer.partitionCount=2", "--spring.cloud.stream.kafka.streams.bindings.process-out-0.producer.streamPartitionerBeanName" + "=streamPartitioner", "--spring.cloud.stream.kafka.streams.binder.configuration.commit.interval.ms=1000", "--spring.cloud.stream.kafka.streams.binder.configuration.default.key.serde" + "=org.apache.kafka.common.serialization.Serdes$StringSerde", "--spring.cloud.stream.kafka.streams.binder.configuration.default.value.serde" + "=org.apache.kafka.common.serialization.Serdes$StringSerde", "--spring.cloud.stream.kafka.streams.binder.brokers=" + embeddedKafka.getBrokersAsString())) { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>(senderProps); try { KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("words-2"); template.sendDefault("foo"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer, "counts-2"); assertThat(cr.value().contains("\"word\":\"foo\",\"count\":1")).isTrue(); assertThat(cr.partition() == 0) .isTrue(); template.sendDefault("bar"); cr = KafkaTestUtils.getSingleRecord(consumer, "counts-2"); assertThat(cr.value().contains("\"word\":\"bar\",\"count\":1")).isTrue(); assertThat(cr.partition() == 1) .isTrue(); } finally { pf.destroy(); } } }
Example 13
Source File: KafkaStreamsBinderPojoInputAndPrimitiveTypeOutputTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
private void receiveAndValidateFoo() { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("foos"); template.sendDefault("{\"id\":\"123\"}"); ConsumerRecord<Integer, Long> cr = KafkaTestUtils.getSingleRecord(consumer, "counts-id"); assertThat(cr.key()).isEqualTo(123); assertThat(cr.value()).isEqualTo(1L); }
Example 14
Source File: KafkaStubMessages.java From spring-cloud-contract with Apache License 2.0 | 5 votes |
Message receive(String topic, long timeout, TimeUnit timeUnit) { Consumer consumer = this.consumers.get(topic); if (consumer == null) { throw new IllegalStateException( "No consumer set up for topic [" + topic + "]"); } ConsumerRecord<String, String> record = KafkaTestUtils.getSingleRecord(consumer, topic, timeUnit.toMillis(timeout)); if (log.isDebugEnabled()) { log.debug("Got a single record for destination [" + topic + "]"); } return new Record(record).toMessage(); }
Example 15
Source File: DeserializationErrorHandlerByKafkaTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
@Test public void test() { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("xyz-DeserializationErrorHandlerByKafkaTests-In"); template.sendDefault(1, null, "foobar"); Map<String, Object> consumerProps = KafkaTestUtils.consumerProps("foobar", "false", embeddedKafka); consumerProps.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); DefaultKafkaConsumerFactory<String, String> cf = new DefaultKafkaConsumerFactory<>( consumerProps); Consumer<String, String> consumer1 = cf.createConsumer(); embeddedKafka.consumeFromAnEmbeddedTopic(consumer1, "error.xyz-DeserializationErrorHandlerByKafkaTests-In.group"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer1, "error.xyz-DeserializationErrorHandlerByKafkaTests-In.group"); assertThat(cr.value()).isEqualTo("foobar"); assertThat(cr.partition()).isEqualTo(0); // custom partition function // Ensuring that the deserialization was indeed done by Kafka natively verify(conversionDelegate, never()).deserializeOnInbound(any(Class.class), any(KStream.class)); verify(conversionDelegate, never()).serializeOnOutbound(any(KStream.class)); }
Example 16
Source File: DeserializationErrorHandlerByKafkaTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
@Test @Ignore public void test() { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("abc-DeserializationErrorHandlerByKafkaTests-In"); template.sendDefault(1, null, "foobar"); Map<String, Object> consumerProps = KafkaTestUtils.consumerProps("foobar", "false", embeddedKafka); consumerProps.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); DefaultKafkaConsumerFactory<String, String> cf = new DefaultKafkaConsumerFactory<>( consumerProps); Consumer<String, String> consumer1 = cf.createConsumer(); embeddedKafka.consumeFromAnEmbeddedTopic(consumer1, "error.abc-DeserializationErrorHandlerByKafkaTests-In.group"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer1, "error.abc-DeserializationErrorHandlerByKafkaTests-In.group"); assertThat(cr.value()).isEqualTo("foobar"); assertThat(cr.partition()).isEqualTo(0); // custom partition function // Ensuring that the deserialization was indeed done by Kafka natively verify(conversionDelegate, never()).deserializeOnInbound(any(Class.class), any(KStream.class)); verify(conversionDelegate, never()).serializeOnOutbound(any(KStream.class)); }
Example 17
Source File: DeserializtionErrorHandlerByBinderTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
@Test @SuppressWarnings("unchecked") public void test() { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("foos1"); template.sendDefault("hello"); template.setDefaultTopic("foos2"); template.sendDefault("hello"); Map<String, Object> consumerProps = KafkaTestUtils.consumerProps("foobar1", "false", embeddedKafka); consumerProps.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); DefaultKafkaConsumerFactory<String, String> cf = new DefaultKafkaConsumerFactory<>( consumerProps); Consumer<String, String> consumer1 = cf.createConsumer(); embeddedKafka.consumeFromEmbeddedTopics(consumer1, "error.foos1.fooz-group", "error.foos2.fooz-group"); ConsumerRecord<String, String> cr1 = KafkaTestUtils.getSingleRecord(consumer1, "error.foos1.fooz-group"); assertThat(cr1.value().equals("hello")).isTrue(); ConsumerRecord<String, String> cr2 = KafkaTestUtils.getSingleRecord(consumer1, "error.foos2.fooz-group"); assertThat(cr2.value().equals("hello")).isTrue(); // Ensuring that the deserialization was indeed done by the binder verify(conversionDelegate).deserializeOnInbound(any(Class.class), any(KStream.class)); }
Example 18
Source File: DeserializtionErrorHandlerByBinderTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 5 votes |
@Test public void test() { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); DefaultKafkaProducerFactory<Integer, String> pf = new DefaultKafkaProducerFactory<>( senderProps); KafkaTemplate<Integer, String> template = new KafkaTemplate<>(pf, true); template.setDefaultTopic("goos"); template.sendDefault(1, 7, "hello"); Map<String, Object> consumerProps = KafkaTestUtils.consumerProps("foobar", "false", embeddedKafka); consumerProps.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); DefaultKafkaConsumerFactory<String, String> cf = new DefaultKafkaConsumerFactory<>( consumerProps); Consumer<String, String> consumer1 = cf.createConsumer(); embeddedKafka.consumeFromAnEmbeddedTopic(consumer1, "error.goos.foobar-group"); ConsumerRecord<String, String> cr = KafkaTestUtils.getSingleRecord(consumer1, "error.goos.foobar-group"); assertThat(cr.value()).isEqualTo("hello"); assertThat(cr.partition()).isEqualTo(0); // Ensuring that the deserialization was indeed done by the binder verify(conversionDelegate).deserializeOnInbound(any(Class.class), any(KStream.class)); }
Example 19
Source File: PerRecordAvroContentTypeTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 4 votes |
@Test @Ignore public void testPerRecordAvroConentTypeAndVerifySerialization() throws Exception { SpringApplication app = new SpringApplication(SensorCountAvroApplication.class); app.setWebApplicationType(WebApplicationType.NONE); try (ConfigurableApplicationContext ignored = app.run("--server.port=0", "--spring.jmx.enabled=false", "--spring.cloud.stream.bindings.input.consumer.useNativeDecoding=false", "--spring.cloud.stream.bindings.output.producer.useNativeEncoding=false", "--spring.cloud.stream.bindings.input.destination=sensors", "--spring.cloud.stream.bindings.output.destination=received-sensors", "--spring.cloud.stream.bindings.output.contentType=application/avro", "--spring.cloud.stream.kafka.streams.bindings.input.consumer.application-id=per-record-avro-contentType-test", "--spring.cloud.stream.kafka.streams.binder.configuration.commit.interval.ms=1000", "--spring.cloud.stream.kafka.streams.binder.brokers=" + embeddedKafka.getBrokersAsString())) { Map<String, Object> senderProps = KafkaTestUtils.producerProps(embeddedKafka); // Use a custom avro test serializer senderProps.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, TestAvroSerializer.class); DefaultKafkaProducerFactory<Integer, Sensor> pf = new DefaultKafkaProducerFactory<>( senderProps); try { KafkaTemplate<Integer, Sensor> template = new KafkaTemplate<>(pf, true); Random random = new Random(); Sensor sensor = new Sensor(); sensor.setId(UUID.randomUUID().toString() + "-v1"); sensor.setAcceleration(random.nextFloat() * 10); sensor.setVelocity(random.nextFloat() * 100); sensor.setTemperature(random.nextFloat() * 50); // Send with avro content type set. Message<?> message = MessageBuilder.withPayload(sensor) .setHeader("contentType", "application/avro").build(); template.setDefaultTopic("sensors"); template.send(message); // Serialized byte[] ^^ is received by the binding process and deserialzed // it using avro converter. // Then finally, the data will be output to a return topic as byte[] // (using the same avro converter). // Receive the byte[] from return topic ConsumerRecord<String, byte[]> cr = KafkaTestUtils .getSingleRecord(consumer, "received-sensors"); final byte[] value = cr.value(); // Convert the byte[] received back to avro object and verify that it is // the same as the one we sent ^^. AvroSchemaMessageConverter avroSchemaMessageConverter = new AvroSchemaMessageConverter(); Message<?> receivedMessage = MessageBuilder.withPayload(value) .setHeader("contentType", MimeTypeUtils.parseMimeType("application/avro")) .build(); Sensor messageConverted = (Sensor) avroSchemaMessageConverter .fromMessage(receivedMessage, Sensor.class); assertThat(messageConverted).isEqualTo(sensor); } finally { pf.destroy(); } } }
Example 20
Source File: KafkaBinderTests.java From spring-cloud-stream-binder-kafka with Apache License 2.0 | 4 votes |
@SuppressWarnings({ "rawtypes", "unchecked" }) @Test public void testPolledConsumerWithDlq() throws Exception { KafkaTestBinder binder = getBinder(); PollableSource<MessageHandler> inboundBindTarget = new DefaultPollableMessageSource( this.messageConverter); ExtendedConsumerProperties<KafkaConsumerProperties> properties = createConsumerProperties(); properties.getExtension().setPollTimeout(1); properties.setMaxAttempts(2); properties.setBackOffInitialInterval(0); properties.getExtension().setEnableDlq(true); Map<String, Object> producerProps = KafkaTestUtils .producerProps(embeddedKafka.getEmbeddedKafka()); Binding<PollableSource<MessageHandler>> binding = binder.bindPollableConsumer( "pollableDlq", "group-pcWithDlq", inboundBindTarget, properties); KafkaTemplate template = new KafkaTemplate( new DefaultKafkaProducerFactory<>(producerProps)); template.send("pollableDlq", "testPollableDLQ"); try { int n = 0; while (n++ < 100) { inboundBindTarget.poll(m -> { throw new RuntimeException("test DLQ"); }); Thread.sleep(100); } } catch (MessageHandlingException e) { assertThat(e.getCause().getMessage()).isEqualTo("test DLQ"); } Map<String, Object> consumerProps = KafkaTestUtils.consumerProps("dlq", "false", embeddedKafka.getEmbeddedKafka()); consumerProps.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); ConsumerFactory cf = new DefaultKafkaConsumerFactory<>(consumerProps); Consumer consumer = cf.createConsumer(); embeddedKafka.getEmbeddedKafka().consumeFromAnEmbeddedTopic(consumer, "error.pollableDlq.group-pcWithDlq"); ConsumerRecord deadLetter = KafkaTestUtils.getSingleRecord(consumer, "error.pollableDlq.group-pcWithDlq"); assertThat(deadLetter).isNotNull(); assertThat(deadLetter.value()).isEqualTo("testPollableDLQ"); binding.unbind(); consumer.close(); }