com.couchbase.client.java.document.ByteArrayDocument Java Examples
The following examples show how to use
com.couchbase.client.java.document.ByteArrayDocument.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example #1
Source File: TestPutCouchbaseKey.java From nifi with Apache License 2.0 | 5 votes |
@Test public void testBinaryDoc() throws Exception { String bucketName = "bucket-1"; String docId = "doc-a"; int expiry = 100; long cas = 200L; String inFileData = "12345"; byte[] inFileDataBytes = inFileData.getBytes(StandardCharsets.UTF_8); Bucket bucket = mock(Bucket.class); when(bucket.upsert(any(ByteArrayDocument.class), eq(PersistTo.NONE), eq(ReplicateTo.NONE))) .thenReturn(ByteArrayDocument.create(docId, expiry, Unpooled.copiedBuffer(inFileData.getBytes(StandardCharsets.UTF_8)).array(), cas)); setupMockBucket(bucket); testRunner.enqueue(inFileDataBytes); testRunner.setProperty(BUCKET_NAME, bucketName); testRunner.setProperty(DOC_ID, docId); testRunner.setProperty(DOCUMENT_TYPE, DocumentType.Binary.name()); testRunner.run(); verify(bucket, times(1)).upsert(any(ByteArrayDocument.class), eq(PersistTo.NONE), eq(ReplicateTo.NONE)); testRunner.assertAllFlowFilesTransferred(REL_SUCCESS); testRunner.assertTransferCount(REL_SUCCESS, 1); testRunner.assertTransferCount(REL_RETRY, 0); testRunner.assertTransferCount(REL_FAILURE, 0); MockFlowFile outFile = testRunner.getFlowFilesForRelationship(REL_SUCCESS).get(0); outFile.assertContentEquals(inFileData); outFile.assertAttributeEquals(CouchbaseAttributes.Cluster.key(), SERVICE_ID); outFile.assertAttributeEquals(CouchbaseAttributes.Bucket.key(), bucketName); outFile.assertAttributeEquals(CouchbaseAttributes.DocId.key(), docId); outFile.assertAttributeEquals(CouchbaseAttributes.Cas.key(), String.valueOf(cas)); outFile.assertAttributeEquals(CouchbaseAttributes.Expiry.key(), String.valueOf(expiry)); }
Example #2
Source File: TestCouchbaseRemoteTableEndToEnd.java From samza with Apache License 2.0 | 4 votes |
@Test public void testEndToEnd() throws Exception { Bucket inputBucket = cluster.openBucket(inputBucketName); inputBucket.upsert(ByteArrayDocument.create("Alice", "20".getBytes())); inputBucket.upsert(ByteArrayDocument.create("Bob", "30".getBytes())); inputBucket.upsert(ByteArrayDocument.create("Chris", "40".getBytes())); inputBucket.upsert(ByteArrayDocument.create("David", "50".getBytes())); inputBucket.close(); String[] users = new String[]{"Alice", "Bob", "Chris", "David"}; int partitionCount = 1; Map<String, String> configs = TestLocalTableEndToEnd.getBaseJobConfig(bootstrapUrl(), zkConnect()); configs.put("streams.User.samza.system", "test"); configs.put("streams.User.source", Base64Serializer.serialize(users)); configs.put("streams.User.partitionCount", String.valueOf(partitionCount)); Config config = new MapConfig(configs); final StreamApplication app = appDesc -> { DelegatingSystemDescriptor inputSystemDescriptor = new DelegatingSystemDescriptor("test"); GenericInputDescriptor<String> inputDescriptor = inputSystemDescriptor.getInputDescriptor("User", new NoOpSerde<>()); CouchbaseTableReadFunction<String> readFunction = new CouchbaseTableReadFunction<>(inputBucketName, String.class, "couchbase://127.0.0.1") .withBootstrapCarrierDirectPort(couchbaseMock.getCarrierPort(inputBucketName)) .withBootstrapHttpDirectPort(couchbaseMock.getHttpPort()) .withSerde(new StringSerde()); CouchbaseTableWriteFunction<JsonObject> writeFunction = new CouchbaseTableWriteFunction<>(outputBucketName, JsonObject.class, "couchbase://127.0.0.1") .withBootstrapCarrierDirectPort(couchbaseMock.getCarrierPort(outputBucketName)) .withBootstrapHttpDirectPort(couchbaseMock.getHttpPort()); RemoteTableDescriptor inputTableDesc = new RemoteTableDescriptor<String, String>("input-table") .withReadFunction(readFunction) .withRateLimiterDisabled(); Table<KV<String, String>> inputTable = appDesc.getTable(inputTableDesc); RemoteTableDescriptor outputTableDesc = new RemoteTableDescriptor<String, JsonObject>("output-table") .withReadFunction(new NoOpTableReadFunction<>()) .withWriteFunction(writeFunction) .withRateLimiterDisabled(); Table<KV<String, JsonObject>> outputTable = appDesc.getTable(outputTableDesc); appDesc.getInputStream(inputDescriptor) .map(k -> KV.of(k, k)) .join(inputTable, new JoinFunction()) .sendTo(outputTable); }; final LocalApplicationRunner runner = new LocalApplicationRunner(app, config); executeRun(runner, config); runner.waitForFinish(); Bucket outputBucket = cluster.openBucket(outputBucketName); Assert.assertEquals("{\"name\":\"Alice\",\"age\":\"20\"}", outputBucket.get("Alice").content().toString()); Assert.assertEquals("{\"name\":\"Bob\",\"age\":\"30\"}", outputBucket.get("Bob").content().toString()); Assert.assertEquals("{\"name\":\"Chris\",\"age\":\"40\"}", outputBucket.get("Chris").content().toString()); Assert.assertEquals("{\"name\":\"David\",\"age\":\"50\"}", outputBucket.get("David").content().toString()); outputBucket.close(); }
Example #3
Source File: CouchbaseTarget.java From datacollector with Apache License 2.0 | 4 votes |
/** * Executes a document write operation. * * @param key the document key * @param ttl the document expiry ttl * @param cas the compare-and-swap (CAS) value to apply to the write * @param baos the raw document content * @param record the record being written * @return an observable for the document write or an empty observable on error */ private Observable<AbstractDocument> writeDoc(String key, int ttl, long cas, ByteArrayOutputStream baos, Record record) { WriteOperationType opType = getOperationFromHeader(record, key); if(opType == null) { return Observable.empty(); } AbstractDocument doc; if(config.dataFormat == DataFormat.JSON) { try { doc = JsonDocument.create(key, ttl, JsonObject.fromJson(baos.toString(config.dataFormatConfig.charset)), cas); } catch(Exception e) { return handleError(record, Errors.COUCHBASE_10, e); } } else { doc = ByteArrayDocument.create(key, ttl, baos.toByteArray(), cas); } switch (opType) { case DELETE: { LOG.debug("DELETE key: {}, TTL: {}, CAS: {}", key, ttl, cas); return connector.bucket().remove(doc, config.persistTo, config.replicateTo) .timeout(config.couchbase.kvTimeout, TimeUnit.MILLISECONDS); } case INSERT: { LOG.debug("INSERT key: {}, TTL: {}, CAS: {}", key, ttl, cas); return connector.bucket().insert(doc, config.persistTo, config.replicateTo) .timeout(config.couchbase.kvTimeout, TimeUnit.MILLISECONDS); } case REPLACE: { LOG.debug("REPLACE key: {}, TTL: {}, CAS: {}", key, ttl, cas); return connector.bucket().replace(doc, config.persistTo, config.replicateTo) .timeout(config.couchbase.kvTimeout, TimeUnit.MILLISECONDS); } case UPSERT: { LOG.debug("UPSERT key: {}, TTL: {}, CAS: {}", key, ttl, cas); return connector.bucket().upsert(doc, config.persistTo, config.replicateTo) .timeout(config.couchbase.kvTimeout, TimeUnit.MILLISECONDS); } default: return Observable.empty(); } }
Example #4
Source File: PutCouchbaseKey.java From nifi with Apache License 2.0 | 4 votes |
@Override public void onTrigger(final ProcessContext context, final ProcessSession session) throws ProcessException { final ComponentLog logger = getLogger(); FlowFile flowFile = session.get(); if (flowFile == null) { return; } final byte[] content = new byte[(int) flowFile.getSize()]; session.read(flowFile, new InputStreamCallback() { @Override public void process(final InputStream in) throws IOException { StreamUtils.fillBuffer(in, content, true); } }); String docId = flowFile.getAttribute(CoreAttributes.UUID.key()); if (context.getProperty(DOC_ID).isSet()) { docId = context.getProperty(DOC_ID).evaluateAttributeExpressions(flowFile).getValue(); } try { Document<?> doc = null; final DocumentType documentType = DocumentType.valueOf(context.getProperty(DOCUMENT_TYPE).getValue()); switch (documentType) { case Json: { doc = RawJsonDocument.create(docId, new String(content, StandardCharsets.UTF_8)); break; } case Binary: { doc = ByteArrayDocument.create(docId, content); break; } } final PersistTo persistTo = PersistTo.valueOf(context.getProperty(PERSIST_TO).getValue()); final ReplicateTo replicateTo = ReplicateTo.valueOf(context.getProperty(REPLICATE_TO).getValue()); final Bucket bucket = openBucket(context); doc = bucket.upsert(doc, persistTo, replicateTo); final Map<String, String> updatedAttrs = new HashMap<>(); updatedAttrs.put(CouchbaseAttributes.Cluster.key(), context.getProperty(COUCHBASE_CLUSTER_SERVICE).getValue()); updatedAttrs.put(CouchbaseAttributes.Bucket.key(), bucket.name()); updatedAttrs.put(CouchbaseAttributes.DocId.key(), docId); updatedAttrs.put(CouchbaseAttributes.Cas.key(), String.valueOf(doc.cas())); updatedAttrs.put(CouchbaseAttributes.Expiry.key(), String.valueOf(doc.expiry())); flowFile = session.putAllAttributes(flowFile, updatedAttrs); session.getProvenanceReporter().send(flowFile, getTransitUrl(bucket, docId)); session.transfer(flowFile, REL_SUCCESS); } catch (final CouchbaseException e) { String errMsg = String.format("Writing document %s to Couchbase Server using %s failed due to %s", docId, flowFile, e); handleCouchbaseException(context, session, logger, flowFile, e, errMsg); } }
Example #5
Source File: TestCouchbaseUtils.java From nifi with Apache License 2.0 | 4 votes |
@Ignore("This test method requires a live Couchbase Server instance") @Test public void testDocumentTypesAndStringConversion() { final CouchbaseCluster cluster = CouchbaseCluster.fromConnectionString("couchbase://192.168.99.100:8091"); final Bucket bucket = cluster.openBucket("b1", "b1password"); bucket.upsert(JsonDocument.create("JsonDocument", JsonObject.create().put("one", 1))); bucket.upsert(JsonArrayDocument.create("JsonArray", JsonArray.create().add(1).add(2).add(3))); bucket.upsert(JsonDoubleDocument.create("JsonDouble", 0.123)); bucket.upsert(JsonStringDocument.create("JsonString", "value")); bucket.upsert(JsonBooleanDocument.create("JsonBoolean", true)); bucket.upsert(JsonLongDocument.create("JsonLong", 123L)); bucket.upsert(RawJsonDocument.create("RawJsonDocument", "value")); bucket.upsert(StringDocument.create("StringDocument", "value")); bucket.upsert(BinaryDocument.create("BinaryDocument", Unpooled.copiedBuffer("value".getBytes(StandardCharsets.UTF_8)))); bucket.upsert(ByteArrayDocument.create("ByteArrayDocument", "value".getBytes(StandardCharsets.UTF_8))); final String[][] expectations = { {"JsonDocument", "String", "{\"one\":1}"}, {"JsonArray", "String", "[1,2,3]"}, {"JsonDouble", "String", "0.123"}, {"JsonString", "String", "\"value\""}, {"JsonBoolean", "String", "true"}, {"JsonLong", "String", "123"}, {"RawJsonDocument", "String", "value"}, {"StringDocument", "String", "value"}, {"BinaryDocument", "byte[]", "value"}, {"ByteArrayDocument", "byte[]", "value"}, }; for (String[] expectation : expectations) { final LegacyDocument document = bucket.get(LegacyDocument.create(expectation[0])); assertEquals(expectation[1], document.content().getClass().getSimpleName()); assertEquals(expectation[2], CouchbaseUtils.getStringContent(document.content())); } final BinaryDocument binaryDocument = bucket.get(BinaryDocument.create("BinaryDocument")); final String stringFromByteBuff = CouchbaseUtils.getStringContent(binaryDocument.content()); assertEquals("value", stringFromByteBuff); try { bucket.get(BinaryDocument.create("JsonDocument")); fail("Getting a JSON document as a BinaryDocument fails"); } catch (TranscodingException e) { assertTrue(e.getMessage().contains("Flags (0x2000000) indicate non-binary document for id JsonDocument")); } }