org.apache.beam.sdk.options.Description Java Examples
The following examples show how to use
org.apache.beam.sdk.options.Description.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example #1
Source File: BulkDecompressor.java From DataflowTemplates with Apache License 2.0 | 5 votes |
@Description( "The output file to write failures during the decompression process " + "(e.g. gs://bucket-name/decompressed/failed.txt). The contents will be one line for " + "each file which failed decompression. Note that this parameter will " + "allow the pipeline to continue processing in the event of a failure.") @Required ValueProvider<String> getOutputFailureFile();
Example #2
Source File: DLPTextToBigQueryStreaming.java From DataflowTemplates with Apache License 2.0 | 5 votes |
@Description( "DLP API has a limit for payload size of 524KB /api call. " + "That's why dataflow process will need to chunk it. User will have to decide " + "on how they would like to batch the request depending on number of rows " + "and how big each row is.") @Required ValueProvider<Integer> getBatchSize();
Example #3
Source File: ExportPipeline.java From DataflowTemplates with Apache License 2.0 | 5 votes |
@Description("If set, specifies the time when the snapshot must be taken." + " String is in the RFC 3339 format in UTC time. " + " Example - 1990-12-31T23:59:60Z" + " Timestamp must be in the past and Maximum timestamp staleness applies." + " https://cloud.google.com/spanner/docs/timestamp-bounds#maximum_timestamp_staleness") @Default.String(value = "") ValueProvider<String> getSnapshotTime();
Example #4
Source File: SinkOptions.java From gcp-ingestion with Mozilla Public License 2.0 | 5 votes |
@Description("When --bqReadMethod=storageapi, all rows of the input table are read by default," + " but this option can take a SQL text filtering statement, similar to a WHERE clause;" + " currently, only a single predicate that is a comparison between a column and a constant" + " value is supported; a likely choice to limit partitions would be something like" + " \"CAST(submission_timestamp AS DATE) BETWEEN '2020-01-10' AND '2020-01-14'\"; see" + " https://cloud.google.com/bigquery/docs/reference/storage/rpc/google.cloud.bigquery.storage.v1beta1#tablereadoptions") String getBqRowRestriction();
Example #5
Source File: TextConverters.java From DataflowTemplates with Apache License 2.0 | 5 votes |
@Description( "The window duration in which data will be written. Defaults to 5m. " + "Allowed formats are: " + "Ns (for seconds, example: 5s), " + "Nm (for minutes, example: 12m), " + "Nh (for hours, example: 2h).") @Default.String("1m") String getWindowDuration();
Example #6
Source File: ImportOptions.java From feast with Apache License 2.0 | 5 votes |
@Description( "Fixed window size in seconds (default 60) to apply before aggregating the numerical value of " + "features and exporting the aggregated values as metrics. Refer to " + "feast/ingestion/transform/metrics/WriteFeatureValueMetricsDoFn.java" + "for the metric nameas and types used.") @Default.Integer(60) int getWindowSizeInSecForFeatureValueMetric();
Example #7
Source File: UserScore.java From deployment-examples with MIT License | 5 votes |
@Description("Path to the data file(s) containing game data.") /* The default maps to two large Google Cloud Storage files (each ~12GB) holding two subsequent day's worth (roughly) of data. Note: You may want to use a small sample dataset to test it locally/quickly : gs://apache-beam-samples/game/small/gaming_data.csv You can also download it via the command line gsutil cp gs://apache-beam-samples/game/small/gaming_data.csv ./destination_folder/gaming_data.csv */ @Default.String("gs://apache-beam-samples/game/gaming_data*.csv") String getInput();
Example #8
Source File: RepublisherOptions.java From gcp-ingestion with Mozilla Public License 2.0 | 4 votes |
@Description("A JSON-formatted map of channel name to sampling ratio; for example," + " {\"nightly\":1.0,\"release\":0.01} would republish 100% of nightly pings to the" + " sampled nightly topic and 1% of release pings to the sampled release topic") Map<String, Double> getPerChannelSampleRatios();
Example #9
Source File: AvroConverters.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("The prefix of the files to write to. Default is: output.") @Default.String("output") String getOutputFilePrefix();
Example #10
Source File: SinkOptions.java From gcp-ingestion with Mozilla Public License 2.0 | 4 votes |
@Description("Compression format for --errorOutputType=file") @Default.Enum("GZIP") Compression getErrorOutputFileCompression();
Example #11
Source File: TextConverters.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("The maximum number of output shards produced when writing.") @Default.Integer(1) Integer getNumShards();
Example #12
Source File: WriteToGCSAvro.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description( "The filename prefix of the files to write to. Default file prefix is set to \"output\". ") String getOutputFilenamePrefix();
Example #13
Source File: CassandraToBigtable.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("RowKeySeparator") @Default.String("#") ValueProvider<String> getRowKeySeparator();
Example #14
Source File: BigQueryMergeValidatorTemplate.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("The table to set as staging table.") @Default.String("") String getStagingTable();
Example #15
Source File: CassandraToBigtable.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("Cassandra keyspace to read from") ValueProvider<String> getCassandraKeyspace();
Example #16
Source File: RepublisherOptions.java From gcp-ingestion with Mozilla Public License 2.0 | 4 votes |
@Description("A pattern for output topic names (assuming --outputType=pubsub) for per-channel" + " sampling; the pattern must contain a placeholder ${channel} that will be filled in" + " to give a distinct publisher per channel configured in --perChannelSampleRatios") String getPerChannelDestination();
Example #17
Source File: IpPrivacyDecoderOptions.java From gcp-ingestion with Mozilla Public License 2.0 | 4 votes |
@Description("Path (local or gs://) to a file containing the bytes that will be used as" + "the hash key for the client IP") ValueProvider<String> getClientIpHashKey();
Example #18
Source File: PubsubToAvro.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("The Cloud Pub/Sub topic to read from.") @Required ValueProvider<String> getInputTopic();
Example #19
Source File: BigQueryConverters.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("Name of the BQ column storing the unique identifier of the row") ValueProvider<String> getReadIdColumn();
Example #20
Source File: JavascriptTextTransformer.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("UDF Javascript Function Name") String getJavascriptTextTransformFunctionName();
Example #21
Source File: StreamingDataGenerator.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("Indicates rate of messages per second to be published to Pub/Sub.") @Required Long getQps();
Example #22
Source File: JdbcConverters.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description( "KMS Encryption Key should be in the format projects/{gcp_project}/locations/{key_region}/keyRings/{key_ring}/cryptoKeys/{kms_key_name}") ValueProvider<String> getKMSEncryptionKey();
Example #23
Source File: TextIOToBigQuery.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("JSON file with BigQuery Schema description") ValueProvider<String> getJSONPath();
Example #24
Source File: ExampleBigQueryTableOptions.java From deployment-examples with MIT License | 4 votes |
@Description("BigQuery table name") @Default.InstanceFactory(BigQueryTableFactory.class) String getBigQueryTable();
Example #25
Source File: PubsubToAvro.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description( "The shard template of the output file. Specified as repeating sequences " + "of the letters 'S' or 'N' (example: SSS-NNN). These are replaced with the " + "shard number, or number of shards respectively") @Default.String("W-P-SS-of-NN") ValueProvider<String> getOutputShardTemplate();
Example #26
Source File: DebuggingWordCount.java From deployment-examples with MIT License | 4 votes |
@Description( "Regex filter pattern to use in DebuggingWordCount. " + "Only words matching this pattern will be counted.") @Default.String("Flourish|stomach") String getFilterPattern();
Example #27
Source File: SinkOptions.java From gcp-ingestion with Mozilla Public License 2.0 | 4 votes |
@Description("A comma-separated list of docTypes for which we will not accumulate an" + " additional_properties field before publishing to BigQuery;" + " this is especially useful for telemetry/main where we expect to send the" + " same payload to multiple tables, each with only a subset of the overall schema;" + " each docType must be qualified with a namespace like 'telemetry/main'") ValueProvider<List<String>> getBqStrictSchemaDocTypes();
Example #28
Source File: WriteToGCSAvro.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("The directory to output files to. Must end with a slash. ") String getOutputDirectory();
Example #29
Source File: JdbcConverters.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description( "The JDBC connection URL string. " + "for example: jdbc:mysql://some-host:3306/sampledb") ValueProvider<String> getConnectionURL();
Example #30
Source File: JdbcConverters.java From DataflowTemplates with Apache License 2.0 | 4 votes |
@Description("JDBC connection user name. ") ValueProvider<String> getUsername();