Java Code Examples for com.amazonaws.services.s3.transfer.TransferManager#upload()
The following examples show how to use
com.amazonaws.services.s3.transfer.TransferManager#upload() .
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example 1
Source File: Uploads.java From jobcacher-plugin with MIT License | 5 votes |
public void startUploading(TransferManager manager, File file, InputStream inputStream, Destination dest, ObjectMetadata metadata) throws AmazonServiceException { final PutObjectRequest request = new PutObjectRequest(dest.bucketName, dest.objectName, inputStream, metadata); // Set the buffer size (ReadLimit) equal to the multipart upload size, // allowing us to resend data if the connection breaks. request.getRequestClientOptions().setReadLimit(MULTIPART_UPLOAD_THRESHOLD); manager.getConfiguration().setMultipartUploadThreshold( (long) MULTIPART_UPLOAD_THRESHOLD); final Upload upload = manager.upload(request); startedUploads.put(file, upload); openedStreams.put(file, inputStream); }
Example 2
Source File: TablonController.java From spring-cloud-aws-sample with Apache License 2.0 | 5 votes |
@RequestMapping("/anuncio/nuevo") public String nuevoAnuncio(Model model, @RequestParam String nombre, @RequestParam String asunto, @RequestParam String comentario, @RequestParam String filename, @RequestParam MultipartFile file) { if (!file.isEmpty()) { try { ObjectMetadata objectMetadata = new ObjectMetadata(); objectMetadata.setContentType(file.getContentType()); TransferManager transferManager = TransferManagerBuilder.defaultTransferManager(); transferManager.upload(bucket, filename, file.getInputStream(), objectMetadata); } catch (Exception e) { model.addAttribute("message", "You failed to upload " + filename + " => " + e.getMessage()); return "error"; } } else { model.addAttribute("message", "You failed to upload " + filename + " because the file was empty."); return "error"; } Anuncio anuncio = new Anuncio(nombre, asunto, comentario); anuncio.setFoto(s3.getUrl(bucket, filename)); repository.save(anuncio); return "anuncio_guardado"; }
Example 3
Source File: XferMgrUpload.java From aws-doc-sdk-examples with Apache License 2.0 | 5 votes |
public static void uploadFile(String file_path, String bucket_name, String key_prefix, boolean pause) { System.out.println("file: " + file_path + (pause ? " (pause)" : "")); String key_name = null; if (key_prefix != null) { key_name = key_prefix + '/' + file_path; } else { key_name = file_path; } // snippet-start:[s3.java1.s3_xfer_mgr_upload.single] File f = new File(file_path); TransferManager xfer_mgr = TransferManagerBuilder.standard().build(); try { Upload xfer = xfer_mgr.upload(bucket_name, key_name, f); // loop with Transfer.isDone() XferMgrProgress.showTransferProgress(xfer); // or block with Transfer.waitForCompletion() XferMgrProgress.waitForCompletion(xfer); } catch (AmazonServiceException e) { System.err.println(e.getErrorMessage()); System.exit(1); } xfer_mgr.shutdownNow(); // snippet-end:[s3.java1.s3_xfer_mgr_upload.single] }
Example 4
Source File: FileController.java From full-teaching with Apache License 2.0 | 5 votes |
private void productionFileSaver(String keyName, String folderName, File f) throws InterruptedException { String bucketName = this.bucketAWS + "/" + folderName; TransferManager tm = new TransferManager(this.amazonS3); // TransferManager processes all transfers asynchronously, so this call will return immediately Upload upload = tm.upload(bucketName, keyName, f); try { // Or you can block and wait for the upload to finish upload.waitForCompletion(); System.out.println("Upload completed"); } catch (AmazonClientException amazonClientException) { System.out.println("Unable to upload file, upload was aborted."); amazonClientException.printStackTrace(); } }
Example 5
Source File: AmazonS3Storage.java From thunderbit with GNU Affero General Public License v3.0 | 5 votes |
@Override public F.Promise<Void> store(Path path, String key, String name) { Promise<Void> promise = Futures.promise(); TransferManager transferManager = new TransferManager(credentials); try { Upload upload = transferManager.upload(bucketName, key, path.toFile()); upload.addProgressListener((ProgressListener) progressEvent -> { if (progressEvent.getEventType().isTransferEvent()) { if (progressEvent.getEventType().equals(ProgressEventType.TRANSFER_COMPLETED_EVENT)) { transferManager.shutdownNow(); promise.success(null); } else if (progressEvent.getEventType().equals(ProgressEventType.TRANSFER_FAILED_EVENT)) { transferManager.shutdownNow(); logger.error(progressEvent.toString()); promise.failure(new Exception(progressEvent.toString())); } } }); } catch (AmazonServiceException ase) { logAmazonServiceException (ase); } catch (AmazonClientException ace) { logAmazonClientException(ace); } return F.Promise.wrap(promise.future()); }
Example 6
Source File: S3OutputStreamWrapper.java From streams with Apache License 2.0 | 5 votes |
private void addFile() throws Exception { InputStream is = new ByteArrayInputStream(this.outputStream.toByteArray()); int contentLength = outputStream.size(); TransferManager transferManager = new TransferManager(amazonS3Client); ObjectMetadata metadata = new ObjectMetadata(); metadata.setExpirationTime(DateTime.now().plusDays(365 * 3).toDate()); metadata.setContentLength(contentLength); metadata.addUserMetadata("writer", "org.apache.streams"); for (String s : metaData.keySet()) { metadata.addUserMetadata(s, metaData.get(s)); } String fileNameToWrite = path + fileName; Upload upload = transferManager.upload(bucketName, fileNameToWrite, is, metadata); try { upload.waitForUploadResult(); is.close(); transferManager.shutdownNow(false); LOGGER.info("S3 File Close[{} kb] - {}", contentLength / 1024, path + fileName); } catch (Exception ignored) { LOGGER.trace("Ignoring", ignored); } }
Example 7
Source File: MultipartUpload.java From tutorials with MIT License | 5 votes |
public static void main(String[] args) throws Exception { String existingBucketName = "baeldung-bucket"; String keyName = "my-picture.jpg"; String filePath = "documents/my-picture.jpg"; AmazonS3 amazonS3 = AmazonS3ClientBuilder .standard() .withCredentials(new DefaultAWSCredentialsProviderChain()) .withRegion(Regions.DEFAULT_REGION) .build(); int maxUploadThreads = 5; TransferManager tm = TransferManagerBuilder .standard() .withS3Client(amazonS3) .withMultipartUploadThreshold((long) (5 * 1024 * 1024)) .withExecutorFactory(() -> Executors.newFixedThreadPool(maxUploadThreads)) .build(); ProgressListener progressListener = progressEvent -> System.out.println("Transferred bytes: " + progressEvent.getBytesTransferred()); PutObjectRequest request = new PutObjectRequest(existingBucketName, keyName, new File(filePath)); request.setGeneralProgressListener(progressListener); Upload upload = tm.upload(request); try { upload.waitForCompletion(); System.out.println("Upload complete."); } catch (AmazonClientException e) { System.out.println("Error occurred while uploading file"); e.printStackTrace(); } }
Example 8
Source File: S3OperationsImpl.java From herd with Apache License 2.0 | 4 votes |
@Override public Upload upload(PutObjectRequest putObjectRequest, TransferManager transferManager) { return transferManager.upload(putObjectRequest); }
Example 9
Source File: AwsS3Storage.java From ecs-sync with Apache License 2.0 | 4 votes |
@Override void putObject(SyncObject obj, String targetKey) { ObjectMetadata om; if (options.isSyncMetadata()) om = s3MetaFromSyncMeta(obj.getMetadata()); else om = new ObjectMetadata(); if (obj.getMetadata().isDirectory()) om.setContentType(TYPE_DIRECTORY); PutObjectRequest req; File file = (File) obj.getProperty(AbstractFilesystemStorage.PROP_FILE); S3ProgressListener progressListener = null; if (obj.getMetadata().isDirectory()) { req = new PutObjectRequest(config.getBucketName(), targetKey, new ByteArrayInputStream(new byte[0]), om); } else if (file != null) { req = new PutObjectRequest(config.getBucketName(), targetKey, file).withMetadata(om); progressListener = new ByteTransferListener(obj); } else { InputStream stream = obj.getDataStream(); if (options.isMonitorPerformance()) stream = new ProgressInputStream(stream, new PerformanceListener(getWriteWindow())); req = new PutObjectRequest(config.getBucketName(), targetKey, stream, om); } if (options.isSyncAcl()) req.setAccessControlList(s3AclFromSyncAcl(obj.getAcl(), options.isIgnoreInvalidAcls())); TransferManager xferManager = null; try { // xfer manager will figure out if MPU is needed (based on threshold), do the MPU if necessary, // and abort if it fails xferManager = TransferManagerBuilder.standard() .withS3Client(s3) .withExecutorFactory(() -> Executors.newFixedThreadPool(config.getMpuThreadCount())) .withMultipartUploadThreshold((long) config.getMpuThresholdMb() * 1024 * 1024) .withMinimumUploadPartSize((long) config.getMpuPartSizeMb() * 1024 * 1024) .withShutDownThreadPools(true) .build(); // directly update final Upload upload = xferManager.upload(req, progressListener); try { String eTag = time((Callable<String>) () -> upload.waitForUploadResult().getETag(), OPERATION_MPU); log.debug("Wrote {}, etag: {}", targetKey, eTag); } catch (Exception e) { log.error("upload exception", e); if (e instanceof RuntimeException) throw (RuntimeException) e; throw new RuntimeException("upload thread was interrupted", e); } } finally { // NOTE: apparently if we do not reference xferManager again after the upload() call (as in this finally // block), the JVM will for some crazy reason determine it is eligible for GC and call finalize(), which // shuts down the thread pool, fails the upload, and gives absolutely no indication of what's going on... if (xferManager != null) xferManager.shutdownNow(false); } }