Requirement is to copy hdfs files from Hadoop cluster(non-aws) to AWS S3 bucket with standalone java application scheduled with daily CRON. Would be using AmazonS3.copyObject() method for copying. How to specify the kerberized server connection details for the source Hadoop cluster so that S3client can access the files from source hdfs folder.
The below command was used earlier but its not the secure way of transferring files.
hadoop distcp -Dfs.s3a.access.key=<<>> -Dfs.s3a.secret.key=<<>> hdfs://nameservice1/test/test1/folder s3a://<>/test/test1/folder