Splitting and uploading extremely large files to Amazon S3

For storing extremely large files on Amazon S3, the configured virtual machine can be used which would be 10+ GB in size.

In HTML5 file API, very large files are divided into small bits on the client. The server has the responsibility to join files together and move the complete file to S3. 

There is no cost of sending the files between EC2 and S3, but for this, we need to maintain 2 apps to send large files. 

In Amazon multipart upload if chunk upload fails, it can be restarted. 

A 5GB data can be divided into 1024 separate parts and upload each one independently.