We are looking to move approximately 1 PB of data to Archive Storage. One of the options we have been considering is GPV2 storage with Standard Performance and Archive access. I have some questions about pricing. I used the pricing calculator at:
https://azure.microsoft.com/en-ca/pricing/calculator/
The data that will be archived is large image files as well as contract documents. From my reading, block blobs would probably be optimum for archival. Is this right?
In a simple scenario, I would like to upload a 100 MB imagery file. I would have to have an empty blob in a container and use set blob tier to set it to either hot or cool. I would then copy the file and then use set blob tier again to change it to archive. How would Azure handle the copy? How many blocks would the file be broken into? I have read that each putblock will be counted as one transaction and the final PutBlocklist will be counted as one transaction. How many operations will be required? If the price is $1.10 for 100,000 operations what kind of cost can I estimate? Also, what would be the cost for changing the tier from hot or cool to archive?
The more expensive task is reading the data. After 180 days and the client wants to read the data. The blob within the container would have to be set from archive to hot or cool right? It will then take time to rehydrate the data. What will be the cost? How will the file be handled when reading the data and how many get blob operations will be required? What other operations will be needed. Looking at the Azure pricing site, it says the cost is $55.00 for a 100,000 operations. As reading is supposed to be much more expensive, I suppose there will be a large number of operations.
For organizing the data containers would be necessary. Any help on container creation/deletion costs would be helpful.