One of the key requirement for every business considering Hadoop on cloud is “How to upload data to BLOB”. The answer lies at the size of the data and time you want to invest during uploading process. It may also vary on the basis of activity you perform. For example whether you want to put all the data at once or prefer to perform uploading it in chunks followed by action to perform same activity multiple times or in-parallel. With all so many variation end result would be to have data available on cloud.
Looking for this reference across internet I decide to put a blog that provide us multiple possible ways to publish data to BLOB.
UPLOAD USING OPEN SOURCE/3RD PARTY UTILITIES
Tools like Cloud Storage Studio, CloudXplorer, Azure Blob Studio 2011, Azure Blob Explorer,CodePlex.SpaceBlock, Gladinet, Azure Storage Utility, Azure Storage Viewer, Windows Azure Web Storage Explorer, CloudBerry Explorer, CyberDuck, CloudCombine, etc. are used to manage Azure storage.
UPLOAD USING API
Microsoft offers API (REST API and Media Services) that allows managing Azure Storage programmatically. The services is available to connect with Azure SDK and can also be leveraged using PowerShell. Provide support for languages like C#, JAVA and Python.
UPLOAD USING MANAGEMENT PORTAL
Ability to upload files directly via Azure Management Portal is not yet provided and under review. We might have it in upcoming releases.
UPLOAD USING IMPORT/EXPORT SERVICE
Microsoft provides an efficient solution for importing large amounts of on- premise data (TBs/PBs) into Windows Azure Blobs. The preparation process includes creation of encrypted hard disk drives and shipping it to Microsoft data centers through FedEx. Microsoft team using high-speed internal network, will upload this data for you. Considering TBs/PBs for data upload, it’s a recommend way to ease overall deployment/upload process.
Consideration includes:
-
-
- One disk is recommend to be of maximum 4TB. So if you have more than 4TB to be uploaded then you have to provide those no. of disks with 4TB’s each.
- Each device must be shipped from FedEx with an BitLocker key encryption
- Drive prepared using WAImportExport.exe is considered to be best for data uploading through Microsoft import/export services
-
For details please visit: http://azure.microsoft.com/en-us/documentation/articles/storage-import-export-service/
RECOMMENDED REFERENCES
-
- http://msdn.microsoft.com/en-us/library/dd179376.aspx
- http://msdn.microsoft.com/en-us/library/ee691964.aspx
- http://www.asp.net/aspnet/overview/developing-apps-with-windows-azure/building-real-world-cloud-apps-with-windows-azure/unstructured-blob-storage
- http://gauravmantri.com/2013/02/16/uploading-large-files-in-windows-azure-blob-storage-using-shared-access-signature-html-and-javascript/
- http://azure.microsoft.com/en-us/documentation/articles/storage-java-how-to-use-blob-storage/ (JAVA Support)
- http://azure.microsoft.com/en-us/documentation/articles/storage-nodejs-how-to-use-blob-storage/ (Node.js)
- http://blogs.msdn.com/b/tconte/archive/2013/04/17/how-to-interact-with-windows-azure-blob-storage-from-linux-using-python.aspx (Linux with Python)
Visit us at datopic, share your business ideas or connect to join datopic ride on data science and cloud and take deep dive into big data solutions.
Visit us at datopic, share your business ideas or connect to join datopic ride on data science and cloud and take deep dive into big data solutions.
Your Comment: