Overview on "Efficient & Secure Cloud Storage for Handling Big Data"

Cloud computing as a storage for Big Data
Cloud Computing

Cloud is a very new and popular topic in the field of IT. It is one of the active application accepted by the enterprises which can take advantage of low cost, fast deployment. The main purpose of Cloud Computing is that," Users can use the services from anywhere, anytime through the Internet, directly through the browser." It is the extension of the distributed computing through the Internet. This approach handles thousand of approaches like a 'Supercomputer'.



Why we need Cloud Storage ? Cloud provides the facility to store huge amounts of data on cloud server without breaking the security. If data resides on the servers, hackers may attack  the servers to steal information,data stored in the server is thus, a primary concern. This paper describes the efficient and cloud storage for handling Big Data by Arjun Kumar (Department of Ubiquitous-IT) & HoonJae Lee (Department of Computer & information Enginnering).



Cloud Storage


They describe many cloud storage security facilities such as Google Storage for developers, Amazon web Services(AWS). Firstly, they describe the Google Storage for developers, it offers the set of features and capabilities such as  Store & access the data from anywhere on the Internet and  manage Metadata. Security and sharing have the following features such as user authentication using the secret key or Google account, Secure access using SSL(Secure Service Layer). Performance and Scalability have the following features such as store data consistence, data replicated in multiple data centres across USA.

Secondly, they talk about Amazon. What amazon provides for cloud storage? Amazon has the features, Write, Read and Delete objects containing millions of terabytes of data of each, the number of objects that can be stored is unlimited. Each Object is stored in a bucket and can be retrieved via a unique developer's key. A bucket can be stored in region that can minimize the cost, address regulatory requirements etc.. Objects that are stored in the region never leaves the region until it is transferred out by the user. Amazon services has the facility so that a protocol or functional layer can be added easily. With respect to the data production, Amazon provides a highly durable storage infrastructure. The author describes in Amazon S3 how the amazon services store the terabytes of data and how to replicated data in multi-servers.


We have read how the Google Storage and Amazon works. Both of these storage, provide external security to Big Data, if we are looking for an internal security in data. The author describes the Online Data Storage using implicit security. They describe the data partitioning scheme for implementing such a security. These partitions are stored in random servers on the INTERNET and they need to recreate the original data. Data reconstruction requires access to each server, login password and knowledge of the servers in which data partition is stored. In this scheme data is partitioned in such a way that each partition is implicitly secure and has no need to be encrypted.


The authors describe several schemes, which include the implicit storage of encryption keys rather than the data. They provide algorithms for secure cloud storage such as Privacy-Preserving Public Auditing for secure Cloud Storage and Achieving Secure, Scalabe, and Fine-grained Data Access Control in cloud computing. Privacy-Preserving Public Auditing algorithm, consist of four steps. First step is KeyGen (KeyGeneration algorithm) that is run by the user to set up the scheme. Second step describes SigGen(Used by the user to generate verification metadata), third step is GenProof(Runs by the cloud to generate a proof of data storage correctness). Fourth Step is VerifyProof.


The authors propose a method for the cloud storage security. They say, "Data store in a cloud is a process where a user stores the data through a main server into a cloud servers". In this system, the network consist of three backup servers for recovery of data from disaster. The backup servers are located at remote areas from the main server. If one fails, it can be replicated from other servers. Users interact with the cloud servers via a main server to retrieve the data and at the same time data will be copied at multiple servers.

In this proposed network, it consists of security services such as Authentication, Encryption, Decryption, Decompression and Compression. Users receive secret key generated by main server  through a registered email address that is used for security purpose.

1. User login

Login to the main server with registered Email-Id and Password.



User Login

2. Authorized Person Login

After successful login, main server asks for a authorized person to login with a login ID and password. Login ID as an email address and password as a secret key.
Server Login

3. Data Encryption and Compression :

Data is to be encrypted and compressed in multi-servers. Data that can be stored in the cloud must be in the encrypted form for the security purpose. Main server deals with these issues before data is being transferred into the cloud server. Data is encrypted and compressed with the help of secret key.
Process of Data Backup

4. Data Decryption and Decompression.


 These techniques are used after the data is backed up in multiple servers. The data which is taken as backup is stored in multi-servers in unrecognized format, which can not be seen by the user. Main server decrypts the data and sends to the user.    

Process of Data Recovery

Conclusion :

The main goal of the author is to securely store and manage Big data on cloud that is not controlled by the owner of the data. They describe Google Storage, Amazon Web Services in detail. They Propose their own method or system for the cloud storage securely. They exploit the mechanism of encryption and decompression of data using the secret key at the main server while uploading or retrieving the data on cloud storage. According to the authors, this proposed system solves the problem of handling Big Data on cloud with security issues. Moreover, two time authentication provides the access to the legitimate user. The proposed mechanism helps in the data backup and recovery from different Cloud storage servers in feasible storage. From my point of view, We Should use third party for authentication such as Kerboros.  kerboros is the authentication protocol which works on ticket, ticket generated by the ticket granting server. In the end, I will complete my views on this paper. I learned about so many new definitions on the cloud storage

Previous
Next Post »

Popular Posts