Professional Documents
Culture Documents
[1]
The Value of a Big Data Archive A Big Data Archive brings three specic value areas to the enterprise. First, similar to a classic archive, it should allow for the reduction of primary storage consumption and support growth. According to studies, well over 80% of le
foundational component in a future big data project. In short by leveraging both initiatives, costs can be contained and ROI be realized sooner. As a result a Big Data Archive has unique requirements that a simple second tier of storage, or even a basic archive
data on most primary storage systems is not in active use solution, typically cannot meet. Whether from machine or and is therefor wasting this high cost resources performance human generated data sources, a Big Data Archive must abilities. If this data was moved to a secondary, high capacity match the compliance capabilities of disk archiving while storage area, but still one with moderate performance, most additional le access could be done without impact to the users. This could have a signicant, positive impact to the IT budget. Secondary disk tiers have been available for years, as have software products to classify and move that data. The cost savings on primary storage alone motivated many users to move to a two-tier storage infrastructure. But many other data centers were not so inclined. Big Data Archive brings three more motivation points to the equation that should interest all data centers to adopt this multi-tier approach to storage. meeting requirements like dense scaling, high throughput and fast retrieval. Requirements For The Big Data Archive Density Scaling Legacy second tier disk systems and even archive systems both have scaling issues when measured against the Big Data Archive challenge. The requirement to scale to Petabytes is now the starting point for many of these systems. This quickly eliminates single box architectures.
Even legacy scale out storage architectures may not be The rst point is that organizations are beginning to suitable for the Big Data Archive challenge. These systems understand the value of this data and to acknowledge theres were designed to add nodes rapidly and as a result their a real desire to retain, categorize and in the future mine, this capacity per node is limited and they quickly consume information to help make better business decisions or speed product development. They are coming to realize that archiving makes practical sense in the data center and its shortcomings are being eliminated by Big Data storage architectures. available data center oor space. The modern Big Data Archive will need a very dense architecture to maximize capacity on a per node basis and not waste that oor space. In these environments storage (disk drives) has practically become less expensive than the sheet metal (the other components in each node) that surrounds them. Thus,
The second motivational point is the need for compliance. making it critical to use each node to its full potential before Organizations and litigators are beginning to understand that adding another. retention is more than just making sure email is saved or that it can be found (discovered). Retention means keeping all the High Throughput les that exist in relation to a case as well. In the past this meant providing boxes of paper documents. Today most documents are digital and are never printed. Retaining Big Data Archives must also have the ability to ingest large amounts of data quickly. Legacy archive solutions were
electronic documents is not only important it may be the only designed to have data trickle into them over the course of evidence of that information that can be retained. time. Big Data Archives may store very large numbers of different sized les on an ongoing basis. There can be Finally a Big Data Archive is complimentary and may even be part of what was previously considered a separate project. This makes the cost to add a Big Data Archive to a current big data project minimal or may allow it to be the millions of small les that are being archived from traditional Big Data project or a relatively few, very large rich media les being archived from user projects.
[2]
In both cases the ingestion of these les requires that the receiving nodes encode the data and then segment it to the other nodes in the cluster. This background work could cripple legacy archive solutions whose nodes are typically interconnected via a 1GbE infrastructure. Instead, a higher speed backbone is required so that additional throughput can be maintained. Solutions like Isilons NL Scale Out NAS connect via an internal Inniband backbone for very high throughput performance, enabling them to sustain ingest rates that match the requirements of a Big Data Archive. Fast Retrieval Retrieval is also different for the Big Data Archive than it is for the traditional archive storage system. It may need to produce thousands or millions of les very quickly or in some cases it may be desirable to actually perform the search and analysis on the Big Data Archive itself. Traditional archive architectures and legacy second tier storage systems are typically found lacking when asked to provide data quickly as the capacity scales over 1 PB. Its important to remember that archive systems were designed to provide performance better than the platform they were replacing, which for most was optical disk. Big Data Archives operate against a different standard. They need to provide consistent performance thats comparable to most primary storage systems, no matter what the capacity level. Again, Isilons NL series surpasses this expectation and provides near primary storage performance but with the throughput and density that Big Data Archiving requires. Protection & Disaster Recovery
data that can be ingested at any time. If a large archive job is submitted, and then later a catastrophic failure occurs, a signicant amount of data could be permanently lost. For example in the case of machine generated sensor data there may be no way to ever recover it. Data protection needs to be integrated and then augmented into the Big Data Archive. First the system should have no single points of failure and the users should be able to set the data protection level by data type. This would accommodate unrecoverable data, like point-in-time sensor data, which might need a higher level of redundancy than traditional le data. Next, the data needs to be transferred in real time to a second location via built-in replications tools. That data again needs be prioritized based on whether it can be replaced. Finally, there are always some organizations that will want to move to an alternate device all together, even tape, in case of a regional disaster. The Big Data Archive should have the ability to add copy out performance when needed. As an example Isilon can add a class of nodes to their cluster called, backup accelerators, that are specically designed to move data to another device. This allows the other nodes to continue to deliver high throughput and fast retrieval while the cluster gets its data copied to alternate storage devices. Summary The Big Data Archive can be a component of a larger Big Data project or it can be an archive designed specically for Big Data. In either case leveraging that investment to also include human generated data that needs to be stored for mining or compliance reasons is an excellent way to achieve
Protecting 1PB+ environments requires a change in thinking. a greater ROI on the Big Data project. It can also help Nightly backups are no longer a reality, not only because of discover new ways to make better decisions by retaining the size of the solution but also because of the amount of and analyzing existing information.
[3]