Symantec Corp. has announced an add-on solution for Symantec’s Cluster File System that enables customers to run Big Data analytics on their existing infrastructure by making it highly available and manageable. Apache Hadoop offers customers significant value to drive revenue by helping analyse data for business insights, however many existing data solutions lack the data management capabilities and built-in resilience to overcome the cost and complexity of increasing storage and server sprawl. By working closely with Hortonworks, the new Symantec Enterprise Solution for Hadoop offering provides a scalable, resilient data management solution for handling Big Data workloads to help make Apache Hadoop ready for enterprise deployment.
Symantec’s Cluster File System is a proven enterprise solution to address Big Data workloads. With Symantec Enterprise Solution for Hadoop, organisations can:
• Leverage their existing infrastructure by scaling up to 16 PB of data including structured and unstructured data
• Avoid over provisioning on both storage and compute capacity
• Run analytics wherever the data sits, eliminating expensive data moves
• Make Hadoop highly available without a potential single point of failure or a performance bottleneck
Leveraging Existing Infrastructure and Avoiding Over Provisioning
IT administrators have spent considerable time and resources consolidating their data centres and reducing their footprint through virtualisation and cloud computing. Taking advantage of Big Data analytics should leverage this consolidation of storage and compute resources. Symantec Enterprise Solution for Hadoop enables customers to run Hadoop while minimising investments in a parallel infrastructure – greatly reducing the storage footprint to reduce cost and complexity.
Analysing Data Where it Resides and Eliminating Expensive Data Moves
The first step in making the Hadoop infrastructure work is to funnel data for analysis. By enabling integration of existing storage assets into the Hadoop processing framework, organisations can avoid time consuming and costly data movement activities. Symantec Enterprise Solution for Hadoop allows administrators to leave the data where it resides and run analytics on it without having to extract, transform and load it to a separate cluster – avoiding expensive and painful data migrations.
Ensuring Hadoop is Highly Available
In an Apache Hadoop environment, data is distributed across nodes with only one metadata server that knows the data location – potentially resulting in a performance bottleneck and single point of failure that could lead to application downtime. To meet the need for timely insights, Symantec Enterprise Solution for Hadoop provides file system high availability to the metadata server while also ensuring analytics applications continue to run as long as there is at least one working node in the cluster. Since the Hadoop file system is replaced with Symantec’s Cluster File System, each node in the cluster can also access data simultaneously, eliminating both the performance bottleneck and single point of failure.