Platform Computing Announces Support for Apache Hadoop MapReduce Programming Model
Platform Computing has announced that the company is now bringing enterprise-class distributed computing to business analytics applications that process “big” data using the Apache Hadoop MapReduce programming model. The company states that, based on more than 18 years of industry leadership in workload management for high performance computing (HPC) applications, Platform Computing’s analytics solutions are a natural expansion of the company’s distributed computing experience built on the company’s core technologies, Platform LSF and Platform Symphony.
“MapReduce is an important technique for handling big data problems,” said Paul Kent, Vice President Platform R&D at Cary, NC-based SAS. “SAS is looking forward to continuing our enterprise-class partnership with Platform Computing as we integrate this technique into our Data Management and Business Analytics software.”
Extending enterprise-class capabilities to MapReduce distributed workloads will provide the ability to scale to thousands of commodity server cores for shared applications. The results include the ability to perform at very high execution rates, offer IT manageability and monitoring while controlling workload policies for multiple lines of business users and applications and obtain built-in, high availability services that ensure quality of service.
“Platform Computing has been providing solutions for distributed computing infrastructures that align well to the MapReduce paradigm,” said Carl Olofson, Research Vice President, IDC. “Analysis of unstructured data provides a competitive advantage to companies looking to understand behaviors and trends. Dynamically defined data can require very rapid analysis in bulk, and sensor data has volumes that swamp conventional data centers. Customers need a robust solution to manage and process their dynamically defined data, their sensor data, and their unstructured data. MapReduce has proven to be a leading tool for analyzing this data, but customers need enterprise-class solutions to ensure manageability and scalability for these environments. Platform is positioned well to provide distributed workload and enterprise class middleware to address these challenges.”
Platform Computing’s solution also provides enterprise-class capabilities to deliver scaled-out MapReduce workload distribution. Designed to support more than 1,000 simultaneous applications, organizations can dramatically increase server utilization for up to 40,000 cores across all resources resulting in a high return on investment. Unlike less sophisticated solutions that lack multiple analytic application support and scalable distributed workload engines, Platform’s distributed workload services are designed for high scalability, fast performance, and extreme application compatibility through its low-latency distributed architecture. MapReduce application workloads can now run with high reliability under powerful central management, thereby meeting IT SLAs with high reliability and consistency.
“Many of Platform’s customers already use our products to run complex analytics and other distributed workload services,” said Ken Hertzler, Vice President, Product Management, Platform Computing. “Platform is perfectly positioned to run enterprise-class distributed workload for MapReduce applications. Our products are architected from the outset to service large-scale parallel processing on commodity infrastructures. The solutions are also designed to work specifically with multiple distributed file systems, avoiding customer lock-in and offering a single, compatible, distributed computing workload solution throughout the enterprise.”