MapR Looks To Speed-Up Hadoop Clusters - InformationWeek

InformationWeek is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them.Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 8860726.

IoT
IoT
Data Management // Big Data Analytics

MapR Looks To Speed-Up Hadoop Clusters

With the release of its Distribution 5.0 management tool and other features, MapR is looking to make Hadoop easier to use and increase its popularity within the enterprise.

6 Causes Of Big Data Discrepancies
6 Causes Of Big Data Discrepancies
(Click image for larger view and slideshow.)

MapR Technologies wants to make Hadoop faster for its users.

First, the company is releasing MapR Distribution 5.0 to speed real-time management of applications. Complementing this is the introduction of automated templates that can shrink the time it takes to deploy a Hadoop cluster from hours to minutes.

"(We're) looking at a trend. It's about helping companies become more agile. It is about data agility," explained Jack Norris, the chief marketing officer for MapR.

Companies that are handling thousands or millions of transactions online need a way to analyze and manage their applications in real time. These could be call centers, online shopping sites, financial service firms, and the like -- anything involving a high volume of transactions.

"Overall, this is going to help the organization compress data to the action cycle," Norris said.

MapR's auto-provisioning template

MapR's auto-provisioning template

Hadoop is usually used to collect, manage, and analyze large pools of big data. MapR's course is to add toolsets to link Hadoop to real-time operations, in effect tapping big data in real-time to sharpen decision-making.

MapR 5.0, which the company officially announced at the Hadoop Summit in San Jose, Calif., on Tuesday, June 9, will enable organizations to auto-synchronize storage, database, and search across the enterprise; improve its Real-time Data Transport framework to enable real-time search; allow rolling upgrades for applications in Hadoop clusters to complement platform-level upgrades; and improve data governance with auditing for all data access through Apache Drill 1.1, adjusting access privileges for specific analysts.

Concurrently, the MapR Auto-Provisioning Template is also being launched, making a big improvement in the deployment of Hadoop clusters.

Previously, "administrators deployed Hadoop on a node-by-node basis," Norris explained, which involved a lot of repetitive processes. Permissions and dependencies had to be set up just right on each node to avoid problems. Auto-provisioning simply automates that process so that Hadoop can be deployed across racks, with the flexibility to do complex provisions, he added. This can knock down a Hadoop deployment from hours to minutes.

MapR's data governance product.

MapR's data governance product.

The Auto-Provisioning Template is replete with control utilities. Common Hadoop services are all bundled together in a feature called "Data Lake." Interactive SQL with Apache Drill provides the data exploration component. NoSQL with MapR-DB provides the operational analytic portion. Auto-layout, rack awareness, and health checks automates distribution and monitoring for the systems administrator.

Hadoop is only 10-years-old. Yes, the child has grown. Over the years, developers have added tool sets and utilities to make Hadoop more commercially useful. But despite Hadoop's compelling utility, adoption has been slow.

"The big issue regarding Hadoop is education and knowledge transfer," Norris said. In January, MapR rolled out free online demos of Hadoop training tailored for administrators, developers, and data analysts. About 20,000 users are participating, Norris added.

[Read about big data in real life.]

Another MapR strategy to speed adoption is "making sure Hadoop integrates into existing environments," Norris continued. That means making sure MapR is not creating a separate environment for Hadoop. The solution needs "to be data centric, not Hadoop-centric," he said.

"We're going through one of the biggest re-platforming exercises in the data center," Norris said. As a result, MapR has to be in the center of enterprise computing. "We have to help organizations respond faster and more accurately to take advantage of changing business conditions," he said.

However, what companies such as MapR and even Salesforce are trying to do, is to make big data itself easier to use for the entire enterprise, not just data scientists or those with deep knowledge of the technology. If successful, these efforts could make big data a true, enterprise-wide technology that anyone can access and use.

William Terdoslavich is an experienced writer with a working understanding of business, information technology, airlines, politics, government, and history, having worked at Mobile Computing & Communications, Computer Reseller News, Tour and Travel News, and Computer Systems ... View Full Bio

We welcome your comments on this topic on our social media channels, or [contact us directly] with questions about the site.
Comment  | 
Print  | 
More Insights
Comments
Newest First  |  Oldest First  |  Threaded View
Brian.Dean
50%
50%
Brian.Dean,
User Rank: Ninja
6/10/2015 | 11:10:54 PM
Saving Resources
Every employee requires data to complete their job in the most efficient way possible but, every employee cannot work with Hadoop. It is tools such as, MapR and Cloudera, etc., that can shorten the gap between data and the employee in the enterprise.

It is achievable because, Google has managed to deliver big-data to consumers. For instance, Google Adword has a Keyword Planner, allowing a user to screen millions of keywords that are being searched for by users in real-time. If it is achievable with the general user then, it should be easier in the enterprise because, database expertise is greater.
Commentary
Study Proposes 5 Primary Traits of Innovation Leaders
Joao-Pierre S. Ruth, Senior Writer,  11/8/2019
Slideshows
Top-Paying U.S. Cities for Data Scientists and Data Analysts
Cynthia Harvey, Freelance Journalist, InformationWeek,  11/5/2019
Slideshows
10 Strategic Technology Trends for 2020
Jessica Davis, Senior Editor, Enterprise Apps,  11/1/2019
White Papers
Register for InformationWeek Newsletters
Video
Current Issue
Getting Started With Emerging Technologies
Looking to help your enterprise IT team ease the stress of putting new/emerging technologies such as AI, machine learning and IoT to work for their organizations? There are a few ways to get off on the right foot. In this report we share some expert advice on how to approach some of these seemingly daunting tech challenges.
Slideshows
Flash Poll