The proliferation of large-scale data sets is just beginning to change business and science around the world, but enterprises need to prepare in order to gain the most advantage from their information, panelists said at a Silicon Valley event this week.
So-called "big data" is both a challenge to manage and a tool for competitive advantage, according to speakers at a Churchill Club event on Wednesday night in Mountain View, California. The discussion at the Computer History Museum followed the launch of EMC Greenplum's Unified Analytics Platform, which lets business and IT staffs analyse both structured and unstructured data.
New networked devices and applications are collecting more data than ever and more organisations are holding on to it, creating huge demands for storage. In the second quarter of this year, storage companies shipped 5,429 petabytes of disk capacity, up 30.7 percent from last year's second quarter, IDC reported last week.
"Data growth is already faster than both Moore's Law and ... network growth," said Anand Rajaraman, senior vice president of Walmart Global E-Commerce and head of @WalmartLabs. His lab has developed tools for Walmart to take advantage of the new types of data being generated, including applications that collect and analyse information from sources such as Twitter and Facebook to gauge trends and individual consumer preferences.
Great benefits to science
The benefits of big data stretch beyond business to earth sciences, biology, psychology and other fields, Rajaraman said.
"Science has become more and more about collecting large amounts of data and doing analysis," he said.
Big data can be any volume of data that requires new tools to analyse, said Luke Lonergan, chief technology officer and co-founder of Greenplum, which EMC acquired last year. For example, it would take 27 hours to run a logistic regression algorithm, which can be used to predict the probability of an event, on 30G bytes of data, Lonergan said. If run on 32 computers, the process takes 60 seconds, he said.
"'Bigger than previous-generation, non-parallel infrastructure could handle' might be a useful definition. Anything that blows you out of the old way of doing things," Lonergan said.
Analysing data also has gotten harder not only because there is more of it but because it comes from new sources, panelists said. Blogs, Web comments and other information comes in the form of unstructured data, which can't be crunched the way relational databases are. The need to mine different types of content has led to new data analysis platforms, most notably the open-source Hadoop framework that was pioneered by Google and Facebook.
The market for new tools to manage and exploit big data is still growing, said Ping Li, who heads the Big Data Fund at venture capital company Accel Partners.
Understand the implications
"A lot of the applications that ride on top of these new data platforms have yet to be invented," Li said. Traditional business intelligence and ERP (enterprise resource planning) platforms are being adapted to deal with big data, but what's needed are native applications developed specifically for the new world, he said.
Developing countries are active participants in this process, sometimes because companies there have skipped over legacy systems that are ingrained in first-world enterprises, Li said.
Trying to get value out of big data today is like creating an online store in the early days of e-commerce, said Walmart's Rajaraman, who helped develop Amazon.com's marketplace business. Amazon had to invent its own systems for payment, fraud detection and other tasks, each of which later spawned independent vendors that specialise in each area, he said.
It's important for an enterprise to understand the implications of big data and how the new tools work before embarking on a big-data initiative, panelists warned.
"Those who are just standing up Hadoop as is, with no management framework, writing directly to it ... there's going to be some real disillusionment there," said Keith Collins, senior vice president and chief technology officer of SAS.
Big-data tools such as Hadoop can't create value out of information by themselves, Collins warned in an interview at the event. Enterprises have to know what they want to find out from their data and then deal with how to get that out of their data. "The data issues come after the question," he said.