- Silicon Valley's 19 Coolest Places to Work
- Is Windows 8 Development Worth the Trouble?
- 8 Books Every IT Leader Should Read This Year
- 10 Hot Hadoop Startups to Watch
IDG News Service - The proliferation of large-scale data sets is just beginning to change business and science around the world, but enterprises need to prepare in order to gain the most advantage from their information, panelists said at a Silicon Valley event this week.
So-called "big data" is both a challenge to manage and a tool for competitive advantage, according to speakers at a Churchill Club event on Wednesday night in Mountain View, California. The discussion at the Computer History Museum followed the launch of EMC Greenplum's Unified Analytics Platform, which lets business and IT staffs analyze both structured and unstructured data.
New networked devices and applications are collecting more data than ever and more organizations are holding on to it, creating huge demands for storage. In the second quarter of this year, storage companies shipped 5,429 petabytes of disk capacity, up 30.7 percent from last year's second quarter, IDC reported last week.
"Data growth is already faster than both Moore's Law and ... network growth," said Anand Rajaraman, senior vice president of Walmart Global E-Commerce and head of @WalmartLabs. His lab has developed tools for Walmart to take advantage of the new types of data being generated, including applications that collect and analyze information from sources such as Twitter and Facebook to gauge trends and individual consumer preferences.
The benefits of big data stretch beyond business to earth sciences, biology, psychology and other fields, Rajaraman said.
"Science has become more and more about collecting large amounts of data and doing analysis," he said.
Big data can be any volume of data that requires new tools to analyze, said Luke Lonergan, chief technology officer and co-founder of Greenplum, which EMC acquired last year. For example, it would take 27 hours to run a logistic regression algorithm, which can be used to predict the probability of an event, on 30G bytes of data, Lonergan said. If run on 32 computers, the process takes 60 seconds, he said.
"'Bigger than previous-generation, non-parallel infrastructure could handle' might be a useful definition. Anything that blows you out of the old way of doing things," Lonergan said.
Analyzing data also has gotten harder not only because there is more of it but because it comes from new sources, panelists said. Blogs, Web comments and other information comes in the form of unstructured data, which can't be crunched the way relational databases are. The need to mine different types of content has led to new data analysis platforms, most notably the open-source Hadoop framework that was pioneered by Google and Facebook.
The market for new tools to manage and exploit big data is still growing, said Ping Li, who heads the Big Data Fund at venture capital company Accel Partners.
"A lot of the applications that ride on top of these new data platforms have yet to be invented," Li said. Traditional business intelligence and ERP (enterprise resource planning) platforms are being adapted to deal with big data, but what's needed are native applications developed specifically for the new world, he said.