Archive for category Big Data
After a (very) brief respite, #theCUBE is back on the conference circuit. Starting tomorrow at 10:20am ET, theCUBE is live for two days at the HP Vertica Big Data Conference (#HPBigData2013) in Boston. Unlike some Big Data shows that focus more on vendors, the Vertica show is heavy on customers and real Big Data end-users.
Everyone knows that the shift from traditional print and broadcast advertising to digital advertising is taking a huge toll on the media industry. Newspapers can’t charge as much for online ads as they do for print ads, and revenues are shrinking precipitously.
But the impact of this shift is being felt not just in the media industry but in the advertising industry itself. Advertising agencies, whose bread and butter is building creative ad campaigns and negotiating print and broadcast ad placement, are likewise under pressure to make the transition to a data-based marketplace.
There’s an old saying in the data management world: garbage in, garbage out, or GIGO. It means that the results of any data analysis project are only as good as the quality of the data being analyzed. Data quality is of critical importance when data sets are relatively small and structured. If you only have a small sample of data on which to perform your analysis, it better be good data. Otherwise, the resulting insights aren’t insights at all.
The show is a mix of both technical and business-centered Hadoop content. There will be a number of major focuses. One such focus is YARN. As I wrote on SiliconANGLE earlier today:
YARN is as a true Hadoop resource manager, allowing multiple applications – MapReduce, SQL, streaming analysis, etc. – to run on a single cluster of machines simultaneously while maintaining high performance levels. With YARN Hadoop is a true multi-application platform that can serve an entire enterprise.
The benefits of Big Data are often spoken of in the future tense. As in, “Big Data will someday provide enterprises of all types critical insights that allow for increased profitability, improved efficiency and other untold riches.”
Same goes for the technology. Hadoop, some say, will be the foundation of data storage and analytics in the enterprise once it’s proven enterprise-ready.
The reality, however, is that Big Data is today – here and now – delivering on its many promises. We at Wikibon have been documenting Big Data in the Real World for the last two years, including publishing a series of vertical-specific research notes highlighting how enterprises in retail, banking, media, utilities and pharma are leveraging Big Data analytics to drive performance.
Our friends at Forbes.com have put together a fantastic new infographic leveraging data from Wikibon’s Big Data Vendor Revenue and Market Forecast, 2012 – 2017 report. It provides a compelling view of the Big Data universe and illustrates the real revenue vendors are deriving from Big Data. They range from the mega-planets (if you’ll go with me on this analogy) IBM, HP and EMC to the smaller but powerful emerging planets like Hortonworks, 10gen and DataStax.
Ok, not the greatest analogy but still a great infographic:
Then-CEO Sam Palmisano launched IBM’s Smarter Planet initiative five years ago during a speech at the Council on Foreign Relations. IBM would focus its energies, Palmisano said, on helping governments and companies understand and analyze the voluminous data streaming off connected devices and industrial equipment to improve operational efficiencies and deliver better services to citizens and customers.
Since then, IBM has largely had the Industrial Internet, as the concept of has come to be called, to itself. The company’s Smarter Planet division has played a key role in making IBM the biggest Big Data company on the planet and was a lone bright spot in IBM’s otherwise disappointing Q1 2013 results.
Recently, IBM announced a $1 billion initiative intended to improve the overall flash storage market and integrate flash storage in the company’s line of enterprise technology equipment, including servers, storage, and other products. The company feels that flash-based storage is an a tipping point in the marketplace and is poised to become much more widely used, thanks to the incredible performance gains offered by the technology. Further, as is the case with any technology, as it approaches a critical mass point, the overall costs of the technology begin to drop and this is certainly happening with flash storage. There are also other significant cost benefits to flash-based storage, such as reduced power consumption. At scale, such power savings can be real and significant.
I’d like to explore the topic of how system and storage architectures are changing and the impact this will have on application delivery and organizational productivity.
Allow me to put forth the following premise:
Today’s enterprise IT infrastructure limits application value.
What does that mean? To answer this, let’s first explore the notion of value. The value IT brings to an organization flows directly from the application to the business and is measured in terms of the productivity of the organization. Infrastructure in-and-of itself delivers no direct value; however the applications, which run on infrastructure directly affect business value. Value comes in many forms but at the highest level it’s about increasing revenue and/or cutting costs; and ultimately delivering bottom line profits.
The first iteration of HP’s new line of low-power servers, known as Moonshot, begins shipping this week. HP plans to release numerous cartridges for the chassis over the next several quarters, each optimized for specific workloads. The first cartridge for the HP Moonshot 1500 chassis utilizes Intel’s Atom chip and is aimed at web hosting; HP promises future versions of the server optimized for Big Data.