27 February, 2014
There’s a lot of confusing jargon and buzzwords in this new field. It helps to know who some of the major players are and…
Publication: Data Science Central
24 February, 2014
A question I receive quite a bit from people who are newly introduced to NoSQL platforms and the overall concept of multi-structured data sets…
Publication: Information Management
30 March, 2015
There is a reason why Hadoop has become the center of the analytics discussion. The framework can process more kinds of information and faster than traditional data warehouses at a fraction of the cost, yet a new study reveals that organizations are in no rush to switch.
Publication: Silicon Angle
24 March, 2015
In a world of real-time data, why are we still so fixated on Hadoop?
Hadoop, architected around batch processing, remains the poster child for big data, though its outsized reputation still outpaces actual adoption, as 451 Research survey data shows.
[ Harness the power of Hadoop — find out how in…
20 March, 2015
Processing of data in a Hadoop cluster can be done in a variety of ways including streaming, micro-batched, batch mode, and by issuing interactive queries. The boundaries between these ways of processing data are not always completely sharp, but the basic distinctions are important to keep in mind. We…
6 March, 2015
A major issue facing Hadoop early adopters is… now what? While all the excitement about Hadoop is well justified, many people don’t fully understand its limitations – such as the difficulty of connecting and linking data across the cluster – and thus don’t use Hadoop to its fullest potential. you…
Publication: Silicon Angle
18 February, 2015
Data projects are now moving from the experimental stage to providing real returns on investment, with tools such as Hadoop and Cassandra forming an integral part of organizations’ enterprise-wide analytics platform. However, big data is a big investment, both in terms of money and time.
The faster companies are…
Publication: Datacenter Dynamics
3 February, 2015
The idea of gaining knowledge through specialized analysis of mass data started with data collection in the 1960s, and has steadily increased both in the amount of data processed and the sophistication of questions businesses try to answer.
Through this progression from static to dynamic and now to proactive provision…
1 December, 2014
We’re being overrun with Big Data and that has created a need to increase computing and networking power to make it possible to manage the vast amount of information available.
Toward that goal, a new generation of Information and communications technology (ICT) systems has been inspired by the operating…
Publication: Science 2.0
17 November, 2014
With data production accelerating to unprecedented rates, many organizations have turned to Hadoop as an inexpensive way to store and process that data. But those new to Hadoop often find themselves confronting a technology as inscrutable as its name. What is Hadoop? What does it do? What’s with the…
8 October, 2014
When we deployed the first production Hadoop cluster in 2006, we were looking to build a more efficient and cost-effective web search index at Yahoo. Very quickly, other groups at Yahoo started using Hadoop for research jobs and revenue-driving applications, such as click prediction for sponsored search. Since…
22 September, 2014
Hadoop, the open source big-data framework, has gradually evolved from being a shiny object in the laboratory to an increasingly serious tool finding its place in the Enterprise. At Gigaom, we’ve covered Hadoop’s increasing maturity, and completeness as an enterprise technology, because that’s been the story…