Inside the Briefcase

FICO Scales with Oracle Cloud

FICO Scales with Oracle Cloud

Doug Clare, Vice President at FICO, describes how Oracle...

Is Your Enterprise IT the Best It Can Be?

Is Your Enterprise IT the Best It Can Be?

Enterprise IT is a driver of the global economy....

The IoT Imperative for Consumer Industries

The IoT Imperative for Consumer Industries

This IDC white paper examines current and future...

How to align your visual brand guidelines and create consistently on-brand content

How to align your visual brand guidelines and create consistently on-brand content

In this ebook, we’ll explore the various themes leading...

Your B2B Content Strategy in 2017: How To Think Like A Movie Studio + 6 Other Tactics

Your B2B Content Strategy in 2017: How To Think Like A Movie Studio + 6 Other Tactics

Jon Lombardo, Creative Lead, LinkedIn, reveals in this presentation...

Large Scale Analytics in the Enterprise

August 9, 2012 No Comments

SOURCE: Think Big Analytics

The growth of Internet businesses led to a whole new scale of data processing
challenges. Companies like Google, Facebook, Yahoo, Twitter, and Quantcast now
routinely collect and process hundreds to thousands of terabytes of data on a daily basis.
This represents a significant change in the volume of data which can be processed, a
major reduction in processing time required, and of the cost required to store data. The
most important of the techniques used at these companies is storing data in a cluster of
servers and using a distributed data processing technique Google invented, called
MapReduce. Facebook, Yahoo, Twitter, and Quantcast all process data with an open
source technology implementation of MapReduce called Hadoop.

Click here to view this white paper

DATA and ANALYTICS , Featured White Papers

Leave a Reply

(required)

(required)


ADVERTISEMENT

Gartner IT Operations

SuperCharge Your Cloud

American CISO

IBC 2018

ITBriefcase Comparison Report