Inside the Briefcase

IT Briefcase Exclusive Interview: Getting the Most Out of Open Source While Managing License Compliance, Risk, and Security

IT Briefcase Exclusive Interview: Getting the Most Out of Open Source While Managing License Compliance, Risk, and Security

with Kendra Morton, Flexera
In this interview, Kendra Morton,...

Why DEM Matters More Than Ever in Financial Services

Why DEM Matters More Than Ever in Financial Services

Remember waiting in line at the bank? Banking customers...

How to Transform Your Website into a Lead Generating Machine

How to Transform Your Website into a Lead Generating Machine

Responsive customer service has become of special importance, as...

Ironclad SaaS Security for Cloud-Forward Enterprises

Ironclad SaaS Security for Cloud-Forward Enterprises

The 2015 Anthem data breach was the result of...

The Key Benefits of Using Social Media for Business

The Key Benefits of Using Social Media for Business

Worldwide, there are more than 2.6 billion social media...

Large Scale Analytics in the Enterprise

August 9, 2012 No Comments

SOURCE: Think Big Analytics

The growth of Internet businesses led to a whole new scale of data processing
challenges. Companies like Google, Facebook, Yahoo, Twitter, and Quantcast now
routinely collect and process hundreds to thousands of terabytes of data on a daily basis.
This represents a significant change in the volume of data which can be processed, a
major reduction in processing time required, and of the cost required to store data. The
most important of the techniques used at these companies is storing data in a cluster of
servers and using a distributed data processing technique Google invented, called
MapReduce. Facebook, Yahoo, Twitter, and Quantcast all process data with an open
source technology implementation of MapReduce called Hadoop.

Click here to view this white paper

DATA and ANALYTICS , Featured White Papers

Leave a Reply

(required)

(required)


ADVERTISEMENT

Gartner