Amazon Covers All Big-Data Bases
Amazon is about as big a big data practitioner as you can get. It's also the leading big data services provider. For starters, it introduced Elastic MapReduce (EMR) more than three years ago. Based on Hadoop, EMR isn't just a service for MapReduce sand boxes; it's being used for day-to-day high-scale production data processing by businesses including Ticketmaster and DNA researcher Ion Flux.
Amazon Web Services upped the big data ante in 2012 with two new services: Amazon DynamoDB, a NoSQL database service, and Amazon Redshift, a scalable data warehousing service now in preview and set for release early next year.
DynamoDB, the service, is based on Dynamo, the NoSQL database that Amazon developed and deployed in 2007 to run big parts of its massive consumer website. Needless to say, it's proven at high scale. Redshift has yet to be generally available, but Amazon is promising ten times faster performance than conventional relational databases at one-tenth the cost of on-premises data warehouses. With costs as low as $1,000 per terabyte, per year, there's no doubt Redshift will see adoption.
These three services are cornerstones for exploiting big data, and don't forget Amazon's scalable S3 storage, EC2 compute capacity and myriad integration and connection options for corporate data centers. In short, Amazon has been a big data pioneer, and its services appeal to more than just startups, SMBs and Internet businesses.