Amazon Web Services has made available to all users its cloud-based data warehouse Redshift, which it pitches as a lower-cost alternative to on-premise deployments.
Amazon Redshift was first announced as a limited preview at the AWS re:Invent conference, but has now been made generally available from its US East (North Virginia) data center and will be rolled out to other centers in the coming months.
[ Explore the current trends and solutions in BI with InfoWorld's interactive Business Intelligence iGuide. | Keep up with the latest approaches to managing information overload and staying compliant in InfoWorld's Enterprise Data Explosion newsletter. ]
Just like with its other cloud-based offerings, Amazon hopes to attract enterprises to Redshift with the speed and cost of setting up a data warehouse in its cloud.
Traditional data warehouse solutions are really expensive and complicated to manage, Amazon Web Services' Andy Jassy said when the product was launched. Redshift, on the other hand, is about a tenth of the cost, and also automates the deployment and administration, according to Jassy.
With the AWS Management Console or the Amazon Redshift APIs, users can provision a single 2TB data warehouse or as a cluster of 16 2TB nodes or 16TB nodes, by default.
The nodes are called High Storage Extra Large (XL) and Storage Eight Extra Large (8XL).In addition to 2TB or 16TB of storage, they also have 15GB or 120GB of RAM.
On-demand pricing starts at $0.85 per hour for an XL node and $6.80 per hour for the 8XL node. Reserved instance pricing lowers the effective price to $0.228 per hour or under $1,000 per terabyte per year, according to Amazon.
Good security and reliability are key ingredients for any data warehouse. Data written to a node in a Redshift cluster is automatically replicated to other nodes within the cluster and all data is continuously backed up to Amazon's Simple Storage Service (S3), according to Amazon.
To secure data, Redshift can use SSL in transit and hardware-acccelerated AES-256 encryption for primary storage as well as backed-up data. Using Virtual Private Cloud, Redshift can be connected with an enterprise's existing data center using encrypted VPN tunnels.
A number of vendors have been testing their big data and analytics products with the SQL-based Redshift. Users can buy Jaspersoft's Reporting and Analytics for AWS on Amazon's Marketplace, for example. The service can also be integrated with business intelligence tools from the likes of Business Object and Cognos, according to Amazon.
Data can be uploaded to Redshift in a number of different ways. Companies that have a lot of data can use AWS Direct Connect to set up a private network connection at 1G bps or 10G bps between their datacenter and Amazon's cloud. They can also use AWS Import/Export to send data on portable storage devices.
Redshift can also use AWS Data Pipeline to import data or load data directly from services such as S3 and DynamoDB.
Send news tips and comments to email@example.com.