WordPress Shortcode

Link

Big Data in the Cloud with Informatica Cloud and Amazon Redshift

Data warehousing costs have been continually rising with the explosion of Big Data. To help you explore the most cost-effective data warehousing techniques, learn from the cloud experts from Amazon
…

Data warehousing costs have been continually rising with the explosion of Big Data. To help you explore the most cost-effective data warehousing techniques, learn from the cloud experts from Amazon and Informatica.

Learn more: http://www.informaticacloud.com/amazon-redshift

Amazon Redshift is a petabyte-scale cloud-based data warehouse that allows you to provision multiple database nodes on demand and offload raw data from on-premise databases for more cost effective data warehousing. Getting this data into Redshift is easy with Informatica Cloud. In this interactive webinar, you’ll learn:

Use Amazon Redshift for easy scalabilityMigrate completely from existing DW to Amazon RedshiftAnalyze data that was previously too expensive to put into a DWDeploy Redshift because provisioning existing DW systems takes monthsReplace HIVE with Amazon Redshift if they were using HIVE to save money

Using Amazon Redshift to power its upcoming SkyVault productFully managed by Infor to enable customers to run business analyticsChose Redshift for performance, cost, ease-of-use, and scalability

Read only the data you need

Read only the data you need

Read only the data you need

Read only the data you need

Read only the data you need

Informatica Cloud is powered by the Vibe, the same technology that powers the virtual data machine that runs the secure agent. Thus, you use Informatica Cloud to store the various metadata mappings, and upon run-time, the data moves directly from source to target through the execution of the Vibe Secure Agent.

Vibe is the industry’s first and only embeddable virtual data machine to access, aggregate and manage data – regardless of data type, source, volume, compute platform or user. It lets you map once, and deploy anywhere. So you can take your logic that may have defined on-premise, then move it to the cloud. And then move it to Hadoop, or embed it in an application– without recoding.This makes your architecture faster, more flexible, and futureproof.Business BenefitFive time faster turn-around from business idea to solutionAdapt the technology to your business, not vice-versaUtilize all your data, regardless of location, type or volumeIT BenefitFive times faster project deliveryEliminate skills gaps for adopting new technologies and approachesReduce cost of maintaining complex assortment of technologies

13.
Amazon Redshift Customer: bit.ly
“When we want to answer a
question with Redshift, we
just write a SQL query and
get an answer within a few
minutes – if not seconds.”
- Sean O’Connor, Engineer at bit.ly
Bit.ly provides social link sharing
analytics, managing over 300
million shortens and 5 billion
clicks each month

14.
14
Amazon Redshift Customer: HasOffers
“Amazon Redshift introduces a
major opportunity to improve
the performance of our real-
time reporting, allowing us to
run queries up to 50 times faster
than our current OLAP solution.”
- Niek Sanders, VP of Engineering, HasOffers
HasOffers records and reports
billions of desktop and mobile
interactions for performance
marketers

15.
Amazon Redshift Customer: Infor
“This is the formula for fast and broad
adoption, where customers can get
consistent, accurate, and useful
data fast - in weeks not months or
years.”
- Ali Shadman, SVP, Business Cloud & Upgrades, Infor
Infor is the world’s third largest
ERP vendor, serving over 70,000
customers in 194 countries

27.
Amazon Redshift continuously backs up your data and
recovers from failures
• Replication within the cluster and backup to Amazon S3 to maintain
multiple copies of data at all times
• Backups to Amazon S3 are continuous, automatic, and incremental
– Designed for eleven nines of durability
• Continuous monitoring and automated recovery from failures of drives
and nodes
• Able to restore snapshots to any Availability Zone within a region
Slides not intended for redistribution.

34.
Best practices to remember…
• The Amazon S3 bucket that holds the data files must be
created in the same region as your cluster
• Files are deleted from Amazon S3 bucket when upload is
complete
• Choose a batch size where the number of batches
matches the number of slices in your cluster
• Each XL node has 2 slices, each 8XL node has 16
• If you have a 2 node XL cluster and 40,000 rows of data,
choose a batch size of 10,000
• The Informatica Cloud Redshift connector can maximize
Amazon’s parallel processing capabilities this way

36.
PowerCenter Mappings and Informatica Cloud
• If you want to reuse your existing PowerCenter mappings
with Informatica Cloud and Redshift you have 2 options:
• Use the PowerCenter Repository Manager to export your
existing workflows and import them into Informatica Cloud
using the PowerCenter Tasks feature
Or…
• Keep your existing mappings in PowerCenter and stage the
data
• Create a DSS task in Informatica Cloud to move the data to
Redshift from the staging area
• This task can be managed from PowerCenter
1
2