Amazon Redshift has been providing scalable, quick-to-access analytics platforms for many years, but the question remains: how do you get the data from your existing datastore into Redshift for processing?
Traditional ETL methods can't keep up with large volumes of data and can require manual reprocessing when an error occurs. Running queries by record change date put a load on your MySQL server and pollutes your cache.
Wouldn't it be great if you could replicate your data in real time, filter on the tables and schemas you need, all without putting any extra load on your MySQL server? Wouldn't it also be great if schema changes just flowed through from MySQL to RedShift, without intervention on your part?
Join us as we explain how you can have it all: real-time, secure replication from MySQL/MariaDB/RDS MySQL/Aurora to RedShift, with schema changes replicated and no replaying of jobs needed when errors occur.
Matthew Lang, Director of Professional Services ? Americas, has over 20 years of experience in database administration, database programming, and system architecture, including the creation of a database replication product that is still in use today. He has designed highly available, scalable systems that have allowed startups to quickly become enterprise organizations, utilizing a variety of technologies including open source projects, virtualization and cloud.