MySQL Applier for Hadoop
Replication via the Hadoop Applier is implemented by connecting to the MySQL master and reading binary log events as soon as they are committed, and writing them into a file in HDFS. "Events" describe database changes such as table creation operations or changes to table data.

The Hadoop Applier uses an API provided by libhdfs, a C library to manipulate files in HDFS. The library comes precompiled with Hadoop distributions.
It connects to the MySQL master to read the binary log and then:
- Fetches the row insert events occurring on the master
- Decodes these events, extracts data inserted into each field of the row, and uses content handlers to get it in the format required
- Appends it to a text file in HDFS.
Databases are mapped as separate directories, with their tables mapped as sub-directories with a Hive data warehouse directory. Data inserted into each table is written into text files (named as datafile1.txt) in Hive / HDFS. Data can be in comma separated format; or any other, that is configurable by command line arguments.
download from http://labs.mysql.com/
Preferences
http://dev.mysql.com/tech-resources/articles/mysql-hadoop-applier.html
http://www.tuicool.com/articles/NfArA3i
a similar project is https://github.com/noplay/python-mysql-replication
本文介绍了一种通过MySQL Applier实现Hadoop数据复制的方法,该方法通过连接MySQL主服务器并读取已提交的二进制日志事件,然后将这些事件写入HDFS文件。数据库更改事件被描述为表创建操作或表数据更改。数据以逗号分隔或其他可配置格式写入Hive/HDFS。
741

被折叠的 条评论
为什么被折叠?



