直接上代码如下:
**
注意:
1、此程序不需要打jar包到Linux上运行
2、集群为高可用
3、resource包需要添加hadoop的四个配置文件,log4j.properties为日志查看
mapper阶段:
**
package HDFSToHbase.answer_base.mapper;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.NullWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Mapper;
import java.io.IOException;
/**
- @description 从hdfs上把数据迁移到hbase
- @author: gavinwxy@foxmail.com
- @create: 2018-12-06 21:20:59
**/
public class HDFSToHbaseMapper extends Mapper<LongWritable, Text,Text,NullWritable> {
@Override
protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
context.write(value,NullWritable.get());//把每一行数据截取直接发送到reduce端
}
}
**
reduce阶段:
**
package HDFSToHbase.answer_base.reducer;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.mapreduce.TableReducer;
import org.apache.hadoop.io.NullWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Reducer;
import java.io.IOException;
/**
- @description 把数据从hdfs迁移到hbase
- @author: gavinwxy@foxmail.com
- @create: 2018-12-06 21:25:40
**/