概述
import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
import java.util.StringTokenizer;
import java.util.Vector;
import org.apache.hadoop.hbase.mapreduce.TableReducer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.mapreduce.TableMapper;
import org.apache.commons.lang.Validate;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.NullWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapred.TextInputFormat;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.eclipse.jdt.internal.codeassist.complete.CompletionOnArgumentName;
import sun.swing.MenuItemLayoutHelper.ColumnAlignment;
import org.apache.hadoop.hbase.client.Result;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
public class MapHbase {
public static class TokenizerMapper_hbase extends TableMapper<Text, Text> {
public void map(ImmutableBytesWritable key, Result value, Context context)
throws IOException, InterruptedException {
List<KeyValue> kvs = value.list();
for (KeyValue val : kvs) {
System.out.println("column " + new String(val.getQualifier()) + " value " + new String(val.getValue()) + " key " + new String(val.getRow()));
String colname = new String(val.getQualifier());
String colvalue = new String(val.getValue());
context.write(new Text(val.getRow()), new Text(colname + "#" + colvalue));
}
}
}
public static class IntSumReducer_hbase extends TableReducer<Text, Text, ImmutableBytesWritable> {
public void reduce(Text key, Iterable<Text> values, Context context)
throws IOException, InterruptedException {
Put put = new Put(Bytes.toBytes(key.toString()));
for (Text i : values) {
String val[] = i.toString().split("#");
if(val.length == 2){
String colname = val[0];
String colvalue = val[1];
put.add(Bytes.toBytes("cf"), Bytes.toBytes(colname), Bytes.toBytes(colvalue));
} else {
String colname = val[0];
put.add(Bytes.toBytes("cf"), Bytes.toBytes(colname), Bytes.toBytes(""));
}
}
context.write(null, put);
}
}
public static void main(String[] args) throws Exception {
Configuration conf = HBaseConfiguration.create();
conf.set("hbase.zookeeper.quorum",
"datanode01.isesol.com,datanode02.isesol.com,datanode03.isesol.com,datanode04.isesol.com,cmserver.isesol.com");
conf.set("hbase.zookeeper.property.clientPort", "2181");
Scan scan = new Scan();
scan.addFamily(Bytes.toBytes("cf"));
//scan.addColumn(Bytes.toBytes("cf"), Bytes.toBytes("content"));
//scan.setMaxVersions();
Job job = new Job(conf, "t_ui_all");
job.setJarByClass(MapHbase.class);
// job.setMapperClass(TokenizerMapper_hbase.class);
// job.setMapOutputKeyClass(ImmutableBytesWritable.class);
// job.setMapOutputValueClass(Text.class);
TableMapReduceUtil.initTableMapperJob("t_ui_all", scan, TokenizerMapper_hbase.class, Text.class,
Text.class, job);
TableMapReduceUtil.initTableReducerJob("test2", IntSumReducer_hbase.class, job);
// FileInputFormat.addInputPath(job, new Path(args[0]));
System.exit(job.waitForCompletion(true) ? 0 : 1);
}
}
来自 “ ITPUB博客 ” ,链接:http://blog.itpub.net/17036462/viewspace-2140435/,如需转载,请注明出处,否则将追究法律责任。
转载于:http://blog.itpub.net/17036462/viewspace-2140435/
最后
以上就是善良心锁为你收集整理的MapReduce读写HBASE的全部内容,希望文章能够帮你解决MapReduce读写HBASE所遇到的程序开发问题。
如果觉得靠谱客网站的内容还不错,欢迎将靠谱客网站推荐给程序员好友。
本图文内容来源于网友提供,作为学习参考使用,或来自网络收集整理,版权属于原作者所有。
发表评论 取消回复