import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;
import java.sql.PreparedStatement;
import java.sql.ResultSet;
import java.sql.SQLException;
import java.util.Iterator;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.io.Writable;
import org.apache.hadoop.mapred.FileOutputFormat;
import org.apache.hadoop.mapred.JobClient;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.MapReduceBase;
import org.apache.hadoop.mapred.Mapper;
import org.apache.hadoop.mapred.OutputCollector;
import org.apache.hadoop.mapred.Reducer;
import org.apache.hadoop.mapred.Reporter;
import org.apache.hadoop.mapred.lib.db.DBConfiguration;
import org.apache.hadoop.mapred.lib.db.DBInputFormat;
import org.apache.hadoop.mapred.lib.db.DBWritable;/**
* @author DELL_pc
* @date 2017年6月27日*/
public classDbaMysql {public static class DBAccessMapper extends MapReduceBase implements Mapper{public void map(LongWritable key, StudentRecord value, OutputCollectoroutput,
Reporter reporter) throws IOException {//TODO Auto-generated method stub
output.collect(new IntWritable(value.id), newText(value.toString()));
}
}public static class DBAccessReduce extends MapReduceBase implements Reducer{public void reduce(IntWritable key, Iterator values, OutputCollectoroutput,
Reporter reporter) throws IOException {while(values.hasNext()) {
output.collect(key, values.next());
}
}
}public static voidmain(String[] args) {
System.setProperty("hadoop.home.dir", "D:\\hadoop-2.7.6");//这一行一定要
Configuration configuration=newConfiguration();
JobConf jobConf=newJobConf(configuration);
jobConf.setOutputKeyClass(IntWritable.class);
jobConf.setOutputValueClass(Text.class);
jobConf.setInputFormat(DBInputFormat.class);//String[] fields={"id,name"};//DBInputFormat.setInput(jobConf, StudentRecord.class, "bbb", "length(name)>2", "",fields );//bbb是表名,读取方式1
DBInputFormat.setInput(jobConf, StudentRecord.class,"select id,name from bbb","select 3 from dual");//读取方式2
DBConfiguration.configureDB(jobConf,"com.mysql.jdbc.Driver", "jdbc:mysql://localhost:3306/test","root","admin");
jobConf.setMapperClass(DBAccessMapper.class);
jobConf.setReducerClass(DBAccessReduce.class);
FileOutputFormat.setOutputPath(jobConf,new Path("output_mysql"));try{
JobClient.runJob(jobConf);
}catch(IOException e) {//TODO Auto-generated catch block
e.printStackTrace();
}
}public static classStudentRecord implements Writable, DBWritable {intid;
String name;//构造方法
publicStudentRecord() { }//Writable接口是对数据流进行操作的,所以输入是DataInput类对象
public void readFields(DataInput in) throws IOException {this.id = in.readInt(); //输入流中的读取下一个整数,并返回
this.name = Text.readString(in);
}publicString toString() {return new String(this.id + " " + this.name);
}//DBWritable负责对数据库进行操作,所以输出格式是PreparedStatement//PreparedStatement接口继承并扩展了Statement接口,用来执行动态的SQL语句,即包含参数的SQL语句
public voidwrite(PreparedStatement stmt) throws SQLException {
stmt.setInt(1, this.id);
stmt.setString(2, this.name);
}//DBWritable负责对数据库进行操作,输入格式是ResultSet//ResultSet接口类似于一张数据表,用来暂时存放从数据库查询操作所获得的结果集
public voidreadFields(ResultSet result) throws SQLException {this.id = result.getInt(1);this.name = result.getString(2);
}//Writable接口是对数据流进行操作的,所以输出是DataOutput类对象
public void write(DataOutput out) throws IOException {out.writeInt(this.id);
Text.writeString(out, this.name);
}
}
}