HDFS之JAVA API

来源:互联网 发布:daum potplayer mac版 编辑:程序博客网 时间:2024/04/30 12:43

一个通过JAVA API操作HDFS文件系统的例子,本例子使用的是hadoop0.20的版本,在windows的eclipse下运行的时候,需要将core-site.xml和hdfs-site.xml放在src/bin目录中。

package com.util;import java.io.BufferedInputStream;import java.io.FileInputStream;import java.io.FileNotFoundException;import java.io.FileOutputStream;import java.io.IOException;import java.io.InputStream;import java.io.OutputStream;import java.net.URI;import org.apache.hadoop.conf.Configuration;import org.apache.hadoop.fs.FSDataInputStream;import org.apache.hadoop.fs.FSDataOutputStream;import org.apache.hadoop.fs.FileStatus;import org.apache.hadoop.fs.FileSystem;import org.apache.hadoop.fs.Path;import org.apache.hadoop.io.IOUtils;import org.apache.hadoop.util.Progressable;public class HdsfAPI {/** 上传文件到HDFS上去 */private static void uploadToHdfs(){String localSrc = "D://mahout.jpg";String dst = "hdfs://wangchen-virtual-machine:9000/input/image2";try {InputStream in = new BufferedInputStream(new FileInputStream(localSrc));Configuration conf = new Configuration();FileSystem fs = FileSystem.get(URI.create(dst), conf);OutputStream out = fs.create(new Path(dst), new Progressable() {public void progress() {System.out.print(".");}});IOUtils.copyBytes(in, out, 4096, true);} catch (FileNotFoundException e) {e.printStackTrace();} catch (IOException e) {e.printStackTrace();}}/** 从HDFS上读取文件到本地*/private static void readFromHdfs() throws FileNotFoundException,IOException {String dst = "hdfs://localhost:9000/user/wangchen/uploadToHdfs.txt";Configuration conf = new Configuration();FileSystem fs = FileSystem.get(URI.create(dst), conf);FSDataInputStream hdfsInStream = fs.open(new Path(dst));OutputStream out = new FileOutputStream("/home/wangchen/getFromHdfs.txt");byte[] ioBuffer = new byte[1024];int readLen = hdfsInStream.read(ioBuffer);while (-1 != readLen) {out.write(ioBuffer, 0, readLen);readLen = hdfsInStream.read(ioBuffer);}out.close();hdfsInStream.close();fs.close();}/** * 以append方式将内容添加到HDFS上文件的末尾; * 注意:文件更新,需要在hdfs-site.xml中添<property><name>dfs. * append.support</name><value>true</value></property> * 注意:对于append操作,从hadoop-0.21版本开始就不支持了 */private static void appendToHdfs() throws FileNotFoundException,IOException {String dst = "hdfs://localhost:9000/user/wangchen/uploadToHdfs.txt";Configuration conf = new Configuration();FileSystem fs = FileSystem.get(URI.create(dst), conf);FSDataOutputStream out = fs.append(new Path(dst));int readLen = "wangchen append this file !".getBytes().length;while (-1 != readLen) {out.write("wangchen append this file !".getBytes(), 0, readLen);}out.close();fs.close();}/** 从HDFS上删除文件 */private static void deleteFromHdfs() throws FileNotFoundException,IOException {String dst = "hdfs://localhost:9000/user/wangchen/uploadToHdfs.txt";Configuration conf = new Configuration();FileSystem fs = FileSystem.get(URI.create(dst), conf);fs.deleteOnExit(new Path(dst));fs.close();}/** 遍历HDFS上的文件和目录 */private static void getDirectoryFromHdfs() throws FileNotFoundException,IOException {String dst = "hdfs://localhost:9000/user/wangchen";Configuration conf = new Configuration();FileSystem fs = FileSystem.get(URI.create(dst), conf);FileStatus fileList[] = fs.listStatus(new Path(dst));int size = fileList.length;for (int i = 0; i < size; i++) {System.out.println("name:" + fileList[i].getPath().getName()+ "\t\tsize:" + fileList[i].getLen());}fs.close();}public static void main(String[] args) {uploadToHdfs();/*try {uploadToHdfs();//readFromHdfs();//appendToHdfs();//deleteFromHdfs();//getDirectoryFromHdfs();} catch (FileNotFoundException e) {e.printStackTrace();} catch (IOException e) {e.printStackTrace();}*/}}
操纵的方法有很多种,下面的例子仅供参考,

[java] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. public class TestHDFSFile {  
  2.   
  3.     private String localPath = "C:/D/JavaWorkSpace/bigdata/temp/";  
  4.     private String hdfsPath = "hdfs://192.168.2.6:9000/user/hadoop/temp/";  
  5.   
  6.     public static void main(String[] args) throws Exception {  
  7.         // new TestHDFSFile().testUpload();  
  8.         // new TestHDFSFile().testCreate();  
  9.         //new TestHDFSFile().testRename();  
  10.         //new TestHDFSFile().testDel();  
  11.         //new TestHDFSFile().testgetModifyTime();  
  12.         //new TestHDFSFile().testExists();  
  13.         //new TestHDFSFile().testFileBlockLocation();  
  14.           
  15.         new TestHDFSFile().testGetHostName();  
  16.           
  17.     }  
  18.   
  19.     // 上传本地文件到HDFS  
  20.     public void testUpload() throws Exception {  
  21.   
  22.         Configuration conf = new Configuration();  
  23.         // conf.addResource(new Path(localPath + "core-site.xml"));  
  24.         FileSystem hdfs = FileSystem.get(conf);  
  25.         Path src = new Path(localPath + "file01.txt");  
  26.         Path dst = new Path(hdfsPath);  
  27.         hdfs.copyFromLocalFile(src, dst);  
  28.   
  29.         System.out.println("Upload to " + conf.get("fs.default.name"));  
  30.         FileStatus files[] = hdfs.listStatus(dst);  
  31.         for (FileStatus file : files) {  
  32.             System.out.println(file.getPath());  
  33.         }  
  34.     }  
  35.   
  36.     // 创建HDFS文件  
  37.     public void testCreate() throws Exception {  
  38.         Configuration conf = new Configuration();  
  39.         byte[] buff = "hello world!".getBytes();  
  40.   
  41.         FileSystem hdfs = FileSystem.get(conf);  
  42.         Path dst = new Path(hdfsPath + "hello.txt");  
  43.         FSDataOutputStream outputStream = null;  
  44.         try {  
  45.             outputStream = hdfs.create(dst);  
  46.             outputStream.write(buff, 0, buff.length);  
  47.         } catch (Exception e) {  
  48.             e.printStackTrace();  
  49.   
  50.         } finally {  
  51.             if (outputStream != null) {  
  52.                 outputStream.close();  
  53.             }  
  54.         }  
  55.   
  56.         FileStatus files[] = hdfs.listStatus(dst);  
  57.         for (FileStatus file : files) {  
  58.             System.out.println(file.getPath());  
  59.         }  
  60.     }  
  61.   
  62.     // 重命名HDFS文件  
  63.   
  64.     public void testRename() throws Exception {  
  65.   
  66.         Configuration conf = new Configuration();  
  67.   
  68.         FileSystem hdfs = FileSystem.get(conf);  
  69.         Path dst = new Path(hdfsPath);  
  70.   
  71.         Path frpath = new Path(hdfsPath + "hello.txt");  
  72.         Path topath = new Path(hdfsPath + "hello2.txt");  
  73.   
  74.         hdfs.rename(frpath, topath);  
  75.   
  76.         FileStatus files[] = hdfs.listStatus(dst);  
  77.         for (FileStatus file : files) {  
  78.             System.out.println(file.getPath());  
  79.         }  
  80.     }  
  81.   
  82.     // 刪除HDFS文件  
  83.     public void testDel() throws Exception {  
  84.   
  85.         Configuration conf = new Configuration();  
  86.   
  87.         FileSystem hdfs = FileSystem.get(conf);  
  88.         Path dst = new Path(hdfsPath);  
  89.   
  90.         Path topath = new Path(hdfsPath+ "hello2.txt");  
  91.   
  92.         boolean ok = hdfs.delete(topath, false);  
  93.         System.out.println(ok ? "删除成功" : "删除失败");  
  94.   
  95.         FileStatus files[] = hdfs.listStatus(dst);  
  96.         for (FileStatus file : files) {  
  97.             System.out.println(file.getPath());  
  98.         }  
  99.     }  
  100.   
  101.     // 查看HDFS文件的最后修改时间  
  102.     public void testgetModifyTime() throws Exception {  
  103.   
  104.         Configuration conf = new Configuration();  
  105.   
  106.         FileSystem hdfs = FileSystem.get(conf);  
  107.         Path dst = new Path(hdfsPath);  
  108.   
  109.         FileStatus files[] = hdfs.listStatus(dst);  
  110.         for (FileStatus file : files) {  
  111.             System.out.println(file.getPath() + "\t"  
  112.                     + file.getModificationTime());  
  113.               
  114.             System.out.println(file.getPath() + "\t"  
  115.                     + new Date(file.getModificationTime()));  
  116.               
  117.         }  
  118.     }  
  119.   
  120.     // 查看HDFS文件是否存在  
  121.   
  122.     public void testExists() throws Exception {  
  123.   
  124.         Configuration conf = new Configuration();  
  125.           
  126.         FileSystem hdfs = FileSystem.get(conf);  
  127.         Path dst = new Path(hdfsPath + "file01.txt");  
  128.   
  129.         boolean ok = hdfs.exists(dst);  
  130.         System.out.println(ok ? "文件存在" : "文件不存在");  
  131.     }  
  132.   
  133.     // 查看某个文件在HDFS集群的位置  
  134.     public void testFileBlockLocation() throws Exception {  
  135.   
  136.         Configuration conf = new Configuration();  
  137.   
  138.         FileSystem hdfs = FileSystem.get(conf);  
  139.         Path dst = new Path(hdfsPath + "file01.txt");  
  140.   
  141.         FileStatus fileStatus = hdfs.getFileStatus(dst);  
  142.         BlockLocation[] blockLocations = hdfs.getFileBlockLocations(fileStatus,  
  143.                 0, fileStatus.getLen());  
  144.         for (BlockLocation block : blockLocations) {  
  145.             System.out.println(Arrays.toString(block.getHosts()) + "\t"  
  146.                     + Arrays.toString(block.getNames()));  
  147.         }  
  148.     }  
  149.   
  150.     // 获取HDFS集群上所有节点名称  
  151.     public void testGetHostName() throws Exception {  
  152.   
  153.         Configuration conf = new Configuration();  
  154.           
  155.         DistributedFileSystem hdfs = (DistributedFileSystem) FileSystem  
  156.                 .get(conf);  
  157.         DatanodeInfo[] dataNodeStats = hdfs.getDataNodeStats();  
  158.   
  159.         for (DatanodeInfo dataNode : dataNodeStats) {  
  160.             System.out.println(dataNode.getHostName() + "\t"  
  161.                     + dataNode.getName());  
  162.         }  
  163.     }  
  164. }  
0 0
原创粉丝点击