HDFS的JAVA客户端编写
HDFS的shell操作很简单,直接查看文档就可以,和Linux指令类似,下面简单总结一下HDFS的JAVA客户端编写。(HA集群的话把core-site.xml,hdfs-site.xml,mapred-site.xml,yarn-site.xml都拷贝到工作空间)
如图建立工程,其中客户端放在hdfs包下面:
需要导包,在hadoop的share文件夹下面,里面会找到不同的jar包。我把我的贴出来供大家参考:
剩下的就是编写客户端代码了。在Linux下面编写不会有那么多问题,但是在windows下面会莫名其妙各种错误,下面会提到。
首先看core-site.xml文件:
<?xml version="1.0" encoding="UTF-8"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. See accompanying LICENSE file. --> <!-- Put site-specific property overrides in this file. --> <configuration> <!-- 指定HADOOP所使用的文件系统schema(URI),HDFS的老大(NameNode)的地址 --> <property> <name>fs.defaultFS</name> <value>hdfs://192.168.230.134:9000</value> </property> <!-- 指定hadoop运行时产生文件的存储目录 --> <property> <name>hadoop.tmp.dir</name> <value>/home/hadoop/app/hadoop-2.4.1/data</value> </property> </configuration>
然后看一下hdfs-site.xml文件:
<?xml version="1.0" encoding="UTF-8"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. See accompanying LICENSE file. --> <!-- Put site-specific property overrides in this file. --> <configuration> <property> <name>dfs.replication</name> <value>1</value> </property> </configuration>
下面写一个工具类,HdfsUtils.java:
package cn.darrenchan.hadoop.hdfs; import java.io.FileInputStream; import java.io.FileNotFoundException; import java.io.IOException; import java.net.URI; import org.apache.commons.io.IOUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.LocatedFileStatus; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.RemoteIterator; import org.junit.Before; import org.junit.Test; /** * 如果涉及权限用户不是hadoop然后报错的话,可以在run as->run * configuration->arguments里面添加-DHADOOP_USER_NAME=hadoop * 还有另外一种终极解决方案,就是在init()方法中 FileSystem.get进行user的指定 * * @author chenchi * */ public class HdfsUtils { private FileSystem fs; /** * 加上@Before就会在test方法之前执行 * * @throws Exception */ @Before public void init() throws Exception { // 读取classpath下的xxx-site.xml 配置文件,并解析其内容,封装到conf对象中 Configuration conf = new Configuration(); // 如果把配置文件放到src下面,这行代码可以不用写 // 也可以在代码中对conf中的配置信息进行手动设置,会覆盖掉配置文件中的读取的值 conf.set("fs.defaultFS", "hdfs://192.168.230.134:9000"); // 这一行是在windows下运行才用到的 System.setProperty("hadoop.home.dir", "E:\\大数据教程Hadoop8天\\hadoop-2.4.1"); // 根据配置信息,去获取一个具体文件系统的客户端操作实例对象 fs = FileSystem.get(new URI("hdfs://weekend110:9000"), conf, "hadoop"); } /** * 上传文件(底层写法) * * @throws IOException */ @Test public void upload() throws IOException { Path dst = new Path("hdfs://192.168.230.134:9000/aa/qingshu.txt"); FSDataOutputStream out = fs.create(dst); FileInputStream in = new FileInputStream("E:/qingshu.txt"); IOUtils.copy(in, out); } /** * 上传文件(简单版) * * @throws IOException */ @Test public void upload2() throws IOException { fs.copyFromLocalFile(new Path("e:/qingshu.txt"), new Path( "hdfs://weekend110:9000/aaa/bbb/ccc/qingshu100.txt")); } /** * 下载文件 * * @throws IOException * @throws IllegalArgumentException */ @Test public void download() throws IllegalArgumentException, IOException { // 上面是这种写法会报空指针异常的错,所以采用下面这种写法 // fs.copyToLocalFile(new Path("hdfs://weekend110:9000/aa/qingshu.txt"), // new Path("e:/chen.txt")); fs.copyToLocalFile(false, new Path( "hdfs://weekend110:9000/aa/qingshu.txt"), new Path( "e:/chen.txt"), true); } /** * 查看文件信息 * * @throws IOException * @throws IllegalArgumentException * @throws FileNotFoundException */ @Test public void listFiles() throws FileNotFoundException, IllegalArgumentException, IOException { // listFiles列出的是文件信息,而且提供递归遍历,可以递归遍历出当前目录以及其子目录下的所有内容 RemoteIterator<LocatedFileStatus> files = fs.listFiles(new Path("/"), true); while (files.hasNext()) { LocatedFileStatus file = files.next(); Path filePath = file.getPath(); String fileName = filePath.getName(); System.out.println(fileName); } System.out.println("---------------------------------"); // listStatus 可以列出文件和文件夹的信息,但是不提供自带的递归遍历,只提供当前目录下的内容
FileStatus[] listStatus = fs.listStatus(new Path("/")); for (FileStatus status : listStatus) { String name = status.getPath().getName(); System.out.println(name + (status.isDirectory() ? " 目录" : " 文件")); } } /** * 创建文件夹 * * @throws IOException * @throws IllegalArgumentException */ @Test public void mkdir() throws IllegalArgumentException, IOException { // 这里也可以简写成/aaa/bbb/ccc fs.mkdirs(new Path("hdfs://weekend110:9000/aaa/bbb/ccc")); } /** * 删除文件或者文件夹 * * @throws IOException * @throws IllegalArgumentException */ @Test public void rm() throws IllegalArgumentException, IOException { // 递归删除,路径名称可以简写 fs.delete(new Path("/aa"), true); // fs.rename(path, path1);// 这是移动文件到不同文件夹 } }
再补充一种通过流的方式访问hdfs:
package cn.darrenchan.hadoop.hdfs; import java.io.File; import java.io.FileInputStream; import java.io.FileOutputStream; import java.io.IOException; import java.net.URI; import java.net.URISyntaxException; import org.apache.commons.io.IOUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.junit.Before; import org.junit.Test; public class StreamAccess { private FileSystem fs; @Before public void init() throws Exception{ Configuration conf = new Configuration(); fs = FileSystem.get(new URI("hdfs://weekend110:9000"), conf, "hadoop"); } /** * 以流的方式进行上传文件 * @throws IOException * @throws IllegalArgumentException */ @Test public void testUpload() throws Exception{ //参数2:是否重写 FSDataOutputStream outputStream = fs.create(new Path("/huiru.txt"), true); FileInputStream inputStream = new FileInputStream(new File("/home/hadoop/case00.txt")); IOUtils.copy(inputStream, outputStream); fs.close(); } /** * 以流的方式进行下载文件 * @throws Exception * @throws */ @Test public void testDownload() throws Exception{ FSDataInputStream inputStream = fs.open(new Path("/huiru.txt")); FileOutputStream outputStream = new FileOutputStream(new File("/home/hadoop/huiru.txt")); IOUtils.copy(inputStream, outputStream); fs.close(); } /** * hdfs支持随机定位进行文件读取,而且可以方便地读取指定长度 * 用于上层分布式运算框架并发处理数据 * @throws Exception * @throws IllegalArgumentException */ @Test public void testRandomDownload() throws Exception{ FSDataInputStream inputStream = fs.open(new Path("/huiru.txt")); //可以将流的起始偏移量进行自定义 inputStream.seek(22); FileOutputStream outputStream = new FileOutputStream(new File("/home/hadoop/randomhuiru.txt")); IOUtils.copy(inputStream, outputStream); fs.close(); } /** * 显示hdfs文件的内容 * @throws Exception * @throws IllegalArgumentException */ @Test public void testCat() throws Exception{ FSDataInputStream inputStream = fs.open(new Path("/huiru.txt")); IOUtils.copy(inputStream, System.out); fs.close(); } }
这里总结一下遇到的两个错误:
1.如果涉及权限用户不是hadoop然后报错的话,可以在run as->run configuration->arguments里面添加-DHADOOP_USER_NAME=hadoop。
还有另外一种终极解决方案,就是在init()方法中 FileSystem.get进行user的指定,即
2.在windows下运行,可能报错:Could not locate executable null\bin\winutils.exe in the Hadoop binaries...
很明显应该是HADOOP_HOME的问题。如果HADOOP_HOME为空,必然fullExeName为null\bin\winutils.exe。解决方法很简单,配置环境变量,不想重启电脑可以在程序里加上:
注:E:\\大数据教程Hadoop8天\\hadoop-2.4.1是我本机解压的hadoop的路径。
稍后再执行,你可能还是会出现同样的错误,因为你进入你的hadoop-x.x.x/bin目录下看,你会发现你压根就没有winutils.exe这个东西。
于是我告诉你,你可以去github下载一个,地球人都知道的地址发你一个。
地址:https://github.com/srccodes/hadoop-common-2.2.0-bin
下载好后,把winutils.exe加入你的hadoop-x.x.x/bin下。