1 package com.xujunqi.month2a; 2 3 import org.apache.hadoop.conf.Configuration; 4 import org.apache.hadoop.fs.FileSystem; 5 import org.apache.hadoop.fs.Path; 6 7 import java.io.IOException; 8 import java.net.URI; 9 import java.net.URISyntaxException; 10 import java.net.URL; 11 12 // <property> 13 // <name>fs.defaultFS</name> 14 // <value>hdfs://xujunqi001:9000</value> 15 // </property> 16 public class JavaApiHdfs { 17 18 public static void main(String[] args) throws IOException, URISyntaxException, InterruptedException { 19 20 Configuration configuration = new Configuration(); 21 22 configuration.set("fs.defaultFS","hdfs://192.168.107.128:9000"); 23 24 FileSystem fileSystem = FileSystem.get(new URI("hdfs://192.168.107.128:9000"),configuration,"root"); 25 fileSystem.copyFromLocalFile(new Path("C:\\Users\\Dell\\Desktop\\项目\\月考练习\\month2a\\hivefile.txt"),new Path("/yuekao/source")); 26 27 //上传到sourse 28 // fileSystem.copyFromLocalFile(new Path(""),dst:""); 29 30 fileSystem.close(); 31 32 } 33 }
使用HDFS的JavaApi,实现文件的新建上传和下载
猜你喜欢
转载自www.cnblogs.com/xjqi/p/12668327.html
今日推荐
周排行