啦啦啦

Posted lijing925

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了啦啦啦相关的知识,希望对你有一定的参考价值。

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.*;
import org.apache.hadoop.io.IOUtils;
import org.apache.hadoop.util.Progressable;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;

import java.io.BufferedInputStream;
import java.io.BufferedReader;
import java.io.BufferedWriter;
import java.io.File;
import java.io.FileInputStream;
import java.io.InputStream;
import java.io.InputStreamReader;
import java.io.OutputStreamWriter;
import java.net.URI;

/**
* Hadoop HDFS Java API 操作
*/
public class HDFSApp {

public static final String HDFS_PATH = "hdfs://192.168.223.132:9000";

FileSystem fileSystem = null;
Configuration configuration = null;


/**
* 创建HDFS目录
*/
@Test
public void mkdir() throws Exception {
fileSystem.mkdirs(new Path("/kr"));
}

/**
* 创建文件
*/
@Test
public void create() throws Exception {
FSDataOutputStream output = fileSystem.create(new Path("/kr/hdfstest1.txt"));
output.write("1605-1 123456 HDFS".getBytes());
output.flush();
output.close();
}

/**
* 查看HDFS文件的内容
*/
@Test
public void cat() throws Exception {
FSDataInputStream fin = fileSystem.open(new Path("/hdfsapi/test/hdfstest2.txt"));
BufferedReader in = new BufferedReader(new InputStreamReader(fin, "UTF-8"));
System.out.println(in.readLine());
in.close();
}


@Test
public void copy() throws Exception {
FSDataInputStream fin = fileSystem.open(new Path("/kr/hdfstest1.txt"));
BufferedReader in = new BufferedReader(new InputStreamReader(fin, "UTF-8"));
FSDataOutputStream fout = fileSystem.create(new Path("/hdfsapi/test/hdfstest2.txt"));
BufferedWriter out = new BufferedWriter(new OutputStreamWriter(fout, "UTF-8"));
out.write(in.readLine());
out.flush();
out.close();
}

/**
* 重命名
*/
@Test
public void rename() throws Exception {
Path oldPath = new Path("/hdfsapi/test/a.txt");
Path newPath = new Path("/hdfsapi/test/b.txt");
fileSystem.rename(oldPath, newPath);
}

/**
* 上传文件到HDFS
*
* @throws Exception
*/
@Test
public void copyFromLocalFile() throws Exception {
Path localPath = new Path("E:/data/input.txt");
Path hdfsPath = new Path("/hdfsapi/test");
fileSystem.copyFromLocalFile(localPath, hdfsPath);
}

/**
* 上传文件到HDFS
*/
@Test
public void copyFromLocalFileWithProgress() throws Exception {
InputStream in = new BufferedInputStream(
new FileInputStream(
new File("/Users/rocky/source/spark-1.6.1/spark-1.6.1-bin-2.6.0-cdh5.5.0.tgz")));

FSDataOutputStream output = fileSystem.create(new Path("/hdfsapi/test/spark-1.6.1.tgz"),
new Progressable() {
public void progress() {
System.out.print("."); //带进度提醒信息
}
});


IOUtils.copyBytes(in, output, 4096);
}


/**
* 下载HDFS文件
*/
@Test
public void copyToLocalFile() throws Exception {
Path localPath = new Path("/Users/rocky/tmp/h.txt");
Path hdfsPath = new Path("/hdfsapi/test/hello.txt");
fileSystem.copyToLocalFile(hdfsPath, localPath);
}

/**
* 查看某个目录下的所有文件
*/
@Test
public void listFiles() throws Exception {
FileStatus[] fileStatuses = fileSystem.listStatus(new Path("/"));

for(FileStatus fileStatus : fileStatuses) {
String isDir = fileStatus.isDirectory() ? "文件夹" : "文件";
short replication = fileStatus.getReplication();
long len = fileStatus.getLen();
String path = fileStatus.getPath().toString();

System.out.println(isDir + " " + replication + " " + len + " " + path);
}

}

/**
* 删除
*/
@Test
public void delete() throws Exception{
fileSystem.delete(new Path("/"), true);
}


@Before
public void setUp() throws Exception {

configuration = new Configuration();
fileSystem = FileSystem.get(new URI(HDFS_PATH), configuration, "keke");
System.out.println("HDFSApp.setUp");
}

@After
public void tearDown() throws Exception {
configuration = null;
fileSystem = null;

System.out.println("HDFSApp.tearDown");
}

}


























































































































以上是关于啦啦啦的主要内容,如果未能解决你的问题,请参考以下文章

mongodb 初学 索引

python 2.x中的中文

MarkDown语法

啦啦啦家

iOS ---------Xcode9代码折叠

ZJOI 2018 历史