From 8e7dcc02b22672c5799ec0b758ba5b8d37b8d8dc Mon Sep 17 00:00:00 2001 From: dailidong Date: Fri, 17 Apr 2020 11:42:54 +0800 Subject: [PATCH] Update HadoopUtilsTest.java --- .../common/utils/HadoopUtilsTest.java | 67 ++++++++++++++----- 1 file changed, 51 insertions(+), 16 deletions(-) diff --git a/dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/utils/HadoopUtilsTest.java b/dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/utils/HadoopUtilsTest.java index b7bf2209d6..2b973895b9 100644 --- a/dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/utils/HadoopUtilsTest.java +++ b/dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/utils/HadoopUtilsTest.java @@ -16,16 +16,19 @@ */ package org.apache.dolphinscheduler.common.utils; -import org.apache.dolphinscheduler.common.enums.ResourceType; +import org.apache.dolphinscheduler.common.Constants; +import org.junit.Assert; import org.junit.Ignore; import org.junit.Test; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import java.io.IOException; +import java.util.Arrays; import java.util.List; @Ignore +//todo there is no hadoop environment public class HadoopUtilsTest { private static final Logger logger = LoggerFactory.getLogger(HadoopUtilsTest.class); @@ -40,6 +43,47 @@ public class HadoopUtilsTest { logger.info(HadoopUtils.getInstance().getApplicationUrl("application_1548381297012_0030")); } + @Test + public void getConfiguration(){ + logger.info(HadoopUtils.getInstance().getConfiguration().get(Constants.HDFS_ROOT_USER)); + } + + @Test + public void mkdir() throws IOException { + boolean result = HadoopUtils.getInstance().mkdir("/dolphinscheduler/hdfs"); + Assert.assertEquals(true, result); + } + + @Test + public void delete() throws IOException { + boolean result = HadoopUtils.getInstance().delete("/dolphinscheduler/hdfs",true); + Assert.assertEquals(true, result); + } + + @Test + public void exists() throws IOException { + boolean result = HadoopUtils.getInstance().exists("/dolphinscheduler/hdfs"); + Assert.assertEquals(true, result); + } + + @Test + public void getHdfsDataBasePath() throws IOException { + String result = HadoopUtils.getInstance().getHdfsDataBasePath(); + Assert.assertEquals("/dolphinscheduler", result); + } + + @Test + public void getHdfsResDir() throws IOException { + String result = HadoopUtils.getInstance().getHdfsResDir("11000"); + Assert.assertEquals("/dolphinscheduler/resources/11000", result); + } + + @Test + public void isYarnEnabled() throws IOException { + boolean result = HadoopUtils.getInstance().isYarnEnabled(); + Assert.assertEquals(false, result); + } + @Test public void test() throws IOException { HadoopUtils.getInstance().copyLocalToHdfs("/root/teamviewer_13.1.8286.x86_64.rpm", "/journey", true, true); @@ -51,7 +95,7 @@ public class HadoopUtilsTest { byte[] bytes = HadoopUtils.getInstance().catFile("/dolphinscheduler/hdfs/resources/35435.sh"); logger.info(new String(bytes)); } catch (Exception e) { - + logger.error(e.getMessage(),e); } } @Test @@ -82,23 +126,14 @@ public class HadoopUtilsTest { } @Test - public void catFileTest()throws Exception{ + public void catFileWithLimitTest()throws Exception{ List stringList = HadoopUtils.getInstance().catFile("/dolphinscheduler/hdfs/resources/WCSparkPython.py", 0, 1000); logger.info(String.join(",",stringList)); } @Test - public void getHdfsFileNameTest(){ - logger.info(HadoopUtils.getHdfsFileName(ResourceType.FILE,"test","/test")); - } - - @Test - public void getHdfsResourceFileNameTest(){ - logger.info(HadoopUtils.getHdfsResourceFileName("test","/test")); - } - - @Test - public void getHdfsUdfFileNameTest(){ - logger.info(HadoopUtils.getHdfsUdfFileName("test","/test.jar")); + public void catFileTest()throws Exception{ + byte[] content = HadoopUtils.getInstance().catFile("/dolphinscheduler/hdfs/resources/WCSparkPython.py"); + logger.info(Arrays.toString(content)); } -} \ No newline at end of file +}