Browse Source

[feat] Add type local turn on resource center by default (#13303)

3.2.0-release
Jay Chung 1 year ago committed by GitHub
parent
commit
78e5569afd
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
  1. 17
      docs/docs/en/guide/resource/configuration.md
  2. 19
      docs/docs/zh/guide/resource/configuration.md
  3. 2
      dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/enums/ResUploadType.java
  4. 6
      dolphinscheduler-common/src/main/resources/common.properties
  5. 2
      dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/utils/PropertyUtilsTest.java
  6. 7
      dolphinscheduler-storage-plugin/dolphinscheduler-storage-api/src/main/java/org/apache/dolphinscheduler/plugin/storage/api/StorageType.java
  7. 45
      dolphinscheduler-storage-plugin/dolphinscheduler-storage-hdfs/src/main/java/org/apache/dolphinscheduler/plugin/storage/hdfs/HdfsStorageOperator.java
  8. 82
      dolphinscheduler-storage-plugin/dolphinscheduler-storage-hdfs/src/main/java/org/apache/dolphinscheduler/plugin/storage/hdfs/HdfsStorageProperties.java
  9. 42
      dolphinscheduler-storage-plugin/dolphinscheduler-storage-hdfs/src/main/java/org/apache/dolphinscheduler/plugin/storage/hdfs/LocalStorageOperatorFactory.java
  10. 15
      dolphinscheduler-storage-plugin/dolphinscheduler-storage-hdfs/src/test/java/org/apache/dolphinscheduler/plugin/storage/hdfs/HdfsStorageOperatorTest.java

17
docs/docs/en/guide/resource/configuration.md

@ -9,13 +9,22 @@
### Configure `common.properties` ### Configure `common.properties`
If you deploy DolphinScheduler in `Cluster` or `Pseudo-Cluster` mode, you need to configure `api-server/conf/common.properties` and `worker-server/conf/common.properties`. DolphinScheduler Resource Center uses local file system by default, and does not require any additional configuration.
If you deploy DolphinScheduler in `Standalone` mode, you only need to configure `standalone-server/conf/common.properties` as follows: But please make sure to change the following configuration at the same time when you need to modify the default value.
- If you deploy DolphinScheduler in `Cluster` or `Pseudo-Cluster` mode, you need to configure `api-server/conf/common.properties` and `worker-server/conf/common.properties`.
- If you deploy DolphinScheduler in `Standalone` mode, you only need to configure `standalone-server/conf/common.properties` as follows:
The configuration you may need to change:
- Change `resource.storage.upload.base.path` to your local directory path. Please make sure the `tenant resource.hdfs.root.user` has read and write permissions for `resource.storage.upload.base.path`, e,g. `/tmp/dolphinscheduler`. `DolphinScheduler` will create the directory you configure if it does not exist. - Change `resource.storage.upload.base.path` to your local directory path. Please make sure the `tenant resource.hdfs.root.user` has read and write permissions for `resource.storage.upload.base.path`, e,g. `/tmp/dolphinscheduler`. `DolphinScheduler` will create the directory you configure if it does not exist.
- Modify `resource.storage.type=HDFS` and `resource.hdfs.fs.defaultFS=file:///`.
> NOTE: Please modify the value of `resource.storage.upload.base.path` if you do not want to use the default value as the base path. > NOTE:
> 1. LOCAL mode does not support reading and writing in distributed mode, which mean you can only use your resource in one machine, unless use shared file mount point
> 2. Please modify the value of `resource.storage.upload.base.path` if you do not want to use the default value as the base path.
> 3. The local config is `resource.storage.type=LOCAL` it has actually configured two setting, `resource.storage.type=HDFS`
> and `resource.hdfs.fs.defaultFS=file:///`, The configuration of `resource.storage.type=LOCAL` is for user-friendly, and enables
> the local resource center to be enabled by default
## Use HDFS or Remote Object Storage ## Use HDFS or Remote Object Storage

19
docs/docs/zh/guide/resource/configuration.md

@ -9,13 +9,20 @@
### 配置 `common.properties` 文件 ### 配置 `common.properties` 文件
如果您以 `集群` 模式或者 `伪集群` 模式部署DolphinScheduler,您需要对以下路径的文件进行配置:`api-server/conf/common.properties` 和 `worker-server/conf/common.properties` Dolphinscheduler 资源中心使用本地系统默认是开启的,不需要用户做任何额外的配置,但是当用户需要对默认配置做修改时,请确保同时完成下面的修改。
若您以 `单机` 模式部署DolphinScheduler,您只需要配置 `standalone-server/conf/common.properties`,具体配置如下:
- 如果您以 `集群` 模式或者 `伪集群` 模式部署DolphinScheduler,您需要对以下路径的文件进行配置:`api-server/conf/common.properties` 和 `worker-server/conf/common.properties`
- 若您以 `单机` 模式部署DolphinScheduler,您只需要配置 `standalone-server/conf/common.properties`,具体配置如下:
您可能需要涉及如下的修改:
- 将 `resource.storage.upload.base.path` 改为本地存储路径,请确保部署 DolphinScheduler 的用户拥有读写权限,例如:`resource.storage.upload.base.path=/tmp/dolphinscheduler`。当路径不存在时会自动创建文件夹 - 将 `resource.storage.upload.base.path` 改为本地存储路径,请确保部署 DolphinScheduler 的用户拥有读写权限,例如:`resource.storage.upload.base.path=/tmp/dolphinscheduler`。当路径不存在时会自动创建文件夹
- 修改 `resource.storage.type=HDFS``resource.hdfs.fs.defaultFS=file:///`
> **注意**:如果您不想用默认值作为资源中心的基础路径,请修改`resource.storage.upload.base.path`的值。 > **注意**
> 1. LOCAL模式不支持分布式模式读写,意味着上传的资源只能在一台机器上使用,除非使用共享文件挂载点
> 2. 如果您不想用默认值作为资源中心的基础路径,请修改`resource.storage.upload.base.path`的值。
> 3. 当配置 `resource.storage.type=LOCAL`,其实您配置了两个配置项,分别是 `resource.storage.type=HDFS``resource.hdfs.fs.defaultFS=file:///` ,我们单独配置 `resource.storage.type=LOCAL` 这个值是为了
> 方便用户,并且能使得本地资源中心默认开启
## 对接分布式或远端对象存储 ## 对接分布式或远端对象存储
@ -46,8 +53,8 @@
# user data local directory path, please make sure the directory exists and have read write permissions # user data local directory path, please make sure the directory exists and have read write permissions
data.basedir.path=/tmp/dolphinscheduler data.basedir.path=/tmp/dolphinscheduler
# resource storage type: HDFS, S3, OSS, NONE # resource storage type: LOCAL, HDFS, S3, OSS
resource.storage.type=HDFS resource.storage.type=LOCAL
# resource store on HDFS/S3/OSS path, resource file will store to this hadoop hdfs path, self configuration, # resource store on HDFS/S3/OSS path, resource file will store to this hadoop hdfs path, self configuration,
# please make sure the directory exists on hdfs and have read write permissions. "/dolphinscheduler" is recommended # please make sure the directory exists on hdfs and have read write permissions. "/dolphinscheduler" is recommended

2
dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/enums/ResUploadType.java

@ -21,5 +21,5 @@ package org.apache.dolphinscheduler.common.enums;
* data base types * data base types
*/ */
public enum ResUploadType { public enum ResUploadType {
HDFS, S3, OSS, NONE LOCAL, HDFS, S3, OSS, NONE
} }

6
dolphinscheduler-common/src/main/resources/common.properties

@ -21,8 +21,10 @@ data.basedir.path=/tmp/dolphinscheduler
# resource view suffixs # resource view suffixs
#resource.view.suffixs=txt,log,sh,bat,conf,cfg,py,java,sql,xml,hql,properties,json,yml,yaml,ini,js #resource.view.suffixs=txt,log,sh,bat,conf,cfg,py,java,sql,xml,hql,properties,json,yml,yaml,ini,js
# resource storage type: HDFS, S3, OSS, NONE # resource storage type: LOCAL, HDFS, S3, OSS, NONE. LOCAL type is a specific type of HDFS with "resource.hdfs.fs.defaultFS = file:///" configuration
resource.storage.type=NONE # please notice that LOCAL mode does not support reading and writing in distributed mode, which mean you can only use your resource in one machine, unless
# use shared file mount point
resource.storage.type=LOCAL
# resource store on HDFS/S3 path, resource file will store to this base path, self configuration, please make sure the directory exists on hdfs and have read write permissions. "/dolphinscheduler" is recommended # resource store on HDFS/S3 path, resource file will store to this base path, self configuration, please make sure the directory exists on hdfs and have read write permissions. "/dolphinscheduler" is recommended
resource.storage.upload.base.path=/dolphinscheduler resource.storage.upload.base.path=/dolphinscheduler

2
dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/utils/PropertyUtilsTest.java

@ -31,6 +31,6 @@ public class PropertyUtilsTest {
@Test @Test
public void getResUploadStartupState() { public void getResUploadStartupState() {
Assertions.assertFalse(PropertyUtils.getResUploadStartupState()); Assertions.assertTrue(PropertyUtils.getResUploadStartupState());
} }
} }

7
dolphinscheduler-storage-plugin/dolphinscheduler-storage-api/src/main/java/org/apache/dolphinscheduler/plugin/storage/api/StorageType.java

@ -21,9 +21,10 @@ import java.util.Optional;
public enum StorageType { public enum StorageType {
HDFS(0, "HDFS"), LOCAL(0, "LOCAL"),
OSS(1, "OSS"), HDFS(1, "HDFS"),
S3(2, "S3"), OSS(2, "OSS"),
S3(3, "S3"),
; ;
private final int code; private final int code;

45
dolphinscheduler-storage-plugin/dolphinscheduler-storage-hdfs/src/main/java/org/apache/dolphinscheduler/plugin/storage/hdfs/HdfsStorageOperator.java

@ -77,22 +77,17 @@ import com.google.common.cache.LoadingCache;
public class HdfsStorageOperator implements Closeable, StorageOperate { public class HdfsStorageOperator implements Closeable, StorageOperate {
private static final Logger logger = LoggerFactory.getLogger(HdfsStorageOperator.class); private static final Logger logger = LoggerFactory.getLogger(HdfsStorageOperator.class);
private String hdfsUser; private static HdfsStorageProperties hdfsProperties;
public static final String RM_HA_IDS = PropertyUtils.getString(Constants.YARN_RESOURCEMANAGER_HA_RM_IDS);
public static final String APP_ADDRESS = PropertyUtils.getString(Constants.YARN_APPLICATION_STATUS_ADDRESS);
public static final String JOB_HISTORY_ADDRESS = PropertyUtils.getString(Constants.YARN_JOB_HISTORY_STATUS_ADDRESS);
public static final int HADOOP_RESOURCE_MANAGER_HTTP_ADDRESS_PORT_VALUE =
PropertyUtils.getInt(Constants.HADOOP_RESOURCE_MANAGER_HTTPADDRESS_PORT, 8088);
private static final String HADOOP_UTILS_KEY = "HADOOP_UTILS_KEY"; private static final String HADOOP_UTILS_KEY = "HADOOP_UTILS_KEY";
private static final LoadingCache<String, HdfsStorageOperator> cache = CacheBuilder private static final LoadingCache<String, HdfsStorageOperator> cache = CacheBuilder
.newBuilder() .newBuilder()
.expireAfterWrite(PropertyUtils.getInt(Constants.KERBEROS_EXPIRE_TIME, 2), TimeUnit.HOURS) .expireAfterWrite(HdfsStorageProperties.getKerberosExpireTime(), TimeUnit.HOURS)
.build(new CacheLoader<String, HdfsStorageOperator>() { .build(new CacheLoader<String, HdfsStorageOperator>() {
@Override @Override
public HdfsStorageOperator load(String key) throws Exception { public HdfsStorageOperator load(String key) throws Exception {
return new HdfsStorageOperator(); return new HdfsStorageOperator(hdfsProperties);
} }
}); });
@ -101,8 +96,13 @@ public class HdfsStorageOperator implements Closeable, StorageOperate {
private Configuration configuration; private Configuration configuration;
private FileSystem fs; private FileSystem fs;
private HdfsStorageOperator() { public HdfsStorageOperator() {
hdfsUser = PropertyUtils.getString(Constants.HDFS_ROOT_USER); this(new HdfsStorageProperties());
}
public HdfsStorageOperator(HdfsStorageProperties hdfsStorageProperties) {
// Overwrite config from passing hdfsStorageProperties
hdfsProperties = hdfsStorageProperties;
init(); init();
initHdfsPath(); initHdfsPath();
} }
@ -133,16 +133,12 @@ public class HdfsStorageOperator implements Closeable, StorageOperate {
try { try {
configuration = new HdfsConfiguration(); configuration = new HdfsConfiguration();
String hdfsUser = hdfsProperties.getUser();
if (CommonUtils.loadKerberosConf(configuration)) { if (CommonUtils.loadKerberosConf(configuration)) {
hdfsUser = ""; hdfsUser = "";
} }
String defaultFS = configuration.get(Constants.FS_DEFAULT_FS); String defaultFS = getDefaultFS();
if (StringUtils.isBlank(defaultFS)) {
defaultFS = PropertyUtils.getString(Constants.FS_DEFAULT_FS);
}
// first get key from core-site.xml hdfs-site.xml ,if null ,then try to get from properties file // first get key from core-site.xml hdfs-site.xml ,if null ,then try to get from properties file
// the default is the local file system // the default is the local file system
if (StringUtils.isNotBlank(defaultFS)) { if (StringUtils.isNotBlank(defaultFS)) {
@ -189,7 +185,7 @@ public class HdfsStorageOperator implements Closeable, StorageOperate {
public String getDefaultFS() { public String getDefaultFS() {
String defaultFS = getConfiguration().get(Constants.FS_DEFAULT_FS); String defaultFS = getConfiguration().get(Constants.FS_DEFAULT_FS);
if (StringUtils.isBlank(defaultFS)) { if (StringUtils.isBlank(defaultFS)) {
defaultFS = PropertyUtils.getString(Constants.FS_DEFAULT_FS); defaultFS = hdfsProperties.getDefaultFS();
} }
return defaultFS; return defaultFS;
} }
@ -207,18 +203,20 @@ public class HdfsStorageOperator implements Closeable, StorageOperate {
public String getApplicationUrl(String applicationId) throws BaseException { public String getApplicationUrl(String applicationId) throws BaseException {
yarnEnabled = true; yarnEnabled = true;
String appUrl = StringUtils.isEmpty(RM_HA_IDS) ? APP_ADDRESS : getAppAddress(APP_ADDRESS, RM_HA_IDS); String appUrl = StringUtils.isEmpty(hdfsProperties.getYarnResourceRmIds())
? hdfsProperties.getYarnAppStatusAddress()
: getAppAddress(hdfsProperties.getYarnAppStatusAddress(), hdfsProperties.getYarnResourceRmIds());
if (StringUtils.isBlank(appUrl)) { if (StringUtils.isBlank(appUrl)) {
throw new BaseException("yarn application url generation failed"); throw new BaseException("yarn application url generation failed");
} }
logger.debug("yarn application url:{}, applicationId:{}", appUrl, applicationId); logger.debug("yarn application url:{}, applicationId:{}", appUrl, applicationId);
return String.format(appUrl, HADOOP_RESOURCE_MANAGER_HTTP_ADDRESS_PORT_VALUE, applicationId); return String.format(appUrl, hdfsProperties.getHadoopResourceManagerHttpAddressPort(), applicationId);
} }
public String getJobHistoryUrl(String applicationId) { public String getJobHistoryUrl(String applicationId) {
// eg:application_1587475402360_712719 -> job_1587475402360_712719 // eg:application_1587475402360_712719 -> job_1587475402360_712719
String jobId = applicationId.replace("application", "job"); String jobId = applicationId.replace("application", "job");
return String.format(JOB_HISTORY_ADDRESS, jobId); return String.format(hdfsProperties.getYarnJobHistoryStatusAddress(), jobId);
} }
/** /**
@ -601,7 +599,7 @@ public class HdfsStorageOperator implements Closeable, StorageOperate {
* @return data hdfs path * @return data hdfs path
*/ */
public static String getHdfsDataBasePath() { public static String getHdfsDataBasePath() {
String defaultFS = PropertyUtils.getString(Constants.FS_DEFAULT_FS); String defaultFS = hdfsProperties.getDefaultFS();
defaultFS = defaultFS.endsWith("/") ? StringUtils.chop(defaultFS) : defaultFS; defaultFS = defaultFS.endsWith("/") ? StringUtils.chop(defaultFS) : defaultFS;
if (FOLDER_SEPARATOR.equals(RESOURCE_UPLOAD_PATH)) { if (FOLDER_SEPARATOR.equals(RESOURCE_UPLOAD_PATH)) {
return defaultFS + ""; return defaultFS + "";
@ -764,7 +762,8 @@ public class HdfsStorageOperator implements Closeable, StorageOperate {
String[] rmIdArr = rmIds.split(Constants.COMMA); String[] rmIdArr = rmIds.split(Constants.COMMA);
String yarnUrl = protocol + "%s:" + HADOOP_RESOURCE_MANAGER_HTTP_ADDRESS_PORT_VALUE + "/ws/v1/cluster/info"; String yarnUrl =
protocol + "%s:" + hdfsProperties.getHadoopResourceManagerHttpAddressPort() + "/ws/v1/cluster/info";
try { try {
@ -791,7 +790,7 @@ public class HdfsStorageOperator implements Closeable, StorageOperate {
public static String getRMState(String url) { public static String getRMState(String url) {
String retStr = Boolean.TRUE String retStr = Boolean.TRUE
.equals(PropertyUtils.getBoolean(Constants.HADOOP_SECURITY_AUTHENTICATION_STARTUP_STATE, false)) .equals(hdfsProperties.isHadoopSecurityAuthStartupState())
? KerberosHttpClient.get(url) ? KerberosHttpClient.get(url)
: HttpUtils.get(url); : HttpUtils.get(url);

82
dolphinscheduler-storage-plugin/dolphinscheduler-storage-hdfs/src/main/java/org/apache/dolphinscheduler/plugin/storage/hdfs/HdfsStorageProperties.java

@ -0,0 +1,82 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.dolphinscheduler.plugin.storage.hdfs;
import static org.apache.dolphinscheduler.common.constants.Constants.FS_DEFAULT_FS;
import static org.apache.dolphinscheduler.common.constants.Constants.HADOOP_RESOURCE_MANAGER_HTTPADDRESS_PORT;
import static org.apache.dolphinscheduler.common.constants.Constants.HADOOP_SECURITY_AUTHENTICATION_STARTUP_STATE;
import static org.apache.dolphinscheduler.common.constants.Constants.HDFS_ROOT_USER;
import static org.apache.dolphinscheduler.common.constants.Constants.KERBEROS_EXPIRE_TIME;
import static org.apache.dolphinscheduler.common.constants.Constants.YARN_APPLICATION_STATUS_ADDRESS;
import static org.apache.dolphinscheduler.common.constants.Constants.YARN_JOB_HISTORY_STATUS_ADDRESS;
import static org.apache.dolphinscheduler.common.constants.Constants.YARN_RESOURCEMANAGER_HA_RM_IDS;
import org.apache.dolphinscheduler.common.utils.PropertyUtils;
import lombok.Data;
import org.springframework.context.annotation.Configuration;
@Data
@Configuration
public class HdfsStorageProperties {
/**
* HDFS storage user
*/
private String user = PropertyUtils.getString(HDFS_ROOT_USER);
/**
* HDFS default fs
*/
private String defaultFS = PropertyUtils.getString(FS_DEFAULT_FS);
/**
* YARN resource manager HA RM ids
*/
private String yarnResourceRmIds = PropertyUtils.getString(YARN_RESOURCEMANAGER_HA_RM_IDS);
/**
* YARN application status address
*/
private String yarnAppStatusAddress = PropertyUtils.getString(YARN_APPLICATION_STATUS_ADDRESS);
/**
* YARN job history status address
*/
private String yarnJobHistoryStatusAddress = PropertyUtils.getString(YARN_JOB_HISTORY_STATUS_ADDRESS);
/**
* Hadoop resouece manager http address port
*/
private String hadoopResourceManagerHttpAddressPort =
PropertyUtils.getString(HADOOP_RESOURCE_MANAGER_HTTPADDRESS_PORT);
/**
* Hadoop security authentication startup state
*/
private boolean hadoopSecurityAuthStartupState =
PropertyUtils.getBoolean(HADOOP_SECURITY_AUTHENTICATION_STARTUP_STATE, false);
/**
* Kerberos expire time
*/
public static int getKerberosExpireTime() {
return PropertyUtils.getInt(KERBEROS_EXPIRE_TIME, 2);
}
}

42
dolphinscheduler-storage-plugin/dolphinscheduler-storage-hdfs/src/main/java/org/apache/dolphinscheduler/plugin/storage/hdfs/LocalStorageOperatorFactory.java

@ -0,0 +1,42 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.dolphinscheduler.plugin.storage.hdfs;
import org.apache.dolphinscheduler.plugin.storage.api.StorageOperate;
import org.apache.dolphinscheduler.plugin.storage.api.StorageOperateFactory;
import org.apache.dolphinscheduler.plugin.storage.api.StorageType;
import com.google.auto.service.AutoService;
@AutoService(StorageOperateFactory.class)
public class LocalStorageOperatorFactory implements StorageOperateFactory {
private static final String LOCAL_DEFAULT_FS = "file:///";
@Override
public StorageOperate createStorageOperate() {
HdfsStorageProperties hdfsStorageProperties = new HdfsStorageProperties();
hdfsStorageProperties.setDefaultFS(LOCAL_DEFAULT_FS);
return new HdfsStorageOperator(hdfsStorageProperties);
}
@Override
public StorageType getStorageOperate() {
return StorageType.LOCAL;
}
}

15
dolphinscheduler-storage-plugin/dolphinscheduler-storage-hdfs/src/test/java/org/apache/dolphinscheduler/plugin/storage/hdfs/HdfsStorageOperatorTest.java

@ -39,34 +39,39 @@ public class HdfsStorageOperatorTest {
@Test @Test
public void getHdfsTenantDir() { public void getHdfsTenantDir() {
logger.info(HdfsStorageOperator.getHdfsTenantDir("1234")); HdfsStorageOperator hdfsStorageOperator = new HdfsStorageOperator();
logger.info(hdfsStorageOperator.getHdfsTenantDir("1234"));
Assertions.assertTrue(true); Assertions.assertTrue(true);
} }
@Test @Test
public void getHdfsUdfFileName() { public void getHdfsUdfFileName() {
logger.info(HdfsStorageOperator.getHdfsUdfFileName("admin", "file_name")); HdfsStorageOperator hdfsStorageOperator = new HdfsStorageOperator();
logger.info(hdfsStorageOperator.getHdfsUdfFileName("admin", "file_name"));
Assertions.assertTrue(true); Assertions.assertTrue(true);
} }
@Test @Test
public void getHdfsResourceFileName() { public void getHdfsResourceFileName() {
logger.info(HdfsStorageOperator.getHdfsResourceFileName("admin", "file_name")); HdfsStorageOperator hdfsStorageOperator = new HdfsStorageOperator();
logger.info(hdfsStorageOperator.getHdfsResourceFileName("admin", "file_name"));
Assertions.assertTrue(true); Assertions.assertTrue(true);
} }
@Test @Test
public void getHdfsFileName() { public void getHdfsFileName() {
logger.info(HdfsStorageOperator.getHdfsFileName(ResourceType.FILE, "admin", "file_name")); HdfsStorageOperator hdfsStorageOperator = new HdfsStorageOperator();
logger.info(hdfsStorageOperator.getHdfsFileName(ResourceType.FILE, "admin", "file_name"));
Assertions.assertTrue(true); Assertions.assertTrue(true);
} }
@Test @Test
public void getAppAddress() { public void getAppAddress() {
HdfsStorageOperator hdfsStorageOperator = new HdfsStorageOperator();
try (MockedStatic<HttpUtils> mockedHttpUtils = Mockito.mockStatic(HttpUtils.class)) { try (MockedStatic<HttpUtils> mockedHttpUtils = Mockito.mockStatic(HttpUtils.class)) {
mockedHttpUtils.when(() -> HttpUtils.get("http://ds1:8088/ws/v1/cluster/info")) mockedHttpUtils.when(() -> HttpUtils.get("http://ds1:8088/ws/v1/cluster/info"))
.thenReturn("{\"clusterInfo\":{\"state\":\"STARTED\",\"haState\":\"ACTIVE\"}}"); .thenReturn("{\"clusterInfo\":{\"state\":\"STARTED\",\"haState\":\"ACTIVE\"}}");
logger.info(HdfsStorageOperator.getAppAddress("http://ds1:8088/ws/v1/cluster/apps/%s", "ds1,ds2")); logger.info(hdfsStorageOperator.getAppAddress("http://ds1:8088/ws/v1/cluster/apps/%s", "ds1,ds2"));
Assertions.assertTrue(true); Assertions.assertTrue(true);
} }
} }

Loading…
Cancel
Save