diff --git a/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/K8SNamespaceServiceImpl.java b/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/K8SNamespaceServiceImpl.java index 760fbf8f5f..9ead1cd306 100644 --- a/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/K8SNamespaceServiceImpl.java +++ b/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/K8SNamespaceServiceImpl.java @@ -18,6 +18,7 @@ package org.apache.dolphinscheduler.api.service.impl; import org.apache.dolphinscheduler.api.enums.Status; +import org.apache.dolphinscheduler.api.k8s.K8sClientService; import org.apache.dolphinscheduler.api.service.K8sNamespaceService; import org.apache.dolphinscheduler.api.utils.PageInfo; import org.apache.dolphinscheduler.api.utils.Result; @@ -29,7 +30,6 @@ import org.apache.dolphinscheduler.dao.entity.User; import org.apache.dolphinscheduler.dao.mapper.ClusterMapper; import org.apache.dolphinscheduler.dao.mapper.K8sNamespaceMapper; import org.apache.dolphinscheduler.remote.exceptions.RemotingException; -import org.apache.dolphinscheduler.api.k8s.K8sClientService; import org.apache.commons.lang3.StringUtils; @@ -58,14 +58,14 @@ public class K8SNamespaceServiceImpl extends BaseServiceImpl implements K8sNames private static final Logger logger = LoggerFactory.getLogger(K8SNamespaceServiceImpl.class); private static String resourceYaml = "apiVersion: v1\n" - + "kind: ResourceQuota\n" - + "metadata:\n" - + " name: ${name}\n" - + " namespace: ${namespace}\n" - + "spec:\n" - + " hard:\n" - + " ${limitCpu}\n" - + " ${limitMemory}\n"; + + "kind: ResourceQuota\n" + + "metadata:\n" + + " name: ${name}\n" + + " namespace: ${namespace}\n" + + "spec:\n" + + " hard:\n" + + " ${limitCpu}\n" + + " ${limitMemory}\n"; @Autowired private K8sNamespaceMapper k8sNamespaceMapper; @@ -118,7 +118,8 @@ public class K8SNamespaceServiceImpl extends BaseServiceImpl implements K8sNames * @return */ @Override - public Map createK8sNamespace(User loginUser, String namespace, Long clusterCode, Double limitsCpu, Integer limitsMemory) { + public Map createK8sNamespace(User loginUser, String namespace, Long clusterCode, Double limitsCpu, + Integer limitsMemory) { Map result = new HashMap<>(); if (isNotAdmin(loginUser, result)) { return result; @@ -209,7 +210,8 @@ public class K8SNamespaceServiceImpl extends BaseServiceImpl implements K8sNames * @return */ @Override - public Map updateK8sNamespace(User loginUser, int id, String userName, Double limitsCpu, Integer limitsMemory) { + public Map updateK8sNamespace(User loginUser, int id, String userName, Double limitsCpu, + Integer limitsMemory) { Map result = new HashMap<>(); if (isNotAdmin(loginUser, result)) { return result; @@ -331,7 +333,7 @@ public class K8SNamespaceServiceImpl extends BaseServiceImpl implements K8sNames * @return yaml file */ private String genDefaultResourceYaml(K8sNamespace k8sNamespace) { - //resource use same name with namespace + // resource use same name with namespace String name = k8sNamespace.getNamespace(); String namespace = k8sNamespace.getNamespace(); String cpuStr = null; @@ -345,7 +347,7 @@ public class K8SNamespaceServiceImpl extends BaseServiceImpl implements K8sNames } String result = resourceYaml.replace("${name}", name) - .replace("${namespace}", namespace); + .replace("${namespace}", namespace); if (cpuStr == null) { result = result.replace("${limitCpu}", ""); } else { @@ -360,7 +362,6 @@ public class K8SNamespaceServiceImpl extends BaseServiceImpl implements K8sNames return result; } - /** * query unauthorized namespace * @@ -432,7 +433,8 @@ public class K8SNamespaceServiceImpl extends BaseServiceImpl implements K8sNames * @param authedNamespaceList authed namespace list * @return namespace list that authorization */ - private List getUnauthorizedNamespaces(Set namespaceSet, List authedNamespaceList) { + private List getUnauthorizedNamespaces(Set namespaceSet, + List authedNamespaceList) { List resultList = new ArrayList<>(); for (K8sNamespace k8sNamespace : namespaceSet) { boolean existAuth = false; @@ -450,4 +452,4 @@ public class K8SNamespaceServiceImpl extends BaseServiceImpl implements K8sNames } return resultList; } -} \ No newline at end of file +} diff --git a/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/ProcessDefinitionServiceImpl.java b/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/ProcessDefinitionServiceImpl.java index e60b32b110..1aad9237ca 100644 --- a/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/ProcessDefinitionServiceImpl.java +++ b/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/ProcessDefinitionServiceImpl.java @@ -17,14 +17,28 @@ package org.apache.dolphinscheduler.api.service.impl; -import com.baomidou.mybatisplus.core.metadata.IPage; -import com.baomidou.mybatisplus.extension.plugins.pagination.Page; -import com.fasterxml.jackson.databind.JsonNode; -import com.fasterxml.jackson.databind.node.ArrayNode; -import com.fasterxml.jackson.databind.node.ObjectNode; -import com.google.common.collect.Lists; -import org.apache.commons.collections4.CollectionUtils; -import org.apache.commons.lang3.StringUtils; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.TASK_DEFINITION_MOVE; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.VERSION_DELETE; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.VERSION_LIST; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_BATCH_COPY; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_CREATE; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_DEFINITION; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_DEFINITION_DELETE; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_DEFINITION_EXPORT; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_EXPORT; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_IMPORT; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_ONLINE_OFFLINE; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_SWITCH_TO_THIS_VERSION; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_TREE_VIEW; +import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_UPDATE; +import static org.apache.dolphinscheduler.common.Constants.CMD_PARAM_SUB_PROCESS_DEFINE_CODE; +import static org.apache.dolphinscheduler.common.Constants.COPY_SUFFIX; +import static org.apache.dolphinscheduler.common.Constants.DEFAULT_WORKER_GROUP; +import static org.apache.dolphinscheduler.common.Constants.EMPTY_STRING; +import static org.apache.dolphinscheduler.common.Constants.IMPORT_SUFFIX; +import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.COMPLEX_TASK_TYPES; +import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.TASK_TYPE_SQL; + import org.apache.dolphinscheduler.api.dto.DagDataSchedule; import org.apache.dolphinscheduler.api.dto.ScheduleParam; import org.apache.dolphinscheduler.api.dto.treeview.Instance; @@ -92,16 +106,10 @@ import org.apache.dolphinscheduler.plugin.task.api.parameters.ParametersNode; import org.apache.dolphinscheduler.plugin.task.api.parameters.SqlParameters; import org.apache.dolphinscheduler.service.process.ProcessService; import org.apache.dolphinscheduler.service.task.TaskPluginManager; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; -import org.springframework.beans.factory.annotation.Autowired; -import org.springframework.http.MediaType; -import org.springframework.stereotype.Service; -import org.springframework.transaction.annotation.Transactional; -import org.springframework.web.multipart.MultipartFile; -import javax.servlet.ServletOutputStream; -import javax.servlet.http.HttpServletResponse; +import org.apache.commons.collections4.CollectionUtils; +import org.apache.commons.lang3.StringUtils; + import java.io.BufferedOutputStream; import java.io.BufferedReader; import java.io.IOException; @@ -124,27 +132,6 @@ import java.util.stream.Collectors; import java.util.zip.ZipEntry; import java.util.zip.ZipInputStream; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.TASK_DEFINITION_MOVE; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.VERSION_DELETE; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.VERSION_LIST; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_BATCH_COPY; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_CREATE; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_DEFINITION; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_DEFINITION_DELETE; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_DEFINITION_EXPORT; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_EXPORT; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_IMPORT; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_ONLINE_OFFLINE; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_SWITCH_TO_THIS_VERSION; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_TREE_VIEW; -import static org.apache.dolphinscheduler.api.constants.ApiFuncIdentificationConstant.WORKFLOW_UPDATE; -import static org.apache.dolphinscheduler.common.Constants.CMD_PARAM_SUB_PROCESS_DEFINE_CODE; -import static org.apache.dolphinscheduler.common.Constants.COPY_SUFFIX; -import static org.apache.dolphinscheduler.common.Constants.DEFAULT_WORKER_GROUP; -import static org.apache.dolphinscheduler.common.Constants.EMPTY_STRING; -import static org.apache.dolphinscheduler.common.Constants.IMPORT_SUFFIX; -import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.COMPLEX_TASK_TYPES; -import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.TASK_TYPE_SQL; import javax.servlet.ServletOutputStream; import javax.servlet.http.HttpServletResponse; @@ -265,7 +252,7 @@ public class ProcessDefinitionServiceImpl extends BaseServiceImpl implements Pro if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } - if(checkDescriptionLength(description)){ + if (checkDescriptionLength(description)) { putMsg(result, Status.DESCRIPTION_TOO_LONG_ERROR); return result; } @@ -616,7 +603,7 @@ public class ProcessDefinitionServiceImpl extends BaseServiceImpl implements Pro if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } - if(checkDescriptionLength(description)){ + if (checkDescriptionLength(description)) { putMsg(result, Status.DESCRIPTION_TOO_LONG_ERROR); return result; } @@ -773,7 +760,8 @@ public class ProcessDefinitionServiceImpl extends BaseServiceImpl implements Pro * @return true if process definition name not exists, otherwise false */ @Override - public Map verifyProcessDefinitionName(User loginUser, long projectCode, String name, long processDefinitionCode) { + public Map verifyProcessDefinitionName(User loginUser, long projectCode, String name, + long processDefinitionCode) { Project project = projectMapper.queryByCode(projectCode); // check user access for project Map result = @@ -2215,7 +2203,7 @@ public class ProcessDefinitionServiceImpl extends BaseServiceImpl implements Pro if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } - if(checkDescriptionLength(description)){ + if (checkDescriptionLength(description)) { putMsg(result, Status.DESCRIPTION_TOO_LONG_ERROR); return result; } @@ -2354,7 +2342,7 @@ public class ProcessDefinitionServiceImpl extends BaseServiceImpl implements Pro if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } - if(checkDescriptionLength(description)){ + if (checkDescriptionLength(description)) { putMsg(result, Status.DESCRIPTION_TOO_LONG_ERROR); return result; } diff --git a/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/ProcessTaskRelationServiceImpl.java b/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/ProcessTaskRelationServiceImpl.java index 7b35244477..0e41c774e3 100644 --- a/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/ProcessTaskRelationServiceImpl.java +++ b/dolphinscheduler-api/src/main/java/org/apache/dolphinscheduler/api/service/impl/ProcessTaskRelationServiceImpl.java @@ -98,10 +98,11 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P */ @Transactional @Override - public Map createProcessTaskRelation(User loginUser, long projectCode, long processDefinitionCode, long preTaskCode, long postTaskCode) { + public Map createProcessTaskRelation(User loginUser, long projectCode, long processDefinitionCode, + long preTaskCode, long postTaskCode) { Project project = projectMapper.queryByCode(projectCode); - //check user access for project - Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode,null); + // check user access for project + Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode, null); if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } @@ -115,11 +116,14 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P return result; } updateProcessDefiniteVersion(loginUser, result, processDefinition); - List processTaskRelationList = processTaskRelationMapper.queryByProcessCode(projectCode, processDefinitionCode); + List processTaskRelationList = + processTaskRelationMapper.queryByProcessCode(projectCode, processDefinitionCode); List processTaskRelations = Lists.newArrayList(processTaskRelationList); if (!processTaskRelations.isEmpty()) { - Map preTaskCodeMap = processTaskRelations.stream().filter(r -> r.getPostTaskCode() == postTaskCode) - .collect(Collectors.toMap(ProcessTaskRelation::getPreTaskCode, processTaskRelation -> processTaskRelation)); + Map preTaskCodeMap = + processTaskRelations.stream().filter(r -> r.getPostTaskCode() == postTaskCode) + .collect(Collectors.toMap(ProcessTaskRelation::getPreTaskCode, + processTaskRelation -> processTaskRelation)); if (!preTaskCodeMap.isEmpty()) { if (preTaskCodeMap.containsKey(preTaskCode) || (!preTaskCodeMap.containsKey(0L) && preTaskCode == 0L)) { putMsg(result, Status.PROCESS_TASK_RELATION_EXIST, String.valueOf(processDefinitionCode)); @@ -135,7 +139,8 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P ProcessTaskRelation processTaskRelation = setRelation(processDefinition, postTaskDefinition); if (preTaskCode != 0L) { TaskDefinition preTaskDefinition = taskDefinitionMapper.queryByCode(preTaskCode); - List upstreamTaskRelationList = processTaskRelations.stream().filter(r -> r.getPostTaskCode() == preTaskCode).collect(Collectors.toList()); + List upstreamTaskRelationList = processTaskRelations.stream() + .filter(r -> r.getPostTaskCode() == preTaskCode).collect(Collectors.toList()); // upstream is or not exist if (upstreamTaskRelationList.isEmpty()) { ProcessTaskRelation preProcessTaskRelation = setRelation(processDefinition, preTaskDefinition); @@ -169,7 +174,8 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P return processTaskRelation; } - private void updateProcessDefiniteVersion(User loginUser, Map result, ProcessDefinition processDefinition) { + private void updateProcessDefiniteVersion(User loginUser, Map result, + ProcessDefinition processDefinition) { int insertVersion = processService.saveProcessDefine(loginUser, processDefinition, Boolean.TRUE, Boolean.TRUE); if (insertVersion <= 0) { putMsg(result, Status.UPDATE_PROCESS_DEFINITION_ERROR); @@ -189,10 +195,11 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P */ @Transactional @Override - public Map deleteTaskProcessRelation(User loginUser, long projectCode, long processDefinitionCode, long taskCode) { + public Map deleteTaskProcessRelation(User loginUser, long projectCode, long processDefinitionCode, + long taskCode) { Project project = projectMapper.queryByCode(projectCode); - //check user access for project - Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode,null); + // check user access for project + Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode, null); if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } @@ -210,7 +217,8 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P putMsg(result, Status.TASK_DEFINE_NOT_EXIST, String.valueOf(taskCode)); return result; } - List processTaskRelations = processTaskRelationMapper.queryByProcessCode(projectCode, processDefinitionCode); + List processTaskRelations = + processTaskRelationMapper.queryByProcessCode(projectCode, processDefinitionCode); List processTaskRelationList = Lists.newArrayList(processTaskRelations); if (CollectionUtils.isEmpty(processTaskRelationList)) { putMsg(result, Status.DATA_IS_NULL, "processTaskRelationList"); @@ -232,8 +240,8 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P updateProcessDefiniteVersion(loginUser, result, processDefinition); updateRelation(loginUser, result, processDefinition, processTaskRelationList); if (TASK_TYPE_CONDITIONS.equals(taskDefinition.getTaskType()) - || TASK_TYPE_DEPENDENT.equals(taskDefinition.getTaskType()) - || TASK_TYPE_SUB_PROCESS.equals(taskDefinition.getTaskType())) { + || TASK_TYPE_DEPENDENT.equals(taskDefinition.getTaskType()) + || TASK_TYPE_SUB_PROCESS.equals(taskDefinition.getTaskType())) { int deleteTaskDefinition = taskDefinitionMapper.deleteByCode(taskCode); if (0 == deleteTaskDefinition) { putMsg(result, Status.DELETE_TASK_DEFINE_BY_CODE_ERROR); @@ -246,9 +254,11 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P private void updateRelation(User loginUser, Map result, ProcessDefinition processDefinition, List processTaskRelationList) { - List relationLogs = processTaskRelationList.stream().map(ProcessTaskRelationLog::new).collect(Collectors.toList()); - int insertResult = processService.saveTaskRelation(loginUser, processDefinition.getProjectCode(), processDefinition.getCode(), - processDefinition.getVersion(), relationLogs, Lists.newArrayList(), Boolean.TRUE); + List relationLogs = + processTaskRelationList.stream().map(ProcessTaskRelationLog::new).collect(Collectors.toList()); + int insertResult = processService.saveTaskRelation(loginUser, processDefinition.getProjectCode(), + processDefinition.getCode(), + processDefinition.getVersion(), relationLogs, Lists.newArrayList(), Boolean.TRUE); if (insertResult == Constants.EXIT_CODE_SUCCESS) { putMsg(result, Status.SUCCESS); result.put(Constants.DATA_LIST, processDefinition); @@ -269,10 +279,11 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P */ @Transactional @Override - public Map deleteUpstreamRelation(User loginUser, long projectCode, String preTaskCodes, long taskCode) { + public Map deleteUpstreamRelation(User loginUser, long projectCode, String preTaskCodes, + long taskCode) { Project project = projectMapper.queryByCode(projectCode); - //check user access for project - Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode,null); + // check user access for project + Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode, null); if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } @@ -286,12 +297,14 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P return result; } - List preTaskCodeList = Lists.newArrayList(preTaskCodes.split(Constants.COMMA)).stream().map(Long::parseLong).collect(Collectors.toList()); + List preTaskCodeList = Lists.newArrayList(preTaskCodes.split(Constants.COMMA)).stream() + .map(Long::parseLong).collect(Collectors.toList()); if (preTaskCodeList.contains(0L)) { putMsg(result, Status.DATA_IS_NULL, "preTaskCodes"); return result; } - List currentUpstreamList = upstreamList.stream().map(ProcessTaskRelation::getPreTaskCode).collect(Collectors.toList()); + List currentUpstreamList = + upstreamList.stream().map(ProcessTaskRelation::getPreTaskCode).collect(Collectors.toList()); if (currentUpstreamList.contains(0L)) { putMsg(result, Status.DATA_IS_NOT_VALID, "currentUpstreamList"); return result; @@ -303,12 +316,15 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P putMsg(result, Status.DATA_IS_NOT_VALID, StringUtils.join(preTaskCodeList, Constants.COMMA)); return result; } - ProcessDefinition processDefinition = processDefinitionMapper.queryByCode(upstreamList.get(0).getProcessDefinitionCode()); + ProcessDefinition processDefinition = + processDefinitionMapper.queryByCode(upstreamList.get(0).getProcessDefinitionCode()); if (processDefinition == null) { - putMsg(result, Status.PROCESS_DEFINE_NOT_EXIST, String.valueOf(upstreamList.get(0).getProcessDefinitionCode())); + putMsg(result, Status.PROCESS_DEFINE_NOT_EXIST, + String.valueOf(upstreamList.get(0).getProcessDefinitionCode())); return result; } - List processTaskRelations = processTaskRelationMapper.queryByProcessCode(projectCode, processDefinition.getCode()); + List processTaskRelations = + processTaskRelationMapper.queryByProcessCode(projectCode, processDefinition.getCode()); List processTaskRelationList = Lists.newArrayList(processTaskRelations); List processTaskRelationWaitRemove = Lists.newArrayList(); for (ProcessTaskRelation processTaskRelation : processTaskRelationList) { @@ -318,7 +334,8 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P processTaskRelationWaitRemove.add(processTaskRelation); } } else { - if (processTaskRelation.getPostTaskCode() == taskCode && (currentUpstreamList.isEmpty() || tmpCurrent.isEmpty())) { + if (processTaskRelation.getPostTaskCode() == taskCode + && (currentUpstreamList.isEmpty() || tmpCurrent.isEmpty())) { processTaskRelation.setPreTaskVersion(0); processTaskRelation.setPreTaskCode(0L); } @@ -341,10 +358,11 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P */ @Transactional @Override - public Map deleteDownstreamRelation(User loginUser, long projectCode, String postTaskCodes, long taskCode) { + public Map deleteDownstreamRelation(User loginUser, long projectCode, String postTaskCodes, + long taskCode) { Project project = projectMapper.queryByCode(projectCode); - //check user access for project - Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode,null); + // check user access for project + Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode, null); if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } @@ -352,24 +370,31 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P putMsg(result, Status.DATA_IS_NULL, "postTaskCodes"); return result; } - List downstreamList = processTaskRelationMapper.queryDownstreamByCode(projectCode, taskCode); + List downstreamList = + processTaskRelationMapper.queryDownstreamByCode(projectCode, taskCode); if (CollectionUtils.isEmpty(downstreamList)) { putMsg(result, Status.DATA_IS_NULL, "taskCode"); return result; } - List postTaskCodeList = Lists.newArrayList(postTaskCodes.split(Constants.COMMA)).stream().map(Long::parseLong).collect(Collectors.toList()); + List postTaskCodeList = Lists.newArrayList(postTaskCodes.split(Constants.COMMA)).stream() + .map(Long::parseLong).collect(Collectors.toList()); if (postTaskCodeList.contains(0L)) { putMsg(result, Status.DATA_IS_NULL, "postTaskCodes"); return result; } - ProcessDefinition processDefinition = processDefinitionMapper.queryByCode(downstreamList.get(0).getProcessDefinitionCode()); + ProcessDefinition processDefinition = + processDefinitionMapper.queryByCode(downstreamList.get(0).getProcessDefinitionCode()); if (processDefinition == null) { - putMsg(result, Status.PROCESS_DEFINE_NOT_EXIST, String.valueOf(downstreamList.get(0).getProcessDefinitionCode())); + putMsg(result, Status.PROCESS_DEFINE_NOT_EXIST, + String.valueOf(downstreamList.get(0).getProcessDefinitionCode())); return result; } - List processTaskRelations = processTaskRelationMapper.queryByProcessCode(projectCode, processDefinition.getCode()); + List processTaskRelations = + processTaskRelationMapper.queryByProcessCode(projectCode, processDefinition.getCode()); List processTaskRelationList = Lists.newArrayList(processTaskRelations); - processTaskRelationList.removeIf(processTaskRelation -> postTaskCodeList.contains(processTaskRelation.getPostTaskCode()) && processTaskRelation.getPreTaskCode() == taskCode); + processTaskRelationList + .removeIf(processTaskRelation -> postTaskCodeList.contains(processTaskRelation.getPostTaskCode()) + && processTaskRelation.getPreTaskCode() == taskCode); updateProcessDefiniteVersion(loginUser, result, processDefinition); updateRelation(loginUser, result, processDefinition, processTaskRelationList); return result; @@ -386,24 +411,25 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P @Override public Map queryUpstreamRelation(User loginUser, long projectCode, long taskCode) { Project project = projectMapper.queryByCode(projectCode); - //check user access for project - Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode,null); + // check user access for project + Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode, null); if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } - List processTaskRelationList = processTaskRelationMapper.queryUpstreamByCode(projectCode, taskCode); + List processTaskRelationList = + processTaskRelationMapper.queryUpstreamByCode(projectCode, taskCode); List taskDefinitionLogList = new ArrayList<>(); if (CollectionUtils.isNotEmpty(processTaskRelationList)) { Set taskDefinitions = processTaskRelationList - .stream() - .map(processTaskRelation -> { - TaskDefinition taskDefinition = buildTaskDefinition(); - taskDefinition.setProjectCode(processTaskRelation.getProjectCode()); - taskDefinition.setCode(processTaskRelation.getPreTaskCode()); - taskDefinition.setVersion(processTaskRelation.getPreTaskVersion()); - return taskDefinition; - }) - .collect(Collectors.toSet()); + .stream() + .map(processTaskRelation -> { + TaskDefinition taskDefinition = buildTaskDefinition(); + taskDefinition.setProjectCode(processTaskRelation.getProjectCode()); + taskDefinition.setCode(processTaskRelation.getPreTaskCode()); + taskDefinition.setVersion(processTaskRelation.getPreTaskVersion()); + return taskDefinition; + }) + .collect(Collectors.toSet()); taskDefinitionLogList = taskDefinitionLogMapper.queryByTaskDefinitions(taskDefinitions); } result.put(Constants.DATA_LIST, taskDefinitionLogList); @@ -422,24 +448,25 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P @Override public Map queryDownstreamRelation(User loginUser, long projectCode, long taskCode) { Project project = projectMapper.queryByCode(projectCode); - //check user access for project - Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode,null); + // check user access for project + Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode, null); if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } - List processTaskRelationList = processTaskRelationMapper.queryDownstreamByCode(projectCode, taskCode); + List processTaskRelationList = + processTaskRelationMapper.queryDownstreamByCode(projectCode, taskCode); List taskDefinitionLogList = new ArrayList<>(); if (CollectionUtils.isNotEmpty(processTaskRelationList)) { Set taskDefinitions = processTaskRelationList - .stream() - .map(processTaskRelation -> { - TaskDefinition taskDefinition = buildTaskDefinition(); - taskDefinition.setProjectCode(processTaskRelation.getProjectCode()); - taskDefinition.setCode(processTaskRelation.getPostTaskCode()); - taskDefinition.setVersion(processTaskRelation.getPostTaskVersion()); - return taskDefinition; - }) - .collect(Collectors.toSet()); + .stream() + .map(processTaskRelation -> { + TaskDefinition taskDefinition = buildTaskDefinition(); + taskDefinition.setProjectCode(processTaskRelation.getProjectCode()); + taskDefinition.setCode(processTaskRelation.getPostTaskCode()); + taskDefinition.setVersion(processTaskRelation.getPostTaskVersion()); + return taskDefinition; + }) + .collect(Collectors.toSet()); taskDefinitionLogList = taskDefinitionLogMapper.queryByTaskDefinitions(taskDefinitions); } result.put(Constants.DATA_LIST, taskDefinitionLogList); @@ -459,10 +486,11 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P */ @Transactional @Override - public Map deleteEdge(User loginUser, long projectCode, long processDefinitionCode, long preTaskCode, long postTaskCode) { + public Map deleteEdge(User loginUser, long projectCode, long processDefinitionCode, + long preTaskCode, long postTaskCode) { Project project = projectMapper.queryByCode(projectCode); - //check user access for project - Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode,null); + // check user access for project + Map result = projectService.checkProjectAndAuth(loginUser, project, projectCode, null); if (result.get(Constants.STATUS) != Status.SUCCESS) { return result; } @@ -471,7 +499,8 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P putMsg(result, Status.PROCESS_DEFINE_NOT_EXIST, String.valueOf(processDefinitionCode)); return result; } - List processTaskRelations = processTaskRelationMapper.queryByProcessCode(projectCode, processDefinitionCode); + List processTaskRelations = + processTaskRelationMapper.queryByProcessCode(projectCode, processDefinitionCode); List processTaskRelationList = Lists.newArrayList(processTaskRelations); if (CollectionUtils.isEmpty(processTaskRelationList)) { putMsg(result, Status.DATA_IS_NULL, "processTaskRelationList"); @@ -522,6 +551,7 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P private TaskDefinition buildTaskDefinition() { return new TaskDefinition() { + @Override public boolean equals(Object o) { if (this == o) { @@ -532,8 +562,8 @@ public class ProcessTaskRelationServiceImpl extends BaseServiceImpl implements P } TaskDefinition that = (TaskDefinition) o; return getCode() == that.getCode() - && getVersion() == that.getVersion() - && getProjectCode() == that.getProjectCode(); + && getVersion() == that.getVersion() + && getProjectCode() == that.getProjectCode(); } @Override diff --git a/dolphinscheduler-bom/pom.xml b/dolphinscheduler-bom/pom.xml index a34b837443..6319eab6d7 100644 --- a/dolphinscheduler-bom/pom.xml +++ b/dolphinscheduler-bom/pom.xml @@ -275,8 +275,8 @@ io.grpc grpc-bom ${io.grpc.version} - import pom + import diff --git a/dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/Constants.java b/dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/Constants.java index 804a2ddaef..c6eecb4212 100644 --- a/dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/Constants.java +++ b/dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/Constants.java @@ -17,11 +17,12 @@ package org.apache.dolphinscheduler.common; -import org.apache.commons.lang3.StringUtils; -import org.apache.commons.lang3.SystemUtils; import org.apache.dolphinscheduler.common.enums.WorkflowExecutionStatus; import org.apache.dolphinscheduler.plugin.task.api.enums.TaskExecutionStatus; +import org.apache.commons.lang3.StringUtils; +import org.apache.commons.lang3.SystemUtils; + import java.time.Duration; import java.util.regex.Pattern; diff --git a/dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/utils/OSUtils.java b/dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/utils/OSUtils.java index 89104e4604..b5203a9245 100644 --- a/dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/utils/OSUtils.java +++ b/dolphinscheduler-common/src/main/java/org/apache/dolphinscheduler/common/utils/OSUtils.java @@ -19,8 +19,13 @@ package org.apache.dolphinscheduler.common.utils; import org.apache.dolphinscheduler.common.shell.ShellExecutor; -import org.apache.commons.lang3.SystemUtils; +import oshi.SystemInfo; +import oshi.hardware.CentralProcessor; +import oshi.hardware.GlobalMemory; +import oshi.hardware.HardwareAbstractionLayer; + import org.apache.commons.lang3.StringUtils; +import org.apache.commons.lang3.SystemUtils; import java.io.BufferedReader; import java.io.File; @@ -42,11 +47,6 @@ import java.util.regex.Pattern; import org.slf4j.Logger; import org.slf4j.LoggerFactory; -import oshi.SystemInfo; -import oshi.hardware.CentralProcessor; -import oshi.hardware.GlobalMemory; -import oshi.hardware.HardwareAbstractionLayer; - /** * os utils */ @@ -101,7 +101,7 @@ public class OSUtils { */ public static double diskAvailable() { File file = new File("."); - long freeSpace = file.getFreeSpace(); //unallocated / free disk space in bytes. + long freeSpace = file.getFreeSpace(); // unallocated / free disk space in bytes. double diskAvailable = freeSpace / 1024.0 / 1024 / 1024; @@ -160,7 +160,7 @@ public class OSUtils { long now = System.currentTimeMillis(); if (now - prevTickTime > 950) { // Enough time has elapsed. - cpuUsage = processor.getSystemCpuLoadBetweenTicks(prevTicks); + cpuUsage = processor.getSystemCpuLoadBetweenTicks(prevTicks); prevTickTime = System.currentTimeMillis(); prevTicks = processor.getSystemCpuLoadTicks(); } @@ -198,8 +198,9 @@ public class OSUtils { private static List getUserListFromLinux() throws IOException { List userList = new ArrayList<>(); - try (BufferedReader bufferedReader = new BufferedReader( - new InputStreamReader(new FileInputStream("/etc/passwd")))) { + try ( + BufferedReader bufferedReader = new BufferedReader( + new InputStreamReader(new FileInputStream("/etc/passwd")))) { String line; while ((line = bufferedReader.readLine()) != null) { @@ -273,14 +274,14 @@ public class OSUtils { * @return boolean */ public static boolean existTenantCodeInLinux(String tenantCode) { - try{ - String result = exeCmd("id "+ tenantCode); - if (!StringUtils.isEmpty(result)){ + try { + String result = exeCmd("id " + tenantCode); + if (!StringUtils.isEmpty(result)) { return result.contains("uid="); } - }catch (Exception e){ - //because ShellExecutor method throws exception to the linux return status is not 0 - //not exist user return status is 1 + } catch (Exception e) { + // because ShellExecutor method throws exception to the linux return status is not 0 + // not exist user return status is 1 logger.error(e.getMessage(), e); } return false; @@ -472,8 +473,9 @@ public class OSUtils { // system available physical memory double availablePhysicalMemorySize = availablePhysicalMemorySize(); if (loadAverage > maxCpuLoadAvg || availablePhysicalMemorySize < reservedMemory) { - logger.warn("Current cpu load average {} is too high or available memory {}G is too low, under max.cpuLoad.avg={} and reserved.memory={}G", - loadAverage, availablePhysicalMemorySize, maxCpuLoadAvg, reservedMemory); + logger.warn( + "Current cpu load average {} is too high or available memory {}G is too low, under max.cpuLoad.avg={} and reserved.memory={}G", + loadAverage, availablePhysicalMemorySize, maxCpuLoadAvg, reservedMemory); return true; } return false; diff --git a/dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/os/OSUtilsTest.java b/dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/os/OSUtilsTest.java index 8f4a3a0f77..f21b58070a 100644 --- a/dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/os/OSUtilsTest.java +++ b/dolphinscheduler-common/src/test/java/org/apache/dolphinscheduler/common/os/OSUtilsTest.java @@ -16,18 +16,17 @@ */ package org.apache.dolphinscheduler.common.os; +import org.apache.dolphinscheduler.common.utils.OSUtils; import org.apache.commons.lang3.SystemUtils; -import org.apache.dolphinscheduler.common.utils.OSUtils; + +import java.util.List; import org.junit.Assert; import org.junit.Test; import org.slf4j.Logger; import org.slf4j.LoggerFactory; -import java.util.List; - - /** * OSUtilsTest */ @@ -64,7 +63,7 @@ public class OSUtilsTest { } @Test - public void availablePhysicalMemorySize(){ + public void availablePhysicalMemorySize() { double physicalMemorySize = OSUtils.availablePhysicalMemorySize(); logger.info("physicalMemorySize : {}", physicalMemorySize); Assert.assertTrue(physicalMemorySize >= 0.0); @@ -72,31 +71,29 @@ public class OSUtilsTest { } @Test - public void existTenantCodeInLinux(){ - if (SystemUtils.IS_OS_LINUX){ + public void existTenantCodeInLinux() { + if (SystemUtils.IS_OS_LINUX) { boolean test = OSUtils.existTenantCodeInLinux("root"); Assert.assertTrue(test); boolean test1 = OSUtils.existTenantCodeInLinux("xxxtt"); Assert.assertFalse(test1); - }else{ - Assert.assertFalse("system must be linux",false); + } else { + Assert.assertFalse("system must be linux", false); } } @Test - public void existOSTenandCode(){ - if (SystemUtils.IS_OS_LINUX){ + public void existOSTenandCode() { + if (SystemUtils.IS_OS_LINUX) { List userList = OSUtils.getUserList(); Assert.assertTrue(userList.contains("root")); Assert.assertFalse(userList.contains("xxxtt")); - }else{ - Assert.assertFalse("system must be linux",false); + } else { + Assert.assertFalse("system must be linux", false); } } - - } diff --git a/dolphinscheduler-dao/src/main/java/org/apache/dolphinscheduler/dao/entity/TaskDefinition.java b/dolphinscheduler-dao/src/main/java/org/apache/dolphinscheduler/dao/entity/TaskDefinition.java index ff6b71be90..16b599df2f 100644 --- a/dolphinscheduler-dao/src/main/java/org/apache/dolphinscheduler/dao/entity/TaskDefinition.java +++ b/dolphinscheduler-dao/src/main/java/org/apache/dolphinscheduler/dao/entity/TaskDefinition.java @@ -34,6 +34,8 @@ import java.util.List; import java.util.Map; import java.util.Objects; +import lombok.Data; + import com.baomidou.mybatisplus.annotation.FieldStrategy; import com.baomidou.mybatisplus.annotation.IdType; import com.baomidou.mybatisplus.annotation.TableField; @@ -43,7 +45,6 @@ import com.fasterxml.jackson.databind.JsonNode; import com.fasterxml.jackson.databind.annotation.JsonDeserialize; import com.fasterxml.jackson.databind.annotation.JsonSerialize; import com.google.common.base.Strings; -import lombok.Data; @Data @TableName("t_ds_task_definition") diff --git a/dolphinscheduler-datasource-plugin/dolphinscheduler-datasource-api/src/main/java/org/apache/dolphinscheduler/plugin/datasource/api/utils/CommonUtils.java b/dolphinscheduler-datasource-plugin/dolphinscheduler-datasource-api/src/main/java/org/apache/dolphinscheduler/plugin/datasource/api/utils/CommonUtils.java index 2ae15a01fd..96bb77bc29 100644 --- a/dolphinscheduler-datasource-plugin/dolphinscheduler-datasource-api/src/main/java/org/apache/dolphinscheduler/plugin/datasource/api/utils/CommonUtils.java +++ b/dolphinscheduler-datasource-plugin/dolphinscheduler-datasource-api/src/main/java/org/apache/dolphinscheduler/plugin/datasource/api/utils/CommonUtils.java @@ -17,17 +17,18 @@ package org.apache.dolphinscheduler.plugin.datasource.api.utils; +import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.*; +import static org.apache.dolphinscheduler.spi.utils.Constants.RESOURCE_STORAGE_TYPE; + import org.apache.dolphinscheduler.spi.enums.ResUploadType; import org.apache.dolphinscheduler.spi.utils.PropertyUtils; import org.apache.dolphinscheduler.spi.utils.StringUtils; + import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.security.UserGroupInformation; import java.io.IOException; -import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.*; -import static org.apache.dolphinscheduler.spi.utils.Constants.RESOURCE_STORAGE_TYPE; - /** * common utils */ @@ -36,7 +37,7 @@ public class CommonUtils { private CommonUtils() { throw new UnsupportedOperationException("Construct CommonUtils"); } - + /** * if upload resource is HDFS and kerberos startup is true , else false * @@ -70,7 +71,8 @@ public class CommonUtils { * @param loginUserKeytabPath loginUserKeytabPath * @throws IOException errors */ - public static void loadKerberosConf(String javaSecurityKrb5Conf, String loginUserKeytabUsername, String loginUserKeytabPath) throws IOException { + public static void loadKerberosConf(String javaSecurityKrb5Conf, String loginUserKeytabUsername, + String loginUserKeytabPath) throws IOException { Configuration configuration = new Configuration(); configuration.setClassLoader(configuration.getClass().getClassLoader()); loadKerberosConf(javaSecurityKrb5Conf, loginUserKeytabUsername, loginUserKeytabPath, configuration); @@ -86,12 +88,16 @@ public class CommonUtils { * @return load kerberos config return true * @throws IOException errors */ - public static boolean loadKerberosConf(String javaSecurityKrb5Conf, String loginUserKeytabUsername, String loginUserKeytabPath, Configuration configuration) throws IOException { + public static boolean loadKerberosConf(String javaSecurityKrb5Conf, String loginUserKeytabUsername, + String loginUserKeytabPath, Configuration configuration) throws IOException { if (CommonUtils.getKerberosStartupState()) { - System.setProperty(JAVA_SECURITY_KRB5_CONF, StringUtils.defaultIfBlank(javaSecurityKrb5Conf, PropertyUtils.getString(JAVA_SECURITY_KRB5_CONF_PATH))); + System.setProperty(JAVA_SECURITY_KRB5_CONF, StringUtils.defaultIfBlank(javaSecurityKrb5Conf, + PropertyUtils.getString(JAVA_SECURITY_KRB5_CONF_PATH))); configuration.set(HADOOP_SECURITY_AUTHENTICATION, KERBEROS); UserGroupInformation.setConfiguration(configuration); - UserGroupInformation.loginUserFromKeytab(StringUtils.defaultIfBlank(loginUserKeytabUsername, PropertyUtils.getString(LOGIN_USER_KEY_TAB_USERNAME)), + UserGroupInformation.loginUserFromKeytab( + StringUtils.defaultIfBlank(loginUserKeytabUsername, + PropertyUtils.getString(LOGIN_USER_KEY_TAB_USERNAME)), StringUtils.defaultIfBlank(loginUserKeytabPath, PropertyUtils.getString(LOGIN_USER_KEY_TAB_PATH))); return true; } @@ -134,7 +140,7 @@ public class CommonUtils { public static String getHdfsDataBasePath() { String resourceUploadPath = PropertyUtils.getString(RESOURCE_UPLOAD_PATH, "/dolphinscheduler"); if ("/".equals(resourceUploadPath)) { - // if basepath is configured to /, the generated url may be //default/resources (with extra leading /) + // if basepath is configured to /, the generated url may be //default/resources (with extra leading /) return ""; } else { return resourceUploadPath; diff --git a/dolphinscheduler-log-server/src/main/java/org/apache/dolphinscheduler/server/log/LoggerRequestProcessor.java b/dolphinscheduler-log-server/src/main/java/org/apache/dolphinscheduler/server/log/LoggerRequestProcessor.java index 490f6d76eb..d921d54000 100644 --- a/dolphinscheduler-log-server/src/main/java/org/apache/dolphinscheduler/server/log/LoggerRequestProcessor.java +++ b/dolphinscheduler-log-server/src/main/java/org/apache/dolphinscheduler/server/log/LoggerRequestProcessor.java @@ -46,8 +46,10 @@ import java.io.InputStream; import java.nio.charset.StandardCharsets; import java.nio.file.Files; import java.nio.file.Paths; +import java.util.ArrayList; import java.util.Collections; import java.util.List; +import java.util.Set; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.stream.Collectors; @@ -166,8 +168,9 @@ public class LoggerRequestProcessor implements NettyRequestProcessor { if (!checkPathSecurity(logPath)) { throw new IllegalArgumentException("Illegal path"); } - List appIds = LogUtils.getAppIdsFromLogFile(logPath); - channel.writeAndFlush(new GetAppIdResponseCommand(appIds).convert2Command(command.getOpaque())); + Set appIds = LogUtils.getAppIdsFromLogFile(logPath); + channel.writeAndFlush( + new GetAppIdResponseCommand(new ArrayList<>(appIds)).convert2Command(command.getOpaque())); break; default: throw new IllegalArgumentException("unknown commandType: " + commandType); diff --git a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/builder/TaskExecutionContextBuilder.java b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/builder/TaskExecutionContextBuilder.java index 65b3952efd..40c07df807 100644 --- a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/builder/TaskExecutionContextBuilder.java +++ b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/builder/TaskExecutionContextBuilder.java @@ -68,6 +68,7 @@ public class TaskExecutionContextBuilder { taskExecutionContext.setCurrentExecutionStatus(TaskExecutionStatus.SUBMITTED_SUCCESS); taskExecutionContext.setCpuQuota(taskInstance.getCpuQuota()); taskExecutionContext.setMemoryMax(taskInstance.getMemoryMax()); + taskExecutionContext.setAppIds(taskInstance.getAppLink()); return this; } diff --git a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/processor/queue/TaskEvent.java b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/processor/queue/TaskEvent.java index 9f6679c9d8..dbbca6cab3 100644 --- a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/processor/queue/TaskEvent.java +++ b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/processor/queue/TaskEvent.java @@ -25,8 +25,8 @@ import org.apache.dolphinscheduler.remote.command.TaskRejectCommand; import java.util.Date; -import io.netty.channel.Channel; import lombok.Data; +import io.netty.channel.Channel; /** * task event @@ -113,6 +113,7 @@ public class TaskEvent { event.setStartTime(command.getStartTime()); event.setExecutePath(command.getExecutePath()); event.setLogPath(command.getLogPath()); + event.setAppIds(command.getAppIds()); event.setChannel(channel); event.setWorkerAddress(workerAddress); event.setEvent(TaskEventType.RUNNING); diff --git a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/registry/ServerNodeManager.java b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/registry/ServerNodeManager.java index 147fff1364..3ec10b3ed7 100644 --- a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/registry/ServerNodeManager.java +++ b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/registry/ServerNodeManager.java @@ -17,8 +17,9 @@ package org.apache.dolphinscheduler.server.master.registry; -import org.apache.commons.collections.CollectionUtils; -import org.apache.commons.lang3.StringUtils; +import static org.apache.dolphinscheduler.common.Constants.REGISTRY_DOLPHINSCHEDULER_MASTERS; +import static org.apache.dolphinscheduler.common.Constants.REGISTRY_DOLPHINSCHEDULER_WORKERS; + import org.apache.dolphinscheduler.common.Constants; import org.apache.dolphinscheduler.common.enums.NodeType; import org.apache.dolphinscheduler.common.model.Server; @@ -34,13 +35,10 @@ import org.apache.dolphinscheduler.remote.utils.NamedThreadFactory; import org.apache.dolphinscheduler.server.master.config.MasterConfig; import org.apache.dolphinscheduler.service.queue.MasterPriorityQueue; import org.apache.dolphinscheduler.service.registry.RegistryClient; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; -import org.springframework.beans.factory.InitializingBean; -import org.springframework.beans.factory.annotation.Autowired; -import org.springframework.stereotype.Service; -import javax.annotation.PreDestroy; +import org.apache.commons.collections.CollectionUtils; +import org.apache.commons.lang3.StringUtils; + import java.util.ArrayList; import java.util.Collection; import java.util.Collections; @@ -57,8 +55,13 @@ import java.util.concurrent.locks.Lock; import java.util.concurrent.locks.ReentrantLock; import java.util.concurrent.locks.ReentrantReadWriteLock; -import static org.apache.dolphinscheduler.common.Constants.REGISTRY_DOLPHINSCHEDULER_MASTERS; -import static org.apache.dolphinscheduler.common.Constants.REGISTRY_DOLPHINSCHEDULER_WORKERS; +import javax.annotation.PreDestroy; + +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; +import org.springframework.beans.factory.InitializingBean; +import org.springframework.beans.factory.annotation.Autowired; +import org.springframework.stereotype.Service; /** * server node manager @@ -74,7 +77,6 @@ public class ServerNodeManager implements InitializingBean { private final ReentrantReadWriteLock.ReadLock workerGroupReadLock = workerGroupLock.readLock(); private final ReentrantReadWriteLock.WriteLock workerGroupWriteLock = workerGroupLock.writeLock(); - private final ReentrantReadWriteLock workerNodeInfoLock = new ReentrantReadWriteLock(); private final ReentrantReadWriteLock.ReadLock workerNodeInfoReadLock = workerNodeInfoLock.readLock(); private final ReentrantReadWriteLock.WriteLock workerNodeInfoWriteLock = workerNodeInfoLock.writeLock(); @@ -138,7 +140,6 @@ public class ServerNodeManager implements InitializingBean { return MASTER_SIZE; } - /** * init listener * @@ -153,7 +154,8 @@ public class ServerNodeManager implements InitializingBean { /** * init executor service */ - executorService = Executors.newSingleThreadScheduledExecutor(new NamedThreadFactory("ServerNodeManagerExecutor")); + executorService = + Executors.newSingleThreadScheduledExecutor(new NamedThreadFactory("ServerNodeManagerExecutor")); executorService.scheduleWithFixedDelay(new WorkerNodeInfoAndGroupDbSyncTask(), 0, 10, TimeUnit.SECONDS); /* * init MasterNodeListener listener @@ -212,7 +214,6 @@ public class ServerNodeManager implements InitializingBean { } } - protected Set getWorkerAddressByWorkerGroup(Map newWorkerNodeInfo, WorkerGroup wg) { Set nodes = new HashSet<>(); String[] addrs = wg.getAddrList().split(Constants.COMMA); @@ -238,7 +239,8 @@ public class ServerNodeManager implements InitializingBean { try { String[] parts = path.split("/"); if (parts.length < WORKER_LISTENER_CHECK_LENGTH) { - throw new IllegalArgumentException(String.format("worker group path : %s is not valid, ignore", path)); + throw new IllegalArgumentException( + String.format("worker group path : %s is not valid, ignore", path)); } final String workerGroupName = parts[parts.length - 2]; final String workerAddress = parts[parts.length - 1]; @@ -272,6 +274,7 @@ public class ServerNodeManager implements InitializingBean { } class MasterDataListener implements SubscribeListener { + @Override public void notify(Event event) { final String path = event.path(); @@ -330,7 +333,8 @@ public class ServerNodeManager implements InitializingBean { } else { logger.warn("current addr:{} is not in active master list", masterConfig.getMasterAddress()); } - logger.info("update master nodes, master size: {}, slot: {}, addr: {}", MASTER_SIZE, MASTER_SLOT, masterConfig.getMasterAddress()); + logger.info("update master nodes, master size: {}, slot: {}, addr: {}", MASTER_SIZE, MASTER_SLOT, + masterConfig.getMasterAddress()); } finally { masterLock.unlock(); } diff --git a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/runner/StreamTaskExecuteRunnable.java b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/runner/StreamTaskExecuteRunnable.java index 2faade40ce..edc815d721 100644 --- a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/runner/StreamTaskExecuteRunnable.java +++ b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/runner/StreamTaskExecuteRunnable.java @@ -73,11 +73,11 @@ import java.util.concurrent.ConcurrentLinkedQueue; import java.util.stream.Collectors; import java.util.stream.Stream; +import lombok.NonNull; + import org.slf4j.Logger; import org.slf4j.LoggerFactory; -import lombok.NonNull; - /** * stream task execute */ @@ -118,7 +118,8 @@ public class StreamTaskExecuteRunnable implements Runnable { this.dispatcher = SpringApplicationContext.getBean(ExecutorDispatcher.class); this.taskPluginManager = SpringApplicationContext.getBean(TaskPluginManager.class); this.processTaskRelationMapper = SpringApplicationContext.getBean(ProcessTaskRelationMapper.class); - this.streamTaskInstanceExecCacheManager = SpringApplicationContext.getBean(StreamTaskInstanceExecCacheManager.class); + this.streamTaskInstanceExecCacheManager = + SpringApplicationContext.getBean(StreamTaskInstanceExecCacheManager.class); this.taskDefinition = taskDefinition; this.taskExecuteStartCommand = taskExecuteStartCommand; } @@ -137,7 +138,8 @@ public class StreamTaskExecuteRunnable implements Runnable { // add cache streamTaskInstanceExecCacheManager.cache(taskInstance.getId(), this); - List processTaskRelationList = processTaskRelationMapper.queryByTaskCode(taskDefinition.getCode()); + List processTaskRelationList = + processTaskRelationMapper.queryByTaskCode(taskDefinition.getCode()); long processDefinitionCode = processTaskRelationList.get(0).getProcessDefinitionCode(); int processDefinitionVersion = processTaskRelationList.get(0).getProcessDefinitionVersion(); processDefinition = processService.findProcessDefinition(processDefinitionCode, processDefinitionVersion); @@ -151,24 +153,25 @@ public class StreamTaskExecuteRunnable implements Runnable { } TaskDispatchCommand dispatchCommand = new TaskDispatchCommand(taskExecutionContext, - masterConfig.getMasterAddress(), - taskExecutionContext.getHost(), - System.currentTimeMillis()); + masterConfig.getMasterAddress(), + taskExecutionContext.getHost(), + System.currentTimeMillis()); - ExecutionContext executionContext = new ExecutionContext(dispatchCommand.convert2Command(), ExecutorType.WORKER, taskExecutionContext.getWorkerGroup(), taskInstance); + ExecutionContext executionContext = new ExecutionContext(dispatchCommand.convert2Command(), ExecutorType.WORKER, + taskExecutionContext.getWorkerGroup(), taskInstance); Boolean dispatchSuccess = false; try { dispatchSuccess = dispatcher.dispatch(executionContext); } catch (ExecuteException e) { logger.error("Master dispatch task to worker error, taskInstanceId: {}, worker: {}", - taskInstance.getId(), - executionContext.getHost(), - e); + taskInstance.getId(), + executionContext.getHost(), + e); } if (!dispatchSuccess) { logger.info("Master failed to dispatch task to worker, taskInstanceId: {}, worker: {}", - taskInstance.getId(), - executionContext.getHost()); + taskInstance.getId(), + executionContext.getHost()); // set task instance fail taskInstance.setState(TaskExecutionStatus.FAILURE); @@ -180,8 +183,8 @@ public class StreamTaskExecuteRunnable implements Runnable { taskRunnableStatus = TaskRunnableStatus.STARTED; logger.info("Master success dispatch task to worker, taskInstanceId: {}, worker: {}", - taskInstance.getId(), - executionContext.getHost()); + taskInstance.getId(), + executionContext.getHost()); } public boolean isStart() { @@ -190,7 +193,8 @@ public class StreamTaskExecuteRunnable implements Runnable { public boolean addTaskEvent(TaskEvent taskEvent) { if (taskInstance.getId() != taskEvent.getTaskInstanceId()) { - logger.info("state event would be abounded, taskInstanceId:{}, eventType:{}, state:{}", taskEvent.getTaskInstanceId(), taskEvent.getEvent(), taskEvent.getState()); + logger.info("state event would be abounded, taskInstanceId:{}, eventType:{}, state:{}", + taskEvent.getTaskInstanceId(), taskEvent.getEvent(), taskEvent.getState()); return false; } taskEvents.add(taskEvent); @@ -207,8 +211,8 @@ public class StreamTaskExecuteRunnable implements Runnable { public void handleEvents() { if (!isStart()) { logger.info( - "The stream task instance is not started, will not handle its state event, current state event size: {}", - taskEvents.size()); + "The stream task instance is not started, will not handle its state event, current state event size: {}", + taskEvents.size()); return; } TaskEvent taskEvent = null; @@ -227,14 +231,15 @@ public class StreamTaskExecuteRunnable implements Runnable { ThreadUtils.sleep(Constants.SLEEP_TIME_MILLIS); } catch (StateEventHandleException stateEventHandleException) { logger.error("State event handle error, will retry this event: {}", - taskEvent, - stateEventHandleException); + taskEvent, + stateEventHandleException); ThreadUtils.sleep(Constants.SLEEP_TIME_MILLIS); } catch (Exception e) { - // we catch the exception here, since if the state event handle failed, the state event will still keep in the stateEvents queue. + // we catch the exception here, since if the state event handle failed, the state event will still keep + // in the stateEvents queue. logger.error("State event handle error, get a unknown exception, will retry this event: {}", - taskEvent, - e); + taskEvent, + e); ThreadUtils.sleep(Constants.SLEEP_TIME_MILLIS); } finally { LoggerUtils.removeWorkflowAndTaskInstanceIdMDC(); @@ -267,14 +272,14 @@ public class StreamTaskExecuteRunnable implements Runnable { taskInstance.setMaxRetryTimes(taskDefinition.getFailRetryTimes()); taskInstance.setRetryInterval(taskDefinition.getFailRetryInterval()); - //set task param + // set task param taskInstance.setTaskParams(taskDefinition.getTaskParams()); - //set task group and priority + // set task group and priority taskInstance.setTaskGroupId(taskDefinition.getTaskGroupId()); taskInstance.setTaskGroupPriority(taskDefinition.getTaskGroupPriority()); - //set task cpu quota and max memory + // set task cpu quota and max memory taskInstance.setCpuQuota(taskDefinition.getCpuQuota()); taskInstance.setMemoryMax(taskDefinition.getMemoryMax()); @@ -290,8 +295,10 @@ public class StreamTaskExecuteRunnable implements Runnable { // task dry run flag taskInstance.setDryRun(taskExecuteStartCommand.getDryRun()); - taskInstance.setWorkerGroup(StringUtils.isBlank(taskDefinition.getWorkerGroup()) ? DEFAULT_WORKER_GROUP : taskDefinition.getWorkerGroup()); - taskInstance.setEnvironmentCode(taskDefinition.getEnvironmentCode() == 0 ? -1 : taskDefinition.getEnvironmentCode()); + taskInstance.setWorkerGroup(StringUtils.isBlank(taskDefinition.getWorkerGroup()) ? DEFAULT_WORKER_GROUP + : taskDefinition.getWorkerGroup()); + taskInstance.setEnvironmentCode( + taskDefinition.getEnvironmentCode() == 0 ? -1 : taskDefinition.getEnvironmentCode()); if (!taskInstance.getEnvironmentCode().equals(-1L)) { Environment environment = processService.findEnvironmentByCode(taskInstance.getEnvironmentCode()); @@ -335,15 +342,16 @@ public class StreamTaskExecuteRunnable implements Runnable { TaskChannel taskChannel = taskPluginManager.getTaskChannel(taskInstance.getTaskType()); ResourceParametersHelper resources = taskChannel.getResources(taskInstance.getTaskParams()); - AbstractParameters baseParam = taskPluginManager.getParameters(ParametersNode.builder().taskType(taskInstance.getTaskType()).taskParams(taskInstance.getTaskParams()).build()); + AbstractParameters baseParam = taskPluginManager.getParameters(ParametersNode.builder() + .taskType(taskInstance.getTaskType()).taskParams(taskInstance.getTaskParams()).build()); Map propertyMap = paramParsingPreparation(taskInstance, baseParam); TaskExecutionContext taskExecutionContext = TaskExecutionContextBuilder.get() - .buildTaskInstanceRelatedInfo(taskInstance) - .buildTaskDefinitionRelatedInfo(taskDefinition) - .buildResourceParametersInfo(resources) - .buildBusinessParamsMap(new HashMap<>()) - .buildParamInfo(propertyMap) - .create(); + .buildTaskInstanceRelatedInfo(taskInstance) + .buildTaskDefinitionRelatedInfo(taskDefinition) + .buildResourceParametersInfo(resources) + .buildBusinessParamsMap(new HashMap<>()) + .buildParamInfo(propertyMap) + .create(); taskExecutionContext.setTenantCode(tenant.getTenantCode()); taskExecutionContext.setProjectCode(processDefinition.getProjectCode()); @@ -360,15 +368,18 @@ public class StreamTaskExecuteRunnable implements Runnable { */ protected Map getResourceFullNames(TaskInstance taskInstance) { Map resourcesMap = new HashMap<>(); - AbstractParameters baseParam = taskPluginManager.getParameters(ParametersNode.builder().taskType(taskInstance.getTaskType()).taskParams(taskInstance.getTaskParams()).build()); + AbstractParameters baseParam = taskPluginManager.getParameters(ParametersNode.builder() + .taskType(taskInstance.getTaskType()).taskParams(taskInstance.getTaskParams()).build()); if (baseParam != null) { List projectResourceFiles = baseParam.getResourceFilesList(); if (CollectionUtils.isNotEmpty(projectResourceFiles)) { // filter the resources that the resource id equals 0 - Set oldVersionResources = projectResourceFiles.stream().filter(t -> t.getId() == null).collect(Collectors.toSet()); + Set oldVersionResources = + projectResourceFiles.stream().filter(t -> t.getId() == null).collect(Collectors.toSet()); if (CollectionUtils.isNotEmpty(oldVersionResources)) { - oldVersionResources.forEach(t -> resourcesMap.put(t.getRes(), processService.queryTenantCodeByResName(t.getRes(), ResourceType.FILE))); + oldVersionResources.forEach(t -> resourcesMap.put(t.getRes(), + processService.queryTenantCodeByResName(t.getRes(), ResourceType.FILE))); } // get the resource id in order to get the resource names in batch @@ -379,7 +390,8 @@ public class StreamTaskExecuteRunnable implements Runnable { Integer[] resourceIds = resourceIdsSet.toArray(new Integer[resourceIdsSet.size()]); List resources = processService.listResourceByIds(resourceIds); - resources.forEach(t -> resourcesMap.put(t.getFullName(), processService.queryTenantCodeByResName(t.getFullName(), ResourceType.FILE))); + resources.forEach(t -> resourcesMap.put(t.getFullName(), + processService.queryTenantCodeByResName(t.getFullName(), ResourceType.FILE))); } } } @@ -387,7 +399,7 @@ public class StreamTaskExecuteRunnable implements Runnable { return resourcesMap; } - protected boolean handleTaskEvent(TaskEvent taskEvent) throws StateEventHandleException, StateEventHandleError { + protected boolean handleTaskEvent(TaskEvent taskEvent) throws StateEventHandleException, StateEventHandleError { measureTaskState(taskEvent); if (taskInstance.getState() == null) { @@ -411,7 +423,8 @@ public class StreamTaskExecuteRunnable implements Runnable { if (taskInstance.getState().isFinished()) { streamTaskInstanceExecCacheManager.removeByTaskInstanceId(taskInstance.getId()); - logger.info("The stream task instance is finish, taskInstanceId:{}, state:{}", taskInstance.getId(), taskEvent.getState()); + logger.info("The stream task instance is finish, taskInstanceId:{}, state:{}", taskInstance.getId(), + taskEvent.getState()); } return true; @@ -441,15 +454,16 @@ public class StreamTaskExecuteRunnable implements Runnable { } } - public Map paramParsingPreparation(@NonNull TaskInstance taskInstance, @NonNull AbstractParameters parameters) { + public Map paramParsingPreparation(@NonNull TaskInstance taskInstance, + @NonNull AbstractParameters parameters) { // assign value to definedParams here - Map globalParamsMap = taskExecuteStartCommand.getStartParams(); + Map globalParamsMap = taskExecuteStartCommand.getStartParams(); Map globalParams = ParamUtils.getUserDefParamsMap(globalParamsMap); // combining local and global parameters Map localParams = parameters.getInputLocalParametersMap(); - //stream pass params + // stream pass params parameters.setVarPool(taskInstance.getVarPool()); Map varParams = parameters.getVarPoolMap(); @@ -470,7 +484,7 @@ public class StreamTaskExecuteRunnable implements Runnable { private void sendAckToWorker(TaskEvent taskEvent) { // If event handle success, send ack to worker to otherwise the worker will retry this event TaskExecuteRunningAckMessage taskExecuteRunningAckMessage = - new TaskExecuteRunningAckMessage(true, taskEvent.getTaskInstanceId()); + new TaskExecuteRunningAckMessage(true, taskEvent.getTaskInstanceId()); taskEvent.getChannel().writeAndFlush(taskExecuteRunningAckMessage.convert2Command()); } diff --git a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/runner/WorkflowExecuteRunnable.java b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/runner/WorkflowExecuteRunnable.java index a064a2b9dc..8e4e64528e 100644 --- a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/runner/WorkflowExecuteRunnable.java +++ b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/runner/WorkflowExecuteRunnable.java @@ -386,7 +386,7 @@ public class WorkflowExecuteRunnable implements Callable { if (!processInstance.isBlocked()) { submitPostNode(Long.toString(taskInstance.getTaskCode())); } - } else if (taskInstance.taskCanRetry() && processInstance.getState().isReadyStop()) { + } else if (taskInstance.taskCanRetry() && !processInstance.getState().isReadyStop()) { // retry task logger.info("Retry taskInstance taskInstance state: {}", taskInstance.getState()); retryTaskInstance(taskInstance); @@ -847,6 +847,9 @@ public class WorkflowExecuteRunnable implements Callable { } if (task.taskCanRetry()) { if (task.getState().isNeedFaultTolerance()) { + task.setFlag(Flag.NO); + processService.updateTaskInstance(task); + // tolerantTaskInstance add to standby list directly TaskInstance tolerantTaskInstance = cloneTolerantTaskInstance(task); addTaskToStandByList(tolerantTaskInstance); @@ -1085,6 +1088,11 @@ public class WorkflowExecuteRunnable implements Callable { // todo relative funtion: TaskInstance.retryTaskIntervalOverTime newTaskInstance.setState(taskInstance.getState()); newTaskInstance.setEndTime(taskInstance.getEndTime()); + + if (taskInstance.getState() == TaskExecutionStatus.NEED_FAULT_TOLERANCE) { + newTaskInstance.setAppLink(taskInstance.getAppLink()); + } + return newTaskInstance; } @@ -1105,6 +1113,7 @@ public class WorkflowExecuteRunnable implements Callable { newTaskInstance.setProcessInstance(processInstance); newTaskInstance.setRetryTimes(taskInstance.getRetryTimes()); newTaskInstance.setState(taskInstance.getState()); + newTaskInstance.setAppLink(taskInstance.getAppLink()); return newTaskInstance; } diff --git a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/service/MasterFailoverService.java b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/service/MasterFailoverService.java index 667589cce3..9e1db701fa 100644 --- a/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/service/MasterFailoverService.java +++ b/dolphinscheduler-master/src/main/java/org/apache/dolphinscheduler/server/master/service/MasterFailoverService.java @@ -17,13 +17,7 @@ package org.apache.dolphinscheduler.server.master.service; -import io.micrometer.core.annotation.Counted; -import io.micrometer.core.annotation.Timed; -import lombok.NonNull; -import org.apache.commons.collections4.CollectionUtils; -import org.apache.commons.lang3.time.StopWatch; import org.apache.dolphinscheduler.common.Constants; -import org.apache.dolphinscheduler.common.enums.Flag; import org.apache.dolphinscheduler.common.enums.NodeType; import org.apache.dolphinscheduler.common.model.Server; import org.apache.dolphinscheduler.common.utils.LoggerUtils; @@ -47,9 +41,9 @@ import org.apache.dolphinscheduler.server.utils.ProcessUtils; import org.apache.dolphinscheduler.service.process.ProcessService; import org.apache.dolphinscheduler.service.registry.RegistryClient; import org.apache.dolphinscheduler.spi.utils.StringUtils; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; -import org.springframework.stereotype.Service; + +import org.apache.commons.collections4.CollectionUtils; +import org.apache.commons.lang3.time.StopWatch; import java.util.Date; import java.util.List; @@ -57,6 +51,15 @@ import java.util.Optional; import java.util.concurrent.TimeUnit; import java.util.stream.Collectors; +import lombok.NonNull; + +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; +import org.springframework.stereotype.Service; + +import io.micrometer.core.annotation.Counted; +import io.micrometer.core.annotation.Timed; + @Service public class MasterFailoverService { @@ -241,7 +244,6 @@ public class MasterFailoverService { } taskInstance.setState(TaskExecutionStatus.NEED_FAULT_TOLERANCE); - taskInstance.setFlag(Flag.NO); processService.saveTaskInstance(taskInstance); } @@ -285,7 +287,8 @@ public class MasterFailoverService { // The processInstance is newly created return false; } - if (processInstance.getRestartTime() != null && processInstance.getRestartTime().after(beFailoveredMasterStartupTime)) { + if (processInstance.getRestartTime() != null + && processInstance.getRestartTime().after(beFailoveredMasterStartupTime)) { // the processInstance is already be failovered. return false; } diff --git a/dolphinscheduler-master/src/test/java/org/apache/dolphinscheduler/server/master/dispatch/ExecutionContextTestUtils.java b/dolphinscheduler-master/src/test/java/org/apache/dolphinscheduler/server/master/dispatch/ExecutionContextTestUtils.java index 6bde573579..ed34fa0603 100644 --- a/dolphinscheduler-master/src/test/java/org/apache/dolphinscheduler/server/master/dispatch/ExecutionContextTestUtils.java +++ b/dolphinscheduler-master/src/test/java/org/apache/dolphinscheduler/server/master/dispatch/ExecutionContextTestUtils.java @@ -36,6 +36,7 @@ import org.mockito.Mockito; * for test use only */ public class ExecutionContextTestUtils { + public static ExecutionContext getExecutionContext(int port) { TaskInstance taskInstance = Mockito.mock(TaskInstance.class); ProcessDefinition processDefinition = Mockito.mock(ProcessDefinition.class); @@ -45,15 +46,15 @@ public class ExecutionContextTestUtils { processInstance.setCommandType(CommandType.COMPLEMENT_DATA); taskInstance.setProcessInstance(processInstance); TaskExecutionContext context = TaskExecutionContextBuilder.get() - .buildTaskInstanceRelatedInfo(taskInstance) - .buildProcessInstanceRelatedInfo(processInstance) - .buildProcessDefinitionRelatedInfo(processDefinition) - .create(); + .buildTaskInstanceRelatedInfo(taskInstance) + .buildProcessInstanceRelatedInfo(processInstance) + .buildProcessDefinitionRelatedInfo(processDefinition) + .create(); TaskDispatchCommand requestCommand = new TaskDispatchCommand(context, - "127.0.0.1:5678", - "127.0.0.1:5678", - System.currentTimeMillis()); + "127.0.0.1:5678", + "127.0.0.1:5678", + System.currentTimeMillis()); Command command = requestCommand.convert2Command(); ExecutionContext executionContext = new ExecutionContext(command, ExecutorType.WORKER, taskInstance); diff --git a/dolphinscheduler-service/src/main/java/org/apache/dolphinscheduler/service/log/LogClientService.java b/dolphinscheduler-service/src/main/java/org/apache/dolphinscheduler/service/log/LogClientService.java index 3088724ed4..735fbab7c4 100644 --- a/dolphinscheduler-service/src/main/java/org/apache/dolphinscheduler/service/log/LogClientService.java +++ b/dolphinscheduler-service/src/main/java/org/apache/dolphinscheduler/service/log/LogClientService.java @@ -17,7 +17,6 @@ package org.apache.dolphinscheduler.service.log; -import lombok.NonNull; import org.apache.dolphinscheduler.common.utils.JSONUtils; import org.apache.dolphinscheduler.common.utils.LoggerUtils; import org.apache.dolphinscheduler.common.utils.NetUtils; @@ -37,12 +36,17 @@ import org.apache.dolphinscheduler.remote.command.log.ViewLogResponseCommand; import org.apache.dolphinscheduler.remote.config.NettyClientConfig; import org.apache.dolphinscheduler.remote.exceptions.RemotingException; import org.apache.dolphinscheduler.remote.utils.Host; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; -import javax.annotation.Nullable; +import java.util.ArrayList; import java.util.List; +import javax.annotation.Nullable; + +import lombok.NonNull; + +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + /** * log client */ @@ -92,7 +96,8 @@ public class LogClientService implements AutoCloseable { * @return log content */ public String rollViewLog(String host, int port, String path, int skipLineNum, int limit) { - logger.info("roll view log, host : {}, port : {}, path {}, skipLineNum {} ,limit {}", host, port, path, skipLineNum, limit); + logger.info("roll view log, host : {}, port : {}, path {}, skipLineNum {} ,limit {}", host, port, path, + skipLineNum, limit); RollViewLogRequestCommand request = new RollViewLogRequestCommand(path, skipLineNum, limit); String result = ""; final Host address = new Host(host, port); @@ -202,18 +207,20 @@ public class LogClientService implements AutoCloseable { return result; } - public @Nullable List getAppIds(@NonNull String host, int port, @NonNull String taskLogFilePath) throws RemotingException, InterruptedException { + public @Nullable List getAppIds(@NonNull String host, int port, + @NonNull String taskLogFilePath) throws RemotingException, InterruptedException { logger.info("Begin to get appIds from worker: {}:{} taskLogPath: {}", host, port, taskLogFilePath); final Host workerAddress = new Host(host, port); List appIds = null; try { if (NetUtils.getHost().equals(host)) { - appIds = LogUtils.getAppIdsFromLogFile(taskLogFilePath); + appIds = new ArrayList<>(LogUtils.getAppIdsFromLogFile(taskLogFilePath)); } else { final Command command = new GetAppIdRequestCommand(taskLogFilePath).convert2Command(); Command response = this.client.sendSync(workerAddress, command, LOG_REQUEST_TIMEOUT); if (response != null) { - GetAppIdResponseCommand responseCommand = JSONUtils.parseObject(response.getBody(), GetAppIdResponseCommand.class); + GetAppIdResponseCommand responseCommand = + JSONUtils.parseObject(response.getBody(), GetAppIdResponseCommand.class); appIds = responseCommand.getAppIds(); } } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractRemoteTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractRemoteTask.java new file mode 100644 index 0000000000..9ee702fd20 --- /dev/null +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractRemoteTask.java @@ -0,0 +1,86 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.dolphinscheduler.plugin.task.api; + +import org.apache.dolphinscheduler.plugin.task.api.model.ApplicationInfo; +import org.apache.dolphinscheduler.spi.utils.StringUtils; + +import java.util.Set; + +public abstract class AbstractRemoteTask extends AbstractTask { + + /** + * constructor + * + * @param taskExecutionContext taskExecutionContext + */ + protected AbstractRemoteTask(TaskExecutionContext taskExecutionContext) { + super(taskExecutionContext); + } + + @Override + public void cancel() throws TaskException { + this.cancelApplication(); + } + + public abstract Set getApplicationIds() throws TaskException; + + public abstract void cancelApplication() throws TaskException; + + /** + * If appIds is empty, submit a new remote application; otherwise, just track application status. + * + * @param taskCallBack + * @throws TaskException + */ + @Override + public void handle(TaskCallBack taskCallBack) throws TaskException { + // if appIds is not empty, just track application status, avoid resubmitting remote task + if (StringUtils.isNotEmpty(taskRequest.getAppIds())) { + setAppIds(taskRequest.getAppIds()); + trackApplicationStatus(); + return; + } + + // submit a remote application + submitApplication(); + + if (StringUtils.isNotEmpty(getAppIds())) { + taskRequest.setAppIds(getAppIds()); + // callback to update remote application info + taskCallBack.updateRemoteApplicationInfo(taskRequest.getTaskInstanceId(), new ApplicationInfo(getAppIds())); + } + + // keep tracking application status + trackApplicationStatus(); + } + + /** + * submit a new remote application and get application info + * + * @return + * @throws TaskException + */ + public abstract void submitApplication() throws TaskException; + + /** + * keep checking application status + * @throws TaskException + */ + public abstract void trackApplicationStatus() throws TaskException; +} diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractTask.java index 0b23905059..47cc7d9c75 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractTask.java @@ -18,6 +18,7 @@ package org.apache.dolphinscheduler.plugin.task.api; import org.apache.dolphinscheduler.plugin.task.api.enums.TaskExecutionStatus; +import org.apache.dolphinscheduler.plugin.task.api.model.Property; import org.apache.dolphinscheduler.plugin.task.api.model.TaskAlertInfo; import org.apache.dolphinscheduler.plugin.task.api.parameters.AbstractParameters; import org.apache.dolphinscheduler.spi.utils.StringUtils; @@ -29,19 +30,28 @@ import java.io.IOException; import java.io.InputStreamReader; import java.nio.charset.StandardCharsets; import java.util.HashSet; +import java.util.Map; import java.util.Set; +import java.util.StringJoiner; +import java.util.concurrent.LinkedBlockingQueue; import java.util.regex.Matcher; import java.util.regex.Pattern; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; +import org.slf4j.Marker; +import org.slf4j.MarkerFactory; + /** * executive task */ public abstract class AbstractTask { - /** - * rules for extracting application ID - */ - protected static final Pattern YARN_APPLICATION_REGEX = Pattern.compile(TaskConstants.YARN_APPLICATION_REGEX); + public static final Marker FINALIZE_SESSION_MARKER = MarkerFactory.getMarker("FINALIZE_SESSION"); + + protected final Logger logger = LoggerFactory.getLogger(String.format(TaskConstants.TASK_LOG_LOGGER_NAME_FORMAT, getClass())); + + public String rgex = "['\"]*\\$\\{(.*?)\\}['\"]*"; /** * varPool string @@ -68,11 +78,6 @@ public abstract class AbstractTask { */ protected String appIds; - /** - * cancel flag - */ - protected volatile boolean cancel = false; - /** * exit code */ @@ -101,60 +106,9 @@ public abstract class AbstractTask { return null; } - public abstract void handle() throws TaskException; - - - /** - * cancel application - * - * @param status status - * @throws Exception exception - */ - public void cancelApplication(boolean status) throws Exception { - this.cancel = status; - } - - /** - * get application ids - * @return - * @throws IOException - */ - public Set getApplicationIds() throws IOException { - Set appIds = new HashSet<>(); - - File file = new File(taskRequest.getLogPath()); - if (!file.exists()) { - return appIds; - } + public abstract void handle(TaskCallBack taskCallBack) throws TaskException; - /* - * analysis log? get submitted yarn application id - */ - try (BufferedReader br = new BufferedReader(new InputStreamReader(new FileInputStream(taskRequest.getLogPath()), StandardCharsets.UTF_8))) { - String line; - while ((line = br.readLine()) != null) { - String appId = findAppId(line); - if (StringUtils.isNotEmpty(appId)) { - appIds.add(appId); - } - } - } - return appIds; - } - - /** - * find app id - * - * @param line line - * @return appid - */ - protected String findAppId(String line) { - Matcher matcher = YARN_APPLICATION_REGEX.matcher(line); - if (matcher.find()) { - return matcher.group(); - } - return null; - } + public abstract void cancel() throws TaskException; public void setVarPool(String varPool) { this.varPool = varPool; @@ -177,14 +131,6 @@ public abstract class AbstractTask { this.exitStatusCode = exitStatusCode; } - public String getAppIds() { - return appIds; - } - - public void setAppIds(String appIds) { - this.appIds = appIds; - } - public int getProcessId() { return processId; } @@ -201,6 +147,14 @@ public abstract class AbstractTask { this.resultString = resultString; } + public String getAppIds() { + return appIds; + } + + public void setAppIds(String appIds) { + this.appIds = appIds; + } + public boolean getNeedAlert() { return needAlert; } @@ -245,4 +199,51 @@ public abstract class AbstractTask { return status; } + /** + * log handle + * + * @param logs log list + */ + public void logHandle(LinkedBlockingQueue logs) { + // note that the "new line" is added here to facilitate log parsing + if (logs.contains(FINALIZE_SESSION_MARKER.toString())) { + logger.info(FINALIZE_SESSION_MARKER, FINALIZE_SESSION_MARKER.toString()); + } else { + StringJoiner joiner = new StringJoiner("\n\t"); + while (!logs.isEmpty()) { + joiner.add(logs.poll()); + } + logger.info(" -> {}", joiner); + } + } + + /** + * regular expressions match the contents between two specified strings + * + * @param content content + * @param rgex rgex + * @param sqlParamsMap sql params map + * @param paramsPropsMap params props map + */ + public void setSqlParamsMap(String content, String rgex, Map sqlParamsMap, + Map paramsPropsMap,int taskInstanceId) { + Pattern pattern = Pattern.compile(rgex); + Matcher m = pattern.matcher(content); + int index = 1; + while (m.find()) { + + String paramName = m.group(1); + Property prop = paramsPropsMap.get(paramName); + + if (prop == null) { + logger.error("setSqlParamsMap: No Property with paramName: {} is found in paramsPropsMap of task instance" + + " with id: {}. So couldn't put Property in sqlParamsMap.", paramName, taskInstanceId); + } else { + sqlParamsMap.put(index, prop); + index++; + logger.info("setSqlParamsMap: Property with paramName: {} put in sqlParamsMap of content {} successfully.", paramName, content); + } + + } + } } \ No newline at end of file diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractTaskExecutor.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractTaskExecutor.java deleted file mode 100644 index 6806d6e2b9..0000000000 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractTaskExecutor.java +++ /dev/null @@ -1,97 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one or more - * contributor license agreements. See the NOTICE file distributed with - * this work for additional information regarding copyright ownership. - * The ASF licenses this file to You under the Apache License, Version 2.0 - * (the "License"); you may not use this file except in compliance with - * the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.dolphinscheduler.plugin.task.api; - -import org.apache.dolphinscheduler.plugin.task.api.model.Property; - -import java.util.Map; -import java.util.StringJoiner; -import java.util.concurrent.LinkedBlockingQueue; -import java.util.regex.Matcher; -import java.util.regex.Pattern; - -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; -import org.slf4j.Marker; -import org.slf4j.MarkerFactory; - -public abstract class AbstractTaskExecutor extends AbstractTask { - - public static final Marker FINALIZE_SESSION_MARKER = MarkerFactory.getMarker("FINALIZE_SESSION"); - - protected final Logger logger = LoggerFactory.getLogger(String.format(TaskConstants.TASK_LOG_LOGGER_NAME_FORMAT, getClass())); - - public String rgex = "['\"]*\\$\\{(.*?)\\}['\"]*"; - /** - * constructor - * - * @param taskRequest taskRequest - */ - protected AbstractTaskExecutor(TaskExecutionContext taskRequest) { - super(taskRequest); - } - - /** - * log handle - * - * @param logs log list - */ - public void logHandle(LinkedBlockingQueue logs) { - // note that the "new line" is added here to facilitate log parsing - if (logs.contains(FINALIZE_SESSION_MARKER.toString())) { - logger.info(FINALIZE_SESSION_MARKER, FINALIZE_SESSION_MARKER.toString()); - } else { - StringJoiner joiner = new StringJoiner("\n\t"); - while (!logs.isEmpty()) { - joiner.add(logs.poll()); - } - logger.info(" -> {}", joiner); - } - } - - /** - * regular expressions match the contents between two specified strings - * - * @param content content - * @param rgex rgex - * @param sqlParamsMap sql params map - * @param paramsPropsMap params props map - */ - public void setSqlParamsMap(String content, String rgex, Map sqlParamsMap, - Map paramsPropsMap,int taskInstanceId) { - Pattern pattern = Pattern.compile(rgex); - Matcher m = pattern.matcher(content); - int index = 1; - while (m.find()) { - - String paramName = m.group(1); - Property prop = paramsPropsMap.get(paramName); - - if (prop == null) { - logger.error("setSqlParamsMap: No Property with paramName: {} is found in paramsPropsMap of task instance" - + " with id: {}. So couldn't put Property in sqlParamsMap.", paramName, taskInstanceId); - } else { - sqlParamsMap.put(index, prop); - index++; - logger.info("setSqlParamsMap: Property with paramName: {} put in sqlParamsMap of content {} successfully.", paramName, content); - } - - } - } - -} diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractYarnTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractYarnTask.java index e354360d80..e51429692f 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractYarnTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/AbstractYarnTask.java @@ -19,13 +19,17 @@ package org.apache.dolphinscheduler.plugin.task.api; import org.apache.dolphinscheduler.plugin.task.api.model.ResourceInfo; import org.apache.dolphinscheduler.plugin.task.api.model.TaskResponse; +import org.apache.dolphinscheduler.plugin.task.api.utils.LogUtils; +import java.util.Set; +import java.util.regex.Matcher; import java.util.regex.Pattern; /** * abstract yarn task */ -public abstract class AbstractYarnTask extends AbstractTaskExecutor { +public abstract class AbstractYarnTask extends AbstractRemoteTask { + /** * process task */ @@ -44,12 +48,13 @@ public abstract class AbstractYarnTask extends AbstractTaskExecutor { public AbstractYarnTask(TaskExecutionContext taskRequest) { super(taskRequest); this.shellCommandExecutor = new ShellCommandExecutor(this::logHandle, - taskRequest, - logger); + taskRequest, + logger); } + // todo split handle to submit and track @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { // SHELL task exit code TaskResponse response = shellCommandExecutor.run(buildCommand()); @@ -69,17 +74,54 @@ public abstract class AbstractYarnTask extends AbstractTaskExecutor { } } + // todo + @Override + public void submitApplication() throws TaskException { + + } + + // todo + @Override + public void trackApplicationStatus() throws TaskException { + + } + /** * cancel application * - * @param status status - * @throws Exception exception + * @throws TaskException exception */ @Override - public void cancelApplication(boolean status) throws Exception { - cancel = true; + public void cancelApplication() throws TaskException { // cancel process - shellCommandExecutor.cancelApplication(); + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } + } + + /** + * get application ids + * @return + * @throws TaskException + */ + public Set getApplicationIds() throws TaskException { + return LogUtils.getAppIdsFromLogFile(taskRequest.getLogPath(), logger); + } + + /** + * find app id + * + * @param line line + * @return appid + */ + protected String findAppId(String line) { + Matcher matcher = YARN_APPLICATION_REGEX.matcher(line); + if (matcher.find()) { + return matcher.group(); + } + return null; } /** @@ -106,8 +148,8 @@ public abstract class AbstractYarnTask extends AbstractTaskExecutor { } return mainJar.getId() == null - ? mainJar.getRes() - // when update resource maybe has error - : mainJar.getResourceName().replaceFirst("/", ""); + ? mainJar.getRes() + // when update resource maybe has error + : mainJar.getResourceName().replaceFirst("/", ""); } } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/TaskCallBack.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/TaskCallBack.java new file mode 100644 index 0000000000..71643d56cd --- /dev/null +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/TaskCallBack.java @@ -0,0 +1,25 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.dolphinscheduler.plugin.task.api; + +import org.apache.dolphinscheduler.plugin.task.api.model.ApplicationInfo; + +public interface TaskCallBack { + + public void updateRemoteApplicationInfo(int taskInstanceId, ApplicationInfo applicationInfo); +} diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/enums/TaskExecutionStatus.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/enums/TaskExecutionStatus.java index a8d297cb0d..d4e9d6d01b 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/enums/TaskExecutionStatus.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/enums/TaskExecutionStatus.java @@ -17,11 +17,11 @@ package org.apache.dolphinscheduler.plugin.task.api.enums; -import com.baomidou.mybatisplus.annotation.EnumValue; - import java.util.HashMap; import java.util.Map; +import com.baomidou.mybatisplus.annotation.EnumValue; + public enum TaskExecutionStatus { SUBMITTED_SUCCESS(0, "submit success"), @@ -80,7 +80,7 @@ public enum TaskExecutionStatus { } public boolean isFailure() { - return this == TaskExecutionStatus.FAILURE; + return this == TaskExecutionStatus.FAILURE || this == NEED_FAULT_TOLERANCE; } public boolean isPause() { diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/k8s/AbstractK8sTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/k8s/AbstractK8sTask.java index a82e5f6bbd..4567076405 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/k8s/AbstractK8sTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/k8s/AbstractK8sTask.java @@ -17,13 +17,15 @@ package org.apache.dolphinscheduler.plugin.task.api.k8s; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.plugin.task.api.k8s.impl.K8sTaskExecutor; import org.apache.dolphinscheduler.plugin.task.api.model.TaskResponse; -public abstract class AbstractK8sTask extends AbstractTaskExecutor { +public abstract class AbstractK8sTask extends AbstractRemoteTask { /** * process task */ @@ -38,8 +40,9 @@ public abstract class AbstractK8sTask extends AbstractTaskExecutor { this.abstractK8sTaskExecutor = new K8sTaskExecutor(logger,taskRequest); } + // todo split handle to submit and track @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { TaskResponse response = abstractK8sTaskExecutor.run(buildCommand()); setExitStatusCode(response.getExitStatusCode()); @@ -50,15 +53,25 @@ public abstract class AbstractK8sTask extends AbstractTaskExecutor { } } + // todo + @Override + public void submitApplication() throws TaskException { + + } + + // todo + @Override + public void trackApplicationStatus() throws TaskException { + + } + /** * cancel application * - * @param status status * @throws Exception exception */ @Override - public void cancelApplication(boolean status) throws Exception { - cancel = true; + public void cancelApplication() throws TaskException { // cancel process abstractK8sTaskExecutor.cancelApplication(buildCommand()); } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/loop/BaseLoopTaskExecutor.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/loop/BaseLoopTaskExecutor.java index 3c4857f2d0..36fcef97c3 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/loop/BaseLoopTaskExecutor.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/loop/BaseLoopTaskExecutor.java @@ -17,10 +17,12 @@ package org.apache.dolphinscheduler.plugin.task.api.loop; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; +import org.apache.dolphinscheduler.plugin.task.api.model.ApplicationInfo; import org.apache.dolphinscheduler.plugin.task.api.utils.RetryUtils; import java.time.Duration; @@ -34,7 +36,12 @@ import lombok.NonNull; *

* The loop task type means, we will submit a task, and loop the task status until the task is finished. */ -public abstract class BaseLoopTaskExecutor extends AbstractTaskExecutor { +public abstract class BaseLoopTaskExecutor extends AbstractRemoteTask { + + /** + * cancel flag + */ + protected volatile boolean cancel = false; /** * The task instance info will be set when task has submitted successful. @@ -46,11 +53,13 @@ public abstract class BaseLoopTaskExecutor extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { final long loopInterval = getTaskInstanceStatusQueryInterval().toMillis(); loopTaskInstanceInfo = submitLoopTask(); - this.appIds = loopTaskInstanceInfo.getTaskInstanceId(); + this.setAppIds(loopTaskInstanceInfo.getTaskInstanceId()); + taskCallBack.updateRemoteApplicationInfo(taskRequest.getTaskInstanceId(), new ApplicationInfo(getAppIds())); + // loop the task status until the task is finished or task has been canceled. // we use retry utils here to avoid the task status query failure due to network failure. // the default retry policy is 3 times, and the interval is 1 second. @@ -94,8 +103,7 @@ public abstract class BaseLoopTaskExecutor extends AbstractTaskExecutor { /** * Query the loop task status, if query failed, directly throw exception */ - public abstract @NonNull LoopTaskInstanceStatus queryTaskInstanceStatus(@NonNull LoopTaskInstanceInfo taskInstanceInfo) - throws TaskException; + public abstract @NonNull LoopTaskInstanceStatus queryTaskInstanceStatus(@NonNull LoopTaskInstanceInfo taskInstanceInfo) throws TaskException; /** * Get the interval time to query the loop task status @@ -110,8 +118,8 @@ public abstract class BaseLoopTaskExecutor extends AbstractTaskExecutor { public abstract void cancelLoopTaskInstance(@Nullable LoopTaskInstanceInfo taskInstanceInfo) throws TaskException; @Override - public void cancelApplication(boolean status) throws Exception { + public void cancelApplication() throws TaskException { + this.cancel = true; cancelLoopTaskInstance(loopTaskInstanceInfo); - super.cancelApplication(status); } } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/model/ApplicationInfo.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/model/ApplicationInfo.java new file mode 100644 index 0000000000..4320cc40b0 --- /dev/null +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/model/ApplicationInfo.java @@ -0,0 +1,30 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.dolphinscheduler.plugin.task.api.model; + +import lombok.AllArgsConstructor; +import lombok.Data; +import lombok.NoArgsConstructor; + +@Data +@AllArgsConstructor +@NoArgsConstructor +public class ApplicationInfo { + + String appIds; +} diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/stream/StreamTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/stream/StreamTask.java index d554278516..51ab17e156 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/stream/StreamTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/stream/StreamTask.java @@ -17,8 +17,6 @@ package org.apache.dolphinscheduler.plugin.task.api.stream; -import java.io.IOException; - public interface StreamTask { public void savePoint() throws Exception; } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/utils/LogUtils.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/utils/LogUtils.java index 220d2583d2..0df5c3a911 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/utils/LogUtils.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/main/java/org/apache/dolphinscheduler/plugin/task/api/utils/LogUtils.java @@ -17,47 +17,46 @@ package org.apache.dolphinscheduler.plugin.task.api.utils; -import lombok.NonNull; -import lombok.experimental.UtilityClass; -import lombok.extern.slf4j.Slf4j; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; -import org.slf4j.Logger; import java.io.File; import java.io.IOException; import java.nio.file.Files; import java.nio.file.Paths; -import java.util.ArrayList; import java.util.Collections; import java.util.HashSet; -import java.util.List; import java.util.Set; import java.util.regex.Matcher; import java.util.regex.Pattern; import java.util.stream.Stream; +import lombok.NonNull; +import lombok.experimental.UtilityClass; +import lombok.extern.slf4j.Slf4j; + +import org.slf4j.Logger; + @Slf4j @UtilityClass public class LogUtils { private static final Pattern APPLICATION_REGEX = Pattern.compile(TaskConstants.YARN_APPLICATION_REGEX); - public List getAppIdsFromLogFile(@NonNull String logPath) { + public Set getAppIdsFromLogFile(@NonNull String logPath) { return getAppIdsFromLogFile(logPath, log); } - public List getAppIdsFromLogFile(@NonNull String logPath, Logger logger) { + public Set getAppIdsFromLogFile(@NonNull String logPath, Logger logger) { File logFile = new File(logPath); if (!logFile.exists() || !logFile.isFile()) { - return Collections.emptyList(); + return Collections.emptySet(); } Set appIds = new HashSet<>(); try (Stream stream = Files.lines(Paths.get(logPath))) { stream.filter(line -> { - Matcher matcher = APPLICATION_REGEX.matcher(line); - return matcher.find(); - } - ).forEach(line -> { + Matcher matcher = APPLICATION_REGEX.matcher(line); + return matcher.find(); + }).forEach(line -> { Matcher matcher = APPLICATION_REGEX.matcher(line); if (matcher.find()) { String appId = matcher.group(); @@ -66,10 +65,10 @@ public class LogUtils { } } }); - return new ArrayList<>(appIds); + return appIds; } catch (IOException e) { logger.error("Get appId from log file erro, logPath: {}", logPath, e); - return Collections.emptyList(); + return Collections.emptySet(); } } } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/test/java/org/apache/dolphinscheduler/plugin/task/api/utils/LogUtilsTest.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/test/java/org/apache/dolphinscheduler/plugin/task/api/utils/LogUtilsTest.java index feb10ad859..bd961d3137 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/test/java/org/apache/dolphinscheduler/plugin/task/api/utils/LogUtilsTest.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-api/src/test/java/org/apache/dolphinscheduler/plugin/task/api/utils/LogUtilsTest.java @@ -17,11 +17,12 @@ package org.apache.dolphinscheduler.plugin.task.api.utils; -import com.google.common.collect.Lists; +import java.util.Set; + import org.junit.Assert; import org.junit.Test; -import java.util.List; +import com.google.common.collect.Sets; public class LogUtilsTest { @@ -30,7 +31,7 @@ public class LogUtilsTest { @Test public void getAppIdsFromLogFile() { - List appIds = LogUtils.getAppIdsFromLogFile(APP_ID_FILE); - Assert.assertEquals(Lists.newArrayList("application_1548381669007_1234"), appIds); + Set appIds = LogUtils.getAppIdsFromLogFile(APP_ID_FILE); + Assert.assertEquals(Sets.newHashSet("application_1548381669007_1234"), appIds); } -} \ No newline at end of file +} diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-chunjun/src/main/java/org/apache/dolphinscheduler/plugin/task/chunjun/ChunJunTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-chunjun/src/main/java/org/apache/dolphinscheduler/plugin/task/chunjun/ChunJunTask.java index 48914ea2e1..a690dcd383 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-chunjun/src/main/java/org/apache/dolphinscheduler/plugin/task/chunjun/ChunJunTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-chunjun/src/main/java/org/apache/dolphinscheduler/plugin/task/chunjun/ChunJunTask.java @@ -18,9 +18,11 @@ package org.apache.dolphinscheduler.plugin.task.chunjun; import org.apache.commons.io.FileUtils; -import org.apache.commons.lang3.SystemUtils; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.commons.lang.SystemUtils; + +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -41,6 +43,7 @@ import java.nio.file.attribute.FileAttribute; import java.nio.file.attribute.PosixFilePermission; import java.nio.file.attribute.PosixFilePermissions; import java.util.ArrayList; +import java.util.Collections; import java.util.List; import java.util.Map; import java.util.Set; @@ -51,7 +54,7 @@ import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.RWXR_XR_ /** * chunjun task */ -public class ChunJunTask extends AbstractTaskExecutor { +public class ChunJunTask extends AbstractTask { /** * chunjun path */ @@ -104,7 +107,7 @@ public class ChunJunTask extends AbstractTaskExecutor { * @throws TaskException exception */ @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { Map paramsMap = taskExecutionContext.getPrepareParamsMap(); @@ -113,7 +116,9 @@ public class ChunJunTask extends AbstractTaskExecutor { TaskResponse commandExecuteResult = shellCommandExecutor.run(shellCommandFilePath); setExitStatusCode(commandExecuteResult.getExitStatusCode()); - setAppIds(String.join(TaskConstants.COMMA, getApplicationIds())); + + // todo get applicationId + setAppIds(String.join(TaskConstants.COMMA, Collections.emptySet())); setProcessId(commandExecuteResult.getProcessId()); } catch (InterruptedException e) { Thread.currentThread().interrupt(); @@ -251,13 +256,16 @@ public class ChunJunTask extends AbstractTaskExecutor { /** * cancel ChunJun process * - * @param cancelApplication cancelApplication * @throws Exception if error throws Exception */ @Override - public void cancelApplication(boolean cancelApplication) throws Exception { + public void cancel() throws TaskException { // cancel process - shellCommandExecutor.cancelApplication(); + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } } } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-datax/src/main/java/org/apache/dolphinscheduler/plugin/task/datax/DataxTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-datax/src/main/java/org/apache/dolphinscheduler/plugin/task/datax/DataxTask.java index f36436176f..bd3bf236d8 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-datax/src/main/java/org/apache/dolphinscheduler/plugin/task/datax/DataxTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-datax/src/main/java/org/apache/dolphinscheduler/plugin/task/datax/DataxTask.java @@ -33,8 +33,9 @@ import org.apache.commons.io.FileUtils; import org.apache.commons.lang3.SystemUtils; import org.apache.dolphinscheduler.plugin.datasource.api.plugin.DataSourceClientProvider; import org.apache.dolphinscheduler.plugin.datasource.api.utils.DataSourceUtils; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -75,7 +76,7 @@ import static org.apache.dolphinscheduler.plugin.datasource.api.utils.PasswordUt import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.EXIT_CODE_FAILURE; import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.RWXR_XR_X; -public class DataxTask extends AbstractTaskExecutor { +public class DataxTask extends AbstractTask { /** * jvm parameters */ @@ -147,7 +148,7 @@ public class DataxTask extends AbstractTaskExecutor { * @throws Exception if error throws Exception */ @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { // replace placeholder,and combine local and global parameters Map paramsMap = taskExecutionContext.getPrepareParamsMap(); @@ -158,7 +159,6 @@ public class DataxTask extends AbstractTaskExecutor { TaskResponse commandExecuteResult = shellCommandExecutor.run(shellCommandFilePath); setExitStatusCode(commandExecuteResult.getExitStatusCode()); - setAppIds(String.join(TaskConstants.COMMA, getApplicationIds())); setProcessId(commandExecuteResult.getProcessId()); } catch (InterruptedException e) { Thread.currentThread().interrupt(); @@ -175,14 +175,16 @@ public class DataxTask extends AbstractTaskExecutor { /** * cancel DataX process * - * @param cancelApplication cancelApplication - * @throws Exception if error throws Exception + * @throws TaskException if error throws Exception */ @Override - public void cancelApplication(boolean cancelApplication) - throws Exception { + public void cancel() throws TaskException { // cancel process - shellCommandExecutor.cancelApplication(); + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } } /** diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-dinky/src/main/java/org/apache/dolphinscheduler/plugin/task/dinky/DinkyTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-dinky/src/main/java/org/apache/dolphinscheduler/plugin/task/dinky/DinkyTask.java index 987126b0a7..51751fe10d 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-dinky/src/main/java/org/apache/dolphinscheduler/plugin/task/dinky/DinkyTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-dinky/src/main/java/org/apache/dolphinscheduler/plugin/task/dinky/DinkyTask.java @@ -21,7 +21,10 @@ import com.fasterxml.jackson.core.JsonProcessingException; import com.fasterxml.jackson.databind.JsonNode; import com.fasterxml.jackson.databind.ObjectMapper; import com.fasterxml.jackson.databind.node.MissingNode; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; + +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -40,12 +43,14 @@ import org.apache.http.util.EntityUtils; import java.net.URI; import java.nio.charset.StandardCharsets; +import java.util.Collections; import java.util.HashMap; import java.util.Map; +import java.util.Set; import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.EXIT_CODE_FAILURE; -public class DinkyTask extends AbstractTaskExecutor { +public class DinkyTask extends AbstractRemoteTask { /** * taskExecutionContext @@ -67,6 +72,11 @@ public class DinkyTask extends AbstractTaskExecutor { this.taskExecutionContext = taskExecutionContext; } + @Override + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); + } + @Override public void init() { final String taskParams = taskExecutionContext.getTaskParams(); @@ -77,8 +87,9 @@ public class DinkyTask extends AbstractTaskExecutor { } } + // todo split handle to submit and track @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { String address = this.dinkyParameters.getAddress(); @@ -130,6 +141,16 @@ public class DinkyTask extends AbstractTaskExecutor { } } + @Override + public void submitApplication() throws TaskException { + + } + + @Override + public void trackApplicationStatus() throws TaskException { + + } + /** * map dinky task status to exitStatusCode * @@ -166,8 +187,7 @@ public class DinkyTask extends AbstractTaskExecutor { } @Override - public void cancelApplication(boolean status) throws Exception { - super.cancelApplication(status); + public void cancelApplication() throws TaskException { String address = this.dinkyParameters.getAddress(); String taskId = this.dinkyParameters.getTaskId(); logger.info("trying terminate dinky task, taskId: {}, address: {}, taskId: {}", diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-dvc/src/main/java/org/apache/dolphinscheduler/plugin/task/dvc/DvcTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-dvc/src/main/java/org/apache/dolphinscheduler/plugin/task/dvc/DvcTask.java index 2e3bece309..eb08e55f1a 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-dvc/src/main/java/org/apache/dolphinscheduler/plugin/task/dvc/DvcTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-dvc/src/main/java/org/apache/dolphinscheduler/plugin/task/dvc/DvcTask.java @@ -17,9 +17,11 @@ package org.apache.dolphinscheduler.plugin.task.dvc; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.EXIT_CODE_FAILURE; + +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; -import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.plugin.task.api.model.TaskResponse; @@ -29,12 +31,10 @@ import org.apache.dolphinscheduler.spi.utils.JSONUtils; import java.util.ArrayList; import java.util.List; -import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.EXIT_CODE_FAILURE; - /** * shell task */ -public class DvcTask extends AbstractTaskExecutor { +public class DvcTask extends AbstractTask { /** * dvc parameters @@ -75,13 +75,12 @@ public class DvcTask extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { // construct process String command = buildCommand(); TaskResponse commandExecuteResult = shellCommandExecutor.run(command); setExitStatusCode(commandExecuteResult.getExitStatusCode()); - setAppIds(String.join(TaskConstants.COMMA, getApplicationIds())); setProcessId(commandExecuteResult.getProcessId()); parameters.dealOutParam(shellCommandExecutor.getVarPool()); } catch (InterruptedException e) { @@ -97,9 +96,13 @@ public class DvcTask extends AbstractTaskExecutor { } @Override - public void cancelApplication(boolean cancelApplication) throws Exception { + public void cancel() throws TaskException { // cancel process - shellCommandExecutor.cancelApplication(); + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } } public String buildCommand() { @@ -107,9 +110,9 @@ public class DvcTask extends AbstractTaskExecutor { TaskTypeEnum taskType = parameters.getDvcTaskType(); if (taskType == TaskTypeEnum.UPLOAD) { command = buildUploadCommond(); - }else if (taskType == TaskTypeEnum.DOWNLOAD){ + } else if (taskType == TaskTypeEnum.DOWNLOAD) { command = buildDownCommond(); - }else if (taskType == TaskTypeEnum.INIT){ + } else if (taskType == TaskTypeEnum.INIT) { command = buildInitDvcCommond(); } logger.info("Run DVC task with command: \n{}", command); @@ -159,12 +162,9 @@ public class DvcTask extends AbstractTaskExecutor { } - @Override public AbstractParameters getParameters() { return parameters; } - } - diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/AbstractEmrTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/AbstractEmrTask.java index 329c3bc3d5..07ff171ba1 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/AbstractEmrTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/AbstractEmrTask.java @@ -22,7 +22,8 @@ import static com.fasterxml.jackson.databind.DeserializationFeature.FAIL_ON_UNKN import static com.fasterxml.jackson.databind.DeserializationFeature.READ_UNKNOWN_ENUM_VALUES_AS_NULL; import static com.fasterxml.jackson.databind.MapperFeature.REQUIRE_SETTERS_FOR_GETTERS; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.plugin.task.api.parameters.AbstractParameters; @@ -44,7 +45,7 @@ import com.fasterxml.jackson.databind.PropertyNamingStrategy; * * @since v3.1.0 */ -public abstract class AbstractEmrTask extends AbstractTaskExecutor { +public abstract class AbstractEmrTask extends AbstractRemoteTask { final TaskExecutionContext taskExecutionContext; EmrParameters emrParameters; diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/EmrAddStepsTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/EmrAddStepsTask.java index 6f4fc95e48..a746185805 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/EmrAddStepsTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/EmrAddStepsTask.java @@ -30,11 +30,15 @@ import com.amazonaws.services.elasticmapreduce.model.StepState; import com.amazonaws.services.elasticmapreduce.model.StepStatus; import com.fasterxml.jackson.core.JsonProcessingException; import com.google.common.collect.Sets; + +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; +import java.util.Collections; import java.util.HashSet; +import java.util.Set; import java.util.concurrent.TimeUnit; /** @@ -62,7 +66,12 @@ public class EmrAddStepsTask extends AbstractEmrTask { } @Override - public void handle() throws TaskException { + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); + } + + @Override + public void submitApplication() throws TaskException { StepStatus stepStatus = null; try { AddJobFlowStepsRequest addJobFlowStepsRequest = createAddJobFlowStepsRequest(); @@ -77,13 +86,27 @@ public class EmrAddStepsTask extends AbstractEmrTask { stepStatus = getStepStatus(); + } catch (EmrTaskException | SdkBaseException e) { + logger.error("emr task submit failed with error", e); + throw new TaskException("emr task submit fail", e); + } finally { + final int exitStatusCode = calculateExitStatusCode(stepStatus); + setExitStatusCode(exitStatusCode); + logger.info("emr task finished with step status : {}", stepStatus); + } + } + + @Override + public void trackApplicationStatus() throws TaskException { + StepStatus stepStatus = getStepStatus(); + + try { while (waitingStateSet.contains(stepStatus.getState())) { TimeUnit.SECONDS.sleep(10); stepStatus = getStepStatus(); } - } catch (EmrTaskException | SdkBaseException e) { - logger.error("emr task submit failed with error", e); + logger.error("emr task failed with error", e); } catch (InterruptedException e) { Thread.currentThread().interrupt(); throw new TaskException("Execute emr task failed", e); @@ -154,8 +177,7 @@ public class EmrAddStepsTask extends AbstractEmrTask { } @Override - public void cancelApplication(boolean status) throws Exception { - super.cancelApplication(status); + public void cancelApplication() throws TaskException { logger.info("trying cancel emr step, taskId:{}, clusterId:{}, stepId:{}", this.taskExecutionContext.getTaskInstanceId(), clusterId, stepId); CancelStepsRequest cancelStepsRequest = new CancelStepsRequest().withClusterId(clusterId).withStepIds(stepId); CancelStepsResult cancelStepsResult = emrClient.cancelSteps(cancelStepsRequest); diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/EmrJobFlowTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/EmrJobFlowTask.java index bc361f6538..770fb9f996 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/EmrJobFlowTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/main/java/org/apache/dolphinscheduler/plugin/task/emr/EmrJobFlowTask.java @@ -21,7 +21,9 @@ import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; +import java.util.Collections; import java.util.HashSet; +import java.util.Set; import java.util.concurrent.TimeUnit; import com.amazonaws.SdkBaseException; @@ -41,10 +43,9 @@ import com.google.common.collect.Sets; public class EmrJobFlowTask extends AbstractEmrTask { private final HashSet waitingStateSet = Sets.newHashSet( - ClusterState.STARTING.toString(), - ClusterState.BOOTSTRAPPING.toString(), - ClusterState.RUNNING.toString() - ); + ClusterState.STARTING.toString(), + ClusterState.BOOTSTRAPPING.toString(), + ClusterState.RUNNING.toString()); /** * constructor @@ -56,7 +57,12 @@ public class EmrJobFlowTask extends AbstractEmrTask { } @Override - public void handle() throws TaskException { + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); + } + + @Override + public void submitApplication() throws TaskException { ClusterStatus clusterStatus = null; try { RunJobFlowRequest runJobFlowRequest = createRunJobFlowRequest(); @@ -65,18 +71,34 @@ public class EmrJobFlowTask extends AbstractEmrTask { RunJobFlowResult result = emrClient.runJobFlow(runJobFlowRequest); clusterId = result.getJobFlowId(); - // Failover on EMR Task type has not been implemented. In this time, DS only supports failover on yarn task type . Other task type, such as EMR task, k8s task not ready yet. + // Failover on EMR Task type has not been implemented. In this time, DS only supports failover on yarn task + // type . Other task type, such as EMR task, k8s task not ready yet. setAppIds(clusterId); clusterStatus = getClusterStatus(); + } catch (EmrTaskException | SdkBaseException e) { + logger.error("emr task submit failed with error", e); + throw new TaskException("emr task submit failed", e); + } finally { + final int exitStatusCode = calculateExitStatusCode(clusterStatus); + setExitStatusCode(exitStatusCode); + logger.info("emr task finished with cluster status : {}", clusterStatus); + } + } + + @Override + public void trackApplicationStatus() throws TaskException { + ClusterStatus clusterStatus = null; + try { + clusterStatus = getClusterStatus(); + while (waitingStateSet.contains(clusterStatus.getState())) { TimeUnit.SECONDS.sleep(10); clusterStatus = getClusterStatus(); } - } catch (EmrTaskException | SdkBaseException e) { - logger.error("emr task submit failed with error", e); + logger.error("emr task failed with error", e); } catch (InterruptedException e) { Thread.currentThread().interrupt(); throw new TaskException("Execute emr task failed", e); @@ -123,7 +145,8 @@ public class EmrJobFlowTask extends AbstractEmrTask { case TERMINATED: case TERMINATING: String code = stateChangeReason.getCode(); - if (code != null && code.equalsIgnoreCase(ClusterStateChangeReasonCode.ALL_STEPS_COMPLETED.toString())) { + if (code != null + && code.equalsIgnoreCase(ClusterStateChangeReasonCode.ALL_STEPS_COMPLETED.toString())) { return TaskConstants.EXIT_CODE_SUCCESS; } else { return TaskConstants.EXIT_CODE_KILL; @@ -148,9 +171,9 @@ public class EmrJobFlowTask extends AbstractEmrTask { } @Override - public void cancelApplication(boolean status) throws Exception { - super.cancelApplication(status); - logger.info("trying terminate job flow, taskId:{}, clusterId:{}", this.taskExecutionContext.getTaskInstanceId(), clusterId); + public void cancelApplication() throws TaskException { + logger.info("trying terminate job flow, taskId:{}, clusterId:{}", this.taskExecutionContext.getTaskInstanceId(), + clusterId); TerminateJobFlowsRequest terminateJobFlowsRequest = new TerminateJobFlowsRequest().withJobFlowIds(clusterId); TerminateJobFlowsResult terminateJobFlowsResult = emrClient.terminateJobFlows(terminateJobFlowsRequest); logger.info("the result of terminate job flow is:{}", terminateJobFlowsResult); diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/test/java/org/apache/dolphinscheduler/plugin/task/emr/EmrAddStepsTaskTest.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/test/java/org/apache/dolphinscheduler/plugin/task/emr/EmrAddStepsTaskTest.java index d74d36fa82..48aee81648 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/test/java/org/apache/dolphinscheduler/plugin/task/emr/EmrAddStepsTaskTest.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/test/java/org/apache/dolphinscheduler/plugin/task/emr/EmrAddStepsTaskTest.java @@ -28,6 +28,8 @@ import static org.powermock.api.mockito.PowerMockito.mockStatic; import static org.powermock.api.mockito.PowerMockito.spy; import static org.powermock.api.mockito.PowerMockito.when; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; +import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.spi.utils.JSONUtils; @@ -88,6 +90,9 @@ public class EmrAddStepsTaskTest { private EmrAddStepsTask emrAddStepsTask; private AmazonElasticMapReduce emrClient; private Step step; + private TaskCallBack taskCallBack = (taskInstanceId, appIds) -> { + + }; @Before public void before() throws Exception { @@ -116,15 +121,14 @@ public class EmrAddStepsTaskTest { emrAddStepsTask.init(); } - @Test + @Test(expected = TaskException.class) public void testCanNotParseJson() throws Exception { mockStatic(JSONUtils.class); when(emrAddStepsTask, "createAddJobFlowStepsRequest").thenThrow(new EmrTaskException("can not parse AddJobFlowStepsRequest from json", new Exception("error"))); - emrAddStepsTask.handle(); - Assert.assertEquals(EXIT_CODE_FAILURE, emrAddStepsTask.getExitStatusCode()); + emrAddStepsTask.handle(taskCallBack); } - @Test + @Test(expected = TaskException.class) public void testDefineJsonStepNotOne() throws Exception { // mock EmrParameters and EmrAddStepsTask EmrParameters emrParameters = buildErrorEmrTaskParameters(); @@ -134,16 +138,14 @@ public class EmrAddStepsTaskTest { emrAddStepsTask = spy(new EmrAddStepsTask(taskExecutionContext)); doReturn(emrClient).when(emrAddStepsTask, "createEmrClient"); emrAddStepsTask.init(); - emrAddStepsTask.handle(); - - Assert.assertEquals(EXIT_CODE_FAILURE, emrAddStepsTask.getExitStatusCode()); + emrAddStepsTask.handle(taskCallBack); } @Test public void testHandle() throws Exception { when(step.getStatus()).thenReturn(pendingState, runningState, completedState); - emrAddStepsTask.handle(); + emrAddStepsTask.handle(taskCallBack); Assert.assertEquals(EXIT_CODE_SUCCESS, emrAddStepsTask.getExitStatusCode()); } @@ -151,19 +153,18 @@ public class EmrAddStepsTaskTest { public void testHandleUserRequestTerminate() throws Exception { when(step.getStatus()).thenReturn(pendingState, runningState, cancelledState); - emrAddStepsTask.handle(); + emrAddStepsTask.handle(taskCallBack); Assert.assertEquals(EXIT_CODE_KILL, emrAddStepsTask.getExitStatusCode()); } - @Test + @Test(expected = TaskException.class) public void testHandleError() throws Exception { when(step.getStatus()).thenReturn(pendingState, runningState, failedState); - emrAddStepsTask.handle(); + emrAddStepsTask.handle(taskCallBack); Assert.assertEquals(EXIT_CODE_FAILURE, emrAddStepsTask.getExitStatusCode()); when(emrClient.addJobFlowSteps(any())).thenThrow(new AmazonElasticMapReduceException("error"), new EmrTaskException()); - emrAddStepsTask.handle(); - Assert.assertEquals(EXIT_CODE_FAILURE, emrAddStepsTask.getExitStatusCode()); + emrAddStepsTask.handle(taskCallBack); } private EmrParameters buildEmrTaskParameters() { diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/test/java/org/apache/dolphinscheduler/plugin/task/emr/EmrJobFlowTaskTest.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/test/java/org/apache/dolphinscheduler/plugin/task/emr/EmrJobFlowTaskTest.java index 65c6c0c239..71596ea53e 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/test/java/org/apache/dolphinscheduler/plugin/task/emr/EmrJobFlowTaskTest.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-emr/src/test/java/org/apache/dolphinscheduler/plugin/task/emr/EmrJobFlowTaskTest.java @@ -28,6 +28,8 @@ import static org.powermock.api.mockito.PowerMockito.mockStatic; import static org.powermock.api.mockito.PowerMockito.spy; import static org.powermock.api.mockito.PowerMockito.when; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; +import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.spi.utils.JSONUtils; @@ -117,6 +119,9 @@ public class EmrJobFlowTaskTest { private EmrJobFlowTask emrJobFlowTask; private AmazonElasticMapReduce emrClient; private Cluster cluster; + private TaskCallBack taskCallBack = (taskInstanceId, appIds) -> { + + }; @Before public void before() throws Exception { @@ -146,7 +151,7 @@ public class EmrJobFlowTaskTest { when(cluster.getStatus()).thenReturn(startingStatus, softwareConfigStatus, runningStatus, terminatingStatus); - emrJobFlowTask.handle(); + emrJobFlowTask.handle(taskCallBack); Assert.assertEquals(EXIT_CODE_SUCCESS, emrJobFlowTask.getExitStatusCode()); } @@ -155,7 +160,7 @@ public class EmrJobFlowTaskTest { public void testHandleAliveWhenNoSteps() throws Exception { when(cluster.getStatus()).thenReturn(startingStatus, softwareConfigStatus, runningStatus, waitingStatus); - emrJobFlowTask.handle(); + emrJobFlowTask.handle(taskCallBack); Assert.assertEquals(EXIT_CODE_SUCCESS, emrJobFlowTask.getExitStatusCode()); } @@ -163,7 +168,7 @@ public class EmrJobFlowTaskTest { public void testHandleUserRequestTerminate() throws Exception { when(cluster.getStatus()).thenReturn(startingStatus, userRequestTerminateStatus); - emrJobFlowTask.handle(); + emrJobFlowTask.handle(taskCallBack); Assert.assertEquals(EXIT_CODE_KILL, emrJobFlowTask.getExitStatusCode()); } @@ -171,36 +176,28 @@ public class EmrJobFlowTaskTest { public void testHandleTerminatedWithError() throws Exception { when(cluster.getStatus()).thenReturn(startingStatus, softwareConfigStatus, runningStatus, terminatedWithErrorsStatus); - emrJobFlowTask.handle(); + emrJobFlowTask.handle(taskCallBack); Assert.assertEquals(EXIT_CODE_FAILURE, emrJobFlowTask.getExitStatusCode()); } - @Test + @Test(expected = TaskException.class) public void testCanNotParseJson() throws Exception { mockStatic(JSONUtils.class); when(emrJobFlowTask, "createRunJobFlowRequest").thenThrow(new EmrTaskException("can not parse RunJobFlowRequest from json", new Exception("error"))); - emrJobFlowTask.handle(); - Assert.assertEquals(EXIT_CODE_FAILURE, emrJobFlowTask.getExitStatusCode()); + emrJobFlowTask.handle(taskCallBack); } - @Test + @Test(expected = TaskException.class) public void testClusterStatusNull() throws Exception { when(emrClient.describeCluster(any())).thenReturn(null); - - emrJobFlowTask.handle(); - Assert.assertEquals(EXIT_CODE_FAILURE, emrJobFlowTask.getExitStatusCode()); + emrJobFlowTask.handle(taskCallBack); } - @Test + @Test(expected = TaskException.class) public void testRunJobFlowError() throws Exception { - when(emrClient.runJobFlow(any())).thenThrow(new AmazonElasticMapReduceException("error"), new EmrTaskException()); - emrJobFlowTask.handle(); - Assert.assertEquals(EXIT_CODE_FAILURE, emrJobFlowTask.getExitStatusCode()); - emrJobFlowTask.handle(); - Assert.assertEquals(EXIT_CODE_FAILURE, emrJobFlowTask.getExitStatusCode()); - + emrJobFlowTask.handle(taskCallBack); } private String buildEmrTaskParameters() { diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-flink-stream/src/main/java/org/apache/dolphinscheduler/plugin/task/flink/FlinkStreamTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-flink-stream/src/main/java/org/apache/dolphinscheduler/plugin/task/flink/FlinkStreamTask.java index a6915ea182..f6d4e56815 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-flink-stream/src/main/java/org/apache/dolphinscheduler/plugin/task/flink/FlinkStreamTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-flink-stream/src/main/java/org/apache/dolphinscheduler/plugin/task/flink/FlinkStreamTask.java @@ -18,6 +18,7 @@ package org.apache.dolphinscheduler.plugin.task.flink; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; +import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.plugin.task.api.model.ResourceInfo; import org.apache.dolphinscheduler.plugin.task.api.parameters.AbstractParameters; @@ -97,7 +98,7 @@ public class FlinkStreamTask extends FlinkTask implements StreamTask { } @Override - public void cancelApplication(boolean status) throws Exception { + public void cancelApplication() throws TaskException { Set appIds = getApplicationIds(); if (CollectionUtils.isEmpty(appIds)) { logger.error("can not get appId, taskInstanceId:{}", taskExecutionContext.getTaskInstanceId()); @@ -110,8 +111,11 @@ public class FlinkStreamTask extends FlinkTask implements StreamTask { ProcessBuilder processBuilder = new ProcessBuilder(); processBuilder.command(args); - processBuilder.start(); - super.cancelApplication(status); + try { + processBuilder.start(); + } catch (IOException e) { + throw new TaskException("cancel application error", e); + } } @Override diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-flink/src/main/java/org/apache/dolphinscheduler/plugin/task/flink/FlinkTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-flink/src/main/java/org/apache/dolphinscheduler/plugin/task/flink/FlinkTask.java index b7dc41fe01..a342b2aac3 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-flink/src/main/java/org/apache/dolphinscheduler/plugin/task/flink/FlinkTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-flink/src/main/java/org/apache/dolphinscheduler/plugin/task/flink/FlinkTask.java @@ -19,6 +19,7 @@ package org.apache.dolphinscheduler.plugin.task.flink; import org.apache.dolphinscheduler.plugin.task.api.AbstractYarnTask; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; +import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.plugin.task.api.model.Property; import org.apache.dolphinscheduler.plugin.task.api.model.ResourceInfo; @@ -32,6 +33,7 @@ import org.apache.dolphinscheduler.spi.utils.StringUtils; import java.io.BufferedReader; import java.io.File; import java.io.FileInputStream; +import java.io.FileNotFoundException; import java.io.IOException; import java.io.InputStreamReader; import java.nio.charset.StandardCharsets; @@ -112,7 +114,7 @@ public class FlinkTask extends AbstractYarnTask { } @Override - public Set getApplicationIds() throws IOException { + public Set getApplicationIds() throws TaskException { Set appIds = new HashSet<>(); File file = new File(taskRequest.getLogPath()); @@ -131,6 +133,10 @@ public class FlinkTask extends AbstractYarnTask { appIds.add(appId); } } + } catch (FileNotFoundException e) { + throw new TaskException("get application id error, file not found, path:" + taskRequest.getLogPath()); + } catch (IOException e) { + throw new TaskException("get application id error, path:" + taskRequest.getLogPath(), e); } return appIds; } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-hivecli/src/main/java/org/apache/dolphinscheduler/plugin/task/hivecli/HiveCliTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-hivecli/src/main/java/org/apache/dolphinscheduler/plugin/task/hivecli/HiveCliTask.java index 2af1bacb38..06ae0b953e 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-hivecli/src/main/java/org/apache/dolphinscheduler/plugin/task/hivecli/HiveCliTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-hivecli/src/main/java/org/apache/dolphinscheduler/plugin/task/hivecli/HiveCliTask.java @@ -19,8 +19,10 @@ package org.apache.dolphinscheduler.plugin.task.hivecli; import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.EXIT_CODE_FAILURE; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.plugin.task.api.model.Property; @@ -34,10 +36,12 @@ import org.apache.dolphinscheduler.spi.utils.JSONUtils; import org.apache.commons.lang3.StringUtils; import java.util.ArrayList; +import java.util.Collections; import java.util.List; import java.util.Map; +import java.util.Set; -public class HiveCliTask extends AbstractTaskExecutor { +public class HiveCliTask extends AbstractRemoteTask { private HiveCliParameters hiveCliParameters; @@ -54,6 +58,11 @@ public class HiveCliTask extends AbstractTaskExecutor { logger); } + @Override + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); + } + @Override public void init() { logger.info("hiveCli task params {}", taskExecutionContext.getTaskParams()); @@ -65,8 +74,9 @@ public class HiveCliTask extends AbstractTaskExecutor { } } + // todo split handle to submit and track @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { final TaskResponse taskResponse = shellCommandExecutor.run(buildCommand()); setExitStatusCode(taskResponse.getExitStatusCode()); @@ -85,6 +95,16 @@ public class HiveCliTask extends AbstractTaskExecutor { } } + @Override + public void submitApplication() throws TaskException { + + } + + @Override + public void trackApplicationStatus() throws TaskException { + + } + protected String buildCommand() { final List args = new ArrayList<>(); @@ -126,8 +146,12 @@ public class HiveCliTask extends AbstractTaskExecutor { } @Override - public void cancelApplication(boolean cancelApplication) throws Exception { - shellCommandExecutor.cancelApplication(); + public void cancelApplication() throws TaskException { + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } } } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-http/src/main/java/org/apache/dolphinscheduler/plugin/task/http/HttpTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-http/src/main/java/org/apache/dolphinscheduler/plugin/task/http/HttpTask.java index df17725cf2..2562d5dcc1 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-http/src/main/java/org/apache/dolphinscheduler/plugin/task/http/HttpTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-http/src/main/java/org/apache/dolphinscheduler/plugin/task/http/HttpTask.java @@ -19,14 +19,14 @@ package org.apache.dolphinscheduler.plugin.task.http; import static org.apache.dolphinscheduler.plugin.task.http.HttpTaskConstants.APPLICATION_JSON; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.plugin.task.api.model.Property; import org.apache.dolphinscheduler.plugin.task.api.parameters.AbstractParameters; import org.apache.dolphinscheduler.plugin.task.api.parser.ParamUtils; import org.apache.dolphinscheduler.plugin.task.api.parser.ParameterUtils; -import org.apache.dolphinscheduler.plugin.task.api.utils.MapUtils; import org.apache.dolphinscheduler.spi.utils.DateUtils; import org.apache.dolphinscheduler.spi.utils.JSONUtils; import org.apache.dolphinscheduler.spi.utils.StringUtils; @@ -48,13 +48,12 @@ import org.apache.http.util.EntityUtils; import java.io.IOException; import java.nio.charset.StandardCharsets; import java.util.ArrayList; -import java.util.HashMap; import java.util.List; import java.util.Map; import com.fasterxml.jackson.databind.node.ObjectNode; -public class HttpTask extends AbstractTaskExecutor { +public class HttpTask extends AbstractTask { /** * output @@ -90,7 +89,7 @@ public class HttpTask extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { long startTime = System.currentTimeMillis(); String formatTimeStamp = DateUtils.formatTimeStamp(startTime); String statusCode = null; @@ -114,6 +113,11 @@ public class HttpTask extends AbstractTaskExecutor { } + @Override + public void cancel() throws TaskException { + + } + /** * send request * diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-http/src/test/java/org/apache/dolphinscheduler/plugin/task/http/HttpTaskTest.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-http/src/test/java/org/apache/dolphinscheduler/plugin/task/http/HttpTaskTest.java index 23b8095f83..2a063e5ac8 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-http/src/test/java/org/apache/dolphinscheduler/plugin/task/http/HttpTaskTest.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-http/src/test/java/org/apache/dolphinscheduler/plugin/task/http/HttpTaskTest.java @@ -76,11 +76,11 @@ public class HttpTaskTest { HttpTask headHttpTask = generateHttpTask(HttpMethod.HEAD, HttpStatus.SC_OK); HttpTask putHttpTask = generateHttpTask(HttpMethod.PUT, HttpStatus.SC_OK); HttpTask deleteHttpTask = generateHttpTask(HttpMethod.DELETE, HttpStatus.SC_OK); - getHttpTask.handle(); - postHttpTask.handle(); - headHttpTask.handle(); - putHttpTask.handle(); - deleteHttpTask.handle(); + getHttpTask.handle(null); + postHttpTask.handle(null); + headHttpTask.handle(null); + putHttpTask.handle(null); + deleteHttpTask.handle(null); Assert.assertEquals(EXIT_CODE_SUCCESS, getHttpTask.getExitStatusCode()); Assert.assertEquals(EXIT_CODE_SUCCESS, postHttpTask.getExitStatusCode()); Assert.assertEquals(EXIT_CODE_SUCCESS, headHttpTask.getExitStatusCode()); @@ -91,7 +91,7 @@ public class HttpTaskTest { @Test public void testHandleCheckCodeDefaultError() throws Exception { HttpTask getHttpTask = generateHttpTask(HttpMethod.GET, HttpStatus.SC_BAD_REQUEST); - getHttpTask.handle(); + getHttpTask.handle(null); Assert.assertEquals(EXIT_CODE_FAILURE, getHttpTask.getExitStatusCode()); } @@ -102,8 +102,8 @@ public class HttpTaskTest { condition, HttpStatus.SC_CREATED, ""); HttpTask httpErrorTask = generateHttpTask(HttpMethod.GET, HttpCheckCondition.STATUS_CODE_CUSTOM, condition, HttpStatus.SC_OK, ""); - httpTask.handle(); - httpErrorTask.handle(); + httpTask.handle(null); + httpErrorTask.handle(null); Assert.assertEquals(EXIT_CODE_SUCCESS, httpTask.getExitStatusCode()); Assert.assertEquals(EXIT_CODE_FAILURE, httpErrorTask.getExitStatusCode()); } @@ -114,8 +114,8 @@ public class HttpTaskTest { "success", HttpStatus.SC_OK, "{\"status\": \"success\"}"); HttpTask httpErrorTask = generateHttpTask(HttpMethod.GET, HttpCheckCondition.BODY_CONTAINS, "success", HttpStatus.SC_OK, "{\"status\": \"failed\"}"); - httpTask.handle(); - httpErrorTask.handle(); + httpTask.handle(null); + httpErrorTask.handle(null); Assert.assertEquals(EXIT_CODE_SUCCESS, httpTask.getExitStatusCode()); Assert.assertEquals(EXIT_CODE_FAILURE, httpErrorTask.getExitStatusCode()); } @@ -126,8 +126,8 @@ public class HttpTaskTest { "failed", HttpStatus.SC_OK, "{\"status\": \"success\"}"); HttpTask httpErrorTask = generateHttpTask(HttpMethod.GET, HttpCheckCondition.BODY_NOT_CONTAINS, "failed", HttpStatus.SC_OK, "{\"status\": \"failed\"}"); - httpTask.handle(); - httpErrorTask.handle(); + httpTask.handle(null); + httpErrorTask.handle(null); Assert.assertEquals(EXIT_CODE_SUCCESS, httpTask.getExitStatusCode()); Assert.assertEquals(EXIT_CODE_FAILURE, httpErrorTask.getExitStatusCode()); } @@ -148,7 +148,7 @@ public class HttpTaskTest { HttpTask httpTask = generateHttpTask(MOCK_DISPATCH_PATH_REQ_BODY_TO_RES_BODY, HttpMethod.POST, httpParams, prepareParamsMap, HttpCheckCondition.BODY_CONTAINS, "20220812", HttpStatus.SC_OK, ""); - httpTask.handle(); + httpTask.handle(null); Assert.assertEquals(EXIT_CODE_SUCCESS, httpTask.getExitStatusCode()); } @@ -168,7 +168,7 @@ public class HttpTaskTest { HttpTask httpTask = generateHttpTask(MOCK_DISPATCH_PATH_REQ_PARAMS_TO_RES_BODY, HttpMethod.POST, httpParams, prepareParamsMap, HttpCheckCondition.BODY_CONTAINS, "20220812", HttpStatus.SC_OK, ""); - httpTask.handle(); + httpTask.handle(null); Assert.assertEquals(EXIT_CODE_SUCCESS, httpTask.getExitStatusCode()); } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-jupyter/src/main/java/org/apache/dolphinscheduler/plugin/task/jupyter/JupyterTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-jupyter/src/main/java/org/apache/dolphinscheduler/plugin/task/jupyter/JupyterTask.java index ab135ac99e..6587fddc31 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-jupyter/src/main/java/org/apache/dolphinscheduler/plugin/task/jupyter/JupyterTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-jupyter/src/main/java/org/apache/dolphinscheduler/plugin/task/jupyter/JupyterTask.java @@ -18,8 +18,11 @@ package org.apache.dolphinscheduler.plugin.task.jupyter; import com.fasterxml.jackson.databind.ObjectMapper; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; + +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -35,10 +38,12 @@ import org.apache.dolphinscheduler.spi.utils.StringUtils; import java.io.IOException; import java.util.ArrayList; +import java.util.Collections; import java.util.List; import java.util.Map; +import java.util.Set; -public class JupyterTask extends AbstractTaskExecutor { +public class JupyterTask extends AbstractRemoteTask { /** * jupyter parameters @@ -60,6 +65,11 @@ public class JupyterTask extends AbstractTaskExecutor { logger); } + @Override + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); + } + @Override public void init() { logger.info("jupyter task params {}", taskExecutionContext.getTaskParams()); @@ -76,8 +86,9 @@ public class JupyterTask extends AbstractTaskExecutor { } } + // todo split handle to submit and track @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { // SHELL task exit code TaskResponse response = shellCommandExecutor.run(buildCommand()); @@ -96,6 +107,16 @@ public class JupyterTask extends AbstractTaskExecutor { } } + @Override + public void submitApplication() throws TaskException { + + } + + @Override + public void trackApplicationStatus() throws TaskException { + + } + /** * create command * @@ -223,9 +244,13 @@ public class JupyterTask extends AbstractTaskExecutor { } @Override - public void cancelApplication(boolean cancelApplication) throws Exception { + public void cancelApplication() throws TaskException { // cancel process - shellCommandExecutor.cancelApplication(); + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } } @Override diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-k8s/src/main/java/org/apache/dolphinscheduler/plugin/task/k8s/K8sTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-k8s/src/main/java/org/apache/dolphinscheduler/plugin/task/k8s/K8sTask.java index 7265e9fda8..b405ae5f17 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-k8s/src/main/java/org/apache/dolphinscheduler/plugin/task/k8s/K8sTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-k8s/src/main/java/org/apache/dolphinscheduler/plugin/task/k8s/K8sTask.java @@ -30,7 +30,9 @@ import org.apache.dolphinscheduler.plugin.task.api.parameters.K8sTaskParameters; import org.apache.dolphinscheduler.plugin.task.api.parser.ParamUtils; import org.apache.dolphinscheduler.spi.utils.JSONUtils; +import java.util.Collections; import java.util.Map; +import java.util.Set; public class K8sTask extends AbstractK8sTask { @@ -56,6 +58,11 @@ public class K8sTask extends AbstractK8sTask { } } + @Override + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); + } + @Override public AbstractParameters getParameters() { return k8sTaskParameters; diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-mlflow/src/main/java/org/apache/dolphinscheduler/plugin/task/mlflow/MlflowTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-mlflow/src/main/java/org/apache/dolphinscheduler/plugin/task/mlflow/MlflowTask.java index 0c42a5c09b..daa916cdea 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-mlflow/src/main/java/org/apache/dolphinscheduler/plugin/task/mlflow/MlflowTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-mlflow/src/main/java/org/apache/dolphinscheduler/plugin/task/mlflow/MlflowTask.java @@ -18,8 +18,9 @@ package org.apache.dolphinscheduler.plugin.task.mlflow; import org.apache.dolphinscheduler.common.thread.ThreadUtils; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -40,7 +41,7 @@ import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.EXIT_COD /** * shell task */ -public class MlflowTask extends AbstractTaskExecutor { +public class MlflowTask extends AbstractTask { /** * shell parameters @@ -81,7 +82,7 @@ public class MlflowTask extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { // construct process String command = buildCommand(); @@ -93,7 +94,6 @@ public class MlflowTask extends AbstractTaskExecutor { exitCode = commandExecuteResult.getExitStatusCode(); } setExitStatusCode(exitCode); - setAppIds(String.join(TaskConstants.COMMA, getApplicationIds())); setProcessId(commandExecuteResult.getProcessId()); mlflowParameters.dealOutParam(shellCommandExecutor.getVarPool()); } catch (InterruptedException e) { @@ -109,9 +109,13 @@ public class MlflowTask extends AbstractTaskExecutor { } @Override - public void cancelApplication(boolean cancelApplication) throws Exception { + public void cancel() throws TaskException { // cancel process - shellCommandExecutor.cancelApplication(); + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } } public String buildCommand() { diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-pigeon/src/main/java/org/apache/dolphinscheduler/plugin/task/pigeon/PigeonTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-pigeon/src/main/java/org/apache/dolphinscheduler/plugin/task/pigeon/PigeonTask.java index e5771d588d..177122847e 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-pigeon/src/main/java/org/apache/dolphinscheduler/plugin/task/pigeon/PigeonTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-pigeon/src/main/java/org/apache/dolphinscheduler/plugin/task/pigeon/PigeonTask.java @@ -18,7 +18,10 @@ package org.apache.dolphinscheduler.plugin.task.pigeon; import org.apache.commons.collections4.CollectionUtils; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; + +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -27,6 +30,7 @@ import org.apache.dolphinscheduler.spi.utils.JSONUtils; import org.apache.dolphinscheduler.spi.utils.StringUtils; import org.apache.http.HttpEntity; import org.apache.http.StatusLine; +import org.apache.http.client.ClientProtocolException; import org.apache.http.client.methods.CloseableHttpResponse; import org.apache.http.client.methods.HttpPost; import org.apache.http.entity.StringEntity; @@ -36,18 +40,21 @@ import org.apache.http.util.EntityUtils; import org.java_websocket.client.WebSocketClient; import org.java_websocket.handshake.ServerHandshake; +import java.io.IOException; import java.net.HttpURLConnection; import java.net.URI; import java.nio.charset.StandardCharsets; +import java.util.Collections; import java.util.List; import java.util.Map; import java.util.Objects; +import java.util.Set; import java.util.stream.Collectors; /** * TIS DataX Task **/ -public class PigeonTask extends AbstractTaskExecutor { +public class PigeonTask extends AbstractRemoteTask { public static final String KEY_POOL_VAR_PIGEON_HOST = "p_host"; private final TaskExecutionContext taskExecutionContext; @@ -62,6 +69,11 @@ public class PigeonTask extends AbstractTaskExecutor { this.config = PigeonConfig.getInstance(); } + @Override + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); + } + @Override public void init() { super.init(); @@ -72,8 +84,9 @@ public class PigeonTask extends AbstractTaskExecutor { } } + // todo split handle to submit and track @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { // Trigger PIGEON DataX pipeline logger.info("start execute PIGEON task"); long startTime = System.currentTimeMillis(); @@ -153,13 +166,22 @@ public class PigeonTask extends AbstractTaskExecutor { } } + @Override + public void submitApplication() throws TaskException { + + } + + @Override + public void trackApplicationStatus() throws TaskException { + + } + private void addFormUrlencoded(HttpPost post) { post.addHeader("content-type", "application/x-www-form-urlencoded"); } @Override - public void cancelApplication(boolean status) throws Exception { - super.cancelApplication(status); + public void cancelApplication() throws TaskException { logger.info("start to cancelApplication"); Objects.requireNonNull(triggerResult, "triggerResult can not be null"); logger.info("start to cancelApplication taskId:{}", triggerResult.getTaskId()); @@ -181,8 +203,12 @@ public class PigeonTask extends AbstractTaskExecutor { if (CollectionUtils.isNotEmpty(errormsg)) { errs.append(",errs:").append(errormsg.stream().collect(Collectors.joining(","))); } - throw new Exception("cancel PIGEON job faild taskId:" + triggerResult.getTaskId() + errs.toString()); + throw new TaskException("cancel PIGEON job faild taskId:" + triggerResult.getTaskId() + errs); } + } catch (ClientProtocolException e) { + throw new TaskException("client protocol error", e); + } catch (Exception e) { + throw new TaskException("pigeon execute error", e); } } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-pigeon/src/test/java/org/apache/dolphinscheduler/plugin/task/pigeon/PigeonTaskTest.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-pigeon/src/test/java/org/apache/dolphinscheduler/plugin/task/pigeon/PigeonTaskTest.java index 07a624a3e6..f748315b65 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-pigeon/src/test/java/org/apache/dolphinscheduler/plugin/task/pigeon/PigeonTaskTest.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-pigeon/src/test/java/org/apache/dolphinscheduler/plugin/task/pigeon/PigeonTaskTest.java @@ -114,7 +114,7 @@ public class PigeonTaskTest { file("src/test/resources/org/apache/dolphinscheduler/plugin/task/pigeon/PigeonTaskTest.json")); running(server, () -> { - pigeonTask.handle(); + pigeonTask.handle(null); Assert.assertEquals("PIGEON execute be success", TaskExecutionStatus.SUCCESS, pigeonTask.getExitStatus()); }); diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-procedure/src/main/java/org/apache/dolphinscheduler/plugin/task/procedure/ProcedureTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-procedure/src/main/java/org/apache/dolphinscheduler/plugin/task/procedure/ProcedureTask.java index 02e960ebdb..743b936e22 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-procedure/src/main/java/org/apache/dolphinscheduler/plugin/task/procedure/ProcedureTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-procedure/src/main/java/org/apache/dolphinscheduler/plugin/task/procedure/ProcedureTask.java @@ -19,7 +19,8 @@ package org.apache.dolphinscheduler.plugin.task.procedure; import org.apache.dolphinscheduler.plugin.datasource.api.plugin.DataSourceClientProvider; import org.apache.dolphinscheduler.plugin.datasource.api.utils.DataSourceUtils; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.plugin.task.api.enums.DataType; @@ -47,7 +48,7 @@ import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.EXIT_COD /** * procedure task */ -public class ProcedureTask extends AbstractTaskExecutor { +public class ProcedureTask extends AbstractTask { /** * procedure parameters @@ -84,7 +85,7 @@ public class ProcedureTask extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { logger.info("procedure type : {}, datasource : {}, method : {} , localParams : {}", procedureParameters.getType(), procedureParameters.getDatasource(), @@ -129,6 +130,11 @@ public class ProcedureTask extends AbstractTaskExecutor { } } + @Override + public void cancel() throws TaskException { + + } + private String formatSql(Map sqlParamsMap, Map paramsMap) { // combining local and global parameters setSqlParamsMap(procedureParameters.getMethod(), rgex, sqlParamsMap, paramsMap, taskExecutionContext.getTaskInstanceId()); diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-python/src/main/java/org/apache/dolphinscheduler/plugin/task/python/PythonTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-python/src/main/java/org/apache/dolphinscheduler/plugin/task/python/PythonTask.java index 6c18d206d9..8c1d26392e 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-python/src/main/java/org/apache/dolphinscheduler/plugin/task/python/PythonTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-python/src/main/java/org/apache/dolphinscheduler/plugin/task/python/PythonTask.java @@ -17,8 +17,9 @@ package org.apache.dolphinscheduler.plugin.task.python; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -27,7 +28,6 @@ import org.apache.dolphinscheduler.plugin.task.api.model.TaskResponse; import org.apache.dolphinscheduler.plugin.task.api.parameters.AbstractParameters; import org.apache.dolphinscheduler.plugin.task.api.parser.ParamUtils; import org.apache.dolphinscheduler.plugin.task.api.parser.ParameterUtils; -import org.apache.dolphinscheduler.plugin.task.api.utils.MapUtils; import org.apache.dolphinscheduler.spi.utils.JSONUtils; import org.apache.commons.io.FileUtils; @@ -37,7 +37,6 @@ import java.io.IOException; import java.nio.charset.StandardCharsets; import java.nio.file.Files; import java.nio.file.Paths; -import java.util.HashMap; import java.util.Map; import com.google.common.base.Preconditions; @@ -45,7 +44,7 @@ import com.google.common.base.Preconditions; /** * python task */ -public class PythonTask extends AbstractTaskExecutor { +public class PythonTask extends AbstractTask { /** * python parameters @@ -100,7 +99,7 @@ public class PythonTask extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { // generate the content of this python script String pythonScriptContent = buildPythonScriptContent(); @@ -113,7 +112,6 @@ public class PythonTask extends AbstractTaskExecutor { TaskResponse taskResponse = shellCommandExecutor.run(command); setExitStatusCode(taskResponse.getExitStatusCode()); - setAppIds(String.join(TaskConstants.COMMA, getApplicationIds())); setProcessId(taskResponse.getProcessId()); setVarPool(shellCommandExecutor.getVarPool()); } catch (Exception e) { @@ -124,9 +122,13 @@ public class PythonTask extends AbstractTaskExecutor { } @Override - public void cancelApplication(boolean cancelApplication) throws Exception { + public void cancel() throws TaskException { // cancel process - shellCommandExecutor.cancelApplication(); + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } } @Override diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-pytorch/src/main/java/org/apache/dolphinscheduler/plugin/task/pytorch/PytorchTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-pytorch/src/main/java/org/apache/dolphinscheduler/plugin/task/pytorch/PytorchTask.java index 3cd7c7c4f8..c362097e38 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-pytorch/src/main/java/org/apache/dolphinscheduler/plugin/task/pytorch/PytorchTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-pytorch/src/main/java/org/apache/dolphinscheduler/plugin/task/pytorch/PytorchTask.java @@ -17,8 +17,9 @@ package org.apache.dolphinscheduler.plugin.task.pytorch; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -33,7 +34,7 @@ import java.util.ArrayList; import java.util.List; import java.util.Map; -public class PytorchTask extends AbstractTaskExecutor { +public class PytorchTask extends AbstractTask { private final ShellCommandExecutor shellCommandExecutor; protected PytorchParameters pytorchParameters; @@ -65,12 +66,11 @@ public class PytorchTask extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { String command = buildPythonExecuteCommand(); TaskResponse taskResponse = shellCommandExecutor.run(command); setExitStatusCode(taskResponse.getExitStatusCode()); - setAppIds(taskResponse.getAppIds()); setProcessId(taskResponse.getProcessId()); setVarPool(shellCommandExecutor.getVarPool()); } catch (InterruptedException e) { @@ -84,6 +84,11 @@ public class PytorchTask extends AbstractTaskExecutor { } } + @Override + public void cancel() throws TaskException { + + } + public String buildPythonExecuteCommand() throws Exception { List args = new ArrayList<>(); diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/main/java/org/apache/dolphinscheduler/plugin/task/sagemaker/PipelineUtils.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/main/java/org/apache/dolphinscheduler/plugin/task/sagemaker/PipelineUtils.java index b5df5b9ea5..a70637e046 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/main/java/org/apache/dolphinscheduler/plugin/task/sagemaker/PipelineUtils.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/main/java/org/apache/dolphinscheduler/plugin/task/sagemaker/PipelineUtils.java @@ -23,6 +23,10 @@ import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import java.util.Collections; import java.util.List; +import lombok.AllArgsConstructor; +import lombok.Data; +import lombok.NoArgsConstructor; + import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -39,75 +43,57 @@ import com.amazonaws.services.sagemaker.model.StopPipelineExecutionResult; public class PipelineUtils { + protected final Logger logger = + LoggerFactory.getLogger(String.format(TaskConstants.TASK_LOG_LOGGER_NAME_FORMAT, getClass())); + private static final String EXECUTING = "Executing"; + private static final String SUCCEEDED = "Succeeded"; - protected final Logger logger = LoggerFactory.getLogger(String.format(TaskConstants.TASK_LOG_LOGGER_NAME_FORMAT, getClass())); - private final AmazonSageMaker client; - private String pipelineExecutionArn; - private String clientRequestToken; - private String pipelineStatus; - - public PipelineUtils(AmazonSageMaker client) { - this.client = client; - } - - public int startPipelineExecution(StartPipelineExecutionRequest request) { - int exitStatusCode = TaskConstants.EXIT_CODE_FAILURE; - try { - StartPipelineExecutionResult result = client.startPipelineExecution(request); - pipelineExecutionArn = result.getPipelineExecutionArn(); - clientRequestToken = request.getClientRequestToken(); - exitStatusCode = TaskConstants.EXIT_CODE_SUCCESS; - logger.info("Start pipeline: {} success", pipelineExecutionArn); - } catch (Exception e) { - logger.error("Start pipeline error: {}", e.getMessage()); - } + public PipelineId startPipelineExecution(AmazonSageMaker client, StartPipelineExecutionRequest request) { + StartPipelineExecutionResult result = client.startPipelineExecution(request); + String pipelineExecutionArn = result.getPipelineExecutionArn(); + String clientRequestToken = request.getClientRequestToken(); + logger.info("Start success, pipeline: {}, token: {}", pipelineExecutionArn, clientRequestToken); - return exitStatusCode; + return new PipelineId(pipelineExecutionArn, clientRequestToken); } - public void stopPipelineExecution() { + public void stopPipelineExecution(AmazonSageMaker client, PipelineId pipelineId) { StopPipelineExecutionRequest request = new StopPipelineExecutionRequest(); - request.setPipelineExecutionArn(pipelineExecutionArn); - request.setClientRequestToken(clientRequestToken); - - try { - StopPipelineExecutionResult result = client.stopPipelineExecution(request); - logger.info("Stop pipeline: {} success", result.getPipelineExecutionArn()); - } catch (Exception e) { - logger.error("Stop pipeline error: {}", e.getMessage()); - } + request.setPipelineExecutionArn(pipelineId.getPipelineExecutionArn()); + request.setClientRequestToken(pipelineId.getClientRequestToken()); + StopPipelineExecutionResult result = client.stopPipelineExecution(request); + logger.info("Stop pipeline: {} success", result.getPipelineExecutionArn()); } - public int checkPipelineExecutionStatus() { - describePipelineExecution(); - while (pipelineStatus.equals("Executing")) { + public int checkPipelineExecutionStatus(AmazonSageMaker client, PipelineId pipelineId) { + String pipelineStatus = describePipelineExecution(client, pipelineId); + while (EXECUTING.equals(pipelineStatus)) { logger.info("check Pipeline Steps running"); - listPipelineExecutionSteps(); + listPipelineExecutionSteps(client, pipelineId); ThreadUtils.sleep(SagemakerConstants.CHECK_PIPELINE_EXECUTION_STATUS_INTERVAL); - describePipelineExecution(); + describePipelineExecution(client, pipelineId); } int exitStatusCode = TaskConstants.EXIT_CODE_FAILURE; - if (pipelineStatus.equals("Succeeded")) { + if (SUCCEEDED.equals(pipelineStatus)) { exitStatusCode = TaskConstants.EXIT_CODE_SUCCESS; } - logger.info("exit : {}", exitStatusCode); - logger.info("PipelineExecutionStatus : {}", pipelineStatus); + logger.info("PipelineExecutionStatus : {}, exitStatusCode: {}", pipelineStatus, exitStatusCode); return exitStatusCode; } - private void describePipelineExecution() { + private String describePipelineExecution(AmazonSageMaker client, PipelineId pipelineId) { DescribePipelineExecutionRequest request = new DescribePipelineExecutionRequest(); - request.setPipelineExecutionArn(pipelineExecutionArn); + request.setPipelineExecutionArn(pipelineId.getPipelineExecutionArn()); DescribePipelineExecutionResult result = client.describePipelineExecution(request); - pipelineStatus = result.getPipelineExecutionStatus(); - logger.info("PipelineExecutionStatus: {}", pipelineStatus); + logger.info("PipelineExecutionStatus: {}", result.getPipelineExecutionStatus()); + return result.getPipelineExecutionStatus(); } - private void listPipelineExecutionSteps() { + private void listPipelineExecutionSteps(AmazonSageMaker client, PipelineId pipelineId) { ListPipelineExecutionStepsRequest request = new ListPipelineExecutionStepsRequest(); - request.setPipelineExecutionArn(pipelineExecutionArn); + request.setPipelineExecutionArn(pipelineId.getPipelineExecutionArn()); request.setMaxResults(SagemakerConstants.PIPELINE_MAX_RESULTS); ListPipelineExecutionStepsResult result = client.listPipelineExecutionSteps(request); List steps = result.getPipelineExecutionSteps(); @@ -119,7 +105,12 @@ public class PipelineUtils { } } - public String getPipelineExecutionArn() { - return pipelineExecutionArn; + @Data + @AllArgsConstructor + @NoArgsConstructor + public static class PipelineId { + + private String pipelineExecutionArn; + private String clientRequestToken; } -} \ No newline at end of file +} diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/main/java/org/apache/dolphinscheduler/plugin/task/sagemaker/SagemakerTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/main/java/org/apache/dolphinscheduler/plugin/task/sagemaker/SagemakerTask.java index 2543e60614..22c898d355 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/main/java/org/apache/dolphinscheduler/plugin/task/sagemaker/SagemakerTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/main/java/org/apache/dolphinscheduler/plugin/task/sagemaker/SagemakerTask.java @@ -22,7 +22,7 @@ import static com.fasterxml.jackson.databind.DeserializationFeature.FAIL_ON_UNKN import static com.fasterxml.jackson.databind.DeserializationFeature.READ_UNKNOWN_ENUM_VALUES_AS_NULL; import static com.fasterxml.jackson.databind.MapperFeature.REQUIRE_SETTERS_FOR_GETTERS; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -31,8 +31,11 @@ import org.apache.dolphinscheduler.plugin.task.api.parser.ParamUtils; import org.apache.dolphinscheduler.plugin.task.api.parser.ParameterUtils; import org.apache.dolphinscheduler.spi.utils.JSONUtils; import org.apache.dolphinscheduler.spi.utils.PropertyUtils; +import org.apache.dolphinscheduler.spi.utils.StringUtils; +import java.util.Collections; import java.util.Map; +import java.util.Set; import com.amazonaws.auth.AWSCredentialsProvider; import com.amazonaws.auth.AWSStaticCredentialsProvider; @@ -46,34 +49,43 @@ import com.fasterxml.jackson.databind.PropertyNamingStrategy; /** * SagemakerTask task, Used to start Sagemaker pipeline */ -public class SagemakerTask extends AbstractTaskExecutor { +public class SagemakerTask extends AbstractRemoteTask { private static final ObjectMapper objectMapper = - new ObjectMapper().configure(FAIL_ON_UNKNOWN_PROPERTIES, false).configure(ACCEPT_EMPTY_ARRAY_AS_NULL_OBJECT, true).configure(READ_UNKNOWN_ENUM_VALUES_AS_NULL, true) - .configure(REQUIRE_SETTERS_FOR_GETTERS, true).setPropertyNamingStrategy(new PropertyNamingStrategy.UpperCamelCaseStrategy()); - /** - * taskExecutionContext - */ - private final TaskExecutionContext taskExecutionContext; + new ObjectMapper().configure(FAIL_ON_UNKNOWN_PROPERTIES, false) + .configure(ACCEPT_EMPTY_ARRAY_AS_NULL_OBJECT, true) + .configure(READ_UNKNOWN_ENUM_VALUES_AS_NULL, true) + .configure(REQUIRE_SETTERS_FOR_GETTERS, true) + .setPropertyNamingStrategy(new PropertyNamingStrategy.UpperCamelCaseStrategy()); /** * SageMaker parameters */ private SagemakerParameters parameters; - private PipelineUtils utils; + + private final AmazonSageMaker client; + private final PipelineUtils utils; + private PipelineUtils.PipelineId pipelineId; public SagemakerTask(TaskExecutionContext taskExecutionContext) { super(taskExecutionContext); + client = createClient(); + utils = new PipelineUtils(); + } - this.taskExecutionContext = taskExecutionContext; - + @Override + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); } @Override public void init() { - logger.info("Sagemaker task params {}", taskExecutionContext.getTaskParams()); + logger.info("Sagemaker task params {}", taskRequest.getTaskParams()); - parameters = JSONUtils.parseObject(taskExecutionContext.getTaskParams(), SagemakerParameters.class); + parameters = JSONUtils.parseObject(taskRequest.getTaskParams(), SagemakerParameters.class); + if (parameters == null) { + throw new SagemakerTaskException("Sagemaker task params is empty"); + } if (!parameters.checkParameters()) { throw new SagemakerTaskException("Sagemaker task params is not valid"); } @@ -81,41 +93,51 @@ public class SagemakerTask extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void submitApplication() throws TaskException { try { - int exitStatusCode = handleStartPipeline(); - setExitStatusCode(exitStatusCode); + StartPipelineExecutionRequest request = createStartPipelineRequest(); + + // Start pipeline + pipelineId = utils.startPipelineExecution(client, request); + + // set AppId + setAppIds(JSONUtils.toJsonString(pipelineId)); } catch (Exception e) { setExitStatusCode(TaskConstants.EXIT_CODE_FAILURE); - throw new TaskException("SageMaker task error", e); + throw new TaskException("SageMaker task submit error", e); } } @Override - public void cancelApplication(boolean cancelApplication) { - // stop pipeline - utils.stopPipelineExecution(); - } - - public int handleStartPipeline() { - int exitStatusCode; - StartPipelineExecutionRequest request = createStartPipelineRequest(); - + public void cancelApplication() { + initPipelineId(); try { - AmazonSageMaker client = createClient(); - utils = new PipelineUtils(client); - setAppIds(utils.getPipelineExecutionArn()); + // stop pipeline + utils.stopPipelineExecution(client, pipelineId); } catch (Exception e) { - throw new SagemakerTaskException("can not connect aws ", e); + throw new TaskException("cancel application error", e); } + } - // Start pipeline - exitStatusCode = utils.startPipelineExecution(request); - if (exitStatusCode == TaskConstants.EXIT_CODE_SUCCESS) { - // Keep checking the health status - exitStatusCode = utils.checkPipelineExecutionStatus(); + @Override + public void trackApplicationStatus() throws TaskException { + initPipelineId(); + // Keep checking the health status + exitStatusCode = utils.checkPipelineExecutionStatus(client, pipelineId); + } + + /** + * init sagemaker applicationId if null + */ + private void initPipelineId() { + if (pipelineId == null) { + if (StringUtils.isNotEmpty(getAppIds())) { + pipelineId = JSONUtils.parseObject(getAppIds(), PipelineUtils.PipelineId.class); + } + } + if (pipelineId == null) { + throw new TaskException("sagemaker applicationID is null"); } - return exitStatusCode; } public StartPipelineExecutionRequest createStartPipelineRequest() throws SagemakerTaskException { @@ -142,11 +164,11 @@ public class SagemakerTask extends AbstractTaskExecutor { private String parseRequstJson(String requestJson) { // combining local and global parameters - Map paramsMap = taskExecutionContext.getPrepareParamsMap(); + Map paramsMap = taskRequest.getPrepareParamsMap(); return ParameterUtils.convertParameterPlaceholders(requestJson, ParamUtils.convert(paramsMap)); } - private AmazonSageMaker createClient() { + protected AmazonSageMaker createClient() { final String awsAccessKeyId = PropertyUtils.getString(TaskConstants.AWS_ACCESS_KEY_ID); final String awsSecretAccessKey = PropertyUtils.getString(TaskConstants.AWS_SECRET_ACCESS_KEY); final String awsRegion = PropertyUtils.getString(TaskConstants.AWS_REGION); @@ -154,9 +176,9 @@ public class SagemakerTask extends AbstractTaskExecutor { final AWSCredentialsProvider awsCredentialsProvider = new AWSStaticCredentialsProvider(basicAWSCredentials); // create a SageMaker client return AmazonSageMakerClientBuilder.standard() - .withCredentials(awsCredentialsProvider) - .withRegion(awsRegion) - .build(); + .withCredentials(awsCredentialsProvider) + .withRegion(awsRegion) + .build(); } } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/test/java/org/apache/dolphinscheduler/plugin/task/sagemaker/SagemakerTaskTest.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/test/java/org/apache/dolphinscheduler/plugin/task/sagemaker/SagemakerTaskTest.java index a7dcdca7bd..01e93b382f 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/test/java/org/apache/dolphinscheduler/plugin/task/sagemaker/SagemakerTaskTest.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-sagemaker/src/test/java/org/apache/dolphinscheduler/plugin/task/sagemaker/SagemakerTaskTest.java @@ -37,10 +37,8 @@ import org.junit.Before; import org.junit.Test; import org.junit.runner.RunWith; import org.mockito.Mockito; -import org.powermock.api.mockito.PowerMockito; import org.powermock.core.classloader.annotations.PowerMockIgnore; import org.powermock.core.classloader.annotations.PrepareForTest; -import org.powermock.core.classloader.annotations.SuppressStaticInitializationFor; import org.powermock.modules.junit4.PowerMockRunner; import com.amazonaws.services.sagemaker.AmazonSageMaker; @@ -54,24 +52,23 @@ import com.amazonaws.services.sagemaker.model.StopPipelineExecutionResult; @RunWith(PowerMockRunner.class) @PrepareForTest({JSONUtils.class, PropertyUtils.class,}) @PowerMockIgnore({"javax.*"}) -@SuppressStaticInitializationFor("org.apache.dolphinscheduler.spi.utils.PropertyUtils") public class SagemakerTaskTest { private final String pipelineExecutionArn = "test-pipeline-arn"; + private final String clientRequestToken = "test-pipeline-token"; private SagemakerTask sagemakerTask; private AmazonSageMaker client; - private PipelineUtils pipelineUtils; + private PipelineUtils pipelineUtils = new PipelineUtils(); @Before public void before() { - PowerMockito.mockStatic(PropertyUtils.class); String parameters = buildParameters(); TaskExecutionContext taskExecutionContext = Mockito.mock(TaskExecutionContext.class); Mockito.when(taskExecutionContext.getTaskParams()).thenReturn(parameters); + + client = mock(AmazonSageMaker.class); sagemakerTask = new SagemakerTask(taskExecutionContext); sagemakerTask.init(); - client = mock(AmazonSageMaker.class); - pipelineUtils = new PipelineUtils(client); StartPipelineExecutionResult startPipelineExecutionResult = mock(StartPipelineExecutionResult.class); when(startPipelineExecutionResult.getPipelineExecutionArn()).thenReturn(pipelineExecutionArn); @@ -82,7 +79,8 @@ public class SagemakerTaskTest { DescribePipelineExecutionResult describePipelineExecutionResult = mock(DescribePipelineExecutionResult.class); when(describePipelineExecutionResult.getPipelineExecutionStatus()).thenReturn("Executing", "Succeeded"); - ListPipelineExecutionStepsResult listPipelineExecutionStepsResult = mock(ListPipelineExecutionStepsResult.class); + ListPipelineExecutionStepsResult listPipelineExecutionStepsResult = + mock(ListPipelineExecutionStepsResult.class); PipelineExecutionStep pipelineExecutionStep = mock(PipelineExecutionStep.class); List pipelineExecutionSteps = new ArrayList<>(); pipelineExecutionSteps.add(pipelineExecutionStep); @@ -110,10 +108,11 @@ public class SagemakerTaskTest { @Test public void testPipelineExecution() throws Exception { - pipelineUtils.startPipelineExecution(sagemakerTask.createStartPipelineRequest()); - Assert.assertEquals(pipelineExecutionArn, pipelineUtils.getPipelineExecutionArn()); - Assert.assertEquals(0, pipelineUtils.checkPipelineExecutionStatus()); - pipelineUtils.stopPipelineExecution(); + PipelineUtils.PipelineId pipelineId = + pipelineUtils.startPipelineExecution(client, sagemakerTask.createStartPipelineRequest()); + Assert.assertEquals(pipelineExecutionArn, pipelineId.getPipelineExecutionArn()); + Assert.assertEquals(0, pipelineUtils.checkPipelineExecutionStatus(client, pipelineId)); + pipelineUtils.stopPipelineExecution(client, pipelineId); } private String buildParameters() { diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-seatunnel/src/main/java/org/apache/dolphinscheduler/plugin/task/seatunnel/SeatunnelTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-seatunnel/src/main/java/org/apache/dolphinscheduler/plugin/task/seatunnel/SeatunnelTask.java index cd3d8ebb22..b0e8aa5454 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-seatunnel/src/main/java/org/apache/dolphinscheduler/plugin/task/seatunnel/SeatunnelTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-seatunnel/src/main/java/org/apache/dolphinscheduler/plugin/task/seatunnel/SeatunnelTask.java @@ -19,8 +19,11 @@ package org.apache.dolphinscheduler.plugin.task.seatunnel; import org.apache.commons.io.FileUtils; import org.apache.commons.lang3.BooleanUtils; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; + +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -36,8 +39,10 @@ import java.nio.charset.StandardCharsets; import java.nio.file.Files; import java.nio.file.Paths; import java.util.ArrayList; +import java.util.Collections; import java.util.List; import java.util.Map; +import java.util.Set; import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.EXIT_CODE_FAILURE; import static org.apache.dolphinscheduler.plugin.task.seatunnel.Constants.CONFIG_OPTIONS; @@ -45,7 +50,7 @@ import static org.apache.dolphinscheduler.plugin.task.seatunnel.Constants.CONFIG /** * seatunnel task */ -public class SeatunnelTask extends AbstractTaskExecutor { +public class SeatunnelTask extends AbstractRemoteTask { /** * seatunnel parameters @@ -76,6 +81,11 @@ public class SeatunnelTask extends AbstractTaskExecutor { logger); } + @Override + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); + } + @Override public void init() { logger.info("SeaTunnel task params {}", taskExecutionContext.getTaskParams()); @@ -84,8 +94,9 @@ public class SeatunnelTask extends AbstractTaskExecutor { } } + // todo split handle to submit and track @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { // construct process String command = buildCommand(); @@ -107,9 +118,23 @@ public class SeatunnelTask extends AbstractTaskExecutor { } @Override - public void cancelApplication(boolean cancelApplication) throws Exception { + public void submitApplication() throws TaskException { + + } + + @Override + public void trackApplicationStatus() throws TaskException { + + } + + @Override + public void cancelApplication() throws TaskException { // cancel process - shellCommandExecutor.cancelApplication(); + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } } private String buildCommand() throws Exception { diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-shell/src/main/java/org/apache/dolphinscheduler/plugin/task/shell/ShellTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-shell/src/main/java/org/apache/dolphinscheduler/plugin/task/shell/ShellTask.java index 7e20bed634..7c7c4b2497 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-shell/src/main/java/org/apache/dolphinscheduler/plugin/task/shell/ShellTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-shell/src/main/java/org/apache/dolphinscheduler/plugin/task/shell/ShellTask.java @@ -18,8 +18,10 @@ package org.apache.dolphinscheduler.plugin.task.shell; import org.apache.commons.lang3.SystemUtils; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; + +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.ShellCommandExecutor; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -47,7 +49,7 @@ import static org.apache.dolphinscheduler.plugin.task.api.TaskConstants.RWXR_XR_ /** * shell task */ -public class ShellTask extends AbstractTaskExecutor { +public class ShellTask extends AbstractTask { /** * shell parameters @@ -90,13 +92,12 @@ public class ShellTask extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { // construct process String command = buildCommand(); TaskResponse commandExecuteResult = shellCommandExecutor.run(command); setExitStatusCode(commandExecuteResult.getExitStatusCode()); - setAppIds(String.join(TaskConstants.COMMA, getApplicationIds())); setProcessId(commandExecuteResult.getProcessId()); shellParameters.dealOutParam(shellCommandExecutor.getVarPool()); } catch (InterruptedException e) { @@ -112,9 +113,13 @@ public class ShellTask extends AbstractTaskExecutor { } @Override - public void cancelApplication(boolean cancelApplication) throws Exception { + public void cancel() throws TaskException { // cancel process - shellCommandExecutor.cancelApplication(); + try { + shellCommandExecutor.cancelApplication(); + } catch (Exception e) { + throw new TaskException("cancel application error", e); + } } /** diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-sql/src/main/java/org/apache/dolphinscheduler/plugin/task/sql/SqlTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-sql/src/main/java/org/apache/dolphinscheduler/plugin/task/sql/SqlTask.java index f884f46e5e..9889d6299b 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-sql/src/main/java/org/apache/dolphinscheduler/plugin/task/sql/SqlTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-sql/src/main/java/org/apache/dolphinscheduler/plugin/task/sql/SqlTask.java @@ -23,8 +23,9 @@ import org.apache.commons.collections4.CollectionUtils; import org.apache.dolphinscheduler.plugin.datasource.api.plugin.DataSourceClientProvider; import org.apache.dolphinscheduler.plugin.datasource.api.utils.CommonUtils; import org.apache.dolphinscheduler.plugin.datasource.api.utils.DataSourceUtils; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; import org.apache.dolphinscheduler.plugin.task.api.SQLTaskExecutionContext; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -60,7 +61,7 @@ import java.util.regex.Matcher; import java.util.regex.Pattern; import java.util.stream.Collectors; -public class SqlTask extends AbstractTaskExecutor { +public class SqlTask extends AbstractTask { /** * taskExecutionContext @@ -114,7 +115,7 @@ public class SqlTask extends AbstractTaskExecutor { } @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { logger.info("Full sql parameters: {}", sqlParameters); logger.info("sql type : {}, datasource : {}, sql : {} , localParams : {},udfs : {},showType : {},connParams : {},varPool : {} ,query max result limit {}", sqlParameters.getType(), @@ -164,6 +165,11 @@ public class SqlTask extends AbstractTaskExecutor { } } + @Override + public void cancel() throws TaskException { + + } + /** * execute function and sql * diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-zeppelin/src/main/java/org/apache/dolphinscheduler/plugin/task/zeppelin/ZeppelinTask.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-zeppelin/src/main/java/org/apache/dolphinscheduler/plugin/task/zeppelin/ZeppelinTask.java index f51b352d9c..75b6155587 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-zeppelin/src/main/java/org/apache/dolphinscheduler/plugin/task/zeppelin/ZeppelinTask.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-zeppelin/src/main/java/org/apache/dolphinscheduler/plugin/task/zeppelin/ZeppelinTask.java @@ -19,7 +19,10 @@ package org.apache.dolphinscheduler.plugin.task.zeppelin; import com.fasterxml.jackson.databind.ObjectMapper; import kong.unirest.Unirest; -import org.apache.dolphinscheduler.plugin.task.api.AbstractTaskExecutor; + +import org.apache.dolphinscheduler.plugin.task.api.AbstractRemoteTask; +import org.apache.dolphinscheduler.plugin.task.api.AbstractTask; +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskConstants; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; @@ -32,11 +35,13 @@ import org.apache.zeppelin.client.ParagraphResult; import org.apache.zeppelin.client.Status; import org.apache.zeppelin.client.ZeppelinClient; +import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Map; +import java.util.Set; -public class ZeppelinTask extends AbstractTaskExecutor { +public class ZeppelinTask extends AbstractRemoteTask { /** * taskExecutionContext @@ -74,8 +79,9 @@ public class ZeppelinTask extends AbstractTaskExecutor { this.zClient = getZeppelinClient(); } + // todo split handle to submit and track @Override - public void handle() throws TaskException { + public void handle(TaskCallBack taskCallBack) throws TaskException { try { final String paragraphId = this.zeppelinParameters.getParagraphId(); final String productionNoteDirectory = this.zeppelinParameters.getProductionNoteDirectory(); @@ -142,6 +148,15 @@ public class ZeppelinTask extends AbstractTaskExecutor { logger.error("zeppelin task submit failed with error", e); throw new TaskException("Execute ZeppelinTask exception"); } + } + + @Override + public void submitApplication() throws TaskException { + + } + + @Override + public void trackApplicationStatus() throws TaskException { } @@ -188,9 +203,8 @@ public class ZeppelinTask extends AbstractTaskExecutor { } @Override - public void cancelApplication(boolean status) throws Exception { + public void cancelApplication() throws TaskException { final String restEndpoint = this.zeppelinParameters.getRestEndpoint(); - super.cancelApplication(status); final String noteId = this.zeppelinParameters.getNoteId(); final String paragraphId = this.zeppelinParameters.getParagraphId(); if (paragraphId == null) { @@ -207,7 +221,11 @@ public class ZeppelinTask extends AbstractTaskExecutor { this.taskExecutionContext.getTaskInstanceId(), noteId, paragraphId); - this.zClient.cancelParagraph(noteId, paragraphId); + try { + this.zClient.cancelParagraph(noteId, paragraphId); + } catch (Exception e) { + throw new TaskException("cancel paragraph error", e); + } logger.info("zeppelin task terminated, taskId: {}, noteId: {}, paragraphId: {}", this.taskExecutionContext.getTaskInstanceId(), noteId, @@ -216,4 +234,9 @@ public class ZeppelinTask extends AbstractTaskExecutor { } + @Override + public Set getApplicationIds() throws TaskException { + return Collections.emptySet(); + } + } diff --git a/dolphinscheduler-task-plugin/dolphinscheduler-task-zeppelin/src/test/java/org/apache/dolphinscheduler/plugin/task/zeppelin/ZeppelinTaskTest.java b/dolphinscheduler-task-plugin/dolphinscheduler-task-zeppelin/src/test/java/org/apache/dolphinscheduler/plugin/task/zeppelin/ZeppelinTaskTest.java index ff70ba5b45..70a0945a94 100644 --- a/dolphinscheduler-task-plugin/dolphinscheduler-task-zeppelin/src/test/java/org/apache/dolphinscheduler/plugin/task/zeppelin/ZeppelinTaskTest.java +++ b/dolphinscheduler-task-plugin/dolphinscheduler-task-zeppelin/src/test/java/org/apache/dolphinscheduler/plugin/task/zeppelin/ZeppelinTaskTest.java @@ -28,6 +28,8 @@ import static org.powermock.api.mockito.PowerMockito.spy; import static org.powermock.api.mockito.PowerMockito.when; import com.fasterxml.jackson.databind.ObjectMapper; + +import org.apache.dolphinscheduler.plugin.task.api.TaskCallBack; import org.apache.dolphinscheduler.plugin.task.api.TaskException; import org.apache.dolphinscheduler.plugin.task.api.TaskExecutionContext; import org.apache.dolphinscheduler.spi.utils.DateUtils; @@ -73,6 +75,9 @@ public class ZeppelinTaskTest { private ZeppelinTask zeppelinTask; private ParagraphResult paragraphResult; private NoteResult noteResult; + private TaskCallBack taskCallBack = (taskInstanceId, appIds) -> { + + }; @Before public void before() throws Exception { @@ -95,7 +100,7 @@ public class ZeppelinTaskTest { @Test public void testHandleWithParagraphExecutionSuccess() throws Exception { when(this.paragraphResult.getStatus()).thenReturn(Status.FINISHED); - this.zeppelinTask.handle(); + this.zeppelinTask.handle(taskCallBack); Mockito.verify(this.zClient).executeParagraph(MOCK_NOTE_ID, MOCK_PARAGRAPH_ID, (Map) mapper.readValue(MOCK_PARAMETERS, Map.class)); @@ -107,7 +112,7 @@ public class ZeppelinTaskTest { @Test public void testHandleWithParagraphExecutionAborted() throws Exception { when(this.paragraphResult.getStatus()).thenReturn(Status.ABORT); - this.zeppelinTask.handle(); + this.zeppelinTask.handle(taskCallBack); Mockito.verify(this.zClient).executeParagraph(MOCK_NOTE_ID, MOCK_PARAGRAPH_ID, (Map) mapper.readValue(MOCK_PARAMETERS, Map.class)); @@ -119,7 +124,7 @@ public class ZeppelinTaskTest { @Test public void testHandleWithParagraphExecutionError() throws Exception { when(this.paragraphResult.getStatus()).thenReturn(Status.ERROR); - this.zeppelinTask.handle(); + this.zeppelinTask.handle(taskCallBack); Mockito.verify(this.zClient).executeParagraph(MOCK_NOTE_ID, MOCK_PARAGRAPH_ID, (Map) mapper.readValue(MOCK_PARAMETERS, Map.class)); @@ -133,7 +138,7 @@ public class ZeppelinTaskTest { when(this.zClient.executeParagraph(any(), any(), any(Map.class))). thenThrow(new TaskException("Something wrong happens from zeppelin side")); // when(this.paragraphResult.getStatus()).thenReturn(Status.ERROR); - this.zeppelinTask.handle(); + this.zeppelinTask.handle(taskCallBack); Mockito.verify(this.zClient).executeParagraph(MOCK_NOTE_ID, MOCK_PARAGRAPH_ID, (Map) mapper.readValue(MOCK_PARAMETERS, Map.class)); @@ -159,7 +164,7 @@ public class ZeppelinTaskTest { when(paragraphResult.getResultInText()).thenReturn("mock-zeppelin-paragraph-execution-result"); this.zeppelinTask.init(); when(this.paragraphResult.getStatus()).thenReturn(Status.FINISHED); - this.zeppelinTask.handle(); + this.zeppelinTask.handle(taskCallBack); Mockito.verify(this.zClient).executeNote(MOCK_NOTE_ID, (Map) mapper.readValue(MOCK_PARAMETERS, Map.class)); Mockito.verify(this.noteResult).getParagraphResultList(); @@ -186,7 +191,7 @@ public class ZeppelinTaskTest { this.zeppelinTask.init(); when(this.paragraphResult.getStatus()).thenReturn(Status.FINISHED); when(DateUtils.getTimestampString()).thenReturn("123456789"); - this.zeppelinTask.handle(); + this.zeppelinTask.handle(taskCallBack); Mockito.verify(this.zClient).cloneNote( MOCK_NOTE_ID, String.format("%s%s_%s", MOCK_PRODUCTION_DIRECTORY, MOCK_NOTE_ID, "123456789")); diff --git a/dolphinscheduler-tools/src/main/java/org/apache/dolphinscheduler/tools/datasource/dao/UpgradeDao.java b/dolphinscheduler-tools/src/main/java/org/apache/dolphinscheduler/tools/datasource/dao/UpgradeDao.java index 3fb639b226..0fb4614889 100644 --- a/dolphinscheduler-tools/src/main/java/org/apache/dolphinscheduler/tools/datasource/dao/UpgradeDao.java +++ b/dolphinscheduler-tools/src/main/java/org/apache/dolphinscheduler/tools/datasource/dao/UpgradeDao.java @@ -78,6 +78,7 @@ import com.google.common.base.Joiner; import com.google.common.base.Strings; public abstract class UpgradeDao { + public static final Logger logger = LoggerFactory.getLogger(UpgradeDao.class); private static final String T_VERSION_NAME = "t_escheduler_version"; private static final String T_NEW_VERSION_NAME = "t_ds_version"; @@ -102,7 +103,7 @@ public abstract class UpgradeDao { * @param dbType db type */ private void runInitSql(DbType dbType) { - String sqlFile = String.format("dolphinscheduler_%s.sql",dbType.getDescp()); + String sqlFile = String.format("dolphinscheduler_%s.sql", dbType.getDescp()); Resource mysqlSQLFilePath = new ClassPathResource("sql/" + sqlFile); try (Connection conn = dataSource.getConnection()) { // Execute the dolphinscheduler_ddl.sql script to create the table structure of dolphinscheduler @@ -198,7 +199,8 @@ public abstract class UpgradeDao { Map replaceProcessDefinitionMap = new HashMap<>(); try { Map oldWorkerGroupMap = workerGroupDao.queryAllOldWorkerGroup(dataSource.getConnection()); - Map processDefinitionJsonMap = processDefinitionDao.queryAllProcessDefinition(dataSource.getConnection()); + Map processDefinitionJsonMap = + processDefinitionDao.queryAllProcessDefinition(dataSource.getConnection()); for (Map.Entry entry : processDefinitionJsonMap.entrySet()) { ObjectNode jsonObject = JSONUtils.parseObject(entry.getValue()); @@ -225,7 +227,8 @@ public abstract class UpgradeDao { replaceProcessDefinitionMap.put(entry.getKey(), jsonObject.toString()); } if (replaceProcessDefinitionMap.size() > 0) { - processDefinitionDao.updateProcessDefinitionJson(dataSource.getConnection(), replaceProcessDefinitionMap); + processDefinitionDao.updateProcessDefinitionJson(dataSource.getConnection(), + replaceProcessDefinitionMap); } } catch (Exception e) { logger.error("update process definition json workergroup error", e); @@ -238,7 +241,8 @@ public abstract class UpgradeDao { Map replaceProcessDefinitionMap = new HashMap<>(); try { Map resourcesMap = resourceDao.listAllResources(dataSource.getConnection()); - Map processDefinitionJsonMap = processDefinitionDao.queryAllProcessDefinition(dataSource.getConnection()); + Map processDefinitionJsonMap = + processDefinitionDao.queryAllProcessDefinition(dataSource.getConnection()); for (Map.Entry entry : processDefinitionJsonMap.entrySet()) { ObjectNode jsonObject = JSONUtils.parseObject(entry.getValue()); @@ -249,10 +253,13 @@ public abstract class UpgradeDao { ObjectNode param = (ObjectNode) task.get("params"); if (param != null) { - List resourceList = JSONUtils.toList(param.get("resourceList").toString(), ResourceInfo.class); - ResourceInfo mainJar = JSONUtils.parseObject(param.get("mainJar").toString(), ResourceInfo.class); + List resourceList = + JSONUtils.toList(param.get("resourceList").toString(), ResourceInfo.class); + ResourceInfo mainJar = + JSONUtils.parseObject(param.get("mainJar").toString(), ResourceInfo.class); if (mainJar != null && mainJar.getId() == null) { - String fullName = mainJar.getRes().startsWith("/") ? mainJar.getRes() : String.format("/%s", mainJar.getRes()); + String fullName = mainJar.getRes().startsWith("/") ? mainJar.getRes() + : String.format("/%s", mainJar.getRes()); if (resourcesMap.containsKey(fullName)) { mainJar.setId(resourcesMap.get(fullName)); param.put("mainJar", JSONUtils.parseObject(JSONUtils.toJsonString(mainJar))); @@ -261,7 +268,8 @@ public abstract class UpgradeDao { if (CollectionUtils.isNotEmpty(resourceList)) { List newResourceList = resourceList.stream().map(resInfo -> { - String fullName = resInfo.getRes().startsWith("/") ? resInfo.getRes() : String.format("/%s", resInfo.getRes()); + String fullName = resInfo.getRes().startsWith("/") ? resInfo.getRes() + : String.format("/%s", resInfo.getRes()); if (resInfo.getId() == null && resourcesMap.containsKey(fullName)) { resInfo.setId(resourcesMap.get(fullName)); } @@ -281,7 +289,8 @@ public abstract class UpgradeDao { replaceProcessDefinitionMap.put(entry.getKey(), jsonObject.toString()); } if (replaceProcessDefinitionMap.size() > 0) { - processDefinitionDao.updateProcessDefinitionJson(dataSource.getConnection(), replaceProcessDefinitionMap); + processDefinitionDao.updateProcessDefinitionJson(dataSource.getConnection(), + replaceProcessDefinitionMap); } } catch (Exception e) { logger.error("update process definition json resource list error", e); @@ -291,7 +300,8 @@ public abstract class UpgradeDao { private void upgradeDolphinSchedulerDML(String schemaDir) { String schemaVersion = schemaDir.split("_")[0]; - Resource sqlFilePath = new ClassPathResource(String.format("sql/upgrade/%s/%s/dolphinscheduler_dml.sql", schemaDir, getDbType().name().toLowerCase())); + Resource sqlFilePath = new ClassPathResource(String.format("sql/upgrade/%s/%s/dolphinscheduler_dml.sql", + schemaDir, getDbType().name().toLowerCase())); logger.info("sqlSQLFilePath: {}", sqlFilePath); Connection conn = null; PreparedStatement pstmt = null; @@ -355,7 +365,8 @@ public abstract class UpgradeDao { * @param schemaDir schemaDir */ private void upgradeDolphinSchedulerDDL(String schemaDir, String scriptFile) { - Resource sqlFilePath = new ClassPathResource(String.format("sql/upgrade/%s/%s/%s", schemaDir, getDbType().name().toLowerCase(), scriptFile)); + Resource sqlFilePath = new ClassPathResource( + String.format("sql/upgrade/%s/%s/%s", schemaDir, getDbType().name().toLowerCase(), scriptFile)); Connection conn = null; PreparedStatement pstmt = null; try { @@ -423,21 +434,26 @@ public abstract class UpgradeDao { projectDao.updateProjectCode(dataSource.getConnection(), projectIdCodeMap); // execute process definition code - List processDefinitions = processDefinitionDao.queryProcessDefinition(dataSource.getConnection()); - processDefinitionDao.updateProcessDefinitionCode(dataSource.getConnection(), processDefinitions, projectIdCodeMap); + List processDefinitions = + processDefinitionDao.queryProcessDefinition(dataSource.getConnection()); + processDefinitionDao.updateProcessDefinitionCode(dataSource.getConnection(), processDefinitions, + projectIdCodeMap); // execute schedule Map allSchedule = scheduleDao.queryAllSchedule(dataSource.getConnection()); - Map processIdCodeMap = processDefinitions.stream().collect(Collectors.toMap(ProcessDefinition::getId, ProcessDefinition::getCode)); + Map processIdCodeMap = processDefinitions.stream() + .collect(Collectors.toMap(ProcessDefinition::getId, ProcessDefinition::getCode)); scheduleDao.updateScheduleCode(dataSource.getConnection(), allSchedule, processIdCodeMap); // json split - Map processDefinitionJsonMap = processDefinitionDao.queryAllProcessDefinition(dataSource.getConnection()); + Map processDefinitionJsonMap = + processDefinitionDao.queryAllProcessDefinition(dataSource.getConnection()); List processDefinitionLogs = new ArrayList<>(); List processTaskRelationLogs = new ArrayList<>(); List taskDefinitionLogs = new ArrayList<>(); Map>> processTaskMap = new HashMap<>(); - splitProcessDefinitionJson(processDefinitions, processDefinitionJsonMap, processDefinitionLogs, processTaskRelationLogs, taskDefinitionLogs, processTaskMap); + splitProcessDefinitionJson(processDefinitions, processDefinitionJsonMap, processDefinitionLogs, + processTaskRelationLogs, taskDefinitionLogs, processTaskMap); convertDependence(taskDefinitionLogs, projectIdCodeMap, processTaskMap); // execute json split @@ -456,7 +472,7 @@ public abstract class UpgradeDao { List taskDefinitionLogs, Map>> processTaskMap) throws Exception { Map processDefinitionMap = processDefinitions.stream() - .collect(Collectors.toMap(ProcessDefinition::getId, processDefinition -> processDefinition)); + .collect(Collectors.toMap(ProcessDefinition::getId, processDefinition -> processDefinition)); Date now = new Date(); for (Map.Entry entry : processDefinitionJsonMap.entrySet()) { if (entry.getValue() == null) { @@ -465,7 +481,8 @@ public abstract class UpgradeDao { ObjectNode jsonObject = JSONUtils.parseObject(entry.getValue()); ProcessDefinition processDefinition = processDefinitionMap.get(entry.getKey()); if (processDefinition != null) { - processDefinition.setTenantId(jsonObject.get("tenantId") == null ? -1 : jsonObject.get("tenantId").asInt()); + processDefinition + .setTenantId(jsonObject.get("tenantId") == null ? -1 : jsonObject.get("tenantId").asInt()); processDefinition.setTimeout(jsonObject.get("timeout").asInt()); processDefinition.setGlobalParams(jsonObject.get("globalParams").toString()); } else { @@ -485,8 +502,10 @@ public abstract class UpgradeDao { if (param != null) { JsonNode resourceJsonNode = param.get("resourceList"); if (resourceJsonNode != null && !resourceJsonNode.isEmpty()) { - List resourceList = JSONUtils.toList(param.get("resourceList").toString(), ResourceInfo.class); - List resourceIds = resourceList.stream().map(ResourceInfo::getId).collect(Collectors.toList()); + List resourceList = + JSONUtils.toList(param.get("resourceList").toString(), ResourceInfo.class); + List resourceIds = + resourceList.stream().map(ResourceInfo::getId).collect(Collectors.toList()); taskDefinitionLog.setResourceIds(Joiner.on(Constants.COMMA).join(resourceIds)); } else { taskDefinitionLog.setResourceIds(""); @@ -494,7 +513,8 @@ public abstract class UpgradeDao { if (TASK_TYPE_SUB_PROCESS.equals(taskType)) { JsonNode jsonNodeDefinitionId = param.get("processDefinitionId"); if (jsonNodeDefinitionId != null) { - param.put("processDefinitionCode", processDefinitionMap.get(jsonNodeDefinitionId.asInt()).getCode()); + param.put("processDefinitionCode", + processDefinitionMap.get(jsonNodeDefinitionId.asInt()).getCode()); param.remove("processDefinitionId"); } } @@ -502,23 +522,29 @@ public abstract class UpgradeDao { param.put("dependence", task.get("dependence")); taskDefinitionLog.setTaskParams(JSONUtils.toJsonString(param)); } - TaskTimeoutParameter timeout = JSONUtils.parseObject(JSONUtils.toJsonString(task.get("timeout")), TaskTimeoutParameter.class); + TaskTimeoutParameter timeout = + JSONUtils.parseObject(JSONUtils.toJsonString(task.get("timeout")), TaskTimeoutParameter.class); if (timeout != null) { taskDefinitionLog.setTimeout(timeout.getInterval()); taskDefinitionLog.setTimeoutFlag(timeout.getEnable() ? TimeoutFlag.OPEN : TimeoutFlag.CLOSE); taskDefinitionLog.setTimeoutNotifyStrategy(timeout.getStrategy()); } - String desc = task.get("description") != null ? task.get("description").asText() : - task.get("desc") != null ? task.get("desc").asText() : ""; + String desc = task.get("description") != null ? task.get("description").asText() + : task.get("desc") != null ? task.get("desc").asText() : ""; taskDefinitionLog.setDescription(desc); - taskDefinitionLog.setFlag(Constants.FLOWNODE_RUN_FLAG_NORMAL.equals(task.get("runFlag").asText()) ? Flag.YES : Flag.NO); + taskDefinitionLog.setFlag( + Constants.FLOWNODE_RUN_FLAG_NORMAL.equals(task.get("runFlag").asText()) ? Flag.YES : Flag.NO); taskDefinitionLog.setTaskType(taskType); - taskDefinitionLog.setFailRetryInterval(TASK_TYPE_SUB_PROCESS.equals(taskType) ? 1 : task.get("retryInterval").asInt()); - taskDefinitionLog.setFailRetryTimes(TASK_TYPE_SUB_PROCESS.equals(taskType) ? 0 : task.get("maxRetryTimes").asInt()); - taskDefinitionLog.setTaskPriority(JSONUtils.parseObject(JSONUtils.toJsonString(task.get("taskInstancePriority")), Priority.class)); + taskDefinitionLog.setFailRetryInterval( + TASK_TYPE_SUB_PROCESS.equals(taskType) ? 1 : task.get("retryInterval").asInt()); + taskDefinitionLog.setFailRetryTimes( + TASK_TYPE_SUB_PROCESS.equals(taskType) ? 0 : task.get("maxRetryTimes").asInt()); + taskDefinitionLog.setTaskPriority(JSONUtils + .parseObject(JSONUtils.toJsonString(task.get("taskInstancePriority")), Priority.class)); String name = task.get("name").asText(); taskDefinitionLog.setName(name); - taskDefinitionLog.setWorkerGroup(task.get("workerGroup") == null ? "default" : task.get("workerGroup").asText()); + taskDefinitionLog + .setWorkerGroup(task.get("workerGroup") == null ? "default" : task.get("workerGroup").asText()); long taskCode = CodeGenerateUtils.getInstance().genCode(); taskDefinitionLog.setCode(taskCode); taskDefinitionLog.setVersion(Constants.VERSION_FIRST); @@ -550,14 +576,16 @@ public abstract class UpgradeDao { } } - public void convertConditions(List taskDefinitionLogList, Map taskNameCodeMap) throws Exception { + public void convertConditions(List taskDefinitionLogList, + Map taskNameCodeMap) throws Exception { for (TaskDefinitionLog taskDefinitionLog : taskDefinitionLogList) { if (TASK_TYPE_CONDITIONS.equals(taskDefinitionLog.getTaskType())) { ObjectMapper objectMapper = new ObjectMapper(); ObjectNode taskParams = JSONUtils.parseObject(taskDefinitionLog.getTaskParams()); // reset conditionResult ObjectNode conditionResult = (ObjectNode) taskParams.get("conditionResult"); - List successNode = JSONUtils.toList(conditionResult.get("successNode").toString(), String.class); + List successNode = + JSONUtils.toList(conditionResult.get("successNode").toString(), String.class); List nodeCode = new ArrayList<>(); successNode.forEach(node -> nodeCode.add(taskNameCodeMap.get(node))); conditionResult.set("successNode", objectMapper.readTree(objectMapper.writeValueAsString(nodeCode))); @@ -567,10 +595,12 @@ public abstract class UpgradeDao { conditionResult.set("failedNode", objectMapper.readTree(objectMapper.writeValueAsString(nodeCode))); // reset dependItemList ObjectNode dependence = (ObjectNode) taskParams.get("dependence"); - ArrayNode dependTaskList = JSONUtils.parseArray(JSONUtils.toJsonString(dependence.get("dependTaskList"))); + ArrayNode dependTaskList = + JSONUtils.parseArray(JSONUtils.toJsonString(dependence.get("dependTaskList"))); for (int i = 0; i < dependTaskList.size(); i++) { ObjectNode dependTask = (ObjectNode) dependTaskList.path(i); - ArrayNode dependItemList = JSONUtils.parseArray(JSONUtils.toJsonString(dependTask.get("dependItemList"))); + ArrayNode dependItemList = + JSONUtils.parseArray(JSONUtils.toJsonString(dependTask.get("dependItemList"))); for (int j = 0; j < dependItemList.size(); j++) { ObjectNode dependItem = (ObjectNode) dependItemList.path(j); JsonNode depTasks = dependItem.get("depTasks"); @@ -591,8 +621,9 @@ public abstract class UpgradeDao { if (Strings.isNullOrEmpty(locations)) { return locations; } - Map locationsMap = JSONUtils.parseObject(locations, new TypeReference>() { - }); + Map locationsMap = + JSONUtils.parseObject(locations, new TypeReference>() { + }); if (locationsMap == null) { return locations; } @@ -615,26 +646,33 @@ public abstract class UpgradeDao { if (TASK_TYPE_DEPENDENT.equals(taskDefinitionLog.getTaskType())) { ObjectNode taskParams = JSONUtils.parseObject(taskDefinitionLog.getTaskParams()); ObjectNode dependence = (ObjectNode) taskParams.get("dependence"); - ArrayNode dependTaskList = JSONUtils.parseArray(JSONUtils.toJsonString(dependence.get("dependTaskList"))); + ArrayNode dependTaskList = + JSONUtils.parseArray(JSONUtils.toJsonString(dependence.get("dependTaskList"))); for (int i = 0; i < dependTaskList.size(); i++) { ObjectNode dependTask = (ObjectNode) dependTaskList.path(i); - ArrayNode dependItemList = JSONUtils.parseArray(JSONUtils.toJsonString(dependTask.get("dependItemList"))); + ArrayNode dependItemList = + JSONUtils.parseArray(JSONUtils.toJsonString(dependTask.get("dependItemList"))); for (int j = 0; j < dependItemList.size(); j++) { ObjectNode dependItem = (ObjectNode) dependItemList.path(j); dependItem.put("projectCode", projectIdCodeMap.get(dependItem.get("projectId").asInt())); int definitionId = dependItem.get("definitionId").asInt(); Map> processCodeTaskNameCodeMap = processTaskMap.get(definitionId); if (processCodeTaskNameCodeMap == null) { - logger.warn("We can't find processDefinition [{}], please check it is not exist, remove this dependence", definitionId); + logger.warn( + "We can't find processDefinition [{}], please check it is not exist, remove this dependence", + definitionId); dependItemList.remove(j); continue; } - Optional>> mapEntry = processCodeTaskNameCodeMap.entrySet().stream().findFirst(); + Optional>> mapEntry = + processCodeTaskNameCodeMap.entrySet().stream().findFirst(); if (mapEntry.isPresent()) { Map.Entry> processCodeTaskNameCodeEntry = mapEntry.get(); dependItem.put("definitionCode", processCodeTaskNameCodeEntry.getKey()); String depTasks = dependItem.get("depTasks").asText(); - long taskCode = "ALL".equals(depTasks) || processCodeTaskNameCodeEntry.getValue() == null ? 0L : processCodeTaskNameCodeEntry.getValue().get(depTasks); + long taskCode = + "ALL".equals(depTasks) || processCodeTaskNameCodeEntry.getValue() == null ? 0L + : processCodeTaskNameCodeEntry.getValue().get(depTasks); dependItem.put("depTaskCode", taskCode); } dependItem.remove("projectId"); diff --git a/dolphinscheduler-worker/src/main/java/org/apache/dolphinscheduler/server/worker/message/TaskExecuteRunningMessageSender.java b/dolphinscheduler-worker/src/main/java/org/apache/dolphinscheduler/server/worker/message/TaskExecuteRunningMessageSender.java index 7ba75efba1..5b874118e5 100644 --- a/dolphinscheduler-worker/src/main/java/org/apache/dolphinscheduler/server/worker/message/TaskExecuteRunningMessageSender.java +++ b/dolphinscheduler-worker/src/main/java/org/apache/dolphinscheduler/server/worker/message/TaskExecuteRunningMessageSender.java @@ -25,11 +25,11 @@ import org.apache.dolphinscheduler.remote.utils.Host; import org.apache.dolphinscheduler.server.worker.config.WorkerConfig; import org.apache.dolphinscheduler.server.worker.rpc.WorkerRpcClient; +import lombok.NonNull; + import org.springframework.beans.factory.annotation.Autowired; import org.springframework.stereotype.Component; -import lombok.NonNull; - @Component public class TaskExecuteRunningMessageSender implements MessageSender { @@ -46,10 +46,10 @@ public class TaskExecuteRunningMessageSender implements MessageSender projectRes = taskExecutionContext.getResources(); if (MapUtils.isEmpty(projectRes)) { @@ -113,7 +118,8 @@ public class TaskExecutionCheckerUtils { String resPath = storageOperate.getResourceFileName(tenantCode, fullName); logger.info("get resource file from path:{}", resPath); long resourceDownloadStartTime = System.currentTimeMillis(); - storageOperate.download(tenantCode, resPath, execLocalPath + File.separator + fullName, false, true); + storageOperate.download(tenantCode, resPath, execLocalPath + File.separator + fullName, false, + true); WorkerServerMetrics .recordWorkerResourceDownloadTime(System.currentTimeMillis() - resourceDownloadStartTime); WorkerServerMetrics.recordWorkerResourceDownloadSize(