You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
155 lines
5.1 KiB
155 lines
5.1 KiB
/* |
|
* Licensed to the Apache Software Foundation (ASF) under one or more |
|
* contributor license agreements. See the NOTICE file distributed with |
|
* this work for additional information regarding copyright ownership. |
|
* The ASF licenses this file to You under the Apache License, Version 2.0 |
|
* (the "License"); you may not use this file except in compliance with |
|
* the License. You may obtain a copy of the License at |
|
* |
|
* http://www.apache.org/licenses/LICENSE-2.0 |
|
* |
|
* Unless required by applicable law or agreed to in writing, software |
|
* distributed under the License is distributed on an "AS IS" BASIS, |
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
* See the License for the specific language governing permissions and |
|
* limitations under the License. |
|
*/ |
|
|
|
package org.apache.dolphinscheduler.server.worker.task.spark; |
|
|
|
import org.apache.dolphinscheduler.common.enums.SparkVersion; |
|
import org.apache.dolphinscheduler.common.process.Property; |
|
import org.apache.dolphinscheduler.common.process.ResourceInfo; |
|
import org.apache.dolphinscheduler.common.task.AbstractParameters; |
|
import org.apache.dolphinscheduler.common.task.spark.SparkParameters; |
|
import org.apache.dolphinscheduler.common.utils.JSONUtils; |
|
import org.apache.dolphinscheduler.common.utils.ParameterUtils; |
|
import org.apache.dolphinscheduler.dao.entity.Resource; |
|
import org.apache.dolphinscheduler.server.entity.TaskExecutionContext; |
|
import org.apache.dolphinscheduler.server.utils.ParamUtils; |
|
import org.apache.dolphinscheduler.server.utils.SparkArgsUtils; |
|
import org.apache.dolphinscheduler.server.worker.task.AbstractYarnTask; |
|
|
|
import java.util.ArrayList; |
|
import java.util.List; |
|
import java.util.Map; |
|
|
|
import org.slf4j.Logger; |
|
|
|
/** |
|
* spark task |
|
*/ |
|
public class SparkTask extends AbstractYarnTask { |
|
|
|
/** |
|
* spark1 command |
|
* usage: spark-submit [options] <app jar | python file> [app arguments] |
|
*/ |
|
private static final String SPARK1_COMMAND = "${SPARK_HOME1}/bin/spark-submit"; |
|
|
|
/** |
|
* spark2 command |
|
* usage: spark-submit [options] <app jar | python file> [app arguments] |
|
*/ |
|
private static final String SPARK2_COMMAND = "${SPARK_HOME2}/bin/spark-submit"; |
|
|
|
/** |
|
* spark parameters |
|
*/ |
|
private SparkParameters sparkParameters; |
|
|
|
/** |
|
* taskExecutionContext |
|
*/ |
|
private TaskExecutionContext taskExecutionContext; |
|
|
|
public SparkTask(TaskExecutionContext taskExecutionContext, Logger logger) { |
|
super(taskExecutionContext, logger); |
|
this.taskExecutionContext = taskExecutionContext; |
|
} |
|
|
|
@Override |
|
public void init() { |
|
|
|
logger.info("spark task params {}", taskExecutionContext.getTaskParams()); |
|
|
|
sparkParameters = JSONUtils.parseObject(taskExecutionContext.getTaskParams(), SparkParameters.class); |
|
|
|
if (null == sparkParameters) { |
|
logger.error("Spark params is null"); |
|
return; |
|
} |
|
|
|
if (!sparkParameters.checkParameters()) { |
|
throw new RuntimeException("spark task params is not valid"); |
|
} |
|
sparkParameters.setQueue(taskExecutionContext.getQueue()); |
|
setMainJarName(); |
|
} |
|
|
|
/** |
|
* create command |
|
* @return command |
|
*/ |
|
@Override |
|
protected String buildCommand() { |
|
// spark-submit [options] <app jar | python file> [app arguments] |
|
List<String> args = new ArrayList<>(); |
|
|
|
// spark version |
|
String sparkCommand = SPARK2_COMMAND; |
|
|
|
if (SparkVersion.SPARK1.name().equals(sparkParameters.getSparkVersion())) { |
|
sparkCommand = SPARK1_COMMAND; |
|
} |
|
|
|
args.add(sparkCommand); |
|
|
|
// other parameters |
|
args.addAll(SparkArgsUtils.buildArgs(sparkParameters)); |
|
|
|
// replace placeholder, and combining local and global parameters |
|
Map<String, Property> paramsMap = ParamUtils.convert(taskExecutionContext,getParameters()); |
|
|
|
String command = null; |
|
|
|
if (null != paramsMap) { |
|
command = ParameterUtils.convertParameterPlaceholders(String.join(" ", args), ParamUtils.convert(paramsMap)); |
|
} |
|
|
|
logger.info("spark task command: {}", command); |
|
|
|
return command; |
|
} |
|
|
|
@Override |
|
protected void setMainJarName() { |
|
// main jar |
|
ResourceInfo mainJar = sparkParameters.getMainJar(); |
|
|
|
if (null == mainJar) { |
|
throw new RuntimeException("Spark task jar params is null"); |
|
} |
|
|
|
int resourceId = mainJar.getId(); |
|
String resourceName; |
|
if (resourceId == 0) { |
|
resourceName = mainJar.getRes(); |
|
} else { |
|
Resource resource = processService.getResourceById(sparkParameters.getMainJar().getId()); |
|
if (resource == null) { |
|
logger.error("resource id: {} not exist", resourceId); |
|
throw new RuntimeException(String.format("resource id: %d not exist", resourceId)); |
|
} |
|
resourceName = resource.getFullName().replaceFirst("/", ""); |
|
} |
|
mainJar.setRes(resourceName); |
|
sparkParameters.setMainJar(mainJar); |
|
|
|
} |
|
|
|
@Override |
|
public AbstractParameters getParameters() { |
|
return sparkParameters; |
|
} |
|
}
|
|
|