Browse Source

Merge branch 'dev-1.3.0' of https://github.com/apache/incubator-dolphinscheduler into dev-1.3.0

pull/3/MERGE
break60 5 years ago
parent
commit
354439f67b
  1. 20
      ambari_plugin/README.md
  2. 5
      dolphinscheduler-server/src/main/java/org/apache/dolphinscheduler/server/master/dispatch/ExecutorDispatcher.java
  3. 3
      dolphinscheduler-server/src/main/resources/worker.properties

20
ambari_plugin/README.md

@ -9,11 +9,25 @@
1. 准备RPM包
- 在源码dolphinscheduler-dist目录下执行命令```mvn -U clean install rpm:attached-rpm -Prpmbuild -Dmaven.test.skip=true -X```即可生成(在目录 dolphinscheduler-dist/target/rpm/apache-dolphinscheduler-incubating/RPMS/noarch 下)
- 在项目根目录下执行命令```mvn -U clean install -Prpmbuild -Dmaven.test.skip=true -X```即可生成(在目录 dolphinscheduler-dist/target/rpm/apache-dolphinscheduler-incubating/RPMS/noarch 下)
2. 创建DS的安装用户--权限
2. 创建DS的安装用户--安装目录(/opt/soft)有读写权限
3. 初始化数据库信息
3. rpm安装
- 手动安装(推荐):
- copy准备的rpm包到集群的各个节点
- 用DS的安装用户执行: ```rpm -ivh apache-dolphinscheduler-incubating-xxx.noarch.rpm```
- ambar自动安装
- 集群的各个节点需要配置本地yum源
- copy准备的rpm包到各个节点本地yum源
4. copy插件目录
- copy目录ambari_plugin/common-services/DOLPHIN到ambari-server/resources/common-services/下
- copy目录ambari_plugin/statcks/DOLPHIN到 ambari-server/resources/stacks/HDP/2.6/services/下--stack版本2.6根据实际情况选择
5. 初始化数据库信息
```
-- 创建Dolphin Scheduler的数据库:dolphinscheduler

5
dolphinscheduler-server/src/main/java/org/apache/dolphinscheduler/server/master/dispatch/ExecutorDispatcher.java

@ -81,9 +81,12 @@ public class ExecutorDispatcher implements InitializingBean {
/**
* host select
*/
Host host = hostManager.select(context);
if (StringUtils.isEmpty(host.getAddress())) {
throw new ExecuteException(String.format("fail to execute : %s due to no worker ", context.getCommand()));
throw new ExecuteException(String.format("fail to execute : %s due to no suitable worker , " +
"current task need to %s worker group execute",
context.getCommand(),context.getWorkerGroup()));
}
context.setHost(host);
executorManager.beforeExecute(context);

3
dolphinscheduler-server/src/main/resources/worker.properties

@ -21,9 +21,6 @@
# worker heartbeat interval
#worker.heartbeat.interval=10
# submit the number of tasks at a time TODO
#worker.fetch.task.num = 3
# only less than cpu avg load, worker server can work. default value -1: the number of cpu cores * 2
#worker.max.cpuload.avg= -1

Loading…
Cancel
Save