diff --git a/dolphinscheduler-server/src/main/resources/config/install_config.conf b/dolphinscheduler-server/src/main/resources/config/install_config.conf index 73282acc00..2e74b1e1a9 100755 --- a/dolphinscheduler-server/src/main/resources/config/install_config.conf +++ b/dolphinscheduler-server/src/main/resources/config/install_config.conf @@ -21,7 +21,7 @@ # A comma separated list of machine hostname or IP would be installed DolphinScheduler, # including master, worker, api, alert. If you want to deploy in pseudo-distributed # mode, just write a pseudo-distributed hostname -# Example for hostnames: ips="ds1,ds2,ds3,ds4,ds5", Example for IP: ips="192.168.8.1,192.168.8.2,192.168.8.3,192.168.8.4,192.168.8.5" +# Example for hostnames: ips="ds1,ds2,ds3,ds4,ds5", Example for IPs: ips="192.168.8.1,192.168.8.2,192.168.8.3,192.168.8.4,192.168.8.5" ips="ds1,ds2,ds3,ds4,ds5" # Port of SSH protocol, default value is 22. For now we only support same port in all `ips` machine @@ -30,29 +30,29 @@ sshPort="22" # A comma separated list of machine hostname or IP would be installed Master server, it # must be a subset of configuration `ips`. -# Example for hostnames: ips="ds1,ds2", Example for IP: ips="192.168.8.1,192.168.8.2" +# Example for hostnames: masters="ds1,ds2", Example for IPs: masters="192.168.8.1,192.168.8.2" masters="ds1,ds2" # A comma separated list of machine : or :.All hostname or IP must be a # subset of configuration `ips`, And workerGroup have default value as `default`, but we recommend you declare behind the hosts -# Example for hostnames: ips="ds1:default,ds2:default,ds3:default", Example for IP: ips="192.168.8.1:default,192.168.8.2:default,192.168.8.3:default" +# Example for hostnames: workers="ds1:default,ds2:default,ds3:default", Example for IPs: workers="192.168.8.1:default,192.168.8.2:default,192.168.8.3:default" workers="ds1:default,ds2:default,ds3:default,ds4:default,ds5:default" # A comma separated list of machine hostname or IP would be installed Alert server, it # must be a subset of configuration `ips`. -# Example for hostnames: ips="ds3", Example for IP: ips="192.168.8.3" +# Example for hostname: alertServer="ds3", Example for IP: alertServer="192.168.8.3" alertServer="ds3" # A comma separated list of machine hostname or IP would be installed API server, it # must be a subset of configuration `ips`. -# Example for hostnames: ips="ds1", Example for IP: ips="192.168.8.1" +# Example for hostname: apiServers="ds1", Example for IP: apiServers="192.168.8.1" apiServers="ds1" -# The directory to install DolphinScheduler for all machine we config above. It will automatically created by `install.sh` script if not exists. -# **DO NOT** set this configuration same as the current path (pwd) +# The directory to install DolphinScheduler for all machine we config above. It will automatically be created by `install.sh` script if not exists. +# Do not set this configuration same as the current path (pwd) installPath="/data1_1T/dolphinscheduler" -# The user to deploy DolphinScheduler for all machine we config above. For now user must create by yourself before run `install.sh` +# The user to deploy DolphinScheduler for all machine we config above. For now user must create by yourself before running `install.sh` # script. The user needs to have sudo privileges and permissions to operate hdfs. If hdfs is enabled than the root directory needs # to be created by this user deployUser="dolphinscheduler" @@ -67,7 +67,7 @@ dataBasedirPath="/tmp/dolphinscheduler" # and this configuration only support one parameter so far. javaHome="/your/java/home/here" -# DolphinScheduler API service port, also this your DolphinScheduler UI component's URL port, default values is 12345 +# DolphinScheduler API service port, also this is your DolphinScheduler UI component's URL port, default value is 12345 apiServerPort="12345" # --------------------------------------------------------- @@ -122,12 +122,12 @@ taskPluginDir="lib/plugin/task" # resource storage type: HDFS, S3, NONE resourceStorageType="NONE" -# resource store on HDFS/S3 path, resource file will store to this hadoop hdfs path, self configuration, please make sure the directory exists on hdfs and have read write permissions. "/dolphinscheduler" is recommended +# resource store on HDFS/S3 path, resource file will store to this hdfs path, self configuration, please make sure the directory exists on hdfs and has read write permissions. "/dolphinscheduler" is recommended resourceUploadPath="/dolphinscheduler" -# if resourceStorageType is HDFS,defaultFS write namenode address,HA you need to put core-site.xml and hdfs-site.xml in the conf directory. +# if resourceStorageType is HDFS,defaultFS write namenode address,HA, you need to put core-site.xml and hdfs-site.xml in the conf directory. # if S3,write S3 address,HA,for example :s3a://dolphinscheduler, -# Note,s3 be sure to create the root directory /dolphinscheduler +# Note,S3 be sure to create the root directory /dolphinscheduler defaultFS="hdfs://mycluster:8020" # if resourceStorageType is S3, the following three configuration is required, otherwise please ignore @@ -138,13 +138,13 @@ s3SecretKey="xxxxxxxxxx" # resourcemanager port, the default value is 8088 if not specified resourceManagerHttpAddressPort="8088" -# if resourcemanager HA is enabled, please set the HA IPs; if resourcemanager is single, keep this value empty +# if resourcemanager HA is enabled, please set the HA IPs; if resourcemanager is single node, keep this value empty yarnHaIps="192.168.xx.xx,192.168.xx.xx" -# if resourcemanager HA is enabled or not use resourcemanager, please keep the default value; If resourcemanager is single, you only need to replace ds1 to actual resourcemanager hostname +# if resourcemanager HA is enabled or not use resourcemanager, please keep the default value; If resourcemanager is single node, you only need to replace 'yarnIp1' to actual resourcemanager hostname singleYarnIp="yarnIp1" -# who have permissions to create directory under HDFS/S3 root path +# who has permission to create directory under HDFS/S3 root path # Note: if kerberos is enabled, please config hdfsRootUser= hdfsRootUser="hdfs"