diff --git a/docs/docs/en/guide/resource/configuration.md b/docs/docs/en/guide/resource/configuration.md index 67c68a22c2..6bee9e5a67 100644 --- a/docs/docs/en/guide/resource/configuration.md +++ b/docs/docs/en/guide/resource/configuration.md @@ -2,8 +2,8 @@ - You could use `Resource Center` to upload text files, UDFs and other task-related files. - You could configure `Resource Center` to use distributed file system like [Hadoop](https://hadoop.apache.org/docs/r2.7.0/) (2.6+), [MinIO](https://github.com/minio/minio) cluster or remote storage products like [AWS S3](https://aws.amazon.com/s3/), [Alibaba Cloud OSS](https://www.aliyun.com/product/oss), [Huawei Cloud OBS](https://support.huaweicloud.com/obs/index.html) etc. -- You could configure `Resource Center` to use local file system. If you deploy `DolphinScheduler` in `Standalone` mode, you could configure it to use local file system for `Resouce Center` without the need of an external `HDFS` system or `S3`. -- Furthermore, if you deploy `DolphinScheduler` in `Cluster` mode, you could use [S3FS-FUSE](https://github.com/s3fs-fuse/s3fs-fuse) to mount `S3` or [JINDO-FUSE](https://help.aliyun.com/document_detail/187410.html) to mount `OSS` to your machines and use the local file system for `Resouce Center`. In this way, you could operate remote files as if on your local machines. +- You could configure `Resource Center` to use local file system. If you deploy `DolphinScheduler` in `Standalone` mode, you could configure it to use local file system for `Resource Center` without the need of an external `HDFS` system or `S3`. +- Furthermore, if you deploy `DolphinScheduler` in `Cluster` mode, you could use [S3FS-FUSE](https://github.com/s3fs-fuse/s3fs-fuse) to mount `S3` or [JINDO-FUSE](https://help.aliyun.com/document_detail/187410.html) to mount `OSS` to your machines and use the local file system for `Resource Center`. In this way, you could operate remote files as if on your local machines. ## Use Local File System diff --git a/docs/docs/en/guide/start/docker.md b/docs/docs/en/guide/start/docker.md index b98d0572ae..f29bcd5c75 100644 --- a/docs/docs/en/guide/start/docker.md +++ b/docs/docs/en/guide/start/docker.md @@ -128,7 +128,7 @@ and use `admin` and `dolphinscheduler123` as default username and password in th ![login](../../../../img/new_ui/dev/quick-start/login.png) > Note: If you start the services by the way [using exists PostgreSQL ZooKeeper](#using-exists-postgresql-zookeeper), and -> strating with multiple machine, you should change URL domain from `localhost` to IP or hostname the api server running. +> starting with multiple machine, you should change URL domain from `localhost` to IP or hostname the api server running. ## Change Environment Variable diff --git a/docs/docs/en/guide/task/datafactory.md b/docs/docs/en/guide/task/datafactory.md index 1936db81f1..67d00c1227 100644 --- a/docs/docs/en/guide/task/datafactory.md +++ b/docs/docs/en/guide/task/datafactory.md @@ -19,11 +19,11 @@ DolphinScheduler DataFactory functions: ### Application Permission Setting -First, visit the `Subcription` page and choose `Access control (IAM)`, then click `Add role assignment` to the authorization page. -![Subcription-IAM](../../../../img/tasks/demo/datafactory_auth1.png) +First, visit the `Subscription` page and choose `Access control (IAM)`, then click `Add role assignment` to the authorization page. +![Subscription-IAM](../../../../img/tasks/demo/datafactory_auth1.png) After that, select `Contributor` role which satisfy functions calls in data factory. Then click `Members` page, and click `Select members`. Search application name or application `Object ID` to assign `Contributor` role to application. -![Subcription-Role](../../../../img/tasks/demo/datafactory_auth2.png) +![Subscription-Role](../../../../img/tasks/demo/datafactory_auth2.png) ## Configurations diff --git a/docs/docs/en/guide/task/kubernetes.md b/docs/docs/en/guide/task/kubernetes.md index 332669a60c..871a020294 100644 --- a/docs/docs/en/guide/task/kubernetes.md +++ b/docs/docs/en/guide/task/kubernetes.md @@ -26,7 +26,7 @@ K8S task type used to execute a batch task. In this task, the worker submits the | Command | The container execution command (yaml-style array), for example: ["printenv"] | | Args | The args of execution command (yaml-style array), for example: ["HOSTNAME", "KUBERNETES_PORT"] | | Custom label | The customized labels for k8s Job. | -| Node selector | The label selectors for running k8s pod. Different value in value set should be seperated by comma, for example: `value1,value2`. You can refer to https://kubernetes.io/docs/reference/kubernetes-api/common-definitions/node-selector-requirement/ for configuration of different operators. | +| Node selector | The label selectors for running k8s pod. Different value in value set should be separated by comma, for example: `value1,value2`. You can refer to https://kubernetes.io/docs/reference/kubernetes-api/common-definitions/node-selector-requirement/ for configuration of different operators. | | Custom parameter | It is a local user-defined parameter for K8S task, these params will pass to container as environment variables. | ## Task Example diff --git a/docs/docs/en/guide/task/mlflow.md b/docs/docs/en/guide/task/mlflow.md index a14ac12483..d500e4cbf6 100644 --- a/docs/docs/en/guide/task/mlflow.md +++ b/docs/docs/en/guide/task/mlflow.md @@ -148,7 +148,7 @@ After this, you can visit the MLflow service (`http://localhost:5000`) page to v ### Preset Algorithm Repository Configuration -If you can't access github, you can modify the following fields in the `commom.properties` configuration file to replace the github address with an accessible address. +If you can't access github, you can modify the following fields in the `common.properties` configuration file to replace the github address with an accessible address. ```yaml # mlflow task plugin preset repository diff --git a/docs/docs/zh/guide/task/datafactory.md b/docs/docs/zh/guide/task/datafactory.md index 0fa3375bc1..0822f60bec 100644 --- a/docs/docs/zh/guide/task/datafactory.md +++ b/docs/docs/zh/guide/task/datafactory.md @@ -19,10 +19,10 @@ DolphinScheduler DataFactory 组件的功能: ### 应用权限设置 -首先打开当前`Subcription`页面,点击`Access control (IAM)`,再点击`Add role assignment`进入授权页面。 -![Subcription-IAM](../../../../img/tasks/demo/datafactory_auth1.png) +首先打开当前`Subscription`页面,点击`Access control (IAM)`,再点击`Add role assignment`进入授权页面。 +![Subscription-IAM](../../../../img/tasks/demo/datafactory_auth1.png) 首先选择`Contributor`角色足够满足调用数据工厂。然后选择`Members`页面,再选择`Select members`,检索APP名称或APP的`Object ID`并添加,从给指定APP添加权限. -![Subcription-Role](../../../../img/tasks/demo/datafactory_auth2.png) +![Subscription-Role](../../../../img/tasks/demo/datafactory_auth2.png) ## 环境配置 diff --git a/docs/docs/zh/guide/task/mlflow.md b/docs/docs/zh/guide/task/mlflow.md index 3c384ce221..12d2bc5089 100644 --- a/docs/docs/zh/guide/task/mlflow.md +++ b/docs/docs/zh/guide/task/mlflow.md @@ -139,7 +139,7 @@ mlflow server -h 0.0.0.0 -p 5000 --serve-artifacts --backend-store-uri sqlite:// ### 内置算法仓库配置 -如果遇到github无法访问的情况,可以修改`commom.properties`配置文件的以下字段,将github地址替换能访问的地址。 +如果遇到github无法访问的情况,可以修改`common.properties`配置文件的以下字段,将github地址替换能访问的地址。 ```yaml # mlflow task plugin preset repository