morrySnow commented on code in PR #1998:
URL: https://github.com/apache/doris-website/pull/1998#discussion_r1947458065
##########
docs/sql-manual/sql-statements/cluster-management/storage-management/ALTER-STORAGE-POLICY.md:
##########
@@ -26,29 +26,42 @@ under the License.
## Description
-This statement is used to modify an existing cold and hot separation migration
strategy. Only root or admin users can modify resources.
+This statement is used to modify an existing hot-cold tiered migration policy.
Only root or admin users can modify resources.
+## Syntax
```sql
-ALTER STORAGE POLICY 'policy_name'
-PROPERTIES ("key"="value", ...);
+ALTER STORAGE POLICY '<policy_name>' PROPERTIE ("<key>"="<value>"[, ... ]);
```
-## Example
+## Required Parameters
+| Parameter Name | Description
|
+|-------------------|--------------------------------------------------------------|
+| `<policy_name>` | The name of the storage policy. This is the unique
identifier of the storage policy you want to modify, and an existing policy
name must be specified. |
-1. Modify the name to coolown_datetime Cold and hot separation data migration
time point:
+## Optional Parameters
+
+| 参数名称 | 描述 |
Review Comment:
需要变成英文
##########
docs/sql-manual/sql-statements/cluster-management/storage-management/CREATE-STORAGE-VAULT.md:
##########
@@ -26,65 +26,71 @@ specific language governing permissions and limitations
under the License.
-->
-## CREATE-STORAGE-VAULT
+## Description
-### Description
+This command is used to create a storage vault. The topic of this document
describes the syntax for creating a self-managed storage vault in Doris.
-This command is used to create a storage vault. The subject of this document
describes the syntax for creating Doris self-maintained storage vault.
+
+## Syntax
```sql
-CREATE STORAGE VAULT [IF NOT EXISTS] vault
-[properties]
+CREATE STORAGE VAULT [IF NOT EXISTS] <`vault_name`> [ <`properties`> ]
```
-
-#### properties
-
-| param | is required | desc
|
-|:-------|:------------|:-------------------------------------------------------|
-| `type` | required | Only two types of vaults are allowed: `S3` and
`HDFS`. |
-
-##### S3 Vault
-
-| param | is required | desc
|
-|:----------------|:------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
-| `s3.endpoint` | required | The endpoint used for object storage.
<br/>**Notice**, please don't provide the endpoint with any `http://` or
`https://`. And for Azure Blob Storage, the endpoint should be
`blob.core.windows.net`. |
-| `s3.region` | required | The region of your bucket.(Not required
when you'r using GCP or AZURE).
|
-| `s3.root.path` | required | The path where the data would be stored.
|
-| `s3.bucket` | required | The bucket of your object storage account.
(StorageAccount if you're using Azure).
|
-| `s3.access_key` | required | The access key of your object storage
account. (AccountName if you're using Azure).
|
-| `s3.secret_key` | required | The secret key of your object storage
account. (AccountKey if you're using Azure).
|
-| `provider` | required | The cloud vendor which provides the object
storage service. The supported values include `COS`, `OSS`, `S3`, `OBS`, `BOS`,
`AZURE`, `GCP`
|
-| `use_path_style` | optional | Indicate using `path-style URL`(private
environment recommended) or `virtual-hosted-style URL`(public cloud
recommended), default `true` (`path-style`)
|
-
-##### HDFS Vault
-
-| param | is required | desc
|
-|:---------------------------------|:------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------|
-| `fs.defaultFS` | required | Hadoop configuration
property that specifies the default file system to use.
|
-| `path_prefix` | optional | The path prefix to where
the data would be stored. It would be the root_path of your Hadoop user if you
don't provide any prefix. |
-| `hadoop.username` | optional | Hadoop configuration
property that specifies the user accessing the file system. It would be the
user starting Hadoop process if you don't provide any user. |
-| `hadoop.security.authentication` | optional | The authentication way used
for hadoop. If you'd like to use kerberos you can provide with `kerboros`.
|
-| `hadoop.kerberos.principal` | optional | The path to your kerberos
principal. |
-| `hadoop.kerberos.keytab` | optional | The path to your kerberos
keytab. |
-
-### Example
-
-1. create a HDFS storage vault.
+## Required Parameters
+
+| Parameter | Description |
+|-------|-----------------------|
+| `vault_name` | The name of the storage vault. This is the unique identifier
for the new storage vault you are creating. |
+
+## Optional Parameters
+| Parameter | Description
|
+|-------------------|--------------------------------------------------------------|
+| `[IF NOT EXISTS]` | If the specified storage vault already exists, the
creation operation will not be executed, and no error will be thrown. This
prevents duplicate creation of the same storage vault. |
+| `PROPERTIES` | A set of key-value pairs used to set or update specific
properties of the storage vault. Each property consists of a key (<key>) and a
value (<value>), separated by an equals sign (=). Multiple key-value pairs are
separated by commas (,). |
+
+### S3 Vault
+
+| Parameter | Required | Description
|
+|:----------------|:-----|:--------------------------------------------------------------------------------------------------------|
+| `s3.endpoint` | Required | The endpoint for object storage.
+Note: Do not provide a link starting with http:// or https://. For Azure Blob
Storage, the endpoint is fixed as blob.core.windows.net.。 |
+| `s3.region` | Required | The region of your storage bucket. (Not
required if using GCP or AZURE). |
+| `s3.root.path` | Required | The path to store data. |
+| `s3.bucket` | Required | The bucket of your object storage account.
(For Azure, this is the StorageAccount). |
+| `s3.access_key` | Required | The access key for your object storage
account. (For Azure, this is the AccountName). |
+| `s3.secret_key` | Required | The secret key for your object storage
account. (For Azure, this is the AccountKey). |
+| `provider` | Required | The cloud provider offering the object
storage service. Supported values are
`COS`,`OSS`,`S3`,`OBS`,`BOS`,`AZURE`,`GCP` |
+| `use_path_style` | Optional | Use `path-style URL (for private deployment
environments) or `virtual-hosted-style URL`(recommended for public cloud
environments). Default value is true (path-style).
|
+
+### HDFS vault
+
+| Parameter | Required | Description
|
+|:---------------------------------|:-----|:------------------------------------------------------|
+| `fs.defaultFS` |Required| Hadoop configuration property
specifying the default file system to use. |
+| `path_prefix` |Optional| The prefix path for storing
data. If not specified, the default path under the user account will be used.
|
+| `hadoop.username` |Optional| Hadoop configuration property
specifying the user to access the file system. If not specified, the user who
started the Hadoop process will be used. |
+| `hadoop.security.authentication` |Optional| The authentication method for
Hadoop. If you want to use Kerberos, you can specify kerberos. |
+| `hadoop.kerberos.principal` |Optional| The path to your Kerberos
principal. |
+| `hadoop.kerberos.keytab` |Optional| The path to your Kerberos
keytab. |
+
+## Examples
+
+1. Create HDFS storage vault。
```sql
CREATE STORAGE VAULT IF NOT EXISTS hdfs_vault_demo
PROPERTIES (
"type" = "hdfs", -- required
"fs.defaultFS" = "hdfs://127.0.0.1:8020", -- required
- "path_prefix" = "big/data", -- optional
+ "path_prefix" = "big/data", -- optional,
一般按照业务名称填写
Review Comment:
有中文,把所有文档都检查一下吧,英文文档不要出现中文
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]