fix(doc): fix some dead link (#2675)

This commit is contained in:
aias00
2025-07-30 21:22:15 +08:00
committed by GitHub
parent 32b5c89c17
commit 1070541f1d
4 changed files with 4 additions and 4 deletions

View File

@@ -147,5 +147,5 @@ curl -X POST \
- 流模式中如果脱敏后的词被多个chunk拆分可能无法进行还原
- 流模式中如果敏感词语被多个chunk拆分可能会有敏感词的一部分返回给用户的情况
- grok 内置规则列表 https://help.aliyun.com/zh/sls/user-guide/grok-patterns
- 内置敏感词库数据来源 https://github.com/houbb/sensitive-word/tree/master/src/main/resources
- 内置敏感词库数据来源 https://github.com/houbb/sensitive-word-data/tree/main/src/main/resources
- 由于敏感词列表是在文本分词后进行匹配的,所以请将 `deny_words` 设置为单个单词,英文多单词情况如 `hello word` 可能无法匹配

View File

@@ -128,5 +128,5 @@ Please note that you need to replace `"key":"value"` with the actual data conten
- In streaming mode, if the masked words are split across multiple chunks, restoration may not be possible
- In streaming mode, if sensitive words are split across multiple chunks, there may be cases where part of the sensitive word is returned to the user
- Grok built-in rule list: https://help.aliyun.com/zh/sls/user-guide/grok-patterns
- Built-in sensitive word library data source: https://github.com/houbb/sensitive-word/tree/master/src/main/resources
- Built-in sensitive word library data source: https://github.com/houbb/sensitive-word-data/tree/main/src/main/resources
- Since the sensitive word list is matched after tokenizing the text, please set `deny_words` to single words. In the case of multiple words in English, such as `hello world`, the match may not be successful.