mirror of
https://github.com/alibaba/higress.git
synced 2026-02-06 23:21:08 +08:00
556 lines
14 KiB
YAML
556 lines
14 KiB
YAML
# Copyright (c) 2025 Alibaba Group Holding Ltd.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-ai360
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.360.cn"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-baichuan
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.baichuan-ai.com"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-baidu
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "qianfan.baidubce.com"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-deepseek
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.deepseek.com"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-doubao
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "ark.cn-beijing.volces.com"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-github
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "models.inference.ai.azure.com"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-groq
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.groq.com"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-minimax-v2-api
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.minimax.chat-v2-api"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-minimax-pro-api
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.minimax.chat-pro-api"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-mistral
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.mistral.ai"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-qwen-compatible-mode
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "dashscope.aliyuncs.com-compatible-mode"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-qwen
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "dashscope.aliyuncs.com"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-stepfun
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.stepfun.com"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-together-ai
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.together.xyz"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-yi
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.lingyiwanwu.com"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-zhipuai
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "open.bigmodel.cn"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-dify
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.dify.ai"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: networking.k8s.io/v1
|
|
kind: Ingress
|
|
metadata:
|
|
name: wasmplugin-ai-proxy-grok
|
|
namespace: higress-conformance-ai-backend
|
|
spec:
|
|
ingressClassName: higress
|
|
rules:
|
|
- host: "api.x.ai"
|
|
http:
|
|
paths:
|
|
- pathType: Prefix
|
|
path: "/"
|
|
backend:
|
|
service:
|
|
name: llm-mock-service
|
|
port:
|
|
number: 3000
|
|
---
|
|
apiVersion: extensions.higress.io/v1alpha1
|
|
kind: WasmPlugin
|
|
metadata:
|
|
name: ai-proxy
|
|
namespace: higress-system
|
|
spec:
|
|
defaultConfigDisable: true
|
|
phase: UNSPECIFIED_PHASE
|
|
priority: 100
|
|
matchRules:
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": 360gpt-turbo
|
|
"*": 360gpt-pro
|
|
type: ai360
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-ai360
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": baichuan2-13b-chat-v1
|
|
"*": baichuan-7b-v1
|
|
type: baichuan
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-baichuan
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": ernie-3.5-8k
|
|
"*": ernie-3.5-8k
|
|
type: baidu
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-baidu
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": deepseek-reasoner
|
|
"*": deepseek-chat
|
|
type: deepseek
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-deepseek
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"*": fake_doubao_endpoint
|
|
type: doubao
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-doubao
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": cohere-command-r-08-2024
|
|
"*": Phi-3.5-MoE-instruct
|
|
type: github
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-github
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": llama3-8b-8192
|
|
"*": llama-3.1-8b-instant
|
|
type: groq
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-groq
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": abab6.5s-chat
|
|
"gpt-4": abab6.5g-chat
|
|
"*": abab6.5t-chat
|
|
type: minimax
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-minimax-v2-api
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": abab6.5s-chat
|
|
"gpt-4": abab6.5g-chat
|
|
"*": abab6.5t-chat
|
|
type: minimax
|
|
minimaxApiType: pro
|
|
minimaxGroupId: 1
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-minimax-pro-api
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": mistral-tiny
|
|
"*": mistral-large-latest
|
|
type: mistral
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-mistral
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": qwen-turbo
|
|
"gpt-35-turbo": qwen-plus
|
|
"gpt-4-*": qwen-max
|
|
"*": qwen-turbo
|
|
type: qwen
|
|
qwenEnableCompatible: true
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-qwen-compatible-mode
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": qwen-turbo
|
|
"gpt-35-turbo": qwen-plus
|
|
"gpt-4-*": qwen-max
|
|
"*": qwen-turbo
|
|
type: qwen
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-qwen
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": step-1-8k
|
|
"*": step-1-32k
|
|
type: stepfun
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-stepfun
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": meta-llama/Meta-Llama-3-8B-Instruct-Turbo
|
|
"*": meta-llama/Llama-3-8b-chat-hf
|
|
type: together-ai
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-together-ai
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": Yi-Medium
|
|
"*": Yi-Large
|
|
type: yi
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-yi
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": glm-4-plus
|
|
"*": glm-4-long
|
|
type: zhipuai
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-zhipuai
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"*": dify
|
|
type: dify
|
|
botType: Completion
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-dify
|
|
- config:
|
|
provider:
|
|
apiTokens:
|
|
- fake_token
|
|
modelMapping:
|
|
"gpt-3": grok-beta
|
|
"gpt-4": grok-beta
|
|
"*": grok-beta
|
|
type: grok
|
|
ingress:
|
|
- higress-conformance-ai-backend/wasmplugin-ai-proxy-grok
|
|
url: file:///opt/plugins/wasm-go/extensions/ai-proxy/plugin.wasm
|