# Copyright (c) 2025 Alibaba Group Holding Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-ai360 namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.360.cn" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-baichuan namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.baichuan-ai.com" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-baidu namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "qianfan.baidubce.com" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-deepseek namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.deepseek.com" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-doubao namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "ark.cn-beijing.volces.com" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-github namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "models.inference.ai.azure.com" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-groq namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.groq.com" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-minimax-v2-api namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.minimax.chat-v2-api" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-minimax-pro-api namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.minimax.chat-pro-api" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-mistral namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.mistral.ai" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-qwen-compatible-mode namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "dashscope.aliyuncs.com-compatible-mode" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-qwen namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "dashscope.aliyuncs.com" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-stepfun namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.stepfun.com" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-together-ai namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.together.xyz" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-yi namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.lingyiwanwu.com" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-zhipuai namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "open.bigmodel.cn" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: wasmplugin-ai-proxy-dify namespace: higress-conformance-ai-backend spec: ingressClassName: higress rules: - host: "api.dify.ai" http: paths: - pathType: Prefix path: "/" backend: service: name: llm-mock-service port: number: 3000 --- apiVersion: extensions.higress.io/v1alpha1 kind: WasmPlugin metadata: name: ai-proxy namespace: higress-system spec: defaultConfigDisable: true phase: UNSPECIFIED_PHASE priority: 100 matchRules: - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': 360gpt-turbo '*': 360gpt-pro type: ai360 ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-ai360 - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': baichuan2-13b-chat-v1 '*': baichuan-7b-v1 type: baichuan ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-baichuan - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': ernie-3.5-8k '*': ernie-3.5-8k type: baidu ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-baidu - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': deepseek-reasoner '*': deepseek-chat type: deepseek ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-deepseek - config: provider: apiTokens: - fake_token modelMapping: '*': fake_doubao_endpoint type: doubao ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-doubao - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': cohere-command-r-08-2024 '*': Phi-3.5-MoE-instruct type: github ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-github - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': llama3-8b-8192 '*': llama-3.1-8b-instant type: groq ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-groq - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': abab6.5s-chat 'gpt-4': abab6.5g-chat '*': abab6.5t-chat type: minimax ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-minimax-v2-api - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': abab6.5s-chat 'gpt-4': abab6.5g-chat '*': abab6.5t-chat type: minimax minimaxApiType: pro minimaxGroupId: 1 ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-minimax-pro-api - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': mistral-tiny '*': mistral-large-latest type: mistral ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-mistral - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': qwen-turbo 'gpt-35-turbo': qwen-plus 'gpt-4-*': qwen-max '*': qwen-turbo type: qwen qwenEnableCompatible: true ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-qwen-compatible-mode - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': qwen-turbo 'gpt-35-turbo': qwen-plus 'gpt-4-*': qwen-max '*': qwen-turbo type: qwen ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-qwen - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': step-1-8k '*': step-1-32k type: stepfun ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-stepfun - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': meta-llama/Meta-Llama-3-8B-Instruct-Turbo '*': meta-llama/Llama-3-8b-chat-hf type: together-ai ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-together-ai - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': Yi-Medium '*': Yi-Large type: yi ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-yi - config: provider: apiTokens: - fake_token modelMapping: 'gpt-3': glm-4-plus '*': glm-4-long type: zhipuai ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-zhipuai - config: provider: apiTokens: - fake_token modelMapping: '*': dify type: dify botType: Completion ingress: - higress-conformance-ai-backend/wasmplugin-ai-proxy-dify url: file:///opt/plugins/wasm-go/extensions/ai-proxy/plugin.wasm