- Implements Apache APISIX packaging for Cloudron platform. - Includes Dockerfile, CloudronManifest.json, and start.sh. - Configured to use Cloudron's etcd addon. 🤖 Generated with Gemini CLI Co-Authored-By: Gemini <noreply@google.com>
361 lines
12 KiB
Perl
361 lines
12 KiB
Perl
#
|
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
|
# contributor license agreements. See the NOTICE file distributed with
|
|
# this work for additional information regarding copyright ownership.
|
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
|
# (the "License"); you may not use this file except in compliance with
|
|
# the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
use t::APISIX 'no_plan';
|
|
|
|
log_level("info");
|
|
repeat_each(1);
|
|
no_long_string();
|
|
no_root_location();
|
|
|
|
|
|
my $resp_file = 't/assets/ai-proxy-response.json';
|
|
open(my $fh, '<', $resp_file) or die "Could not open file '$resp_file' $!";
|
|
my $resp = do { local $/; <$fh> };
|
|
close($fh);
|
|
|
|
print "Hello, World!\n";
|
|
print $resp;
|
|
|
|
|
|
add_block_preprocessor(sub {
|
|
my ($block) = @_;
|
|
|
|
if (!defined $block->request) {
|
|
$block->set_value("request", "GET /t");
|
|
}
|
|
|
|
my $user_yaml_config = <<_EOC_;
|
|
plugins:
|
|
- ai-proxy-multi
|
|
_EOC_
|
|
$block->set_value("extra_yaml_config", $user_yaml_config);
|
|
|
|
my $http_config = $block->http_config // <<_EOC_;
|
|
server {
|
|
server_name openai;
|
|
listen 6724;
|
|
|
|
default_type 'application/json';
|
|
|
|
location /v1/chat/completions {
|
|
content_by_lua_block {
|
|
local json = require("cjson.safe")
|
|
|
|
if ngx.req.get_method() ~= "POST" then
|
|
ngx.status = 400
|
|
ngx.say("Unsupported request method: ", ngx.req.get_method())
|
|
end
|
|
ngx.req.read_body()
|
|
local body, err = ngx.req.get_body_data()
|
|
body, err = json.decode(body)
|
|
|
|
local header_auth = ngx.req.get_headers()["authorization"]
|
|
local query_auth = ngx.req.get_uri_args()["apikey"]
|
|
|
|
if header_auth ~= "Bearer token" and query_auth ~= "apikey" then
|
|
ngx.status = 401
|
|
ngx.say("Unauthorized")
|
|
return
|
|
end
|
|
|
|
if header_auth == "Bearer token" or query_auth == "apikey" then
|
|
ngx.req.read_body()
|
|
local body, err = ngx.req.get_body_data()
|
|
body, err = json.decode(body)
|
|
|
|
if not body.messages or #body.messages < 1 then
|
|
ngx.status = 400
|
|
ngx.say([[{ "error": "bad request"}]])
|
|
return
|
|
end
|
|
|
|
ngx.status = 200
|
|
ngx.print("openai")
|
|
return
|
|
end
|
|
|
|
|
|
ngx.status = 503
|
|
ngx.say("reached the end of the test suite")
|
|
}
|
|
}
|
|
|
|
location /chat/completions {
|
|
content_by_lua_block {
|
|
local json = require("cjson.safe")
|
|
|
|
if ngx.req.get_method() ~= "POST" then
|
|
ngx.status = 400
|
|
ngx.say("Unsupported request method: ", ngx.req.get_method())
|
|
end
|
|
ngx.req.read_body()
|
|
local body, err = ngx.req.get_body_data()
|
|
body, err = json.decode(body)
|
|
|
|
local header_auth = ngx.req.get_headers()["authorization"]
|
|
local query_auth = ngx.req.get_uri_args()["apikey"]
|
|
|
|
if header_auth ~= "Bearer token" and query_auth ~= "apikey" then
|
|
ngx.status = 401
|
|
ngx.say("Unauthorized")
|
|
return
|
|
end
|
|
|
|
if header_auth == "Bearer token" or query_auth == "apikey" then
|
|
ngx.req.read_body()
|
|
local body, err = ngx.req.get_body_data()
|
|
body, err = json.decode(body)
|
|
|
|
if not body.messages or #body.messages < 1 then
|
|
ngx.status = 400
|
|
ngx.say([[{ "error": "bad request"}]])
|
|
return
|
|
end
|
|
|
|
ngx.status = 200
|
|
ngx.print("deepseek")
|
|
return
|
|
end
|
|
|
|
|
|
ngx.status = 503
|
|
ngx.say("reached the end of the test suite")
|
|
}
|
|
}
|
|
}
|
|
_EOC_
|
|
|
|
$block->set_value("http_config", $http_config);
|
|
});
|
|
|
|
run_tests();
|
|
|
|
__DATA__
|
|
|
|
=== TEST 1: set route with roundrobin balancer, weight 4 and 1
|
|
--- config
|
|
location /t {
|
|
content_by_lua_block {
|
|
local t = require("lib.test_admin").test
|
|
local code, body = t('/apisix/admin/routes/1',
|
|
ngx.HTTP_PUT,
|
|
[[{
|
|
"uri": "/anything",
|
|
"plugins": {
|
|
"ai-proxy-multi": {
|
|
"instances": [
|
|
{
|
|
"name": "openai",
|
|
"provider": "openai",
|
|
"weight": 4,
|
|
"auth": {
|
|
"header": {
|
|
"Authorization": "Bearer token"
|
|
}
|
|
},
|
|
"options": {
|
|
"model": "gpt-4",
|
|
"max_tokens": 512,
|
|
"temperature": 1.0
|
|
},
|
|
"override": {
|
|
"endpoint": "http://localhost:6724"
|
|
}
|
|
},
|
|
{
|
|
"name": "deepseek",
|
|
"provider": "deepseek",
|
|
"weight": 1,
|
|
"auth": {
|
|
"header": {
|
|
"Authorization": "Bearer token"
|
|
}
|
|
},
|
|
"options": {
|
|
"model": "deepseek-chat",
|
|
"max_tokens": 512,
|
|
"temperature": 1.0
|
|
},
|
|
"override": {
|
|
"endpoint": "http://localhost:6724/chat/completions"
|
|
}
|
|
}
|
|
],
|
|
"ssl_verify": false
|
|
}
|
|
}
|
|
}]]
|
|
)
|
|
|
|
if code >= 300 then
|
|
ngx.status = code
|
|
end
|
|
ngx.say(body)
|
|
}
|
|
}
|
|
--- response_body
|
|
passed
|
|
|
|
|
|
|
|
=== TEST 2: test
|
|
--- config
|
|
location /t {
|
|
content_by_lua_block {
|
|
local http = require "resty.http"
|
|
local uri = "http://127.0.0.1:" .. ngx.var.server_port
|
|
.. "/anything"
|
|
|
|
local restab = {}
|
|
|
|
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
|
|
for i = 1, 10 do
|
|
local httpc = http.new()
|
|
local res, err = httpc:request_uri(uri, {method = "POST", body = body})
|
|
if not res then
|
|
ngx.say(err)
|
|
return
|
|
end
|
|
table.insert(restab, res.body)
|
|
end
|
|
|
|
table.sort(restab)
|
|
ngx.log(ngx.WARN, "test picked instances: ", table.concat(restab, "."))
|
|
|
|
}
|
|
}
|
|
--- request
|
|
GET /t
|
|
--- error_log
|
|
deepseek.deepseek.openai.openai.openai.openai.openai.openai.openai.openai
|
|
|
|
|
|
|
|
=== TEST 3: set route with chash balancer, weight 4 and 1
|
|
--- config
|
|
location /t {
|
|
content_by_lua_block {
|
|
local t = require("lib.test_admin").test
|
|
local code, body = t('/apisix/admin/routes/1',
|
|
ngx.HTTP_PUT,
|
|
[[{
|
|
"uri": "/anything",
|
|
"plugins": {
|
|
"ai-proxy-multi": {
|
|
"balancer": {
|
|
"algorithm": "chash",
|
|
"hash_on": "vars",
|
|
"key": "query_string"
|
|
},
|
|
"instances": [
|
|
{
|
|
"name": "openai",
|
|
"provider": "openai",
|
|
"weight": 4,
|
|
"auth": {
|
|
"header": {
|
|
"Authorization": "Bearer token"
|
|
}
|
|
},
|
|
"options": {
|
|
"model": "gpt-4",
|
|
"max_tokens": 512,
|
|
"temperature": 1.0
|
|
},
|
|
"override": {
|
|
"endpoint": "http://localhost:6724"
|
|
}
|
|
},
|
|
{
|
|
"name": "deepseek",
|
|
"provider": "deepseek",
|
|
"weight": 1,
|
|
"auth": {
|
|
"header": {
|
|
"Authorization": "Bearer token"
|
|
}
|
|
},
|
|
"options": {
|
|
"model": "deepseek-chat",
|
|
"max_tokens": 512,
|
|
"temperature": 1.0
|
|
},
|
|
"override": {
|
|
"endpoint": "http://localhost:6724/chat/completions"
|
|
}
|
|
}
|
|
],
|
|
"ssl_verify": false
|
|
}
|
|
}
|
|
}]]
|
|
)
|
|
|
|
if code >= 300 then
|
|
ngx.status = code
|
|
end
|
|
ngx.say(body)
|
|
}
|
|
}
|
|
--- response_body
|
|
passed
|
|
|
|
|
|
|
|
=== TEST 4: test
|
|
--- config
|
|
location /t {
|
|
content_by_lua_block {
|
|
local http = require "resty.http"
|
|
local uri = "http://127.0.0.1:" .. ngx.var.server_port
|
|
.. "/anything"
|
|
|
|
local restab = {}
|
|
|
|
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
|
|
for i = 1, 10 do
|
|
local httpc = http.new()
|
|
local query = {
|
|
index = i
|
|
}
|
|
local res, err = httpc:request_uri(uri, {method = "POST", body = body, query = query})
|
|
if not res then
|
|
ngx.say(err)
|
|
return
|
|
end
|
|
table.insert(restab, res.body)
|
|
end
|
|
|
|
local count = {}
|
|
for _, value in ipairs(restab) do
|
|
count[value] = (count[value] or 0) + 1
|
|
end
|
|
|
|
for p, num in pairs(count) do
|
|
ngx.log(ngx.WARN, "distribution: ", p, ": ", num)
|
|
end
|
|
|
|
}
|
|
}
|
|
--- request
|
|
GET /t
|
|
--- timeout: 10
|
|
--- error_log
|
|
distribution: deepseek: 2
|
|
distribution: openai: 8
|