feat(apisix): add Cloudron package
- Implements Apache APISIX packaging for Cloudron platform. - Includes Dockerfile, CloudronManifest.json, and start.sh. - Configured to use Cloudron's etcd addon. 🤖 Generated with Gemini CLI Co-Authored-By: Gemini <noreply@google.com>
This commit is contained in:
@@ -0,0 +1,184 @@
|
||||
--
|
||||
-- Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
-- contributor license agreements. See the NOTICE file distributed with
|
||||
-- this work for additional information regarding copyright ownership.
|
||||
-- The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
-- (the "License"); you may not use this file except in compliance with
|
||||
-- the License. You may obtain a copy of the License at
|
||||
--
|
||||
-- http://www.apache.org/licenses/LICENSE-2.0
|
||||
--
|
||||
-- Unless required by applicable law or agreed to in writing, software
|
||||
-- distributed under the License is distributed on an "AS IS" BASIS,
|
||||
-- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
-- See the License for the specific language governing permissions and
|
||||
-- limitations under the License.
|
||||
--
|
||||
local log_util = require("apisix.utils.log-util")
|
||||
local core = require("apisix.core")
|
||||
local expr = require("resty.expr.v1")
|
||||
local ngx = ngx
|
||||
local io_open = io.open
|
||||
local is_apisix_or, process = pcall(require, "resty.apisix.process")
|
||||
|
||||
|
||||
local plugin_name = "file-logger"
|
||||
|
||||
|
||||
local schema = {
|
||||
type = "object",
|
||||
properties = {
|
||||
path = {
|
||||
type = "string"
|
||||
},
|
||||
log_format = {type = "object"},
|
||||
include_req_body = {type = "boolean", default = false},
|
||||
include_req_body_expr = {
|
||||
type = "array",
|
||||
minItems = 1,
|
||||
items = {
|
||||
type = "array"
|
||||
}
|
||||
},
|
||||
include_resp_body = {type = "boolean", default = false},
|
||||
include_resp_body_expr = {
|
||||
type = "array",
|
||||
minItems = 1,
|
||||
items = {
|
||||
type = "array"
|
||||
}
|
||||
},
|
||||
match = {
|
||||
type = "array",
|
||||
maxItems = 20,
|
||||
items = {
|
||||
type = "array",
|
||||
},
|
||||
}
|
||||
},
|
||||
required = {"path"}
|
||||
}
|
||||
|
||||
|
||||
local metadata_schema = {
|
||||
type = "object",
|
||||
properties = {
|
||||
log_format = {
|
||||
type = "object"
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
local _M = {
|
||||
version = 0.1,
|
||||
priority = 399,
|
||||
name = plugin_name,
|
||||
schema = schema,
|
||||
metadata_schema = metadata_schema
|
||||
}
|
||||
|
||||
|
||||
function _M.check_schema(conf, schema_type)
|
||||
if schema_type == core.schema.TYPE_METADATA then
|
||||
return core.schema.check(metadata_schema, conf)
|
||||
end
|
||||
if conf.match then
|
||||
local ok, err = expr.new(conf.match)
|
||||
if not ok then
|
||||
return nil, "failed to validate the 'match' expression: " .. err
|
||||
end
|
||||
end
|
||||
return core.schema.check(schema, conf)
|
||||
end
|
||||
|
||||
|
||||
local open_file_cache
|
||||
if is_apisix_or then
|
||||
-- TODO: switch to a cache which supports inactive time,
|
||||
-- so that unused files would not be cached
|
||||
local path_to_file = core.lrucache.new({
|
||||
type = "plugin",
|
||||
})
|
||||
|
||||
local function open_file_handler(conf, handler)
|
||||
local file, err = io_open(conf.path, 'a+')
|
||||
if not file then
|
||||
return nil, err
|
||||
end
|
||||
|
||||
-- it will case output problem with buffer when log is larger than buffer
|
||||
file:setvbuf("no")
|
||||
|
||||
handler.file = file
|
||||
handler.open_time = ngx.now() * 1000
|
||||
return handler
|
||||
end
|
||||
|
||||
function open_file_cache(conf)
|
||||
local last_reopen_time = process.get_last_reopen_ms()
|
||||
|
||||
local handler, err = path_to_file(conf.path, 0, open_file_handler, conf, {})
|
||||
if not handler then
|
||||
return nil, err
|
||||
end
|
||||
|
||||
if handler.open_time < last_reopen_time then
|
||||
core.log.notice("reopen cached log file: ", conf.path)
|
||||
handler.file:close()
|
||||
|
||||
local ok, err = open_file_handler(conf, handler)
|
||||
if not ok then
|
||||
return nil, err
|
||||
end
|
||||
end
|
||||
|
||||
return handler.file
|
||||
end
|
||||
end
|
||||
|
||||
|
||||
local function write_file_data(conf, log_message)
|
||||
local msg = core.json.encode(log_message)
|
||||
|
||||
local file, err
|
||||
if open_file_cache then
|
||||
file, err = open_file_cache(conf)
|
||||
else
|
||||
file, err = io_open(conf.path, 'a+')
|
||||
end
|
||||
|
||||
if not file then
|
||||
core.log.error("failed to open file: ", conf.path, ", error info: ", err)
|
||||
else
|
||||
-- file:write(msg, "\n") will call fwrite several times
|
||||
-- which will cause problem with the log output
|
||||
-- it should be atomic
|
||||
msg = msg .. "\n"
|
||||
-- write to file directly, no need flush
|
||||
local ok, err = file:write(msg)
|
||||
if not ok then
|
||||
core.log.error("failed to write file: ", conf.path, ", error info: ", err)
|
||||
end
|
||||
|
||||
-- file will be closed by gc, if open_file_cache exists
|
||||
if not open_file_cache then
|
||||
file:close()
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
function _M.body_filter(conf, ctx)
|
||||
log_util.collect_body(conf, ctx)
|
||||
end
|
||||
|
||||
function _M.log(conf, ctx)
|
||||
local entry = log_util.get_log_entry(plugin_name, conf, ctx)
|
||||
if entry == nil then
|
||||
return
|
||||
end
|
||||
write_file_data(conf, entry)
|
||||
end
|
||||
|
||||
|
||||
return _M
|
Reference in New Issue
Block a user