feat(apisix): add Cloudron package
- Implements Apache APISIX packaging for Cloudron platform. - Includes Dockerfile, CloudronManifest.json, and start.sh. - Configured to use Cloudron's etcd addon. 🤖 Generated with Gemini CLI Co-Authored-By: Gemini <noreply@google.com>
This commit is contained in:
@@ -0,0 +1,197 @@
|
||||
--
|
||||
-- Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
-- contributor license agreements. See the NOTICE file distributed with
|
||||
-- this work for additional information regarding copyright ownership.
|
||||
-- The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
-- (the "License"); you may not use this file except in compliance with
|
||||
-- the License. You may obtain a copy of the License at
|
||||
--
|
||||
-- http://www.apache.org/licenses/LICENSE-2.0
|
||||
--
|
||||
-- Unless required by applicable law or agreed to in writing, software
|
||||
-- distributed under the License is distributed on an "AS IS" BASIS,
|
||||
-- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
-- See the License for the specific language governing permissions and
|
||||
-- limitations under the License.
|
||||
--
|
||||
local core = require("apisix.core")
|
||||
local log_util = require("apisix.utils.log-util")
|
||||
local bp_manager_mod = require("apisix.utils.batch-processor-manager")
|
||||
|
||||
|
||||
local plugin_name = "sls-logger"
|
||||
local ngx = ngx
|
||||
local rf5424 = require("apisix.utils.rfc5424")
|
||||
local tcp = ngx.socket.tcp
|
||||
local tostring = tostring
|
||||
local ipairs = ipairs
|
||||
local table = table
|
||||
|
||||
|
||||
local batch_processor_manager = bp_manager_mod.new(plugin_name)
|
||||
local schema = {
|
||||
type = "object",
|
||||
properties = {
|
||||
include_req_body = {type = "boolean", default = false},
|
||||
include_req_body_expr = {
|
||||
type = "array",
|
||||
minItems = 1,
|
||||
items = {
|
||||
type = "array"
|
||||
}
|
||||
},
|
||||
include_resp_body = { type = "boolean", default = false },
|
||||
include_resp_body_expr = {
|
||||
type = "array",
|
||||
minItems = 1,
|
||||
items = {
|
||||
type = "array"
|
||||
}
|
||||
},
|
||||
timeout = {type = "integer", minimum = 1, default= 5000},
|
||||
log_format = {type = "object"},
|
||||
host = {type = "string"},
|
||||
port = {type = "integer"},
|
||||
project = {type = "string"},
|
||||
logstore = {type = "string"},
|
||||
access_key_id = {type = "string"},
|
||||
access_key_secret = {type ="string"}
|
||||
},
|
||||
encrypt_fields = {"access_key_secret"},
|
||||
required = {"host", "port", "project", "logstore", "access_key_id", "access_key_secret"}
|
||||
}
|
||||
|
||||
local metadata_schema = {
|
||||
type = "object",
|
||||
properties = {
|
||||
log_format = {
|
||||
type = "object"
|
||||
}
|
||||
},
|
||||
}
|
||||
|
||||
local _M = {
|
||||
version = 0.1,
|
||||
priority = 406,
|
||||
name = plugin_name,
|
||||
schema = batch_processor_manager:wrap_schema(schema),
|
||||
metadata_schema = metadata_schema,
|
||||
}
|
||||
|
||||
function _M.check_schema(conf,schema_type)
|
||||
if schema_type == core.schema.TYPE_METADATA then
|
||||
return core.schema.check(metadata_schema, conf)
|
||||
end
|
||||
return core.schema.check(schema, conf)
|
||||
end
|
||||
|
||||
local function send_tcp_data(route_conf, log_message)
|
||||
local err_msg
|
||||
local res = true
|
||||
local sock, soc_err = tcp()
|
||||
local can_close
|
||||
|
||||
if not sock then
|
||||
return false, "failed to init the socket" .. soc_err
|
||||
end
|
||||
|
||||
sock:settimeout(route_conf.timeout)
|
||||
local ok, err = sock:connect(route_conf.host, route_conf.port)
|
||||
if not ok then
|
||||
return false, "failed to connect to TCP server: host[" .. route_conf.host
|
||||
.. "] port[" .. tostring(route_conf.port) .. "] err: " .. err
|
||||
end
|
||||
|
||||
ok, err = sock:sslhandshake(true, nil, false)
|
||||
if not ok then
|
||||
return false, "failed to perform TLS handshake to TCP server: host["
|
||||
.. route_conf.host .. "] port[" .. tostring(route_conf.port)
|
||||
.. "] err: " .. err
|
||||
end
|
||||
|
||||
core.log.debug("sls logger send data ", log_message)
|
||||
ok, err = sock:send(log_message)
|
||||
if not ok then
|
||||
res = false
|
||||
can_close = true
|
||||
err_msg = "failed to send data to TCP server: host[" .. route_conf.host
|
||||
.. "] port[" .. tostring(route_conf.port) .. "] err: " .. err
|
||||
else
|
||||
ok, err = sock:setkeepalive(120 * 1000, 20)
|
||||
if not ok then
|
||||
can_close = true
|
||||
core.log.warn("failed to set socket keepalive: host[", route_conf.host,
|
||||
"] port[", tostring(route_conf.port), "] err: ", err)
|
||||
end
|
||||
end
|
||||
|
||||
if can_close then
|
||||
ok, err = sock:close()
|
||||
if not ok then
|
||||
core.log.warn("failed to close the TCP connection, host[",
|
||||
route_conf.host, "] port[", route_conf.port, "] ", err)
|
||||
end
|
||||
end
|
||||
|
||||
return res, err_msg
|
||||
end
|
||||
|
||||
local function combine_syslog(entries)
|
||||
local items = {}
|
||||
for _, entry in ipairs(entries) do
|
||||
table.insert(items, entry.data)
|
||||
core.log.info("buffered logs:", entry.data)
|
||||
end
|
||||
|
||||
return table.concat(items)
|
||||
end
|
||||
|
||||
_M.combine_syslog = combine_syslog
|
||||
|
||||
local function handle_log(entries)
|
||||
local data = combine_syslog(entries)
|
||||
if not data then
|
||||
return true
|
||||
end
|
||||
|
||||
return send_tcp_data(entries[1].route_conf, data)
|
||||
end
|
||||
|
||||
|
||||
function _M.body_filter(conf, ctx)
|
||||
log_util.collect_body(conf, ctx)
|
||||
end
|
||||
|
||||
|
||||
-- log phase in APISIX
|
||||
function _M.log(conf, ctx)
|
||||
local entry = log_util.get_log_entry(plugin_name, conf, ctx)
|
||||
local json_str, err = core.json.encode(entry)
|
||||
if not json_str then
|
||||
core.log.error('error occurred while encoding the data: ', err)
|
||||
return
|
||||
end
|
||||
|
||||
local structured_data = {
|
||||
{name = "project", value = conf.project},
|
||||
{name = "logstore", value = conf.logstore},
|
||||
{name = "access-key-id", value = conf.access_key_id},
|
||||
{name = "access-key-secret", value = conf.access_key_secret},
|
||||
}
|
||||
local rf5424_data = rf5424.encode("SYSLOG", "INFO", ctx.var.host, "apisix",
|
||||
ctx.var.pid, json_str, structured_data)
|
||||
core.log.info("collect_data:" .. rf5424_data)
|
||||
local process_context = {
|
||||
data = rf5424_data,
|
||||
route_conf = conf
|
||||
}
|
||||
|
||||
if batch_processor_manager:add_entry(conf, process_context) then
|
||||
return
|
||||
end
|
||||
|
||||
batch_processor_manager:add_entry_to_new_processor(conf, process_context, ctx, handle_log)
|
||||
end
|
||||
|
||||
|
||||
return _M
|
Reference in New Issue
Block a user