feat(apisix): add Cloudron package
- Implements Apache APISIX packaging for Cloudron platform. - Includes Dockerfile, CloudronManifest.json, and start.sh. - Configured to use Cloudron's etcd addon. 🤖 Generated with Gemini CLI Co-Authored-By: Gemini <noreply@google.com>
This commit is contained in:
@@ -0,0 +1,327 @@
|
||||
--
|
||||
-- Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
-- contributor license agreements. See the NOTICE file distributed with
|
||||
-- this work for additional information regarding copyright ownership.
|
||||
-- The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
-- (the "License"); you may not use this file except in compliance with
|
||||
-- the License. You may obtain a copy of the License at
|
||||
--
|
||||
-- http://www.apache.org/licenses/LICENSE-2.0
|
||||
--
|
||||
-- Unless required by applicable law or agreed to in writing, software
|
||||
-- distributed under the License is distributed on an "AS IS" BASIS,
|
||||
-- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
-- See the License for the specific language governing permissions and
|
||||
-- limitations under the License.
|
||||
--
|
||||
local expr = require("resty.expr.v1")
|
||||
local core = require("apisix.core")
|
||||
local log_util = require("apisix.utils.log-util")
|
||||
local producer = require ("resty.kafka.producer")
|
||||
local bp_manager_mod = require("apisix.utils.batch-processor-manager")
|
||||
local plugin = require("apisix.plugin")
|
||||
|
||||
local math = math
|
||||
local pairs = pairs
|
||||
local type = type
|
||||
local req_read_body = ngx.req.read_body
|
||||
local plugin_name = "kafka-logger"
|
||||
local batch_processor_manager = bp_manager_mod.new("kafka logger")
|
||||
|
||||
local lrucache = core.lrucache.new({
|
||||
type = "plugin",
|
||||
})
|
||||
|
||||
local schema = {
|
||||
type = "object",
|
||||
properties = {
|
||||
meta_format = {
|
||||
type = "string",
|
||||
default = "default",
|
||||
enum = {"default", "origin"},
|
||||
},
|
||||
log_format = {type = "object"},
|
||||
-- deprecated, use "brokers" instead
|
||||
broker_list = {
|
||||
type = "object",
|
||||
minProperties = 1,
|
||||
patternProperties = {
|
||||
[".*"] = {
|
||||
description = "the port of kafka broker",
|
||||
type = "integer",
|
||||
minimum = 1,
|
||||
maximum = 65535,
|
||||
},
|
||||
},
|
||||
},
|
||||
brokers = {
|
||||
type = "array",
|
||||
minItems = 1,
|
||||
items = {
|
||||
type = "object",
|
||||
properties = {
|
||||
host = {
|
||||
type = "string",
|
||||
description = "the host of kafka broker",
|
||||
},
|
||||
port = {
|
||||
type = "integer",
|
||||
minimum = 1,
|
||||
maximum = 65535,
|
||||
description = "the port of kafka broker",
|
||||
},
|
||||
sasl_config = {
|
||||
type = "object",
|
||||
description = "sasl config",
|
||||
properties = {
|
||||
mechanism = {
|
||||
type = "string",
|
||||
default = "PLAIN",
|
||||
enum = {"PLAIN"},
|
||||
},
|
||||
user = { type = "string", description = "user" },
|
||||
password = { type = "string", description = "password" },
|
||||
},
|
||||
required = {"user", "password"},
|
||||
},
|
||||
},
|
||||
required = {"host", "port"},
|
||||
},
|
||||
uniqueItems = true,
|
||||
},
|
||||
kafka_topic = {type = "string"},
|
||||
producer_type = {
|
||||
type = "string",
|
||||
default = "async",
|
||||
enum = {"async", "sync"},
|
||||
},
|
||||
required_acks = {
|
||||
type = "integer",
|
||||
default = 1,
|
||||
enum = { 1, -1 },
|
||||
},
|
||||
key = {type = "string"},
|
||||
timeout = {type = "integer", minimum = 1, default = 3},
|
||||
include_req_body = {type = "boolean", default = false},
|
||||
include_req_body_expr = {
|
||||
type = "array",
|
||||
minItems = 1,
|
||||
items = {
|
||||
type = "array"
|
||||
}
|
||||
},
|
||||
include_resp_body = {type = "boolean", default = false},
|
||||
include_resp_body_expr = {
|
||||
type = "array",
|
||||
minItems = 1,
|
||||
items = {
|
||||
type = "array"
|
||||
}
|
||||
},
|
||||
max_req_body_bytes = {type = "integer", minimum = 1, default = 524288},
|
||||
max_resp_body_bytes = {type = "integer", minimum = 1, default = 524288},
|
||||
-- in lua-resty-kafka, cluster_name is defined as number
|
||||
-- see https://github.com/doujiang24/lua-resty-kafka#new-1
|
||||
cluster_name = {type = "integer", minimum = 1, default = 1},
|
||||
-- config for lua-resty-kafka, default value is same as lua-resty-kafka
|
||||
producer_batch_num = {type = "integer", minimum = 1, default = 200},
|
||||
producer_batch_size = {type = "integer", minimum = 0, default = 1048576},
|
||||
producer_max_buffering = {type = "integer", minimum = 1, default = 50000},
|
||||
producer_time_linger = {type = "integer", minimum = 1, default = 1},
|
||||
meta_refresh_interval = {type = "integer", minimum = 1, default = 30},
|
||||
},
|
||||
oneOf = {
|
||||
{ required = {"broker_list", "kafka_topic"},},
|
||||
{ required = {"brokers", "kafka_topic"},},
|
||||
}
|
||||
}
|
||||
|
||||
local metadata_schema = {
|
||||
type = "object",
|
||||
properties = {
|
||||
log_format = {
|
||||
type = "object"
|
||||
},
|
||||
max_pending_entries = {
|
||||
type = "integer",
|
||||
description = "maximum number of pending entries in the batch processor",
|
||||
minimum = 1,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
local _M = {
|
||||
version = 0.1,
|
||||
priority = 403,
|
||||
name = plugin_name,
|
||||
schema = batch_processor_manager:wrap_schema(schema),
|
||||
metadata_schema = metadata_schema,
|
||||
}
|
||||
|
||||
|
||||
function _M.check_schema(conf, schema_type)
|
||||
if schema_type == core.schema.TYPE_METADATA then
|
||||
return core.schema.check(metadata_schema, conf)
|
||||
end
|
||||
|
||||
local ok, err = core.schema.check(schema, conf)
|
||||
if not ok then
|
||||
return nil, err
|
||||
end
|
||||
return log_util.check_log_schema(conf)
|
||||
end
|
||||
|
||||
|
||||
local function get_partition_id(prod, topic, log_message)
|
||||
if prod.async then
|
||||
local ringbuffer = prod.ringbuffer
|
||||
for i = 1, ringbuffer.size, 3 do
|
||||
if ringbuffer.queue[i] == topic and
|
||||
ringbuffer.queue[i+2] == log_message then
|
||||
return math.floor(i / 3)
|
||||
end
|
||||
end
|
||||
core.log.info("current topic in ringbuffer has no message")
|
||||
return nil
|
||||
end
|
||||
|
||||
-- sync mode
|
||||
local sendbuffer = prod.sendbuffer
|
||||
if not sendbuffer.topics[topic] then
|
||||
core.log.info("current topic in sendbuffer has no message")
|
||||
return nil
|
||||
end
|
||||
for i, message in pairs(sendbuffer.topics[topic]) do
|
||||
if log_message == message.queue[2] then
|
||||
return i
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
|
||||
local function create_producer(broker_list, broker_config, cluster_name)
|
||||
core.log.info("create new kafka producer instance")
|
||||
return producer:new(broker_list, broker_config, cluster_name)
|
||||
end
|
||||
|
||||
|
||||
local function send_kafka_data(conf, log_message, prod)
|
||||
local ok, err = prod:send(conf.kafka_topic, conf.key, log_message)
|
||||
core.log.info("partition_id: ",
|
||||
core.log.delay_exec(get_partition_id,
|
||||
prod, conf.kafka_topic, log_message))
|
||||
|
||||
if not ok then
|
||||
return false, "failed to send data to Kafka topic: " .. err ..
|
||||
", brokers: " .. core.json.encode(conf.broker_list)
|
||||
end
|
||||
|
||||
return true
|
||||
end
|
||||
|
||||
|
||||
function _M.access(conf, ctx)
|
||||
if conf.include_req_body then
|
||||
local should_read_body = true
|
||||
if conf.include_req_body_expr then
|
||||
if not conf.request_expr then
|
||||
local request_expr, err = expr.new(conf.include_req_body_expr)
|
||||
if not request_expr then
|
||||
core.log.error('generate request expr err ', err)
|
||||
return
|
||||
end
|
||||
conf.request_expr = request_expr
|
||||
end
|
||||
|
||||
local result = conf.request_expr:eval(ctx.var)
|
||||
|
||||
if not result then
|
||||
should_read_body = false
|
||||
end
|
||||
end
|
||||
if should_read_body then
|
||||
req_read_body()
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
|
||||
function _M.body_filter(conf, ctx)
|
||||
log_util.collect_body(conf, ctx)
|
||||
end
|
||||
|
||||
|
||||
function _M.log(conf, ctx)
|
||||
local metadata = plugin.plugin_metadata(plugin_name)
|
||||
local max_pending_entries = metadata and metadata.value and
|
||||
metadata.value.max_pending_entries or nil
|
||||
local entry
|
||||
if conf.meta_format == "origin" then
|
||||
entry = log_util.get_req_original(ctx, conf)
|
||||
-- core.log.info("origin entry: ", entry)
|
||||
|
||||
else
|
||||
entry = log_util.get_log_entry(plugin_name, conf, ctx)
|
||||
end
|
||||
|
||||
if batch_processor_manager:add_entry(conf, entry, max_pending_entries) then
|
||||
return
|
||||
end
|
||||
|
||||
-- reuse producer via lrucache to avoid unbalanced partitions of messages in kafka
|
||||
local broker_list = core.table.clone(conf.brokers or {})
|
||||
local broker_config = {}
|
||||
|
||||
if conf.broker_list then
|
||||
for host, port in pairs(conf.broker_list) do
|
||||
local broker = {
|
||||
host = host,
|
||||
port = port
|
||||
}
|
||||
core.table.insert(broker_list, broker)
|
||||
end
|
||||
end
|
||||
|
||||
broker_config["request_timeout"] = conf.timeout * 1000
|
||||
broker_config["producer_type"] = conf.producer_type
|
||||
broker_config["required_acks"] = conf.required_acks
|
||||
broker_config["batch_num"] = conf.producer_batch_num
|
||||
broker_config["batch_size"] = conf.producer_batch_size
|
||||
broker_config["max_buffering"] = conf.producer_max_buffering
|
||||
broker_config["flush_time"] = conf.producer_time_linger * 1000
|
||||
broker_config["refresh_interval"] = conf.meta_refresh_interval * 1000
|
||||
|
||||
local prod, err = core.lrucache.plugin_ctx(lrucache, ctx, nil, create_producer,
|
||||
broker_list, broker_config, conf.cluster_name)
|
||||
core.log.info("kafka cluster name ", conf.cluster_name, ", broker_list[1] port ",
|
||||
prod.client.broker_list[1].port)
|
||||
if err then
|
||||
return nil, "failed to identify the broker specified: " .. err
|
||||
end
|
||||
|
||||
-- Generate a function to be executed by the batch processor
|
||||
local func = function(entries, batch_max_size)
|
||||
local data, err
|
||||
if batch_max_size == 1 then
|
||||
data = entries[1]
|
||||
if type(data) ~= "string" then
|
||||
data, err = core.json.encode(data) -- encode as single {}
|
||||
end
|
||||
else
|
||||
data, err = core.json.encode(entries) -- encode as array [{}]
|
||||
end
|
||||
|
||||
if not data then
|
||||
return false, 'error occurred while encoding the data: ' .. err
|
||||
end
|
||||
|
||||
core.log.info("send data to kafka: ", data)
|
||||
|
||||
return send_kafka_data(conf, data, prod)
|
||||
end
|
||||
|
||||
batch_processor_manager:add_entry_to_new_processor(conf, entry, ctx, func, max_pending_entries)
|
||||
end
|
||||
|
||||
|
||||
return _M
|
Reference in New Issue
Block a user