projects
/
collectd.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Merge pull request #2701 from elfiesmelfie/feat_virt_notify_lifecycle
[collectd.git]
/
src
/
write_kafka.c
diff --git
a/src/write_kafka.c
b/src/write_kafka.c
index
ca400ac
..
c120d15
100644
(file)
--- a/
src/write_kafka.c
+++ b/
src/write_kafka.c
@@
-43,7
+43,7
@@
struct kafka_topic_context {
#define KAFKA_FORMAT_GRAPHITE 2
uint8_t format;
unsigned int graphite_flags;
#define KAFKA_FORMAT_GRAPHITE 2
uint8_t format;
unsigned int graphite_flags;
-
_B
ool store_rates;
+
b
ool store_rates;
rd_kafka_topic_conf_t *conf;
rd_kafka_topic_t *topic;
rd_kafka_conf_t *kafka_conf;
rd_kafka_topic_conf_t *conf;
rd_kafka_topic_t *topic;
rd_kafka_conf_t *kafka_conf;
@@
-77,6
+77,14
@@
static void kafka_log(const rd_kafka_t *rkt, int level, const char *fac,
}
#endif
}
#endif
+static rd_kafka_resp_err_t kafka_error() {
+#if RD_KAFKA_VERSION >= 0x000b00ff
+ return rd_kafka_last_error();
+#else
+ return rd_kafka_errno2err(errno);
+#endif
+}
+
static uint32_t kafka_hash(const char *keydata, size_t keylen) {
uint32_t hash = 5381;
for (; keylen > 0; keylen--)
static uint32_t kafka_hash(const char *keydata, size_t keylen) {
uint32_t hash = 5381;
for (; keylen > 0; keylen--)
@@
-89,7
+97,7
@@
static uint32_t kafka_hash(const char *keydata, size_t keylen) {
#define KAFKA_RANDOM_KEY_BUFFER \
(char[KAFKA_RANDOM_KEY_SIZE]) { "" }
static char *kafka_random_key(char buffer[static KAFKA_RANDOM_KEY_SIZE]) {
#define KAFKA_RANDOM_KEY_BUFFER \
(char[KAFKA_RANDOM_KEY_SIZE]) { "" }
static char *kafka_random_key(char buffer[static KAFKA_RANDOM_KEY_SIZE]) {
- s
s
nprintf(buffer, KAFKA_RANDOM_KEY_SIZE, "%08" PRIX32, cdrand_u());
+ snprintf(buffer, KAFKA_RANDOM_KEY_SIZE, "%08" PRIX32, cdrand_u());
return buffer;
}
return buffer;
}
@@
-147,7
+155,7
@@
static int kafka_handle(struct kafka_topic_context *ctx) /* {{{ */
if ((ctx->topic = rd_kafka_topic_new(ctx->kafka, ctx->topic_name,
topic_conf)) == NULL) {
ERROR("write_kafka plugin: cannot create topic : %s\n",
if ((ctx->topic = rd_kafka_topic_new(ctx->kafka, ctx->topic_name,
topic_conf)) == NULL) {
ERROR("write_kafka plugin: cannot create topic : %s\n",
- rd_kafka_err2str(
rd_kafka_errno2err(errno
)));
+ rd_kafka_err2str(
kafka_error(
)));
return errno;
}
return errno;
}
@@
-265,7
+273,7
@@
static void kafka_config_topic(rd_kafka_conf_t *conf,
}
tctx->escape_char = '.';
}
tctx->escape_char = '.';
- tctx->store_rates =
1
;
+ tctx->store_rates =
true
;
tctx->format = KAFKA_FORMAT_JSON;
tctx->key = NULL;
tctx->format = KAFKA_FORMAT_JSON;
tctx->key = NULL;
@@
-398,8
+406,8
@@
static void kafka_config_topic(rd_kafka_conf_t *conf,
rd_kafka_topic_conf_set_partitioner_cb(tctx->conf, kafka_partition);
rd_kafka_topic_conf_set_opaque(tctx->conf, tctx);
rd_kafka_topic_conf_set_partitioner_cb(tctx->conf, kafka_partition);
rd_kafka_topic_conf_set_opaque(tctx->conf, tctx);
- s
s
nprintf(callback_name, sizeof(callback_name), "write_kafka/%s",
-
tctx->topic_name);
+ snprintf(callback_name, sizeof(callback_name), "write_kafka/%s",
+ tctx->topic_name);
status = plugin_register_write(
callback_name, kafka_write,
status = plugin_register_write(
callback_name, kafka_write,