1939 lines
62 KiB
C
1939 lines
62 KiB
C
// SPDX-License-Identifier: GPL-3.0-or-later
|
|
|
|
#include <sys/resource.h>
|
|
|
|
#include "ebpf.h"
|
|
#include "ebpf_socket.h"
|
|
|
|
/*****************************************************************
|
|
*
|
|
* GLOBAL VARIABLES
|
|
*
|
|
*****************************************************************/
|
|
|
|
static char *socket_dimension_names[NETDATA_MAX_SOCKET_VECTOR] = { "sent", "received", "close", "sent",
|
|
"received", "retransmitted" };
|
|
static char *socket_id_names[NETDATA_MAX_SOCKET_VECTOR] = { "tcp_sendmsg", "tcp_cleanup_rbuf", "tcp_close",
|
|
"udp_sendmsg", "udp_recvmsg", "tcp_retransmit_skb" };
|
|
|
|
static netdata_idx_t *socket_hash_values = NULL;
|
|
static netdata_syscall_stat_t *socket_aggregated_data = NULL;
|
|
static netdata_publish_syscall_t *socket_publish_aggregated = NULL;
|
|
|
|
static ebpf_data_t socket_data;
|
|
|
|
ebpf_socket_publish_apps_t **socket_bandwidth_curr = NULL;
|
|
ebpf_socket_publish_apps_t **socket_bandwidth_prev = NULL;
|
|
static ebpf_bandwidth_t *bandwidth_vector = NULL;
|
|
|
|
static int socket_apps_created = 0;
|
|
pthread_mutex_t nv_mutex;
|
|
int wait_to_plot = 0;
|
|
int read_thread_closed = 1;
|
|
|
|
netdata_vector_plot_t inbound_vectors = { .plot = NULL, .next = 0, .last = 0 };
|
|
netdata_vector_plot_t outbound_vectors = { .plot = NULL, .next = 0, .last = 0 };
|
|
netdata_socket_t *socket_values;
|
|
|
|
ebpf_network_viewer_port_list_t *listen_ports = NULL;
|
|
|
|
static int *map_fd = NULL;
|
|
static struct bpf_object *objects = NULL;
|
|
static struct bpf_link **probe_links = NULL;
|
|
|
|
/*****************************************************************
|
|
*
|
|
* PROCESS DATA AND SEND TO NETDATA
|
|
*
|
|
*****************************************************************/
|
|
|
|
/**
|
|
* Update publish structure before to send data to Netdata.
|
|
*
|
|
* @param publish the first output structure with independent dimensions
|
|
* @param tcp structure to store IO from tcp sockets
|
|
* @param udp structure to store IO from udp sockets
|
|
* @param input the structure with the input data.
|
|
*/
|
|
static void ebpf_update_global_publish(
|
|
netdata_publish_syscall_t *publish, netdata_publish_vfs_common_t *tcp, netdata_publish_vfs_common_t *udp,
|
|
netdata_syscall_stat_t *input)
|
|
{
|
|
netdata_publish_syscall_t *move = publish;
|
|
while (move) {
|
|
if (input->call != move->pcall) {
|
|
// This condition happens to avoid initial values with dimensions higher than normal values.
|
|
if (move->pcall) {
|
|
move->ncall = (input->call > move->pcall) ? input->call - move->pcall : move->pcall - input->call;
|
|
move->nbyte = (input->bytes > move->pbyte) ? input->bytes - move->pbyte : move->pbyte - input->bytes;
|
|
move->nerr = (input->ecall > move->nerr) ? input->ecall - move->perr : move->perr - input->ecall;
|
|
} else {
|
|
move->ncall = 0;
|
|
move->nbyte = 0;
|
|
move->nerr = 0;
|
|
}
|
|
|
|
move->pcall = input->call;
|
|
move->pbyte = input->bytes;
|
|
move->perr = input->ecall;
|
|
} else {
|
|
move->ncall = 0;
|
|
move->nbyte = 0;
|
|
move->nerr = 0;
|
|
}
|
|
|
|
input = input->next;
|
|
move = move->next;
|
|
}
|
|
|
|
tcp->write = -((long)publish[0].nbyte);
|
|
tcp->read = (long)publish[1].nbyte;
|
|
|
|
udp->write = -((long)publish[3].nbyte);
|
|
udp->read = (long)publish[4].nbyte;
|
|
}
|
|
|
|
/**
|
|
* Update Network Viewer plot data
|
|
*
|
|
* @param plot the structure where the data will be stored
|
|
* @param sock the last update from the socket
|
|
*/
|
|
static inline void update_nv_plot_data(netdata_plot_values_t *plot, netdata_socket_t *sock)
|
|
{
|
|
if (sock->ct > plot->last_time) {
|
|
plot->last_time = sock->ct;
|
|
plot->plot_recv_packets = sock->recv_packets;
|
|
plot->plot_sent_packets = sock->sent_packets;
|
|
plot->plot_recv_bytes = sock->recv_bytes;
|
|
plot->plot_sent_bytes = sock->sent_bytes;
|
|
plot->plot_retransmit = sock->retransmit;
|
|
}
|
|
|
|
sock->recv_packets = 0;
|
|
sock->sent_packets = 0;
|
|
sock->recv_bytes = 0;
|
|
sock->sent_bytes = 0;
|
|
sock->retransmit = 0;
|
|
}
|
|
|
|
/**
|
|
* Calculate Network Viewer Plot
|
|
*
|
|
* Do math with collected values before to plot data.
|
|
*/
|
|
static inline void calculate_nv_plot()
|
|
{
|
|
uint32_t i;
|
|
uint32_t end = inbound_vectors.next;
|
|
for (i = 0; i < end; i++) {
|
|
update_nv_plot_data(&inbound_vectors.plot[i].plot, &inbound_vectors.plot[i].sock);
|
|
}
|
|
inbound_vectors.max_plot = end;
|
|
|
|
// The 'Other' dimension is always calculated for the chart to have at least one dimension
|
|
update_nv_plot_data(&inbound_vectors.plot[inbound_vectors.last].plot,
|
|
&inbound_vectors.plot[inbound_vectors.last].sock);
|
|
|
|
end = outbound_vectors.next;
|
|
for (i = 0; i < end; i++) {
|
|
update_nv_plot_data(&outbound_vectors.plot[i].plot, &outbound_vectors.plot[i].sock);
|
|
}
|
|
outbound_vectors.max_plot = end;
|
|
|
|
// The 'Other' dimension is always calculated for the chart to have at least one dimension
|
|
update_nv_plot_data(&outbound_vectors.plot[outbound_vectors.last].plot,
|
|
&outbound_vectors.plot[outbound_vectors.last].sock);
|
|
}
|
|
|
|
/**
|
|
* Network viewer send bytes
|
|
*
|
|
* @param ptr the structure with values to plot
|
|
* @param chart the chart name.
|
|
*/
|
|
static inline void ebpf_socket_nv_send_bytes(netdata_vector_plot_t *ptr, char *chart)
|
|
{
|
|
uint32_t i;
|
|
uint32_t end = ptr->last_plot;
|
|
netdata_socket_plot_t *w = ptr->plot;
|
|
collected_number value;
|
|
|
|
write_begin_chart(NETDATA_EBPF_FAMILY, chart);
|
|
for (i = 0; i < end; i++) {
|
|
value = ((collected_number) w[i].plot.plot_sent_bytes);
|
|
write_chart_dimension(w[i].dimension_sent, value);
|
|
value = (collected_number) w[i].plot.plot_recv_bytes;
|
|
write_chart_dimension(w[i].dimension_recv, value);
|
|
}
|
|
|
|
i = ptr->last;
|
|
value = ((collected_number) w[i].plot.plot_sent_bytes);
|
|
write_chart_dimension(w[i].dimension_sent, value);
|
|
value = (collected_number) w[i].plot.plot_recv_bytes;
|
|
write_chart_dimension(w[i].dimension_recv, value);
|
|
write_end_chart();
|
|
}
|
|
|
|
/**
|
|
* Network Viewer Send packets
|
|
*
|
|
* @param ptr the structure with values to plot
|
|
* @param chart the chart name.
|
|
*/
|
|
static inline void ebpf_socket_nv_send_packets(netdata_vector_plot_t *ptr, char *chart)
|
|
{
|
|
uint32_t i;
|
|
uint32_t end = ptr->last_plot;
|
|
netdata_socket_plot_t *w = ptr->plot;
|
|
collected_number value;
|
|
|
|
write_begin_chart(NETDATA_EBPF_FAMILY, chart);
|
|
for (i = 0; i < end; i++) {
|
|
value = ((collected_number)w[i].plot.plot_sent_packets);
|
|
write_chart_dimension(w[i].dimension_sent, value);
|
|
value = (collected_number) w[i].plot.plot_recv_packets;
|
|
write_chart_dimension(w[i].dimension_recv, value);
|
|
}
|
|
|
|
i = ptr->last;
|
|
value = ((collected_number)w[i].plot.plot_sent_packets);
|
|
write_chart_dimension(w[i].dimension_sent, value);
|
|
value = (collected_number)w[i].plot.plot_recv_packets;
|
|
write_chart_dimension(w[i].dimension_recv, value);
|
|
write_end_chart();
|
|
}
|
|
|
|
/**
|
|
* Network Viewer Send Retransmit
|
|
*
|
|
* @param ptr the structure with values to plot
|
|
* @param chart the chart name.
|
|
*/
|
|
static inline void ebpf_socket_nv_send_retransmit(netdata_vector_plot_t *ptr, char *chart)
|
|
{
|
|
uint32_t i;
|
|
uint32_t end = ptr->last_plot;
|
|
netdata_socket_plot_t *w = ptr->plot;
|
|
collected_number value;
|
|
|
|
write_begin_chart(NETDATA_EBPF_FAMILY, chart);
|
|
for (i = 0; i < end; i++) {
|
|
value = (collected_number) w[i].plot.plot_retransmit;
|
|
write_chart_dimension(w[i].dimension_retransmit, value);
|
|
}
|
|
|
|
i = ptr->last;
|
|
value = (collected_number)w[i].plot.plot_retransmit;
|
|
write_chart_dimension(w[i].dimension_retransmit, value);
|
|
write_end_chart();
|
|
}
|
|
|
|
/**
|
|
* Send network viewer data
|
|
*
|
|
* @param ptr the pointer to plot data
|
|
*/
|
|
static void ebpf_socket_send_nv_data(netdata_vector_plot_t *ptr)
|
|
{
|
|
if (!ptr->flags)
|
|
return;
|
|
|
|
if (ptr == (netdata_vector_plot_t *)&outbound_vectors) {
|
|
ebpf_socket_nv_send_bytes(ptr, NETDATA_NV_OUTBOUND_BYTES);
|
|
fflush(stdout);
|
|
|
|
ebpf_socket_nv_send_packets(ptr, NETDATA_NV_OUTBOUND_PACKETS);
|
|
fflush(stdout);
|
|
|
|
ebpf_socket_nv_send_retransmit(ptr, NETDATA_NV_OUTBOUND_RETRANSMIT);
|
|
fflush(stdout);
|
|
} else {
|
|
ebpf_socket_nv_send_bytes(ptr, NETDATA_NV_INBOUND_BYTES);
|
|
fflush(stdout);
|
|
|
|
ebpf_socket_nv_send_packets(ptr, NETDATA_NV_INBOUND_PACKETS);
|
|
fflush(stdout);
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
* Update the publish strctures to create the dimenssions
|
|
*
|
|
* @param curr Last values read from memory.
|
|
* @param prev Previous values read from memory.
|
|
*/
|
|
static void ebpf_socket_update_apps_publish(ebpf_socket_publish_apps_t *curr, ebpf_socket_publish_apps_t *prev)
|
|
{
|
|
curr->publish_received_bytes = curr->bytes_received - prev->bytes_received;
|
|
curr->publish_sent_bytes = curr->bytes_sent - prev->bytes_sent;
|
|
curr->publish_tcp_sent = curr->call_tcp_sent - prev->call_tcp_sent;
|
|
curr->publish_tcp_received = curr->call_tcp_received - prev->call_tcp_received;
|
|
curr->publish_retransmit = curr->retransmit - prev->retransmit;
|
|
curr->publish_udp_sent = curr->call_udp_sent - prev->call_udp_sent;
|
|
curr->publish_udp_received = curr->call_udp_received - prev->call_udp_received;
|
|
}
|
|
|
|
/**
|
|
* Send data to Netdata calling auxiliar functions.
|
|
*
|
|
* @param em the structure with thread information
|
|
*/
|
|
static void ebpf_socket_send_data(ebpf_module_t *em)
|
|
{
|
|
netdata_publish_vfs_common_t common_tcp;
|
|
netdata_publish_vfs_common_t common_udp;
|
|
ebpf_update_global_publish(socket_publish_aggregated, &common_tcp, &common_udp, socket_aggregated_data);
|
|
|
|
write_count_chart(
|
|
NETDATA_TCP_FUNCTION_COUNT, NETDATA_EBPF_FAMILY, socket_publish_aggregated, 3);
|
|
write_io_chart(
|
|
NETDATA_TCP_FUNCTION_BYTES, NETDATA_EBPF_FAMILY, socket_id_names[0], socket_id_names[1], &common_tcp);
|
|
if (em->mode < MODE_ENTRY) {
|
|
write_err_chart(
|
|
NETDATA_TCP_FUNCTION_ERROR, NETDATA_EBPF_FAMILY, socket_publish_aggregated, 2);
|
|
}
|
|
write_count_chart(
|
|
NETDATA_TCP_RETRANSMIT, NETDATA_EBPF_FAMILY, &socket_publish_aggregated[NETDATA_RETRANSMIT_START], 1);
|
|
|
|
write_count_chart(
|
|
NETDATA_UDP_FUNCTION_COUNT, NETDATA_EBPF_FAMILY, &socket_publish_aggregated[NETDATA_UDP_START], 2);
|
|
write_io_chart(
|
|
NETDATA_UDP_FUNCTION_BYTES, NETDATA_EBPF_FAMILY, socket_id_names[3], socket_id_names[4], &common_udp);
|
|
if (em->mode < MODE_ENTRY) {
|
|
write_err_chart(
|
|
NETDATA_UDP_FUNCTION_ERROR, NETDATA_EBPF_FAMILY, &socket_publish_aggregated[NETDATA_UDP_START], 2);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Sum values for pid
|
|
*
|
|
* @param root the structure with all available PIDs
|
|
*
|
|
* @param offset the address that we are reading
|
|
*
|
|
* @return it returns the sum of all PIDs
|
|
*/
|
|
long long ebpf_socket_sum_values_for_pids(struct pid_on_target *root, size_t offset)
|
|
{
|
|
long long ret = 0;
|
|
while (root) {
|
|
int32_t pid = root->pid;
|
|
ebpf_socket_publish_apps_t *w = socket_bandwidth_curr[pid];
|
|
if (w) {
|
|
ret += get_value_from_structure((char *)w, offset);
|
|
}
|
|
|
|
root = root->next;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Send data to Netdata calling auxiliar functions.
|
|
*
|
|
* @param em the structure with thread information
|
|
* @param root the target list.
|
|
*/
|
|
void ebpf_socket_send_apps_data(ebpf_module_t *em, struct target *root)
|
|
{
|
|
UNUSED(em);
|
|
if (!socket_apps_created)
|
|
return;
|
|
|
|
struct target *w;
|
|
collected_number value;
|
|
|
|
write_begin_chart(NETDATA_APPS_FAMILY, NETDATA_NET_APPS_BANDWIDTH_SENT);
|
|
for (w = root; w; w = w->next) {
|
|
if (unlikely(w->exposed && w->processes)) {
|
|
value = ebpf_socket_sum_values_for_pids(w->root_pid, offsetof(ebpf_socket_publish_apps_t,
|
|
publish_sent_bytes));
|
|
write_chart_dimension(w->name, value);
|
|
}
|
|
}
|
|
write_end_chart();
|
|
|
|
write_begin_chart(NETDATA_APPS_FAMILY, NETDATA_NET_APPS_BANDWIDTH_RECV);
|
|
for (w = root; w; w = w->next) {
|
|
if (unlikely(w->exposed && w->processes)) {
|
|
value = ebpf_socket_sum_values_for_pids(w->root_pid, offsetof(ebpf_socket_publish_apps_t,
|
|
publish_received_bytes));
|
|
write_chart_dimension(w->name, value);
|
|
}
|
|
}
|
|
write_end_chart();
|
|
|
|
write_begin_chart(NETDATA_APPS_FAMILY, NETDATA_NET_APPS_BANDWIDTH_TCP_SEND_CALLS);
|
|
for (w = root; w; w = w->next) {
|
|
if (unlikely(w->exposed && w->processes)) {
|
|
value = ebpf_socket_sum_values_for_pids(w->root_pid, offsetof(ebpf_socket_publish_apps_t,
|
|
publish_tcp_sent));
|
|
write_chart_dimension(w->name, value);
|
|
}
|
|
}
|
|
write_end_chart();
|
|
|
|
write_begin_chart(NETDATA_APPS_FAMILY, NETDATA_NET_APPS_BANDWIDTH_TCP_RECV_CALLS);
|
|
for (w = root; w; w = w->next) {
|
|
if (unlikely(w->exposed && w->processes)) {
|
|
value = ebpf_socket_sum_values_for_pids(w->root_pid, offsetof(ebpf_socket_publish_apps_t,
|
|
publish_tcp_received));
|
|
write_chart_dimension(w->name, value);
|
|
}
|
|
}
|
|
write_end_chart();
|
|
|
|
write_begin_chart(NETDATA_APPS_FAMILY, NETDATA_NET_APPS_BANDWIDTH_TCP_RETRANSMIT);
|
|
for (w = root; w; w = w->next) {
|
|
if (unlikely(w->exposed && w->processes)) {
|
|
value = ebpf_socket_sum_values_for_pids(w->root_pid, offsetof(ebpf_socket_publish_apps_t,
|
|
publish_retransmit));
|
|
write_chart_dimension(w->name, value);
|
|
}
|
|
}
|
|
write_end_chart();
|
|
|
|
write_begin_chart(NETDATA_APPS_FAMILY, NETDATA_NET_APPS_BANDWIDTH_UDP_SEND_CALLS);
|
|
for (w = root; w; w = w->next) {
|
|
if (unlikely(w->exposed && w->processes)) {
|
|
value = ebpf_socket_sum_values_for_pids(w->root_pid, offsetof(ebpf_socket_publish_apps_t,
|
|
publish_udp_sent));
|
|
write_chart_dimension(w->name, value);
|
|
}
|
|
}
|
|
write_end_chart();
|
|
|
|
write_begin_chart(NETDATA_APPS_FAMILY, NETDATA_NET_APPS_BANDWIDTH_UDP_RECV_CALLS);
|
|
for (w = root; w; w = w->next) {
|
|
if (unlikely(w->exposed && w->processes)) {
|
|
value = ebpf_socket_sum_values_for_pids(w->root_pid, offsetof(ebpf_socket_publish_apps_t,
|
|
publish_udp_received));
|
|
write_chart_dimension(w->name, value);
|
|
}
|
|
}
|
|
write_end_chart();
|
|
|
|
}
|
|
|
|
/*****************************************************************
|
|
*
|
|
* FUNCTIONS TO CREATE CHARTS
|
|
*
|
|
*****************************************************************/
|
|
|
|
/**
|
|
* Create global charts
|
|
*
|
|
* Call ebpf_create_chart to create the charts for the collector.
|
|
*
|
|
* @param em a pointer to the structure with the default values.
|
|
*/
|
|
static void ebpf_create_global_charts(ebpf_module_t *em)
|
|
{
|
|
ebpf_create_chart(NETDATA_EBPF_FAMILY,
|
|
NETDATA_TCP_FUNCTION_COUNT,
|
|
"Calls to internal functions",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_SOCKET_GROUP,
|
|
21070,
|
|
ebpf_create_global_dimension,
|
|
socket_publish_aggregated,
|
|
3);
|
|
|
|
ebpf_create_chart(NETDATA_EBPF_FAMILY,
|
|
NETDATA_TCP_FUNCTION_BYTES,
|
|
"TCP bandwidth",
|
|
EBPF_COMMON_DIMENSION_BYTESS,
|
|
NETDATA_SOCKET_GROUP,
|
|
21071,
|
|
ebpf_create_global_dimension,
|
|
socket_publish_aggregated,
|
|
3);
|
|
|
|
if (em->mode < MODE_ENTRY) {
|
|
ebpf_create_chart(NETDATA_EBPF_FAMILY,
|
|
NETDATA_TCP_FUNCTION_ERROR,
|
|
"TCP errors",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_SOCKET_GROUP,
|
|
21072,
|
|
ebpf_create_global_dimension,
|
|
socket_publish_aggregated,
|
|
2);
|
|
}
|
|
|
|
ebpf_create_chart(NETDATA_EBPF_FAMILY,
|
|
NETDATA_TCP_RETRANSMIT,
|
|
"Packages retransmitted",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_SOCKET_GROUP,
|
|
21073,
|
|
ebpf_create_global_dimension,
|
|
&socket_publish_aggregated[NETDATA_RETRANSMIT_START],
|
|
1);
|
|
|
|
ebpf_create_chart(NETDATA_EBPF_FAMILY,
|
|
NETDATA_UDP_FUNCTION_COUNT,
|
|
"UDP calls",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_SOCKET_GROUP,
|
|
21074,
|
|
ebpf_create_global_dimension,
|
|
&socket_publish_aggregated[NETDATA_UDP_START],
|
|
2);
|
|
|
|
ebpf_create_chart(NETDATA_EBPF_FAMILY,
|
|
NETDATA_UDP_FUNCTION_BYTES,
|
|
"UDP bandwidth",
|
|
EBPF_COMMON_DIMENSION_BYTESS,
|
|
NETDATA_SOCKET_GROUP,
|
|
21075,
|
|
ebpf_create_global_dimension,
|
|
&socket_publish_aggregated[NETDATA_UDP_START],
|
|
2);
|
|
|
|
if (em->mode < MODE_ENTRY) {
|
|
ebpf_create_chart(NETDATA_EBPF_FAMILY,
|
|
NETDATA_UDP_FUNCTION_ERROR,
|
|
"UDP errors",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_SOCKET_GROUP,
|
|
21076,
|
|
ebpf_create_global_dimension,
|
|
&socket_publish_aggregated[NETDATA_UDP_START],
|
|
2);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Create apps charts
|
|
*
|
|
* Call ebpf_create_chart to create the charts on apps submenu.
|
|
*
|
|
* @param em a pointer to the structure with the default values.
|
|
*/
|
|
void ebpf_socket_create_apps_charts(ebpf_module_t *em, struct target *root)
|
|
{
|
|
UNUSED(em);
|
|
ebpf_create_charts_on_apps(NETDATA_NET_APPS_BANDWIDTH_SENT,
|
|
"Bytes sent",
|
|
EBPF_COMMON_DIMENSION_BYTESS,
|
|
NETDATA_APPS_NET_GROUP,
|
|
20080,
|
|
root);
|
|
|
|
ebpf_create_charts_on_apps(NETDATA_NET_APPS_BANDWIDTH_RECV,
|
|
"bytes received",
|
|
EBPF_COMMON_DIMENSION_BYTESS,
|
|
NETDATA_APPS_NET_GROUP,
|
|
20081,
|
|
root);
|
|
|
|
ebpf_create_charts_on_apps(NETDATA_NET_APPS_BANDWIDTH_TCP_SEND_CALLS,
|
|
"Calls for tcp_sendmsg",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_APPS_NET_GROUP,
|
|
20082,
|
|
root);
|
|
|
|
ebpf_create_charts_on_apps(NETDATA_NET_APPS_BANDWIDTH_TCP_RECV_CALLS,
|
|
"Calls for tcp_cleanup_rbuf",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_APPS_NET_GROUP,
|
|
20083,
|
|
root);
|
|
|
|
ebpf_create_charts_on_apps(NETDATA_NET_APPS_BANDWIDTH_TCP_RETRANSMIT,
|
|
"Calls for tcp_retransmit",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_APPS_NET_GROUP,
|
|
20084,
|
|
root);
|
|
|
|
ebpf_create_charts_on_apps(NETDATA_NET_APPS_BANDWIDTH_UDP_SEND_CALLS,
|
|
"Calls for udp_sendmsg",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_APPS_NET_GROUP,
|
|
20085,
|
|
root);
|
|
|
|
ebpf_create_charts_on_apps(NETDATA_NET_APPS_BANDWIDTH_UDP_RECV_CALLS,
|
|
"Calls for udp_recvmsg",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_APPS_NET_GROUP,
|
|
20086,
|
|
root);
|
|
|
|
socket_apps_created = 1;
|
|
}
|
|
|
|
/**
|
|
* Create network viewer chart
|
|
*
|
|
* Create common charts.
|
|
*
|
|
* @param id the chart id
|
|
* @param title the chart title
|
|
* @param units the units label
|
|
* @param family the group name used to attach the chart on dashaboard
|
|
* @param order the chart order
|
|
* @param ptr the plot structure with values.
|
|
*/
|
|
static void ebpf_socket_create_nv_chart(char *id, char *title, char *units,
|
|
char *family, int order, netdata_vector_plot_t *ptr)
|
|
{
|
|
ebpf_write_chart_cmd(NETDATA_EBPF_FAMILY,
|
|
id,
|
|
title,
|
|
units,
|
|
family,
|
|
"stacked",
|
|
order);
|
|
|
|
uint32_t i;
|
|
uint32_t end = ptr->last_plot;
|
|
netdata_socket_plot_t *w = ptr->plot;
|
|
for (i = 0; i < end; i++) {
|
|
fprintf(stdout, "DIMENSION %s '' incremental -1 1\n", w[i].dimension_sent);
|
|
fprintf(stdout, "DIMENSION %s '' incremental 1 1\n", w[i].dimension_recv);
|
|
}
|
|
|
|
end = ptr->last;
|
|
fprintf(stdout, "DIMENSION %s '' incremental -1 1\n", w[end].dimension_sent);
|
|
fprintf(stdout, "DIMENSION %s '' incremental 1 1\n", w[end].dimension_recv);
|
|
}
|
|
|
|
/**
|
|
* Create network viewer retransmit
|
|
*
|
|
* Create a specific chart.
|
|
*
|
|
* @param id the chart id
|
|
* @param title the chart title
|
|
* @param units the units label
|
|
* @param family the group name used to attach the chart on dashaboard
|
|
* @param order the chart order
|
|
* @param ptr the plot structure with values.
|
|
*/
|
|
static void ebpf_socket_create_nv_retransmit(char *id, char *title, char *units,
|
|
char *family, int order, netdata_vector_plot_t *ptr)
|
|
{
|
|
ebpf_write_chart_cmd(NETDATA_EBPF_FAMILY,
|
|
id,
|
|
title,
|
|
units,
|
|
family,
|
|
"stacked",
|
|
order);
|
|
|
|
uint32_t i;
|
|
uint32_t end = ptr->last_plot;
|
|
netdata_socket_plot_t *w = ptr->plot;
|
|
for (i = 0; i < end; i++) {
|
|
fprintf(stdout, "DIMENSION %s '' incremental 1 1\n", w[i].dimension_retransmit);
|
|
}
|
|
|
|
end = ptr->last;
|
|
fprintf(stdout, "DIMENSION %s '' incremental 1 1\n", w[end].dimension_retransmit);
|
|
}
|
|
|
|
/**
|
|
* Create Network Viewer charts
|
|
*
|
|
* Recreate the charts when new sockets are created.
|
|
*
|
|
* @param ptr a pointer for inbound or outbound vectors.
|
|
*/
|
|
static void ebpf_socket_create_nv_charts(netdata_vector_plot_t *ptr)
|
|
{
|
|
// We do not have new sockets, so we do not need move forward
|
|
if (ptr->max_plot == ptr->last_plot)
|
|
return;
|
|
|
|
ptr->last_plot = ptr->max_plot;
|
|
|
|
if (ptr == (netdata_vector_plot_t *)&outbound_vectors) {
|
|
ebpf_socket_create_nv_chart(NETDATA_NV_OUTBOUND_BYTES,
|
|
"Outbound connections (bytes).",
|
|
EBPF_COMMON_DIMENSION_BYTESS,
|
|
NETDATA_NETWORK_CONNECTIONS_GROUP,
|
|
21080,
|
|
ptr);
|
|
|
|
ebpf_socket_create_nv_chart(NETDATA_NV_OUTBOUND_PACKETS,
|
|
"Outbound connections (packets)",
|
|
EBPF_COMMON_DIMENSION_PACKETS,
|
|
NETDATA_NETWORK_CONNECTIONS_GROUP,
|
|
21082,
|
|
ptr);
|
|
|
|
ebpf_socket_create_nv_retransmit(NETDATA_NV_OUTBOUND_RETRANSMIT,
|
|
"Retransmitted packets",
|
|
EBPF_COMMON_DIMENSION_CALL,
|
|
NETDATA_NETWORK_CONNECTIONS_GROUP,
|
|
21083,
|
|
ptr);
|
|
} else {
|
|
ebpf_socket_create_nv_chart(NETDATA_NV_INBOUND_BYTES,
|
|
"Inbound connections (bytes)",
|
|
EBPF_COMMON_DIMENSION_BYTESS,
|
|
NETDATA_NETWORK_CONNECTIONS_GROUP,
|
|
21084,
|
|
ptr);
|
|
|
|
ebpf_socket_create_nv_chart(NETDATA_NV_INBOUND_PACKETS,
|
|
"Inbound connections (packets)",
|
|
EBPF_COMMON_DIMENSION_PACKETS,
|
|
NETDATA_NETWORK_CONNECTIONS_GROUP,
|
|
21085,
|
|
ptr);
|
|
}
|
|
|
|
ptr->flags |= NETWORK_VIEWER_CHARTS_CREATED;
|
|
}
|
|
|
|
/*****************************************************************
|
|
*
|
|
* READ INFORMATION FROM KERNEL RING
|
|
*
|
|
*****************************************************************/
|
|
|
|
/**
|
|
* Is specific ip inside the range
|
|
*
|
|
* Check if the ip is inside a IP range previously defined
|
|
*
|
|
* @param cmp the IP to compare
|
|
* @param family the IP family
|
|
*
|
|
* @return It returns 1 if the IP is inside the range and 0 otherwise
|
|
*/
|
|
static int is_specific_ip_inside_range(union netdata_ip_t *cmp, int family)
|
|
{
|
|
if (!network_viewer_opt.excluded_ips && !network_viewer_opt.included_ips)
|
|
return 1;
|
|
|
|
uint32_t ipv4_test = ntohl(cmp->addr32[0]);
|
|
ebpf_network_viewer_ip_list_t *move = network_viewer_opt.excluded_ips;
|
|
while (move) {
|
|
if (family == AF_INET) {
|
|
if (ntohl(move->first.addr32[0]) <= ipv4_test &&
|
|
ipv4_test <= ntohl(move->last.addr32[0]) )
|
|
return 0;
|
|
} else {
|
|
if (memcmp(move->first.addr8, cmp->addr8, sizeof(union netdata_ip_t)) <= 0 &&
|
|
memcmp(move->last.addr8, cmp->addr8, sizeof(union netdata_ip_t)) >= 0) {
|
|
return 0;
|
|
}
|
|
}
|
|
move = move->next;
|
|
}
|
|
|
|
move = network_viewer_opt.included_ips;
|
|
while (move) {
|
|
if (family == AF_INET) {
|
|
if (ntohl(move->first.addr32[0]) <= ipv4_test &&
|
|
ntohl(move->last.addr32[0]) >= ipv4_test)
|
|
return 1;
|
|
} else {
|
|
if (memcmp(move->first.addr8, cmp->addr8, sizeof(union netdata_ip_t)) <= 0 &&
|
|
memcmp(move->last.addr8, cmp->addr8, sizeof(union netdata_ip_t)) >= 0) {
|
|
return 1;
|
|
}
|
|
}
|
|
move = move->next;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Is port inside range
|
|
*
|
|
* Verify if the cmp port is inside the range [first, last].
|
|
* This function expects only the last parameter as big endian.
|
|
*
|
|
* @param cmp the value to compare
|
|
*
|
|
* @return It returns 1 when cmp is inside and 0 otherwise.
|
|
*/
|
|
static int is_port_inside_range(uint16_t cmp)
|
|
{
|
|
// We do not have restrictions for ports.
|
|
if (!network_viewer_opt.excluded_port && !network_viewer_opt.included_port)
|
|
return 1;
|
|
|
|
// Test if port is excluded
|
|
ebpf_network_viewer_port_list_t *move = network_viewer_opt.excluded_port;
|
|
cmp = htons(cmp);
|
|
while (move) {
|
|
if (move->cmp_first <= cmp && cmp <= move->cmp_last)
|
|
return 0;
|
|
|
|
move = move->next;
|
|
}
|
|
|
|
// Test if the port is inside allowed range
|
|
move = network_viewer_opt.included_port;
|
|
while (move) {
|
|
if (move->cmp_first <= cmp && cmp <= move->cmp_last)
|
|
return 1;
|
|
|
|
move = move->next;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Hostname matches pattern
|
|
*
|
|
* @param cmp the value to compare
|
|
*
|
|
* @return It returns 1 when the value matches and zero otherwise.
|
|
*/
|
|
int hostname_matches_pattern(char *cmp)
|
|
{
|
|
if (!network_viewer_opt.included_hostnames && !network_viewer_opt.excluded_hostnames)
|
|
return 1;
|
|
|
|
ebpf_network_viewer_hostname_list_t *move = network_viewer_opt.excluded_hostnames;
|
|
while (move) {
|
|
if (simple_pattern_matches(move->value_pattern, cmp))
|
|
return 0;
|
|
|
|
move = move->next;
|
|
}
|
|
|
|
move = network_viewer_opt.included_hostnames;
|
|
while (move) {
|
|
if (simple_pattern_matches(move->value_pattern, cmp))
|
|
return 1;
|
|
|
|
move = move->next;
|
|
}
|
|
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Is socket allowed?
|
|
*
|
|
* Compare destination addresses and destination ports to define next steps
|
|
*
|
|
* @param key the socket read from kernel ring
|
|
* @param family the family used to compare IPs (AF_INET and AF_INET6)
|
|
*
|
|
* @return It returns 1 if this socket is inside the ranges and 0 otherwise.
|
|
*/
|
|
int is_socket_allowed(netdata_socket_idx_t *key, int family)
|
|
{
|
|
if (!is_port_inside_range(key->dport))
|
|
return 0;
|
|
|
|
return is_specific_ip_inside_range(&key->daddr, family);
|
|
}
|
|
|
|
/**
|
|
* Compare sockets
|
|
*
|
|
* Compare destination address and destination port.
|
|
* We do not compare source port, because it is random.
|
|
* We also do not compare source address, because inbound and outbound connections are stored in separated AVL trees.
|
|
*
|
|
* @param a pointer to netdata_socket_plot
|
|
* @param b pointer to netdata_socket_plot
|
|
*
|
|
* @return It returns 0 case the values are equal, 1 case a is bigger than b and -1 case a is smaller than b.
|
|
*/
|
|
static int compare_sockets(void *a, void *b)
|
|
{
|
|
struct netdata_socket_plot *val1 = a;
|
|
struct netdata_socket_plot *val2 = b;
|
|
int cmp;
|
|
|
|
// We do not need to compare val2 family, because data inside hash table is always from the same family
|
|
if (val1->family == AF_INET) { //IPV4
|
|
if (val1->flags & NETDATA_INBOUND_DIRECTION) {
|
|
if (val1->index.sport == val2->index.sport)
|
|
cmp = 0;
|
|
else {
|
|
cmp = (val1->index.sport > val2->index.sport)?1:-1;
|
|
}
|
|
} else {
|
|
cmp = memcmp(&val1->index.dport, &val2->index.dport, sizeof(uint16_t));
|
|
if (!cmp) {
|
|
cmp = memcmp(&val1->index.daddr.addr32[0], &val2->index.daddr.addr32[0], sizeof(uint32_t));
|
|
}
|
|
}
|
|
} else {
|
|
if (val1->flags & NETDATA_INBOUND_DIRECTION) {
|
|
if (val1->index.sport == val2->index.sport)
|
|
cmp = 0;
|
|
else {
|
|
cmp = (val1->index.sport > val2->index.sport)?1:-1;
|
|
}
|
|
} else {
|
|
cmp = memcmp(&val1->index.dport, &val2->index.dport, sizeof(uint16_t));
|
|
if (!cmp) {
|
|
cmp = memcmp(&val1->index.daddr.addr32, &val2->index.daddr.addr32, 4*sizeof(uint32_t));
|
|
}
|
|
}
|
|
}
|
|
|
|
return cmp;
|
|
}
|
|
|
|
/**
|
|
* Build dimension name
|
|
*
|
|
* Fill dimension name vector with values given
|
|
*
|
|
* @param dimname the output vector
|
|
* @param hostname the hostname for the socket.
|
|
* @param service_name the service used to connect.
|
|
* @param proto the protocol used in this connection
|
|
* @param family is this IPV4(AF_INET) or IPV6(AF_INET6)
|
|
*
|
|
* @return it returns the size of the data copied on success and -1 otherwise.
|
|
*/
|
|
static inline int build_outbound_dimension_name(char *dimname, char *hostname, char *service_name,
|
|
char *proto, int family)
|
|
{
|
|
return snprintf(dimname, CONFIG_MAX_NAME - 7, (family == AF_INET)?"%s:%s:%s_":"%s:%s:[%s]_",
|
|
service_name, proto,
|
|
hostname);
|
|
}
|
|
|
|
/**
|
|
* Fill inbound dimension name
|
|
*
|
|
* Mount the dimension name with the input given
|
|
*
|
|
* @param dimname the output vector
|
|
* @param service_name the service used to connect.
|
|
* @param proto the protocol used in this connection
|
|
*
|
|
* @return it returns the size of the data copied on success and -1 otherwise.
|
|
*/
|
|
static inline int build_inbound_dimension_name(char *dimname, char *service_name, char *proto)
|
|
{
|
|
return snprintf(dimname, CONFIG_MAX_NAME - 7, "%s:%s_", service_name,
|
|
proto);
|
|
}
|
|
|
|
/**
|
|
* Fill Resolved Name
|
|
*
|
|
* Fill the resolved name structure with the value given.
|
|
* The hostname is the largest value possible, if it is necessary to cut some value, it must be cut.
|
|
*
|
|
* @param ptr the output vector
|
|
* @param hostname the hostname resolved or IP.
|
|
* @param length the length for the hostname.
|
|
* @param service_name the service name associated to the connection
|
|
* @param is_outbound the is this an outbound connection
|
|
*/
|
|
static inline void fill_resolved_name(netdata_socket_plot_t *ptr, char *hostname, size_t length,
|
|
char *service_name, int is_outbound)
|
|
{
|
|
if (length < NETDATA_MAX_NETWORK_COMBINED_LENGTH)
|
|
ptr->resolved_name = strdupz(hostname);
|
|
else {
|
|
length = NETDATA_MAX_NETWORK_COMBINED_LENGTH;
|
|
ptr->resolved_name = mallocz( NETDATA_MAX_NETWORK_COMBINED_LENGTH + 1);
|
|
memcpy(ptr->resolved_name, hostname, length);
|
|
ptr->resolved_name[length] = '\0';
|
|
}
|
|
|
|
char dimname[CONFIG_MAX_NAME];
|
|
int size;
|
|
char *protocol;
|
|
if (ptr->sock.protocol == IPPROTO_UDP) {
|
|
protocol = "UDP";
|
|
} else if (ptr->sock.protocol == IPPROTO_TCP) {
|
|
protocol = "TCP";
|
|
} else {
|
|
protocol = "ALL";
|
|
}
|
|
|
|
if (is_outbound)
|
|
size = build_outbound_dimension_name(dimname, hostname, service_name, protocol, ptr->family);
|
|
else
|
|
size = build_inbound_dimension_name(dimname,service_name, protocol);
|
|
|
|
if (size > 0) {
|
|
strcpy(&dimname[size], "sent");
|
|
dimname[size + 4] = '\0';
|
|
ptr->dimension_sent = strdupz(dimname);
|
|
|
|
strcpy(&dimname[size], "recv");
|
|
ptr->dimension_recv = strdupz(dimname);
|
|
|
|
dimname[size - 1] = '\0';
|
|
ptr->dimension_retransmit = strdupz(dimname);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Mount dimension names
|
|
*
|
|
* Fill the vector names after to resolve the addresses
|
|
*
|
|
* @param ptr a pointer to the structure where the values are stored.
|
|
* @param is_outbound is a outbound ptr value?
|
|
*
|
|
* @return It returns 1 if the name is valid and 0 otherwise.
|
|
*/
|
|
int fill_names(netdata_socket_plot_t *ptr, int is_outbound)
|
|
{
|
|
char hostname[NI_MAXHOST], service_name[NI_MAXSERV];
|
|
if (ptr->resolved)
|
|
return 1;
|
|
|
|
int ret;
|
|
static int resolve_name = -1;
|
|
static int resolve_service = -1;
|
|
if (resolve_name == -1)
|
|
resolve_name = network_viewer_opt.hostname_resolution_enabled;
|
|
|
|
if (resolve_service == -1)
|
|
resolve_service = network_viewer_opt.service_resolution_enabled;
|
|
|
|
netdata_socket_idx_t *idx = &ptr->index;
|
|
|
|
char *errname = { "Not resolved" };
|
|
// Resolve Name
|
|
if (ptr->family == AF_INET) { //IPV4
|
|
struct sockaddr_in myaddr;
|
|
memset(&myaddr, 0 , sizeof(myaddr));
|
|
|
|
myaddr.sin_family = ptr->family;
|
|
if (is_outbound) {
|
|
myaddr.sin_port = idx->dport;
|
|
myaddr.sin_addr.s_addr = idx->daddr.addr32[0];
|
|
} else {
|
|
myaddr.sin_port = idx->sport;
|
|
myaddr.sin_addr.s_addr = idx->saddr.addr32[0];
|
|
}
|
|
|
|
ret = (!resolve_name)?-1:getnameinfo((struct sockaddr *)&myaddr, sizeof(myaddr), hostname,
|
|
sizeof(hostname), service_name, sizeof(service_name), NI_NAMEREQD);
|
|
|
|
if (!ret && !resolve_service) {
|
|
snprintf(service_name, sizeof(service_name), "%u", ntohs(myaddr.sin_port));
|
|
}
|
|
|
|
if (ret) {
|
|
// I cannot resolve the name, I will use the IP
|
|
if (!inet_ntop(AF_INET, &myaddr.sin_addr.s_addr, hostname, NI_MAXHOST)) {
|
|
strncpy(hostname, errname, 13);
|
|
}
|
|
|
|
snprintf(service_name, sizeof(service_name), "%u", ntohs(myaddr.sin_port));
|
|
ret = 1;
|
|
}
|
|
} else { // IPV6
|
|
struct sockaddr_in6 myaddr6;
|
|
memset(&myaddr6, 0 , sizeof(myaddr6));
|
|
|
|
myaddr6.sin6_family = AF_INET6;
|
|
if (is_outbound) {
|
|
myaddr6.sin6_port = idx->dport;
|
|
memcpy(myaddr6.sin6_addr.s6_addr, idx->daddr.addr8, sizeof(union netdata_ip_t));
|
|
} else {
|
|
myaddr6.sin6_port = idx->sport;
|
|
memcpy(myaddr6.sin6_addr.s6_addr, idx->saddr.addr8, sizeof(union netdata_ip_t));
|
|
}
|
|
|
|
ret = (!resolve_name)?-1:getnameinfo((struct sockaddr *)&myaddr6, sizeof(myaddr6), hostname,
|
|
sizeof(hostname), service_name, sizeof(service_name), NI_NAMEREQD);
|
|
|
|
if (!ret && !resolve_service) {
|
|
snprintf(service_name, sizeof(service_name), "%u", ntohs(myaddr6.sin6_port));
|
|
}
|
|
|
|
if (ret) {
|
|
// I cannot resolve the name, I will use the IP
|
|
if (!inet_ntop(AF_INET6, myaddr6.sin6_addr.s6_addr, hostname, NI_MAXHOST)) {
|
|
strncpy(hostname, errname, 13);
|
|
}
|
|
|
|
snprintf(service_name, sizeof(service_name), "%u", ntohs(myaddr6.sin6_port));
|
|
|
|
ret = 1;
|
|
}
|
|
}
|
|
|
|
fill_resolved_name(ptr, hostname,
|
|
strlen(hostname) + strlen(service_name)+ NETDATA_DOTS_PROTOCOL_COMBINED_LENGTH,
|
|
service_name, is_outbound);
|
|
|
|
if (resolve_name && !ret)
|
|
ret = hostname_matches_pattern(hostname);
|
|
|
|
ptr->resolved++;
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Fill last Network Viewer Dimension
|
|
*
|
|
* Fill the unique dimension that is always plotted.
|
|
*
|
|
* @param ptr the pointer for the last dimension
|
|
* @param is_outbound is this an inbound structure?
|
|
*/
|
|
static void fill_last_nv_dimension(netdata_socket_plot_t *ptr, int is_outbound)
|
|
{
|
|
char hostname[NI_MAXHOST], service_name[NI_MAXSERV];
|
|
char *other = { "other" };
|
|
// We are also copying the NULL bytes to avoid warnings in new compilers
|
|
strncpy(hostname, other, 6);
|
|
strncpy(service_name, other, 6);
|
|
|
|
ptr->family = AF_INET;
|
|
ptr->sock.protocol = 255;
|
|
ptr->flags = (!is_outbound)?NETDATA_INBOUND_DIRECTION:NETDATA_OUTBOUND_DIRECTION;
|
|
|
|
fill_resolved_name(ptr, hostname, 10 + NETDATA_DOTS_PROTOCOL_COMBINED_LENGTH, service_name, is_outbound);
|
|
|
|
#ifdef NETDATA_INTERNAL_CHECKS
|
|
info("Last %s dimension added: ID = %u, IP = OTHER, NAME = %s, DIM1 = %s, DIM2 = %s, DIM3 = %s",
|
|
(is_outbound)?"outbound":"inbound", network_viewer_opt.max_dim - 1, ptr->resolved_name,
|
|
ptr->dimension_recv, ptr->dimension_sent, ptr->dimension_retransmit);
|
|
#endif
|
|
}
|
|
|
|
/**
|
|
* Update Socket Data
|
|
*
|
|
* Update the socket information with last collected data
|
|
*
|
|
* @param sock
|
|
* @param lvalues
|
|
*/
|
|
static inline void update_socket_data(netdata_socket_t *sock, netdata_socket_t *lvalues)
|
|
{
|
|
sock->recv_packets += lvalues->recv_packets;
|
|
sock->sent_packets += lvalues->sent_packets;
|
|
sock->recv_bytes += lvalues->recv_bytes;
|
|
sock->sent_bytes += lvalues->sent_bytes;
|
|
sock->retransmit += lvalues->retransmit;
|
|
|
|
if (lvalues->ct > sock->ct)
|
|
sock->ct = lvalues->ct;
|
|
}
|
|
|
|
/**
|
|
* Store socket inside avl
|
|
*
|
|
* Store the socket values inside the avl tree.
|
|
*
|
|
* @param out the structure with information used to plot charts.
|
|
* @param lvalues Values read from socket ring.
|
|
* @param lindex the index information, the real socket.
|
|
* @param family the family associated to the socket
|
|
* @param flags the connection flags
|
|
*/
|
|
static void store_socket_inside_avl(netdata_vector_plot_t *out, netdata_socket_t *lvalues,
|
|
netdata_socket_idx_t *lindex, int family, uint32_t flags)
|
|
{
|
|
netdata_socket_plot_t test, *ret ;
|
|
|
|
memcpy(&test.index, lindex, sizeof(netdata_socket_idx_t));
|
|
test.flags = flags;
|
|
|
|
ret = (netdata_socket_plot_t *) avl_search_lock(&out->tree, (avl *)&test);
|
|
if (ret) {
|
|
if (lvalues->ct > ret->plot.last_time) {
|
|
update_socket_data(&ret->sock, lvalues);
|
|
}
|
|
} else {
|
|
uint32_t curr = out->next;
|
|
uint32_t last = out->last;
|
|
|
|
netdata_socket_plot_t *w = &out->plot[curr];
|
|
|
|
int resolved;
|
|
if (curr == last) {
|
|
if (lvalues->ct > w->plot.last_time) {
|
|
update_socket_data(&w->sock, lvalues);
|
|
}
|
|
return;
|
|
} else {
|
|
memcpy(&w->sock, lvalues, sizeof(netdata_socket_t));
|
|
memcpy(&w->index, lindex, sizeof(netdata_socket_idx_t));
|
|
w->family = family;
|
|
|
|
resolved = fill_names(w, out != (netdata_vector_plot_t *)&inbound_vectors);
|
|
}
|
|
|
|
if (!resolved) {
|
|
freez(w->resolved_name);
|
|
freez(w->dimension_sent);
|
|
freez(w->dimension_recv);
|
|
freez(w->dimension_retransmit);
|
|
|
|
memset(w, 0, sizeof(netdata_socket_plot_t));
|
|
|
|
return;
|
|
}
|
|
|
|
w->flags = flags;
|
|
netdata_socket_plot_t *check ;
|
|
check = (netdata_socket_plot_t *) avl_insert_lock(&out->tree, (avl *)w);
|
|
if (check != w)
|
|
error("Internal error, cannot insert the AVL tree.");
|
|
|
|
#ifdef NETDATA_INTERNAL_CHECKS
|
|
char iptext[INET6_ADDRSTRLEN];
|
|
if (inet_ntop(family, &w->index.daddr.addr8, iptext, sizeof(iptext)))
|
|
info("New %s dimension added: ID = %u, IP = %s, NAME = %s, DIM1 = %s, DIM2 = %s, DIM3 = %s",
|
|
(out == &inbound_vectors)?"inbound":"outbound", curr, iptext, w->resolved_name,
|
|
w->dimension_recv, w->dimension_sent, w->dimension_retransmit);
|
|
#endif
|
|
curr++;
|
|
if (curr > last)
|
|
curr = last;
|
|
out->next = curr;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Compare Vector to store
|
|
*
|
|
* Compare input values with local address to select table to store.
|
|
*
|
|
* @param direction store inbound and outbound direction.
|
|
* @param cmp index read from hash table.
|
|
* @param proto the protocol read.
|
|
*
|
|
* @return It returns the structure with address to compare.
|
|
*/
|
|
netdata_vector_plot_t * select_vector_to_store(uint32_t *direction, netdata_socket_idx_t *cmp, uint8_t proto)
|
|
{
|
|
if (!listen_ports) {
|
|
*direction = NETDATA_OUTBOUND_DIRECTION;
|
|
return &outbound_vectors;
|
|
}
|
|
|
|
ebpf_network_viewer_port_list_t *move_ports = listen_ports;
|
|
while (move_ports) {
|
|
if (move_ports->protocol == proto && move_ports->first == cmp->sport) {
|
|
*direction = NETDATA_INBOUND_DIRECTION;
|
|
return &inbound_vectors;
|
|
}
|
|
|
|
move_ports = move_ports->next;
|
|
}
|
|
|
|
*direction = NETDATA_OUTBOUND_DIRECTION;
|
|
return &outbound_vectors;
|
|
}
|
|
|
|
/**
|
|
* Hash accumulator
|
|
*
|
|
* @param values the values used to calculate the data.
|
|
* @param key the key to store data.
|
|
* @param removesock check if this socket must be removed .
|
|
* @param family the connection family
|
|
* @param end the values size.
|
|
*/
|
|
static void hash_accumulator(netdata_socket_t *values, netdata_socket_idx_t *key, int *removesock, int family, int end)
|
|
{
|
|
uint64_t bsent = 0, brecv = 0, psent = 0, precv = 0;
|
|
uint16_t retransmit = 0;
|
|
int i;
|
|
uint8_t protocol = values[0].protocol;
|
|
uint64_t ct = values[0].ct;
|
|
for (i = 1; i < end; i++) {
|
|
netdata_socket_t *w = &values[i];
|
|
|
|
precv += w->recv_packets;
|
|
psent += w->sent_packets;
|
|
brecv += w->recv_bytes;
|
|
bsent += w->sent_bytes;
|
|
retransmit += w->retransmit;
|
|
|
|
if (!protocol)
|
|
protocol = w->protocol;
|
|
|
|
if (w->ct > ct)
|
|
ct = w->ct;
|
|
|
|
*removesock += (int)w->removeme;
|
|
}
|
|
|
|
values[0].recv_packets += precv;
|
|
values[0].sent_packets += psent;
|
|
values[0].recv_bytes += brecv;
|
|
values[0].sent_bytes += bsent;
|
|
values[0].retransmit += retransmit;
|
|
values[0].removeme += (uint8_t)*removesock;
|
|
values[0].protocol = (!protocol)?IPPROTO_TCP:protocol;
|
|
values[0].ct = ct;
|
|
|
|
if (is_socket_allowed(key, family)) {
|
|
uint32_t dir;
|
|
netdata_vector_plot_t *table = select_vector_to_store(&dir, key, protocol);
|
|
store_socket_inside_avl(table, &values[0], key, family, dir);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Read socket hash table
|
|
*
|
|
* Read data from hash tables created on kernel ring.
|
|
*
|
|
* @param fd the hash table with data.
|
|
* @param family the family associated to the hash table
|
|
*
|
|
* @return it returns 0 on success and -1 otherwise.
|
|
*/
|
|
static void read_socket_hash_table(int fd, int family, int network_connection)
|
|
{
|
|
if (wait_to_plot)
|
|
return;
|
|
|
|
netdata_socket_idx_t key = {};
|
|
netdata_socket_idx_t next_key;
|
|
netdata_socket_idx_t removeme;
|
|
int removesock = 0;
|
|
|
|
netdata_socket_t *values = socket_values;
|
|
size_t length = ebpf_nprocs*sizeof(netdata_socket_t);
|
|
int test, end = (running_on_kernel < NETDATA_KERNEL_V4_15) ? 1 : ebpf_nprocs;
|
|
|
|
while (bpf_map_get_next_key(fd, &key, &next_key) == 0) {
|
|
// We need to reset the values when we are working on kernel 4.15 or newer, because kernel does not create
|
|
// values for specific processor unless it is used to store data. As result of this behavior one the next socket
|
|
// can have values from the previous one.
|
|
memset(values, 0, length);
|
|
test = bpf_map_lookup_elem(fd, &key, values);
|
|
if (test < 0) {
|
|
key = next_key;
|
|
continue;
|
|
}
|
|
|
|
if (removesock)
|
|
bpf_map_delete_elem(fd, &removeme);
|
|
|
|
if (network_connection) {
|
|
removesock = 0;
|
|
hash_accumulator(values, &key, &removesock, family, end);
|
|
}
|
|
|
|
if (removesock)
|
|
removeme = key;
|
|
|
|
key = next_key;
|
|
}
|
|
|
|
if (removesock)
|
|
bpf_map_delete_elem(fd, &removeme);
|
|
|
|
test = bpf_map_lookup_elem(fd, &next_key, values);
|
|
if (test < 0) {
|
|
return;
|
|
}
|
|
|
|
if (network_connection) {
|
|
removesock = 0;
|
|
hash_accumulator(values, &next_key, &removesock, family, end);
|
|
}
|
|
|
|
if (removesock)
|
|
bpf_map_delete_elem(fd, &next_key);
|
|
}
|
|
|
|
/**
|
|
* Update listen table
|
|
*
|
|
* Update link list when it is necessary.
|
|
*
|
|
* @param value the ports we are listen to.
|
|
* @param proto the protocol used with port connection.
|
|
*/
|
|
void update_listen_table(uint16_t value, uint8_t proto)
|
|
{
|
|
ebpf_network_viewer_port_list_t *w;
|
|
if (likely(listen_ports)) {
|
|
ebpf_network_viewer_port_list_t *move = listen_ports, *store = listen_ports;
|
|
while (move) {
|
|
if (move->protocol == proto && move->first == value)
|
|
return;
|
|
|
|
store = move;
|
|
move = move->next;
|
|
}
|
|
|
|
w = callocz(1, sizeof(ebpf_network_viewer_port_list_t));
|
|
w->first = value;
|
|
w->protocol = proto;
|
|
store->next = w;
|
|
} else {
|
|
w = callocz(1, sizeof(ebpf_network_viewer_port_list_t));
|
|
w->first = value;
|
|
w->protocol = proto;
|
|
|
|
listen_ports = w;
|
|
}
|
|
|
|
#ifdef NETDATA_INTERNAL_CHECKS
|
|
info("The network viewer is monitoring inbound connections for port %u", ntohs(value));
|
|
#endif
|
|
}
|
|
|
|
/**
|
|
* Read listen table
|
|
*
|
|
* Read the table with all ports that we are listen on host.
|
|
*/
|
|
static void read_listen_table()
|
|
{
|
|
uint16_t key = 0;
|
|
uint16_t next_key;
|
|
|
|
int fd = map_fd[NETDATA_SOCKET_LISTEN_TABLE];
|
|
uint8_t value;
|
|
while (bpf_map_get_next_key(fd, &key, &next_key) == 0) {
|
|
int test = bpf_map_lookup_elem(fd, &key, &value);
|
|
if (test < 0) {
|
|
key = next_key;
|
|
continue;
|
|
}
|
|
|
|
// The correct protocol must come from kernel
|
|
update_listen_table(htons(key), (key == 53)?IPPROTO_UDP:IPPROTO_TCP);
|
|
|
|
key = next_key;
|
|
}
|
|
|
|
if (next_key) {
|
|
// The correct protocol must come from kernel
|
|
update_listen_table(htons(next_key), (key == 53)?IPPROTO_UDP:IPPROTO_TCP);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Socket read hash
|
|
*
|
|
* This is the thread callback.
|
|
* This thread is necessary, because we cannot freeze the whole plugin to read the data on very busy socket.
|
|
*
|
|
* @param ptr It is a NULL value for this thread.
|
|
*
|
|
* @return It always returns NULL.
|
|
*/
|
|
void *ebpf_socket_read_hash(void *ptr)
|
|
{
|
|
ebpf_module_t *em = (ebpf_module_t *)ptr;
|
|
|
|
read_thread_closed = 0;
|
|
heartbeat_t hb;
|
|
heartbeat_init(&hb);
|
|
usec_t step = NETDATA_SOCKET_READ_SLEEP_MS;
|
|
int fd_ipv4 = map_fd[NETDATA_SOCKET_IPV4_HASH_TABLE];
|
|
int fd_ipv6 = map_fd[NETDATA_SOCKET_IPV6_HASH_TABLE];
|
|
int network_connection = em->optional;
|
|
while (!close_ebpf_plugin) {
|
|
usec_t dt = heartbeat_next(&hb, step);
|
|
(void)dt;
|
|
|
|
pthread_mutex_lock(&nv_mutex);
|
|
read_listen_table();
|
|
read_socket_hash_table(fd_ipv4, AF_INET, network_connection);
|
|
read_socket_hash_table(fd_ipv6, AF_INET6, network_connection);
|
|
wait_to_plot = 1;
|
|
pthread_mutex_unlock(&nv_mutex);
|
|
}
|
|
|
|
read_thread_closed = 1;
|
|
return NULL;
|
|
}
|
|
|
|
/**
|
|
* Read the hash table and store data to allocated vectors.
|
|
*/
|
|
static void read_hash_global_tables()
|
|
{
|
|
uint64_t idx;
|
|
netdata_idx_t res[NETDATA_SOCKET_COUNTER];
|
|
|
|
netdata_idx_t *val = socket_hash_values;
|
|
int fd = map_fd[NETDATA_SOCKET_GLOBAL_HASH_TABLE];
|
|
for (idx = 0; idx < NETDATA_SOCKET_COUNTER; idx++) {
|
|
if (!bpf_map_lookup_elem(fd, &idx, val)) {
|
|
uint64_t total = 0;
|
|
int i;
|
|
int end = (running_on_kernel < NETDATA_KERNEL_V4_15) ? 1 : ebpf_nprocs;
|
|
for (i = 0; i < end; i++)
|
|
total += val[i];
|
|
|
|
res[idx] = total;
|
|
} else {
|
|
res[idx] = 0;
|
|
}
|
|
}
|
|
|
|
socket_aggregated_data[0].call = res[NETDATA_KEY_CALLS_TCP_SENDMSG];
|
|
socket_aggregated_data[1].call = res[NETDATA_KEY_CALLS_TCP_CLEANUP_RBUF];
|
|
socket_aggregated_data[2].call = res[NETDATA_KEY_CALLS_TCP_CLOSE];
|
|
socket_aggregated_data[3].call = res[NETDATA_KEY_CALLS_UDP_RECVMSG];
|
|
socket_aggregated_data[4].call = res[NETDATA_KEY_CALLS_UDP_SENDMSG];
|
|
socket_aggregated_data[5].call = res[NETDATA_KEY_TCP_RETRANSMIT];
|
|
|
|
socket_aggregated_data[0].ecall = res[NETDATA_KEY_ERROR_TCP_SENDMSG];
|
|
socket_aggregated_data[1].ecall = res[NETDATA_KEY_ERROR_TCP_CLEANUP_RBUF];
|
|
socket_aggregated_data[3].ecall = res[NETDATA_KEY_ERROR_UDP_RECVMSG];
|
|
socket_aggregated_data[4].ecall = res[NETDATA_KEY_ERROR_UDP_SENDMSG];
|
|
|
|
socket_aggregated_data[0].bytes = res[NETDATA_KEY_BYTES_TCP_SENDMSG];
|
|
socket_aggregated_data[1].bytes = res[NETDATA_KEY_BYTES_TCP_CLEANUP_RBUF];
|
|
socket_aggregated_data[3].bytes = res[NETDATA_KEY_BYTES_UDP_RECVMSG];
|
|
socket_aggregated_data[4].bytes = res[NETDATA_KEY_BYTES_UDP_SENDMSG];
|
|
}
|
|
|
|
/**
|
|
* Fill publish apps when necessary.
|
|
*
|
|
* @param current_pid the PID that I am updating
|
|
* @param eb the structure with data read from memory.
|
|
*/
|
|
void ebpf_socket_fill_publish_apps(uint32_t current_pid, ebpf_bandwidth_t *eb)
|
|
{
|
|
ebpf_socket_publish_apps_t *curr = socket_bandwidth_curr[current_pid];
|
|
ebpf_socket_publish_apps_t *prev = socket_bandwidth_prev[current_pid];
|
|
if (!curr) {
|
|
ebpf_socket_publish_apps_t *ptr = callocz(2, sizeof(ebpf_socket_publish_apps_t));
|
|
curr = &ptr[0];
|
|
socket_bandwidth_curr[current_pid] = curr;
|
|
prev = &ptr[1];
|
|
socket_bandwidth_prev[current_pid] = prev;
|
|
} else {
|
|
memcpy(prev, curr, sizeof(ebpf_socket_publish_apps_t));
|
|
}
|
|
|
|
curr->bytes_sent = eb->bytes_sent;
|
|
curr->bytes_received = eb->bytes_received;
|
|
curr->call_tcp_sent = eb->call_tcp_sent;
|
|
curr->call_tcp_received = eb->call_tcp_received;
|
|
curr->retransmit = eb->retransmit;
|
|
curr->call_udp_sent = eb->call_udp_sent;
|
|
curr->call_udp_received = eb->call_udp_received;
|
|
|
|
ebpf_socket_update_apps_publish(curr, prev);
|
|
}
|
|
|
|
/**
|
|
* Bandwidth accumulator.
|
|
*
|
|
* @param out the vector with the values to sum
|
|
*/
|
|
void ebpf_socket_bandwidth_accumulator(ebpf_bandwidth_t *out)
|
|
{
|
|
int i, end = (running_on_kernel >= NETDATA_KERNEL_V4_15) ? ebpf_nprocs : 1;
|
|
ebpf_bandwidth_t *total = &out[0];
|
|
for (i = 1; i < end; i++) {
|
|
ebpf_bandwidth_t *move = &out[i];
|
|
total->bytes_sent += move->bytes_sent;
|
|
total->bytes_received += move->bytes_received;
|
|
total->call_tcp_sent += move->call_tcp_sent;
|
|
total->call_tcp_received += move->call_tcp_received;
|
|
total->retransmit += move->retransmit;
|
|
total->call_udp_sent += move->call_udp_sent;
|
|
total->call_udp_received += move->call_udp_received;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Update the apps data reading information from the hash table
|
|
*/
|
|
static void ebpf_socket_update_apps_data()
|
|
{
|
|
int fd = map_fd[NETDATA_SOCKET_APPS_HASH_TABLE];
|
|
ebpf_bandwidth_t *eb = bandwidth_vector;
|
|
uint32_t key;
|
|
struct pid_stat *pids = root_of_pids;
|
|
while (pids) {
|
|
key = pids->pid;
|
|
|
|
if (bpf_map_lookup_elem(fd, &key, eb)) {
|
|
pids = pids->next;
|
|
continue;
|
|
}
|
|
|
|
ebpf_socket_bandwidth_accumulator(eb);
|
|
|
|
ebpf_socket_fill_publish_apps(key, eb);
|
|
|
|
pids = pids->next;
|
|
}
|
|
}
|
|
|
|
/*****************************************************************
|
|
*
|
|
* FUNCTIONS WITH THE MAIN LOOP
|
|
*
|
|
*****************************************************************/
|
|
|
|
struct netdata_static_thread socket_threads = {"EBPF SOCKET READ",
|
|
NULL, NULL, 1, NULL,
|
|
NULL, ebpf_socket_read_hash };
|
|
|
|
/**
|
|
* Main loop for this collector.
|
|
*
|
|
* @param step the number of microseconds used with heart beat
|
|
* @param em the structure with thread information
|
|
*/
|
|
static void socket_collector(usec_t step, ebpf_module_t *em)
|
|
{
|
|
UNUSED(em);
|
|
UNUSED(step);
|
|
heartbeat_t hb;
|
|
heartbeat_init(&hb);
|
|
|
|
socket_threads.thread = mallocz(sizeof(netdata_thread_t));
|
|
|
|
netdata_thread_create(socket_threads.thread, socket_threads.name,
|
|
NETDATA_THREAD_OPTION_JOINABLE, ebpf_socket_read_hash, em);
|
|
|
|
int socket_apps_enabled = ebpf_modules[EBPF_MODULE_SOCKET_IDX].apps_charts;
|
|
int socket_global_enabled = ebpf_modules[EBPF_MODULE_SOCKET_IDX].global_charts;
|
|
int network_connection = em->optional;
|
|
while (!close_ebpf_plugin) {
|
|
pthread_mutex_lock(&collect_data_mutex);
|
|
pthread_cond_wait(&collect_data_cond_var, &collect_data_mutex);
|
|
|
|
if (socket_global_enabled)
|
|
read_hash_global_tables();
|
|
|
|
if (socket_apps_enabled)
|
|
ebpf_socket_update_apps_data();
|
|
|
|
calculate_nv_plot();
|
|
|
|
pthread_mutex_lock(&lock);
|
|
if (socket_global_enabled)
|
|
ebpf_socket_send_data(em);
|
|
|
|
if (socket_apps_enabled)
|
|
ebpf_socket_send_apps_data(em, apps_groups_root_target);
|
|
|
|
fflush(stdout);
|
|
|
|
if (network_connection) {
|
|
// We are calling fflush many times, because when we have a lot of dimensions
|
|
// we began to have not expected outputs and Netdata closed the plugin.
|
|
pthread_mutex_lock(&nv_mutex);
|
|
ebpf_socket_create_nv_charts(&inbound_vectors);
|
|
fflush(stdout);
|
|
ebpf_socket_send_nv_data(&inbound_vectors);
|
|
|
|
ebpf_socket_create_nv_charts(&outbound_vectors);
|
|
fflush(stdout);
|
|
ebpf_socket_send_nv_data(&outbound_vectors);
|
|
wait_to_plot = 0;
|
|
pthread_mutex_unlock(&nv_mutex);
|
|
|
|
}
|
|
|
|
pthread_mutex_unlock(&collect_data_mutex);
|
|
pthread_mutex_unlock(&lock);
|
|
|
|
}
|
|
}
|
|
|
|
/*****************************************************************
|
|
*
|
|
* FUNCTIONS TO CLOSE THE THREAD
|
|
*
|
|
*****************************************************************/
|
|
|
|
|
|
/**
|
|
* Clean internal socket plot
|
|
*
|
|
* Clean all structures allocated with strdupz.
|
|
*
|
|
* @param ptr the pointer with addresses to clean.
|
|
*/
|
|
static inline void clean_internal_socket_plot(netdata_socket_plot_t *ptr)
|
|
{
|
|
freez(ptr->dimension_recv);
|
|
freez(ptr->dimension_sent);
|
|
freez(ptr->resolved_name);
|
|
freez(ptr->dimension_retransmit);
|
|
}
|
|
|
|
/**
|
|
* Clean socket plot
|
|
*
|
|
* Clean the allocated data for inbound and outbound vectors.
|
|
*/
|
|
static void clean_allocated_socket_plot()
|
|
{
|
|
uint32_t i;
|
|
uint32_t end = inbound_vectors.last;
|
|
netdata_socket_plot_t *plot = inbound_vectors.plot;
|
|
for (i = 0; i < end; i++) {
|
|
clean_internal_socket_plot(&plot[i]);
|
|
}
|
|
|
|
clean_internal_socket_plot(&plot[inbound_vectors.last]);
|
|
|
|
end = outbound_vectors.last;
|
|
plot = outbound_vectors.plot;
|
|
for (i = 0; i < end; i++) {
|
|
clean_internal_socket_plot(&plot[i]);
|
|
}
|
|
clean_internal_socket_plot(&plot[outbound_vectors.last]);
|
|
}
|
|
|
|
/**
|
|
* Clean netowrk ports allocated during initializaion.
|
|
*
|
|
* @param ptr a pointer to the link list.
|
|
*/
|
|
static void clean_network_ports(ebpf_network_viewer_port_list_t *ptr)
|
|
{
|
|
if (unlikely(!ptr))
|
|
return;
|
|
|
|
while (ptr) {
|
|
ebpf_network_viewer_port_list_t *next = ptr->next;
|
|
freez(ptr->value);
|
|
freez(ptr);
|
|
ptr = next;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Clean service names
|
|
*
|
|
* Clean the allocated link list that stores names.
|
|
*
|
|
* @param names the link list.
|
|
*/
|
|
static void clean_service_names(ebpf_network_viewer_dim_name_t *names)
|
|
{
|
|
if (unlikely(!names))
|
|
return;
|
|
|
|
while (names) {
|
|
ebpf_network_viewer_dim_name_t *next = names->next;
|
|
freez(names->name);
|
|
freez(names);
|
|
names = next;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Clean hostnames
|
|
*
|
|
* @param hostnames the hostnames to clean
|
|
*/
|
|
static void clean_hostnames(ebpf_network_viewer_hostname_list_t *hostnames)
|
|
{
|
|
if (unlikely(!hostnames))
|
|
return;
|
|
|
|
while (hostnames) {
|
|
ebpf_network_viewer_hostname_list_t *next = hostnames->next;
|
|
freez(hostnames->value);
|
|
simple_pattern_free(hostnames->value_pattern);
|
|
freez(hostnames);
|
|
hostnames = next;
|
|
}
|
|
}
|
|
|
|
void clean_thread_structures() {
|
|
struct pid_stat *pids = root_of_pids;
|
|
while (pids) {
|
|
freez(socket_bandwidth_curr[pids->pid]);
|
|
|
|
pids = pids->next;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Clean up the main thread.
|
|
*
|
|
* @param ptr thread data.
|
|
*/
|
|
static void ebpf_socket_cleanup(void *ptr)
|
|
{
|
|
ebpf_module_t *em = (ebpf_module_t *)ptr;
|
|
if (!em->enabled)
|
|
return;
|
|
|
|
heartbeat_t hb;
|
|
heartbeat_init(&hb);
|
|
uint32_t tick = 200*USEC_PER_MS;
|
|
while (!read_thread_closed) {
|
|
usec_t dt = heartbeat_next(&hb, tick);
|
|
UNUSED(dt);
|
|
}
|
|
|
|
freez(socket_aggregated_data);
|
|
freez(socket_publish_aggregated);
|
|
freez(socket_hash_values);
|
|
|
|
clean_thread_structures();
|
|
freez(socket_bandwidth_curr);
|
|
freez(socket_bandwidth_prev);
|
|
freez(bandwidth_vector);
|
|
|
|
freez(socket_values);
|
|
clean_allocated_socket_plot();
|
|
freez(inbound_vectors.plot);
|
|
freez(outbound_vectors.plot);
|
|
|
|
clean_port_structure(&listen_ports);
|
|
|
|
ebpf_modules[EBPF_MODULE_SOCKET_IDX].enabled = 0;
|
|
|
|
clean_network_ports(network_viewer_opt.included_port);
|
|
clean_network_ports(network_viewer_opt.excluded_port);
|
|
clean_service_names(network_viewer_opt.names);
|
|
clean_hostnames(network_viewer_opt.included_hostnames);
|
|
clean_hostnames(network_viewer_opt.excluded_hostnames);
|
|
|
|
pthread_mutex_destroy(&nv_mutex);
|
|
freez(socket_data.map_fd);
|
|
|
|
freez(socket_threads.thread);
|
|
|
|
struct bpf_program *prog;
|
|
size_t i = 0 ;
|
|
bpf_object__for_each_program(prog, objects) {
|
|
bpf_link__destroy(probe_links[i]);
|
|
i++;
|
|
}
|
|
bpf_object__close(objects);
|
|
finalized_threads = 1;
|
|
}
|
|
|
|
/*****************************************************************
|
|
*
|
|
* FUNCTIONS TO START THREAD
|
|
*
|
|
*****************************************************************/
|
|
|
|
/**
|
|
* Allocate vectors used with this thread.
|
|
* We are not testing the return, because callocz does this and shutdown the software
|
|
* case it was not possible to allocate.
|
|
*
|
|
* @param length is the length for the vectors used inside the collector.
|
|
*/
|
|
static void ebpf_socket_allocate_global_vectors(size_t length)
|
|
{
|
|
socket_aggregated_data = callocz(length, sizeof(netdata_syscall_stat_t));
|
|
socket_publish_aggregated = callocz(length, sizeof(netdata_publish_syscall_t));
|
|
socket_hash_values = callocz(ebpf_nprocs, sizeof(netdata_idx_t));
|
|
|
|
socket_bandwidth_curr = callocz((size_t)pid_max, sizeof(ebpf_socket_publish_apps_t *));
|
|
socket_bandwidth_prev = callocz((size_t)pid_max, sizeof(ebpf_socket_publish_apps_t *));
|
|
bandwidth_vector = callocz((size_t)ebpf_nprocs, sizeof(ebpf_bandwidth_t));
|
|
|
|
socket_values = callocz((size_t)ebpf_nprocs, sizeof(netdata_socket_t));
|
|
inbound_vectors.plot = callocz(network_viewer_opt.max_dim, sizeof(netdata_socket_plot_t));
|
|
outbound_vectors.plot = callocz(network_viewer_opt.max_dim, sizeof(netdata_socket_plot_t));
|
|
}
|
|
|
|
/**
|
|
* Set local function pointers, this function will never be compiled with static libraries
|
|
*/
|
|
static void set_local_pointers()
|
|
{
|
|
map_fd = socket_data.map_fd;
|
|
}
|
|
|
|
/**
|
|
* Initialize Inbound and Outbound
|
|
*
|
|
* Initialize the common outbound and inbound sockets.
|
|
*/
|
|
static void initialize_inbound_outbound()
|
|
{
|
|
inbound_vectors.last = network_viewer_opt.max_dim - 1;
|
|
outbound_vectors.last = inbound_vectors.last;
|
|
fill_last_nv_dimension(&inbound_vectors.plot[inbound_vectors.last], 0);
|
|
fill_last_nv_dimension(&outbound_vectors.plot[outbound_vectors.last], 1);
|
|
}
|
|
|
|
/*****************************************************************
|
|
*
|
|
* EBPF SOCKET THREAD
|
|
*
|
|
*****************************************************************/
|
|
|
|
/**
|
|
* Socket thread
|
|
*
|
|
* Thread used to generate socket charts.
|
|
*
|
|
* @param ptr a pointer to `struct ebpf_module`
|
|
*
|
|
* @return It always return NULL
|
|
*/
|
|
void *ebpf_socket_thread(void *ptr)
|
|
{
|
|
netdata_thread_cleanup_push(ebpf_socket_cleanup, ptr);
|
|
|
|
avl_init_lock(&inbound_vectors.tree, compare_sockets);
|
|
avl_init_lock(&outbound_vectors.tree, compare_sockets);
|
|
|
|
ebpf_module_t *em = (ebpf_module_t *)ptr;
|
|
fill_ebpf_data(&socket_data);
|
|
|
|
if (!em->enabled)
|
|
goto endsocket;
|
|
|
|
if (pthread_mutex_init(&nv_mutex, NULL)) {
|
|
error("Cannot initialize local mutex");
|
|
goto endsocket;
|
|
}
|
|
pthread_mutex_lock(&lock);
|
|
|
|
ebpf_socket_allocate_global_vectors(NETDATA_MAX_SOCKET_VECTOR);
|
|
initialize_inbound_outbound();
|
|
|
|
if (ebpf_update_kernel(&socket_data)) {
|
|
pthread_mutex_unlock(&lock);
|
|
goto endsocket;
|
|
}
|
|
|
|
set_local_pointers();
|
|
probe_links = ebpf_load_program(ebpf_plugin_dir, em, kernel_string, &objects, socket_data.map_fd);
|
|
if (!probe_links) {
|
|
pthread_mutex_unlock(&lock);
|
|
goto endsocket;
|
|
}
|
|
|
|
ebpf_global_labels(
|
|
socket_aggregated_data, socket_publish_aggregated, socket_dimension_names, socket_id_names,
|
|
NETDATA_MAX_SOCKET_VECTOR);
|
|
|
|
ebpf_create_global_charts(em);
|
|
|
|
finalized_threads = 0;
|
|
pthread_mutex_unlock(&lock);
|
|
|
|
socket_collector((usec_t)(em->update_time * USEC_PER_SEC), em);
|
|
|
|
endsocket:
|
|
netdata_thread_cleanup_pop(1);
|
|
return NULL;
|
|
}
|