1
0
Fork 0
mirror of https://github.com/Ysurac/openmptcprouter-feeds.git synced 2025-03-09 15:40:03 +00:00
openmptcprouter-feeds/mptcp-bpf-minrtt/src/mptcp_bpf_minrtt.c
Ycarus (Yannick Chabanois) def156816f Add MPTCP bpf minRTT scheduler
2025-01-02 17:27:15 +01:00

167 lines
4.3 KiB
C

// SPDX-License-Identifier: GPL-2.0
/* Copyright (c) 2023, SUSE. */
#include <linux/bpf.h>
#include <limits.h>
#include "bpf_tcp_helpers.h"
char _license[] SEC("license") = "GPL";
#define MPTCP_SEND_minrtt_SIZE 65428
struct subflow_send_info {
__u8 subflow_id;
__u64 linger_time;
};
extern bool mptcp_subflow_active(struct mptcp_subflow_context *subflow) __ksym;
extern void mptcp_set_timeout(struct sock *sk) __ksym;
extern __u64 mptcp_wnd_end(const struct mptcp_sock *msk) __ksym;
extern bool tcp_stream_memory_free(const struct sock *sk, int wake) __ksym;
extern bool bpf_mptcp_subflow_queues_empty(struct sock *sk) __ksym;
extern void mptcp_pm_subflow_chk_stale(const struct mptcp_sock *msk, struct sock *ssk) __ksym;
#define SSK_MODE_ACTIVE 0
#define SSK_MODE_BACKUP 1
#define SSK_MODE_MAX 2
static __always_inline bool __sk_stream_memory_free(const struct sock *sk, int wake)
{
if (sk->sk_wmem_queued >= sk->sk_sndbuf)
return false;
return tcp_stream_memory_free(sk, wake);
}
static __always_inline bool sk_stream_memory_free(const struct sock *sk)
{
return __sk_stream_memory_free(sk, 0);
}
SEC("struct_ops/mptcp_sched_minrtt_init")
void BPF_PROG(mptcp_sched_minrtt_init, struct mptcp_sock *msk)
{
bpf_printk("load mptcp_sched_minrtt_init \n");
}
SEC("struct_ops/mptcp_sched_minrtt_release")
void BPF_PROG(mptcp_sched_minrtt_release, struct mptcp_sock *msk)
{
}
static int bpf_minrtt_get_send(struct mptcp_sock *msk,
struct mptcp_sched_data *data)
{
struct mptcp_subflow_context *subflow;
struct sock *sk = (struct sock *)msk;
__u32 selected_minrtt = 0;
__u32 selected_subflow_id = 0;
__u32 minrtt = 0;
__u64 linger_time;
struct sock *ssk;
int i;
for (i = 0; i < data->subflows && i < MPTCP_SUBFLOWS_MAX; i++) {
subflow = bpf_mptcp_subflow_ctx_by_pos(data, i);
if (!subflow)
break;
ssk = mptcp_subflow_tcp_sock(subflow);
if (!mptcp_subflow_active(subflow))
continue;
const struct tcp_sock *tp = bpf_skc_to_tcp_sock(ssk);
if (!tp){
continue;
}
minrtt = tp->srtt_us;
if (minrtt < selected_minrtt || (selected_minrtt == 0 && selected_subflow_id == 0)){
selected_minrtt = tp->srtt_us;
selected_subflow_id = i;
}
}
mptcp_set_timeout(sk);
subflow = bpf_mptcp_subflow_ctx_by_pos(data, selected_subflow_id);
if (!subflow){
return -1;
}
out:
mptcp_subflow_set_scheduled(subflow, true);
return 0;
}
static __always_inline bool tcp_write_queue_empty(struct sock *sk)
{
const struct tcp_sock *tp = bpf_skc_to_tcp_sock(sk);
return tp ? tp->write_seq == tp->snd_nxt : true;
}
static __always_inline bool tcp_rtx_and_write_queues_empty(struct sock *sk)
{
return bpf_mptcp_subflow_queues_empty(sk) && tcp_write_queue_empty(sk);
}
static int bpf_burst_get_retrans(struct mptcp_sock *msk,
struct mptcp_sched_data *data)
{
int backup = MPTCP_SUBFLOWS_MAX, pick = MPTCP_SUBFLOWS_MAX, subflow_id;
struct mptcp_subflow_context *subflow;
int min_stale_count = INT_MAX;
struct sock *ssk;
for (int i = 0; i < data->subflows && i < MPTCP_SUBFLOWS_MAX; i++) {
subflow = bpf_mptcp_subflow_ctx_by_pos(data, i);
if (!subflow)
break;
if (!mptcp_subflow_active(subflow))
continue;
ssk = mptcp_subflow_tcp_sock(subflow);
/* still data outstanding at TCP level? skip this */
if (!tcp_rtx_and_write_queues_empty(ssk)) {
mptcp_pm_subflow_chk_stale(msk, ssk);
min_stale_count = min(min_stale_count, subflow->stale_count);
continue;
}
if (subflow->backup) {
if (backup == MPTCP_SUBFLOWS_MAX)
backup = i;
continue;
}
if (pick == MPTCP_SUBFLOWS_MAX)
pick = i;
}
if (pick < MPTCP_SUBFLOWS_MAX) {
subflow_id = pick;
goto out;
}
subflow_id = min_stale_count > 1 ? backup : MPTCP_SUBFLOWS_MAX;
out:
subflow = bpf_mptcp_subflow_ctx_by_pos(data, subflow_id);
if (!subflow)
return -1;
mptcp_subflow_set_scheduled(subflow, true);
return 0;
}
int BPF_STRUCT_OPS(bpf_minrtt_get_subflow, struct mptcp_sock *msk,
struct mptcp_sched_data *data)
{
if (data->reinject)
return bpf_burst_get_retrans(msk, data);
return bpf_minrtt_get_send(msk, data);
}
SEC(".struct_ops")
struct mptcp_sched_ops minrtt = {
.init = (void *)mptcp_sched_minrtt_init,
.release = (void *)mptcp_sched_minrtt_release,
.get_subflow = (void *)bpf_minrtt_get_subflow,
.name = "bpf_minrtt",
};