Skip to content

Commit

Permalink
Squash to "selftests/bpf: Add bpf_burst scheduler & test"
Browse files Browse the repository at this point in the history
Use the newly added bpf_for_each() helper to walk the conn_list.

Use bpf_mptcp_send_info_to_ssk() helper.

Drop bpf_subflow_send_info, use subflow_send_info instead.

Signed-off-by: Geliang Tang <tanggeliang@kylinos.cn>
Message-Id: <66a4b211db4f06b8c6fddf62bc900d886d8b707b.1734947922.git.tanggeliang@kylinos.cn>
  • Loading branch information
Geliang Tang authored and Patchew Applier committed Dec 23, 2024
1 parent e62cf3a commit dcde300
Showing 1 changed file with 26 additions and 42 deletions.
68 changes: 26 additions & 42 deletions tools/testing/selftests/bpf/progs/mptcp_bpf_burst.c
Original file line number Diff line number Diff line change
Expand Up @@ -11,11 +11,6 @@ char _license[] SEC("license") = "GPL";

#define min(a, b) ((a) < (b) ? (a) : (b))

struct bpf_subflow_send_info {
__u8 subflow_id;
__u64 linger_time;
};

extern bool mptcp_subflow_active(struct mptcp_subflow_context *subflow) __ksym;
extern void mptcp_set_timeout(struct sock *sk) __ksym;
extern __u64 mptcp_wnd_end(const struct mptcp_sock *msk) __ksym;
Expand Down Expand Up @@ -68,10 +63,9 @@ void BPF_PROG(mptcp_sched_burst_release, struct mptcp_sock *msk)
}

SEC("struct_ops")
int BPF_PROG(bpf_burst_get_send, struct mptcp_sock *msk,
struct mptcp_sched_data *data)
int BPF_PROG(bpf_burst_get_send, struct mptcp_sock *msk)
{
struct bpf_subflow_send_info send_info[SSK_MODE_MAX];
struct subflow_send_info send_info[SSK_MODE_MAX];
struct mptcp_subflow_context *subflow;
struct sock *sk = (struct sock *)msk;
__u32 pace, burst, wmem;
Expand All @@ -81,18 +75,12 @@ int BPF_PROG(bpf_burst_get_send, struct mptcp_sock *msk,

/* pick the subflow with the lower wmem/wspace ratio */
for (i = 0; i < SSK_MODE_MAX; ++i) {
send_info[i].subflow_id = MPTCP_SUBFLOWS_MAX;
send_info[i].ssk = NULL;
send_info[i].linger_time = -1;
}

for (i = 0; i < data->subflows && i < MPTCP_SUBFLOWS_MAX; i++) {
bool backup;

subflow = bpf_mptcp_subflow_ctx_by_pos(data, i);
if (!subflow)
break;

backup = subflow->backup || subflow->request_bkup;
bpf_for_each(mptcp_subflow, subflow, msk) {
bool backup = subflow->backup || subflow->request_bkup;

ssk = mptcp_subflow_tcp_sock(subflow);
if (!mptcp_subflow_active(subflow))
Expand All @@ -110,23 +98,24 @@ int BPF_PROG(bpf_burst_get_send, struct mptcp_sock *msk,

linger_time = div_u64((__u64)ssk->sk_wmem_queued << 32, pace);
if (linger_time < send_info[backup].linger_time) {
send_info[backup].subflow_id = i;
send_info[backup].ssk = ssk;
send_info[backup].linger_time = linger_time;
}
}
mptcp_set_timeout(sk);

/* pick the best backup if no other subflow is active */
if (!nr_active)
send_info[SSK_MODE_ACTIVE].subflow_id = send_info[SSK_MODE_BACKUP].subflow_id;
send_info[SSK_MODE_ACTIVE].ssk = send_info[SSK_MODE_BACKUP].ssk;

subflow = bpf_mptcp_subflow_ctx_by_pos(data, send_info[SSK_MODE_ACTIVE].subflow_id);
if (!subflow)
return -1;
ssk = mptcp_subflow_tcp_sock(subflow);
ssk = bpf_mptcp_send_info_to_ssk(&send_info[SSK_MODE_ACTIVE]);
if (!ssk || !sk_stream_memory_free(ssk))
return -1;

subflow = bpf_mptcp_subflow_ctx(ssk);
if (!subflow)
return -1;

burst = min(MPTCP_SEND_BURST_SIZE, mptcp_wnd_end(msk) - msk->snd_nxt);
wmem = ssk->sk_wmem_queued;
if (!burst)
Expand All @@ -143,23 +132,18 @@ int BPF_PROG(bpf_burst_get_send, struct mptcp_sock *msk,
}

SEC("struct_ops")
int BPF_PROG(bpf_burst_get_retrans, struct mptcp_sock *msk,
struct mptcp_sched_data *data)
int BPF_PROG(bpf_burst_get_retrans, struct mptcp_sock *msk)
{
int backup = MPTCP_SUBFLOWS_MAX, pick = MPTCP_SUBFLOWS_MAX, subflow_id;
struct sock *backup = NULL, *pick = NULL;
struct mptcp_subflow_context *subflow;
int min_stale_count = INT_MAX;
struct sock *ssk;

for (int i = 0; i < data->subflows && i < MPTCP_SUBFLOWS_MAX; i++) {
subflow = bpf_mptcp_subflow_ctx_by_pos(data, i);
if (!subflow)
break;
bpf_for_each(mptcp_subflow, subflow, msk) {
struct sock *ssk = bpf_mptcp_subflow_tcp_sock(subflow);

if (!mptcp_subflow_active(subflow))
if (!ssk || !mptcp_subflow_active(subflow))
continue;

ssk = mptcp_subflow_tcp_sock(subflow);
/* still data outstanding at TCP level? skip this */
if (!tcp_rtx_and_write_queues_empty(ssk)) {
mptcp_pm_subflow_chk_stale(msk, ssk);
Expand All @@ -168,23 +152,23 @@ int BPF_PROG(bpf_burst_get_retrans, struct mptcp_sock *msk,
}

if (subflow->backup || subflow->request_bkup) {
if (backup == MPTCP_SUBFLOWS_MAX)
backup = i;
if (!backup)
backup = ssk;
continue;
}

if (pick == MPTCP_SUBFLOWS_MAX)
pick = i;
if (!pick)
pick = ssk;
}

if (pick < MPTCP_SUBFLOWS_MAX) {
subflow_id = pick;
if (pick)
goto out;
}
subflow_id = min_stale_count > 1 ? backup : MPTCP_SUBFLOWS_MAX;
pick = min_stale_count > 1 ? backup : NULL;

out:
subflow = bpf_mptcp_subflow_ctx_by_pos(data, subflow_id);
if (!pick)
return -1;
subflow = bpf_mptcp_subflow_ctx(pick);
if (!subflow)
return -1;
mptcp_subflow_set_scheduled(subflow, true);
Expand Down

0 comments on commit dcde300

Please sign in to comment.