From patchwork Tue Nov 7 21:40:41 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: David Wei X-Patchwork-Id: 13449360 X-Patchwork-Delegate: kuba@kernel.org Received: from lindbergh.monkeyblade.net (lindbergh.monkeyblade.net [23.128.96.19]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 8A7B92BD01 for ; Tue, 7 Nov 2023 21:41:11 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=davidwei-uk.20230601.gappssmtp.com header.i=@davidwei-uk.20230601.gappssmtp.com header.b="0qfxYbvB" Received: from mail-pl1-x630.google.com (mail-pl1-x630.google.com [IPv6:2607:f8b0:4864:20::630]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 2712510EA for ; Tue, 7 Nov 2023 13:41:11 -0800 (PST) Received: by mail-pl1-x630.google.com with SMTP id d9443c01a7336-1cc0e78ec92so41917055ad.3 for ; Tue, 07 Nov 2023 13:41:11 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=davidwei-uk.20230601.gappssmtp.com; s=20230601; t=1699393270; x=1699998070; darn=vger.kernel.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=DEYoqijzZJW9k3dNn5XO9gR9ltrXX/l1ozIXtxfwbU8=; b=0qfxYbvBi+FA5VLTVj5BUubHqOc0REVtaihsVsYyZEw8gMK3yYSvFhiRDwQTldCckd mUEF6RbihQb2N3f18LEW+YZesQBu8NkmWAImGfMOdB+K1+AR1DimFMenOlMnda1DtEPC wgPko3l31ONOiit4m2cn/oAG/0g8rx/xlsNHXj1QJtqybpFrBvAPHyqpe2dFAD6sy1LM VylFuHQXLqaOapxPH4Dd07SF8BtMZUjz+I8uW8PgObEAZtkHl0mSkHU8D2wqHO7fO1FK wj4eeK02VFgmxqYgSdOvTOQN4tSTfFXl+5pBOlmb/nR1Yg6bwIb/w/yddAiSqZuDN1x7 wyoA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1699393270; x=1699998070; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=DEYoqijzZJW9k3dNn5XO9gR9ltrXX/l1ozIXtxfwbU8=; b=cRFcbGeCRBTrsZ7vpGhW7O6+xGDwuUj2oD2Ch8gHALQPDBUu44yv8kxhhWxuCjN8G+ pIS1HoflS1hNjhuZJi6yEYKB9P42WEoUzemvMOLjkRN3xrJ6FIzdq7eBzQAbm7p8k7I/ 4tCTeUUBw4fY8cCCzP9RQ6hB07KHorMGJXax1w6rAajZ6gw99dxLx4QSXpRX4gXWmREw T2mfEdSgC6wU8TtiHhlGSDNfMR9Jr+ygYaDcY8mfzBwPc0as4j3E+jOPf2bauP32sox0 Cja/FCOqMrDSYgq1wFKzmobqlLriFG7XOFe3he+1OB4ZFuDyJHzIHFgbbBMbg0AKnZ+t aojg== X-Gm-Message-State: AOJu0YyV74AmF/n8NztwJhBqbDRa3f8KDX/yPO6EM4uhaHrXbmj8Szl/ 9WqiucTZVXoWab16UElgULvIiQ== X-Google-Smtp-Source: AGHT+IEKSuqqWQvMVmIb/Hqh7B+Gxlpgb+OwjhNZAxtVTPcTPHKznyOvGf1WAmMO7Igy9ntVidWomg== X-Received: by 2002:a17:903:230c:b0:1cc:6e42:1413 with SMTP id d12-20020a170903230c00b001cc6e421413mr256995plh.57.1699393270702; Tue, 07 Nov 2023 13:41:10 -0800 (PST) Received: from localhost (fwdproxy-prn-001.fbsv.net. [2a03:2880:ff:1::face:b00c]) by smtp.gmail.com with ESMTPSA id u1-20020a170902e80100b001cc3a6813f8sm268781plg.154.2023.11.07.13.41.10 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 07 Nov 2023 13:41:10 -0800 (PST) From: David Wei To: io-uring@vger.kernel.org, netdev@vger.kernel.org Cc: Jens Axboe , Pavel Begunkov , Jakub Kicinski , Paolo Abeni , "David S. Miller" , Eric Dumazet , Jesper Dangaard Brouer , David Ahern , Mina Almasry , Willem de Bruijn , Dragos Tatulea Subject: [PATCH 16/20] net: execute custom callback from napi Date: Tue, 7 Nov 2023 13:40:41 -0800 Message-Id: <20231107214045.2172393-17-dw@davidwei.uk> X-Mailer: git-send-email 2.39.3 In-Reply-To: <20231107214045.2172393-1-dw@davidwei.uk> References: <20231107214045.2172393-1-dw@davidwei.uk> Precedence: bulk X-Mailing-List: netdev@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 X-Patchwork-Delegate: kuba@kernel.org From: Pavel Begunkov Sometimes we want to access a napi protected resource from task context like in the case of io_uring zc falling back to copy and accessing the buffer ring. Add a helper function that allows to execute a custom function from napi context by first stopping it similarly to napi_busy_loop(). Experimental and might go away after convertion to custom page pools. It has to share more code with napi_busy_loop(). It also might be spinning too long a better breaking mechanism. Signed-off-by: Pavel Begunkov Signed-off-by: David Wei --- include/net/busy_poll.h | 2 ++ net/core/dev.c | 51 +++++++++++++++++++++++++++++++++++++++++ 2 files changed, 53 insertions(+) diff --git a/include/net/busy_poll.h b/include/net/busy_poll.h index 4dabeb6c76d3..292c3b4eaa7a 100644 --- a/include/net/busy_poll.h +++ b/include/net/busy_poll.h @@ -47,6 +47,8 @@ bool sk_busy_loop_end(void *p, unsigned long start_time); void napi_busy_loop(unsigned int napi_id, bool (*loop_end)(void *, unsigned long), void *loop_end_arg, bool prefer_busy_poll, u16 budget); +void napi_execute(unsigned int napi_id, + bool (*cb)(void *), void *cb_arg); #else /* CONFIG_NET_RX_BUSY_POLL */ static inline unsigned long net_busy_loop_on(void) diff --git a/net/core/dev.c b/net/core/dev.c index 02949a929e7f..66397ac1d8fc 100644 --- a/net/core/dev.c +++ b/net/core/dev.c @@ -6291,6 +6291,57 @@ void napi_busy_loop(unsigned int napi_id, } EXPORT_SYMBOL(napi_busy_loop); +void napi_execute(unsigned int napi_id, + bool (*cb)(void *), void *cb_arg) +{ + bool done = false; + unsigned long val; + void *have_poll_lock = NULL; + struct napi_struct *napi; + + rcu_read_lock(); + napi = napi_by_id(napi_id); + if (!napi) + goto out; + + if (!IS_ENABLED(CONFIG_PREEMPT_RT)) + preempt_disable(); + for (;;) { + local_bh_disable(); + val = READ_ONCE(napi->state); + + /* If multiple threads are competing for this napi, + * we avoid dirtying napi->state as much as we can. + */ + if (val & (NAPIF_STATE_DISABLE | NAPIF_STATE_SCHED | + NAPIF_STATE_IN_BUSY_POLL)) + goto restart; + + if (cmpxchg(&napi->state, val, + val | NAPIF_STATE_IN_BUSY_POLL | + NAPIF_STATE_SCHED) != val) + goto restart; + + have_poll_lock = netpoll_poll_lock(napi); + cb(cb_arg); + done = true; + gro_normal_list(napi); + local_bh_enable(); + break; +restart: + local_bh_enable(); + if (unlikely(need_resched())) + break; + cpu_relax(); + } + if (done) + busy_poll_stop(napi, have_poll_lock, false, 1); + if (!IS_ENABLED(CONFIG_PREEMPT_RT)) + preempt_enable(); +out: + rcu_read_unlock(); +} + #endif /* CONFIG_NET_RX_BUSY_POLL */ static void napi_hash_add(struct napi_struct *napi)