From patchwork Mon Oct 28 21:35:34 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: John Ousterhout X-Patchwork-Id: 13854177 X-Patchwork-Delegate: kuba@kernel.org Received: from smtp1.cs.Stanford.EDU (smtp1.cs.stanford.edu [171.64.64.25]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id C30C51E22E6 for ; Mon, 28 Oct 2024 21:36:20 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=171.64.64.25 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1730151384; cv=none; b=acSczFAHmkyUJd+pksRtkenID6oSOJUPgR1Ezyt5WgcYjqWIJrUUkPyHgpfilZXkupiVpUGV9qC2tUn7kQiwwUSfb9yyumyfRmI6aJowxss40JM/XDD8h1/Sc/uRLGNITNa7M7uTTsQgbJFalOj5x6ulXIHAPQcp2Vnr1Ege+y8= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1730151384; c=relaxed/simple; bh=ZctzJyZbMZ94uCQFWLNgp2QdoPfjEIksWB8LzeH2z0I=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=fDto41WI9RqA6gVfm/4vqZ9xtYGQAn6Jno5QkRnldzWORtQLDUM5N4DVv7qONCEDxdDV7a1+Fd6F0OO1nL2yPea4oghOlZJU5f8xjL4aA+yxTnSSPGIv9H5GAM8adDbcqof39vlm2e6/mgJEyDQ2NyyjY68EoabyR8vvsdRuPj4= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=cs.stanford.edu; spf=pass smtp.mailfrom=cs.stanford.edu; dkim=pass (2048-bit key) header.d=cs.stanford.edu header.i=@cs.stanford.edu header.b=TxpIqGv0; arc=none smtp.client-ip=171.64.64.25 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=cs.stanford.edu Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=cs.stanford.edu Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=cs.stanford.edu header.i=@cs.stanford.edu header.b="TxpIqGv0" DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=cs.stanford.edu; s=cs2308; h=Content-Transfer-Encoding:MIME-Version: References:In-Reply-To:Message-ID:Date:Subject:Cc:To:From:Sender:Reply-To: Content-Type:Content-ID:Content-Description:Resent-Date:Resent-From: Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Id:List-Help: List-Unsubscribe:List-Subscribe:List-Post:List-Owner:List-Archive; bh=a6tBn/tV56L1/dnkfZQauPx/bh2px+sv91sHFyxVfkU=; t=1730151380; x=1731015380; b=TxpIqGv0UY5fVg+ieHwz8cOQ8wUEUIR8Fog73VmoZLUzpWYuZBvLjrlb2e0H4kGhFctRtPZ0l+2 ddA3rmZ2tfVPt38jajcI2wxaDLwHpJFSU96VHeXwLjGaneiYNQI/tLdphu4V9ubYfjGAqWsTUxpYB YYfL+UR9wjSJyhX4Xfz/6DCMSbCRT+lTHxUq5ghDIgGFN+ULczItV+ZvQkCfYP0neSFlscJlFX4fv 48LxkasxuAdyh1I6IqDs9g04yREQJovYmP8fAJmHkTsGrgvSL0bBUR4gv3WNec6djvQZPf6hkCLIo HQfleEC/wtjpEhrUWhhRWVEG4tQdh2Ll0gGQ==; Received: from ouster2016.stanford.edu ([172.24.72.71]:54106 helo=localhost.localdomain) by smtp1.cs.Stanford.EDU with esmtpsa (TLS1.2) tls TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 (Exim 4.94.2) (envelope-from ) id 1t5XPP-0005xj-7v; Mon, 28 Oct 2024 14:36:20 -0700 From: John Ousterhout To: netdev@vger.kernel.org Cc: John Ousterhout Subject: [PATCH net-next 07/12] net: homa: create homa_sock.h and homa_sock.c Date: Mon, 28 Oct 2024 14:35:34 -0700 Message-ID: <20241028213541.1529-8-ouster@cs.stanford.edu> X-Mailer: git-send-email 2.45.1 In-Reply-To: <20241028213541.1529-1-ouster@cs.stanford.edu> References: <20241028213541.1529-1-ouster@cs.stanford.edu> Precedence: bulk X-Mailing-List: netdev@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 X-Spam-Score: -101.0 X-Scan-Signature: 77334d8d872153c9350eeec3077c0d46 X-Patchwork-Delegate: kuba@kernel.org These files provide functions for managing the state that Homa keeps for each open Homa socket. Signed-off-by: John Ousterhout --- net/homa/homa_sock.c | 329 +++++++++++++++++++++++++++++++++++ net/homa/homa_sock.h | 399 +++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 728 insertions(+) create mode 100644 net/homa/homa_sock.c create mode 100644 net/homa/homa_sock.h diff --git a/net/homa/homa_sock.c b/net/homa/homa_sock.c new file mode 100644 index 000000000000..f02a0394fb00 --- /dev/null +++ b/net/homa/homa_sock.c @@ -0,0 +1,329 @@ +// SPDX-License-Identifier: BSD-2-Clause + +/* This file manages homa_sock and homa_socktab objects. */ + +#include "homa_impl.h" +#include "homa_peer.h" +#include "homa_pool.h" + +/** + * homa_socktab_init() - Constructor for homa_socktabs. + * @socktab: The object to initialize; previous contents are discarded. + */ +void homa_socktab_init(struct homa_socktab *socktab) +{ + int i; + + spin_lock_init(&socktab->write_lock); + for (i = 0; i < HOMA_SOCKTAB_BUCKETS; i++) + INIT_HLIST_HEAD(&socktab->buckets[i]); +} + +/** + * homa_socktab_destroy() - Destructor for homa_socktabs. + * @socktab: The object to destroy. + */ +void homa_socktab_destroy(struct homa_socktab *socktab) +{ + struct homa_socktab_scan scan; + struct homa_sock *hsk; + + for (hsk = homa_socktab_start_scan(socktab, &scan); hsk; + hsk = homa_socktab_next(&scan)) { + homa_sock_destroy(hsk); + } +} + +/** + * homa_socktab_start_scan() - Begin an iteration over all of the sockets + * in a socktab. + * @socktab: Socktab to scan. + * @scan: Will hold the current state of the scan; any existing + * contents are discarded. + * + * Return: The first socket in the table, or NULL if the table is + * empty. + * + * Each call to homa_socktab_next will return the next socket in the table. + * All sockets that are present in the table at the time this function is + * invoked will eventually be returned, as long as they are not removed + * from the table. It is safe to remove sockets from the table and/or + * delete them while the scan is in progress. If a socket is removed from + * the table during the scan, it may or may not be returned by + * homa_socktab_next. New entries added during the scan may or may not be + * returned. The caller should use RCU to prevent socket storage from + * being reclaimed during the scan. + */ +struct homa_sock *homa_socktab_start_scan(struct homa_socktab *socktab, + struct homa_socktab_scan *scan) +{ + scan->socktab = socktab; + scan->current_bucket = -1; + scan->next = NULL; + return homa_socktab_next(scan); +} + +/** + * homa_socktab_next() - Return the next socket in an iteration over a socktab. + * @scan: State of the scan. + * + * Return: The next socket in the table, or NULL if the iteration has + * returned all of the sockets in the table. Sockets are not + * returned in any particular order. It's possible that the + * returned socket has been destroyed. + */ +struct homa_sock *homa_socktab_next(struct homa_socktab_scan *scan) +{ + struct homa_socktab_links *links; + struct homa_sock *hsk; + + while (1) { + while (!scan->next) { + scan->current_bucket++; + if (scan->current_bucket >= HOMA_SOCKTAB_BUCKETS) + return NULL; + scan->next = (struct homa_socktab_links *) + hlist_first_rcu(&scan->socktab->buckets + [scan->current_bucket]); + } + links = scan->next; + hsk = links->sock; + scan->next = (struct homa_socktab_links *)hlist_next_rcu(&links + ->hash_links); + return hsk; + } +} + +/** + * homa_sock_init() - Constructor for homa_sock objects. This function + * initializes only the parts of the socket that are owned by Homa. + * @hsk: Object to initialize. + * @homa: Homa implementation that will manage the socket. + * + * Return: always 0 (success). + */ +void homa_sock_init(struct homa_sock *hsk, struct homa *homa) +{ + struct homa_socktab *socktab = homa->port_map; + int i; + + spin_lock_bh(&socktab->write_lock); + atomic_set(&hsk->protect_count, 0); + spin_lock_init(&hsk->lock); + hsk->last_locker = "none"; + atomic_set(&hsk->protect_count, 0); + hsk->homa = homa; + hsk->ip_header_length = (hsk->inet.sk.sk_family == AF_INET) + ? HOMA_IPV4_HEADER_LENGTH : HOMA_IPV6_HEADER_LENGTH; + hsk->shutdown = false; + while (1) { + if (homa->next_client_port < HOMA_MIN_DEFAULT_PORT) + homa->next_client_port = HOMA_MIN_DEFAULT_PORT; + if (!homa_sock_find(socktab, homa->next_client_port)) + break; + homa->next_client_port++; + } + hsk->port = homa->next_client_port; + hsk->inet.inet_num = hsk->port; + hsk->inet.inet_sport = htons(hsk->port); + homa->next_client_port++; + hsk->socktab_links.sock = hsk; + hlist_add_head_rcu(&hsk->socktab_links.hash_links, + &socktab->buckets[homa_port_hash(hsk->port)]); + INIT_LIST_HEAD(&hsk->active_rpcs); + INIT_LIST_HEAD(&hsk->dead_rpcs); + hsk->dead_skbs = 0; + INIT_LIST_HEAD(&hsk->waiting_for_bufs); + INIT_LIST_HEAD(&hsk->ready_requests); + INIT_LIST_HEAD(&hsk->ready_responses); + INIT_LIST_HEAD(&hsk->request_interests); + INIT_LIST_HEAD(&hsk->response_interests); + for (i = 0; i < HOMA_CLIENT_RPC_BUCKETS; i++) { + struct homa_rpc_bucket *bucket = &hsk->client_rpc_buckets[i]; + + spin_lock_init(&bucket->lock); + INIT_HLIST_HEAD(&bucket->rpcs); + bucket->id = i; + } + for (i = 0; i < HOMA_SERVER_RPC_BUCKETS; i++) { + struct homa_rpc_bucket *bucket = &hsk->server_rpc_buckets[i]; + + spin_lock_init(&bucket->lock); + INIT_HLIST_HEAD(&bucket->rpcs); + bucket->id = i + 1000000; + } + hsk->buffer_pool = kzalloc(sizeof(*hsk->buffer_pool), GFP_KERNEL); + spin_unlock_bh(&socktab->write_lock); +} + +/** + * homa_sock_shutdown() - Disable a socket so that it can no longer + * be used for either sending or receiving messages. Any system calls + * currently waiting to send or receive messages will be aborted. + * @hsk: Socket to shut down. + */ +void homa_sock_shutdown(struct homa_sock *hsk) +{ + struct homa_interest *interest; + struct homa_rpc *rpc; + int i; + + homa_sock_lock(hsk, "homa_socket_shutdown"); + if (hsk->shutdown) { + homa_sock_unlock(hsk); + return; + } + + /* The order of cleanup is very important, because there could be + * active operations that hold RPC locks but not the socket lock. + * 1. Set @shutdown; this ensures that no new RPCs will be created for + * this socket (though some creations might already be in progress). + * 2. Remove the socket from the port map: this ensures that + * incoming packets for the socket will be dropped. + * 3. Go through all of the RPCs and delete them; this will + * synchronize with any operations in progress. + * 4. Perform other socket cleanup: at this point we know that + * there will be no concurrent activities on individual RPCs. + * 5. Don't delete the buffer pool until after all of the RPCs + * have been reaped. + * See sync.txt for additional information about locking. + */ + hsk->shutdown = true; + spin_lock_bh(&hsk->homa->port_map->write_lock); + hlist_del_rcu(&hsk->socktab_links.hash_links); + spin_unlock_bh(&hsk->homa->port_map->write_lock); + homa_sock_unlock(hsk); + + list_for_each_entry_rcu(rpc, &hsk->active_rpcs, active_links) { + homa_rpc_lock(rpc, "homa_sock_shutdown"); + homa_rpc_free(rpc); + homa_rpc_unlock(rpc); + } + + homa_sock_lock(hsk, "homa_socket_shutdown #2"); + list_for_each_entry(interest, &hsk->request_interests, request_links) + wake_up_process(interest->thread); + list_for_each_entry(interest, &hsk->response_interests, response_links) + wake_up_process(interest->thread); + homa_sock_unlock(hsk); + + i = 0; + while (!list_empty(&hsk->dead_rpcs)) { + homa_rpc_reap(hsk, 1000); + i++; + } + + homa_pool_destroy(hsk->buffer_pool); + kfree(hsk->buffer_pool); + hsk->buffer_pool = NULL; +} + +/** + * homa_sock_destroy() - Destructor for homa_sock objects. This function + * only cleans up the parts of the object that are owned by Homa. + * @hsk: Socket to destroy. + */ +void homa_sock_destroy(struct homa_sock *hsk) +{ + homa_sock_shutdown(hsk); + sock_set_flag(&hsk->inet.sk, SOCK_RCU_FREE); +} + +/** + * homa_sock_bind() - Associates a server port with a socket; if there + * was a previous server port assignment for @hsk, it is abandoned. + * @socktab: Hash table in which the binding will be recorded. + * @hsk: Homa socket. + * @port: Desired server port for @hsk. If 0, then this call + * becomes a no-op: the socket will continue to use + * its randomly assigned client port. + * + * Return: 0 for success, otherwise a negative errno. + */ +int homa_sock_bind(struct homa_socktab *socktab, struct homa_sock *hsk, + __u16 port) +{ + struct homa_sock *owner; + int result = 0; + + if (port == 0) + return result; + if (port >= HOMA_MIN_DEFAULT_PORT) + return -EINVAL; + homa_sock_lock(hsk, "homa_sock_bind"); + spin_lock_bh(&socktab->write_lock); + if (hsk->shutdown) { + result = -ESHUTDOWN; + goto done; + } + + owner = homa_sock_find(socktab, port); + if (owner) { + if (owner != hsk) + result = -EADDRINUSE; + goto done; + } + hlist_del_rcu(&hsk->socktab_links.hash_links); + hsk->port = port; + hsk->inet.inet_num = port; + hsk->inet.inet_sport = htons(hsk->port); + hlist_add_head_rcu(&hsk->socktab_links.hash_links, + &socktab->buckets[homa_port_hash(port)]); +done: + spin_unlock_bh(&socktab->write_lock); + homa_sock_unlock(hsk); + return result; +} + +/** + * homa_sock_find() - Returns the socket associated with a given port. + * @socktab: Hash table in which to perform lookup. + * @port: The port of interest. + * Return: The socket that owns @port, or NULL if none. + * + * Note: this function uses RCU list-searching facilities, but it doesn't + * call rcu_read_lock. The caller should do that, if the caller cares (this + * way, the caller's use of the socket will also be protected). + */ +struct homa_sock *homa_sock_find(struct homa_socktab *socktab, __u16 port) +{ + struct homa_socktab_links *link; + struct homa_sock *result = NULL; + + hlist_for_each_entry_rcu(link, &socktab->buckets[homa_port_hash(port)], + hash_links) { + struct homa_sock *hsk = link->sock; + + if (hsk->port == port) { + result = hsk; + break; + } + } + return result; +} + +/** + * homa_sock_lock_slow() - This function implements the slow path for + * acquiring a socketC lock. It is invoked when a socket lock isn't immediately + * available. It waits for the lock, but also records statistics about + * the waiting time. + * @hsk: socket to lock. + */ +void homa_sock_lock_slow(struct homa_sock *hsk) +{ + spin_lock_bh(&hsk->lock); +} + +/** + * homa_bucket_lock_slow() - This function implements the slow path for + * locking a bucket in one of the hash tables of RPCs. It is invoked when a + * lock isn't immediately available. It waits for the lock, but also records + * statistics about the waiting time. + * @bucket: The hash table bucket to lock. + * @id: ID of the particular RPC being locked (multiple RPCs may + * share a single bucket lock). + */ +void homa_bucket_lock_slow(struct homa_rpc_bucket *bucket, __u64 id) +{ + spin_lock_bh(&bucket->lock); +} diff --git a/net/homa/homa_sock.h b/net/homa/homa_sock.h new file mode 100644 index 000000000000..bd56c7345a99 --- /dev/null +++ b/net/homa/homa_sock.h @@ -0,0 +1,399 @@ +/* SPDX-License-Identifier: BSD-2-Clause */ + +/* This file defines structs and other things related to Homa sockets. */ + +#ifndef _HOMA_SOCK_H +#define _HOMA_SOCK_H + +/* Forward declarations. */ +struct homa; +struct homa_pool; + +void homa_sock_lock_slow(struct homa_sock *hsk); + +/** + * define HOMA_SOCKTAB_BUCKETS - Number of hash buckets in a homa_socktab. + * Must be a power of 2. + */ +#define HOMA_SOCKTAB_BUCKETS 1024 + +/** + * struct homa_socktab - A hash table that maps from port numbers (either + * client or server) to homa_sock objects. + * + * This table is managed exclusively by homa_socktab.c, using RCU to + * minimize synchronization during lookups. + */ +struct homa_socktab { + /** + * @mutex: Controls all modifications to this object; not needed + * for socket lookups (RCU is used instead). Also used to + * synchronize port allocation. + */ + spinlock_t write_lock; + + /** + * @buckets: Heads of chains for hash table buckets. Chains + * consist of homa_socktab_link objects. + */ + struct hlist_head buckets[HOMA_SOCKTAB_BUCKETS]; +}; + +/** + * struct homa_socktab_links - Used to link homa_socks into the hash chains + * of a homa_socktab. + */ +struct homa_socktab_links { + /* Must be the first element of the struct! */ + struct hlist_node hash_links; + struct homa_sock *sock; +}; + +/** + * struct homa_socktab_scan - Records the state of an iteration over all + * the entries in a homa_socktab, in a way that permits RCU-safe deletion + * of entries. + */ +struct homa_socktab_scan { + /** @socktab: The table that is being scanned. */ + struct homa_socktab *socktab; + + /** + * @current_bucket: the index of the bucket in socktab->buckets + * currently being scanned. If >= HOMA_SOCKTAB_BUCKETS, the scan + * is complete. + */ + int current_bucket; + + /** + * @next: the next socket to return from homa_socktab_next (this + * socket has not yet been returned). NULL means there are no + * more sockets in the current bucket. + */ + struct homa_socktab_links *next; +}; + +/** + * struct homa_rpc_bucket - One bucket in a hash table of RPCs. + */ + +struct homa_rpc_bucket { + /** + * @lock: serves as a lock both for this bucket (e.g., when + * adding and removing RPCs) and also for all of the RPCs in + * the bucket. Must be held whenever manipulating an RPC in + * this bucket. This dual purpose permits clean and safe + * deletion and garbage collection of RPCs. + */ + spinlock_t lock; + + /** @rpcs: list of RPCs that hash to this bucket. */ + struct hlist_head rpcs; + + /** @id: identifier for this bucket, used in error messages etc. + * It's the index of the bucket within its hash table bucket + * array, with an additional offset to separate server and + * client RPCs. + */ + int id; +}; + +/** + * define HOMA_CLIENT_RPC_BUCKETS - Number of buckets in hash tables for + * client RPCs. Must be a power of 2. + */ +#define HOMA_CLIENT_RPC_BUCKETS 1024 + +/** + * define HOMA_SERVER_RPC_BUCKETS - Number of buckets in hash tables for + * server RPCs. Must be a power of 2. + */ +#define HOMA_SERVER_RPC_BUCKETS 1024 + +/** + * struct homa_sock - Information about an open socket. + */ +struct homa_sock { + /* Info for other network layers. Note: IPv6 info (struct ipv6_pinfo + * comes at the very end of the struct, *after* Homa's data, if this + * socket uses IPv6). + */ + union { + /** @sock: generic socket data; must be the first field. */ + struct sock sock; + + /** + * @inet: generic Internet socket data; must also be the + first field (contains sock as its first member). + */ + struct inet_sock inet; + }; + + /** + * @lock: Must be held when modifying fields such as interests + * and lists of RPCs. This lock is used in place of sk->sk_lock + * because it's used differently (it's always used as a simple + * spin lock). See sync.txt for more on Homa's synchronization + * strategy. + */ + spinlock_t lock; + + /** + * @last_locker: identifies the code that most recently acquired + * @lock successfully. Occasionally used for debugging. + */ + char *last_locker; + + /** + * @protect_count: counts the number of calls to homa_protect_rpcs + * for which there have not yet been calls to homa_unprotect_rpcs. + * See sync.txt for more info. + */ + atomic_t protect_count; + + /** + * @homa: Overall state about the Homa implementation. NULL + * means this socket has been deleted. + */ + struct homa *homa; + + /** @shutdown: True means the socket is no longer usable. */ + bool shutdown; + + /** + * @port: Port number: identifies this socket uniquely among all + * those on this node. + */ + __u16 port; + + /** + * @ip_header_length: Length of IP headers for this socket (depends + * on IPv4 vs. IPv6). + */ + int ip_header_length; + + /** + * @client_socktab_links: Links this socket into the homa_socktab + * based on @port. + */ + struct homa_socktab_links socktab_links; + + /** + * @active_rpcs: List of all existing RPCs related to this socket, + * including both client and server RPCs. This list isn't strictly + * needed, since RPCs are already in one of the hash tables below, + * but it's more efficient for homa_timer to have this list + * (so it doesn't have to scan large numbers of hash buckets). + * The list is sorted, with the oldest RPC first. Manipulate with + * RCU so timer can access without locking. + */ + struct list_head active_rpcs; + + /** + * @dead_rpcs: Contains RPCs for which homa_rpc_free has been + * called, but their packet buffers haven't yet been freed. + */ + struct list_head dead_rpcs; + + /** @dead_skbs: Total number of socket buffers in RPCs on dead_rpcs. */ + int dead_skbs; + + /** + * @waiting_for_bufs: Contains RPCs that are blocked because there + * wasn't enough space in the buffer pool region for their incoming + * messages. Sorted in increasing order of message length. + */ + struct list_head waiting_for_bufs; + + /** + * @ready_requests: Contains server RPCs whose request message is + * in a state requiring attention from a user process. The head is + * oldest, i.e. next to return. + */ + struct list_head ready_requests; + + /** + * @ready_responses: Contains client RPCs whose response message is + * in a state requiring attention from a user process. The head is + * oldest, i.e. next to return. + */ + struct list_head ready_responses; + + /** + * @request_interests: List of threads that want to receive incoming + * request messages. + */ + struct list_head request_interests; + + /** + * @response_interests: List of threads that want to receive incoming + * response messages. + */ + struct list_head response_interests; + + /** + * @client_rpc_buckets: Hash table for fast lookup of client RPCs. + * Modifications are synchronized with bucket locks, not + * the socket lock. + */ + struct homa_rpc_bucket client_rpc_buckets[HOMA_CLIENT_RPC_BUCKETS]; + + /** + * @server_rpc_buckets: Hash table for fast lookup of server RPCs. + * Modifications are synchronized with bucket locks, not + * the socket lock. + */ + struct homa_rpc_bucket server_rpc_buckets[HOMA_SERVER_RPC_BUCKETS]; + + /** + * @buffer_pool: used to allocate buffer space for incoming messages. + * Storage is dynamically allocated. + */ + struct homa_pool *buffer_pool; +}; + +void homa_bucket_lock_slow(struct homa_rpc_bucket *bucket, __u64 id); +int homa_sock_bind(struct homa_socktab *socktab, + struct homa_sock *hsk, __u16 port); +void homa_sock_destroy(struct homa_sock *hsk); +struct homa_sock * + homa_sock_find(struct homa_socktab *socktab, __u16 port); +void homa_sock_init(struct homa_sock *hsk, struct homa *homa); +void homa_sock_shutdown(struct homa_sock *hsk); +int homa_socket(struct sock *sk); +void homa_socktab_destroy(struct homa_socktab *socktab); +void homa_socktab_init(struct homa_socktab *socktab); +struct homa_sock + *homa_socktab_next(struct homa_socktab_scan *scan); +struct homa_sock + *homa_socktab_start_scan(struct homa_socktab *socktab, + struct homa_socktab_scan *scan); + +/** + * homa_sock_lock() - Acquire the lock for a socket. If the socket + * isn't immediately available, record stats on the waiting time. + * @hsk: Socket to lock. + * @locker: Static string identifying where the socket was locked; + * used to track down deadlocks. + */ +static inline void homa_sock_lock(struct homa_sock *hsk, const char *locker) +{ + if (!spin_trylock_bh(&hsk->lock)) { +// printk(KERN_NOTICE "Slow path for socket %d, last locker %s", +// hsk->client_port, hsk->last_locker); + homa_sock_lock_slow(hsk); + } +// hsk->last_locker = locker; +} + +/** + * homa_sock_unlock() - Release the lock for a socket. + * @hsk: Socket to lock. + */ +static inline void homa_sock_unlock(struct homa_sock *hsk) +{ + spin_unlock_bh(&hsk->lock); +} + +/** + * port_hash() - Hash function for port numbers. + * @port: Port number being looked up. + * + * Return: The index of the bucket in which this port will be found (if + * it exists. + */ +static inline int homa_port_hash(__u16 port) +{ + /* We can use a really simple hash function here because client + * port numbers are allocated sequentially and server port numbers + * are unpredictable. + */ + return port & (HOMA_SOCKTAB_BUCKETS - 1); +} + +/** + * homa_client_rpc_bucket() - Find the bucket containing a given + * client RPC. + * @hsk: Socket associated with the RPC. + * @id: Id of the desired RPC. + * + * Return: The bucket in which this RPC will appear, if the RPC exists. + */ +static inline struct homa_rpc_bucket *homa_client_rpc_bucket(struct homa_sock *hsk, + __u64 id) +{ + /* We can use a really simple hash function here because RPC ids + * are allocated sequentially. + */ + return &hsk->client_rpc_buckets[(id >> 1) + & (HOMA_CLIENT_RPC_BUCKETS - 1)]; +} + +/** + * homa_server_rpc_bucket() - Find the bucket containing a given + * server RPC. + * @hsk: Socket associated with the RPC. + * @id: Id of the desired RPC. + * + * Return: The bucket in which this RPC will appear, if the RPC exists. + */ +static inline struct homa_rpc_bucket *homa_server_rpc_bucket(struct homa_sock *hsk, + __u64 id) +{ + /* Each client allocates RPC ids sequentially, so they will + * naturally distribute themselves across the hash space. + * Thus we can use the id directly as hash. + */ + return &hsk->server_rpc_buckets[(id >> 1) + & (HOMA_SERVER_RPC_BUCKETS - 1)]; +} + +/** + * homa_bucket_lock() - Acquire the lock for an RPC hash table bucket. + * @bucket: Bucket to lock + * @id: ID of the RPC that is requesting the lock. Normally ignored, + * but used occasionally for diagnostics and debugging. + * @locker: Static string identifying the locking code. Normally ignored, + * but used occasionally for diagnostics and debugging. + */ +static inline void homa_bucket_lock(struct homa_rpc_bucket *bucket, + __u64 id, const char *locker) +{ + if (!spin_trylock_bh(&bucket->lock)) + homa_bucket_lock_slow(bucket, id); +} + +/** + * homa_bucket_try_lock() - Acquire the lock for an RPC hash table bucket if + * it is available. + * @bucket: Bucket to lock + * @id: ID of the RPC that is requesting the lock. + * @locker: Static string identifying the locking code. Normally ignored, + * but used when debugging deadlocks. + * Return: Nonzero if lock was successfully acquired, zero if it is + * currently owned by someone else. + */ +static inline int homa_bucket_try_lock(struct homa_rpc_bucket *bucket, + __u64 id, const char *locker) +{ + if (!spin_trylock_bh(&bucket->lock)) + return 0; + return 1; +} + +/** + * homa_bucket_unlock() - Release the lock for an RPC hash table bucket. + * @bucket: Bucket to unlock. + * @id: ID of the RPC that was using the lock. + */ +static inline void homa_bucket_unlock(struct homa_rpc_bucket *bucket, __u64 id) +{ + spin_unlock_bh(&bucket->lock); +} + +static inline struct homa_sock *homa_sk(const struct sock *sk) +{ + return (struct homa_sock *)sk; +} + +#endif /* _HOMA_SOCK_H */