From patchwork Tue Oct 30 18:46:06 2012 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Sasha Levin X-Patchwork-Id: 1672761 Return-Path: X-Original-To: patchwork-linux-nfs@patchwork.kernel.org Delivered-To: patchwork-process-083081@patchwork1.kernel.org Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by patchwork1.kernel.org (Postfix) with ESMTP id A56933FD2B for ; Tue, 30 Oct 2012 18:50:24 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S965520Ab2J3Sry (ORCPT ); Tue, 30 Oct 2012 14:47:54 -0400 Received: from mail-qa0-f53.google.com ([209.85.216.53]:56829 "EHLO mail-qa0-f53.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S965249Ab2J3Srw (ORCPT ); Tue, 30 Oct 2012 14:47:52 -0400 Received: by mail-qa0-f53.google.com with SMTP id c1so390325qae.19 for ; Tue, 30 Oct 2012 11:47:52 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=from:to:cc:subject:date:message-id:x-mailer:in-reply-to:references; bh=YzGbKLosPR+p7IHedS2LLSdV0HOPmzywimuZkfk5NVk=; b=En/v6MpgUinzwCnfBCscfGSpg8ey+w7Lu3SN7Z/o1PcRyejY4hMRwG/L8z/FKRflSh 1LXF95DgWMNME1ZL5jJRluaoZZ0NLm7WfGxfMHOQns5Ujx2wvqoPgMnmfs2vUbDwBgJD ab5n4c5Xq2xRpTwkuk9IDQwxNaO1zk7RAULDRr+I4V/V1hen4k5ugMxdR8H3aW7zhkqU bm4nq/EJU8opmztaRZqCkkjtqTuumvzyb2gC8uBBdoTkK7Nkw5Lg88G6F9oUFp6uqLBY 4+lNWjuHKAeHM+FVtQoWzNE6f5+8Lz8HD+7+3oILBza4afZTYo4dDcqfJv+bKSWcC70d yJbw== Received: by 10.49.51.7 with SMTP id g7mr26442439qeo.34.1351622872187; Tue, 30 Oct 2012 11:47:52 -0700 (PDT) Received: from localhost.localdomain ([50.145.164.180]) by mx.google.com with ESMTPS id ga9sm682790qab.22.2012.10.30.11.47.46 (version=TLSv1/SSLv3 cipher=OTHER); Tue, 30 Oct 2012 11:47:51 -0700 (PDT) From: Sasha Levin To: torvalds@linux-foundation.org Cc: tj@kernel.org, akpm@linux-foundation.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, paul.gortmaker@windriver.com, davem@davemloft.net, rostedt@goodmis.org, mingo@elte.hu, ebiederm@xmission.com, aarcange@redhat.com, ericvh@gmail.com, netdev@vger.kernel.org, josh@joshtriplett.org, eric.dumazet@gmail.com, mathieu.desnoyers@efficios.com, axboe@kernel.dk, agk@redhat.com, dm-devel@redhat.com, neilb@suse.de, ccaulfie@redhat.com, teigland@redhat.com, Trond.Myklebust@netapp.com, bfields@fieldses.org, fweisbec@gmail.com, jesse@nicira.com, venkat.x.venkatsubra@oracle.com, ejt@redhat.com, snitzer@redhat.com, edumazet@google.com, linux-nfs@vger.kernel.org, dev@openvswitch.org, rds-devel@oss.oracle.com, lw@cn.fujitsu.com, Sasha Levin Subject: [PATCH v8 10/16] dlm: use new hashtable implementation Date: Tue, 30 Oct 2012 14:46:06 -0400 Message-Id: <1351622772-16400-10-git-send-email-levinsasha928@gmail.com> X-Mailer: git-send-email 1.7.12.4 In-Reply-To: <1351622772-16400-1-git-send-email-levinsasha928@gmail.com> References: <1351622772-16400-1-git-send-email-levinsasha928@gmail.com> Sender: linux-nfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org Switch dlm to use the new hashtable implementation. This reduces the amount of generic unrelated code in the dlm. Signed-off-by: Sasha Levin --- fs/dlm/lowcomms.c | 53 ++++++++++++++++++----------------------------------- 1 file changed, 18 insertions(+), 35 deletions(-) diff --git a/fs/dlm/lowcomms.c b/fs/dlm/lowcomms.c index 331ea4f..19acdfe 100644 --- a/fs/dlm/lowcomms.c +++ b/fs/dlm/lowcomms.c @@ -55,6 +55,7 @@ #include #include #include +#include #include "dlm_internal.h" #include "lowcomms.h" @@ -62,7 +63,7 @@ #include "config.h" #define NEEDED_RMEM (4*1024*1024) -#define CONN_HASH_SIZE 32 +#define CONN_HASH_BITS 5 /* Number of messages to send before rescheduling */ #define MAX_SEND_MSG_COUNT 25 @@ -158,34 +159,27 @@ static int dlm_allow_conn; static struct workqueue_struct *recv_workqueue; static struct workqueue_struct *send_workqueue; -static struct hlist_head connection_hash[CONN_HASH_SIZE]; +/* + * On a sidenote, hash function could be very simple because most clusters + * have simple sequential nodeids, so we should be able to go straight to + * a connection struct in the array. We don't utilize it at the moment, + * but it's something work to keep in mind. + */ +static DEFINE_HASHTABLE(connection_hash, CONN_HASH_BITS); static DEFINE_MUTEX(connections_lock); static struct kmem_cache *con_cache; static void process_recv_sockets(struct work_struct *work); static void process_send_sockets(struct work_struct *work); - -/* This is deliberately very simple because most clusters have simple - sequential nodeids, so we should be able to go straight to a connection - struct in the array */ -static inline int nodeid_hash(int nodeid) -{ - return nodeid & (CONN_HASH_SIZE-1); -} - static struct connection *__find_con(int nodeid) { - int r; struct hlist_node *h; struct connection *con; - r = nodeid_hash(nodeid); - - hlist_for_each_entry(con, h, &connection_hash[r], list) { + hash_for_each_possible(connection_hash, con, h, list, nodeid) if (con->nodeid == nodeid) return con; - } return NULL; } @@ -196,7 +190,6 @@ static struct connection *__find_con(int nodeid) static struct connection *__nodeid2con(int nodeid, gfp_t alloc) { struct connection *con = NULL; - int r; con = __find_con(nodeid); if (con || !alloc) @@ -206,8 +199,7 @@ static struct connection *__nodeid2con(int nodeid, gfp_t alloc) if (!con) return NULL; - r = nodeid_hash(nodeid); - hlist_add_head(&con->list, &connection_hash[r]); + hash_add(connection_hash, &con->list, nodeid); con->nodeid = nodeid; mutex_init(&con->sock_mutex); @@ -235,11 +227,8 @@ static void foreach_conn(void (*conn_func)(struct connection *c)) struct hlist_node *h, *n; struct connection *con; - for (i = 0; i < CONN_HASH_SIZE; i++) { - hlist_for_each_entry_safe(con, h, n, &connection_hash[i], list){ - conn_func(con); - } - } + hash_for_each_safe(connection_hash, i, h, n, con, list) + conn_func(con); } static struct connection *nodeid2con(int nodeid, gfp_t allocation) @@ -262,12 +251,10 @@ static struct connection *assoc2con(int assoc_id) mutex_lock(&connections_lock); - for (i = 0 ; i < CONN_HASH_SIZE; i++) { - hlist_for_each_entry(con, h, &connection_hash[i], list) { - if (con->sctp_assoc == assoc_id) { - mutex_unlock(&connections_lock); - return con; - } + hash_for_each(connection_hash, i, h, con, list) { + if (con->sctp_assoc == assoc_id) { + mutex_unlock(&connections_lock); + return con; } } mutex_unlock(&connections_lock); @@ -1638,7 +1625,7 @@ static void free_conn(struct connection *con) close_connection(con, true); if (con->othercon) kmem_cache_free(con_cache, con->othercon); - hlist_del(&con->list); + hash_del(&con->list); kmem_cache_free(con_cache, con); } @@ -1667,10 +1654,6 @@ int dlm_lowcomms_start(void) { int error = -EINVAL; struct connection *con; - int i; - - for (i = 0; i < CONN_HASH_SIZE; i++) - INIT_HLIST_HEAD(&connection_hash[i]); init_local(); if (!dlm_local_count) {