* [Ocfs2-devel] [PATCH 1/1] OCFS2: fasten dlm_lock_resource hash_table lookups
@ 2009-04-22 6:04 wengang wang
2009-04-23 2:27 ` Sunil Mushran
0 siblings, 1 reply; 3+ messages in thread
From: wengang wang @ 2009-04-22 6:04 UTC (permalink / raw)
To: ocfs2-devel
#backporting the 3 patches at http://kernel.us.oracle.com/~smushran/srini/ to 1.2.
enlarge hash_table capacity to fasten hash_table lookups.
Signed-off-by: Wengang wang <wen.gang.wang@oracle.com>
--
diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c
--- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c.orig 2009-04-22 11:00:37.000000000 +0800
+++ ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c 2009-04-22 11:08:27.000000000 +0800
@@ -547,7 +547,7 @@ void dlm_dump_lock_resources(struct dlm_
spin_lock(&dlm->spinlock);
for (i=0; i<DLM_HASH_BUCKETS; i++) {
- bucket = &(dlm->lockres_hash[i]);
+ bucket = dlm_lockres_hash(dlm, i);
hlist_for_each_entry(res, iter, bucket, hash_node)
dlm_print_one_lock_resource(res);
}
diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmrecovery.c.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmrecovery.c
--- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmrecovery.c.orig 2009-04-22 11:01:18.000000000 +0800
+++ ./svnocfs2-1.2/fs/ocfs2/dlm/dlmrecovery.c 2009-04-22 11:21:12.000000000 +0800
@@ -2064,7 +2064,7 @@ static void dlm_finish_local_lockres_rec
* the RECOVERING state and set the owner
* if necessary */
for (i = 0; i < DLM_HASH_BUCKETS; i++) {
- bucket = &(dlm->lockres_hash[i]);
+ bucket = dlm_lockres_hash(dlm, i);
hlist_for_each_entry(res, hash_iter, bucket, hash_node) {
if (res->state & DLM_LOCK_RES_RECOVERING) {
if (res->owner == dead_node) {
@@ -2259,7 +2259,7 @@ static void dlm_do_local_recovery_cleanu
* need to be fired as a result.
*/
for (i = 0; i < DLM_HASH_BUCKETS; i++) {
- bucket = &(dlm->lockres_hash[i]);
+ bucket = dlm_lockres_hash(dlm, i);
hlist_for_each_entry(res, iter, bucket, hash_node) {
/* always prune any $RECOVERY entries for dead nodes,
* otherwise hangs can occur during later recovery */
diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c
--- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c.orig 2009-04-22 11:00:47.000000000 +0800
+++ ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c 2009-04-22 11:33:26.000000000 +0800
@@ -49,6 +49,34 @@
#define MLOG_MASK_PREFIX (ML_DLM|ML_DLM_DOMAIN)
#include "cluster/masklog.h"
+static void dlm_free_pagevec(void **vec, int pages)
+{
+ while (pages--)
+ free_page((unsigned long)vec[pages]);
+ kfree(vec);
+}
+
+static void **dlm_alloc_pagevec(int pages)
+{
+ void **vec = kmalloc(pages * sizeof(void *), GFP_KERNEL);
+ int i;
+
+ if (!vec)
+ return NULL;
+
+ for (i = 0; i < pages; i++) {
+ vec[i] = (void *)__get_free_page(GFP_KERNEL);
+ if (!vec[i])
+ goto out_free;
+ }
+ mlog(0, "Allocated DLM hash pagevec; %d pages (%lu expected), %Zd "
+ "buckets per page\n", pages, DLM_HASH_PAGES, DLM_BUCKETS_PER_PAGE);
+ return vec;
+out_free:
+ dlm_free_pagevec(vec, i);
+ return NULL;
+}
+
/*
* ocfs2 node maps are array of long int, which limits to send them freely
* across the wire due to endianness issues. To workaround this, we convert
@@ -127,7 +155,7 @@ void __dlm_insert_lockres(struct dlm_ctx
q = &res->lockname;
q->hash = full_name_hash(q->name, q->len);
- bucket = &(dlm->lockres_hash[q->hash % DLM_HASH_BUCKETS]);
+ bucket = dlm_lockres_hash(dlm, q->hash);
/* get a reference for our hashtable */
dlm_lockres_get(res);
@@ -151,7 +179,7 @@ struct dlm_lock_resource * __dlm_lookup_
hash = full_name_hash(name, len);
- bucket = &(dlm->lockres_hash[hash % DLM_HASH_BUCKETS]);
+ bucket = dlm_lockres_hash(dlm, hash);
/* check for pre-existing lock */
hlist_for_each(iter, bucket) {
@@ -394,7 +422,7 @@ static int dlm_migrate_all_locks(struct
for (i = 0; i < DLM_HASH_BUCKETS; i++) {
redo_bucket:
n = 0;
- bucket = &dlm->lockres_hash[i];
+ bucket = dlm_lockres_hash(dlm, i);
iter = bucket->first;
while (iter) {
n++;
@@ -1356,7 +1384,8 @@ static struct dlm_ctxt *dlm_alloc_ctxt(c
goto leave;
}
- dlm->lockres_hash = (struct hlist_head *) __get_free_page(GFP_KERNEL);
+ dlm->lockres_hash = (struct hlist_head **)
+ dlm_alloc_pagevec(DLM_HASH_PAGES);
if (!dlm->lockres_hash) {
mlog_errno(-ENOMEM);
kfree(dlm->name);
@@ -1366,7 +1395,7 @@ static struct dlm_ctxt *dlm_alloc_ctxt(c
}
for (i=0; i<DLM_HASH_BUCKETS; i++)
- INIT_HLIST_HEAD(&dlm->lockres_hash[i]);
+ INIT_HLIST_HEAD(dlm_lockres_hash(dlm, i));
strcpy(dlm->name, domain);
dlm->key = key;
diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmcommon.h.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmcommon.h
--- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmcommon.h.orig 2009-04-22 10:59:51.000000000 +0800
+++ ./svnocfs2-1.2/fs/ocfs2/dlm/dlmcommon.h 2009-04-22 11:26:06.000000000 +0800
@@ -37,7 +37,14 @@
#define DLM_THREAD_SHUFFLE_INTERVAL 5 // flush everything every 5 passes
#define DLM_THREAD_MS 200 // flush at least every 200 ms
-#define DLM_HASH_BUCKETS (PAGE_SIZE / sizeof(struct hlist_head))
+#define DLM_HASH_SIZE_DEFAULT (1 << 17)
+#if DLM_HASH_SIZE_DEFAULT < PAGE_SIZE
+# define DLM_HASH_PAGES 1
+#else
+# define DLM_HASH_PAGES (DLM_HASH_SIZE_DEFAULT / PAGE_SIZE)
+#endif
+#define DLM_BUCKETS_PER_PAGE (PAGE_SIZE / sizeof(struct hlist_head))
+#define DLM_HASH_BUCKETS (DLM_HASH_PAGES * DLM_BUCKETS_PER_PAGE)
enum dlm_ast_type {
DLM_AST = 0,
@@ -86,7 +93,7 @@ enum dlm_ctxt_state {
struct dlm_ctxt
{
struct list_head list;
- struct hlist_head *lockres_hash;
+ struct hlist_head **lockres_hash;
struct list_head dirty_list;
struct list_head purge_list;
struct list_head pending_asts;
@@ -136,6 +143,13 @@ struct dlm_ctxt
struct list_head dlm_eviction_callbacks;
};
+static inline
+struct hlist_head *dlm_lockres_hash(struct dlm_ctxt *dlm, unsigned i)
+{
+ return dlm->lockres_hash[(i / DLM_BUCKETS_PER_PAGE) % DLM_HASH_PAGES]
+ + (i % DLM_BUCKETS_PER_PAGE);
+}
+
/* these keventd work queue items are for less-frequently
* called functions that cannot be directly called from the
* net message handlers for some reason, usually because
^ permalink raw reply [flat|nested] 3+ messages in thread
* [Ocfs2-devel] [PATCH 1/1] OCFS2: fasten dlm_lock_resource hash_table lookups
2009-04-22 6:04 [Ocfs2-devel] [PATCH 1/1] OCFS2: fasten dlm_lock_resource hash_table lookups wengang wang
@ 2009-04-23 2:27 ` Sunil Mushran
2009-04-23 11:06 ` Wengang Wang
0 siblings, 1 reply; 3+ messages in thread
From: Sunil Mushran @ 2009-04-23 2:27 UTC (permalink / raw)
To: ocfs2-devel
Wengang,
When we backport, we like to add bread crumbs so that we can track back
to the original patch. This helps us in cases where we, say, forget to
include a patch. Also, we want to retain original authorship.
For example:
http://oss.oracle.com/git/?p=ocfs2-1.4.git;a=commitdiff;h=174f54db9af7c67a1afaa947aed934b9c80f588c
So don't merge the patches. Add the mainline commit at the top of the
patch. As 1.2 is svn, ensure the authors name is included in an Authored-by
tag.
Has it been tested in 1.2?
Thanks
Sunil
wengang wang wrote:
> #backporting the 3 patches at http://kernel.us.oracle.com/~smushran/srini/ to 1.2.
>
> enlarge hash_table capacity to fasten hash_table lookups.
>
> Signed-off-by: Wengang wang <wen.gang.wang@oracle.com>
> --
> diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c
> --- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c.orig 2009-04-22 11:00:37.000000000 +0800
> +++ ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c 2009-04-22 11:08:27.000000000 +0800
> @@ -547,7 +547,7 @@ void dlm_dump_lock_resources(struct dlm_
>
> spin_lock(&dlm->spinlock);
> for (i=0; i<DLM_HASH_BUCKETS; i++) {
> - bucket = &(dlm->lockres_hash[i]);
> + bucket = dlm_lockres_hash(dlm, i);
> hlist_for_each_entry(res, iter, bucket, hash_node)
> dlm_print_one_lock_resource(res);
> }
> diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmrecovery.c.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmrecovery.c
> --- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmrecovery.c.orig 2009-04-22 11:01:18.000000000 +0800
> +++ ./svnocfs2-1.2/fs/ocfs2/dlm/dlmrecovery.c 2009-04-22 11:21:12.000000000 +0800
> @@ -2064,7 +2064,7 @@ static void dlm_finish_local_lockres_rec
> * the RECOVERING state and set the owner
> * if necessary */
> for (i = 0; i < DLM_HASH_BUCKETS; i++) {
> - bucket = &(dlm->lockres_hash[i]);
> + bucket = dlm_lockres_hash(dlm, i);
> hlist_for_each_entry(res, hash_iter, bucket, hash_node) {
> if (res->state & DLM_LOCK_RES_RECOVERING) {
> if (res->owner == dead_node) {
> @@ -2259,7 +2259,7 @@ static void dlm_do_local_recovery_cleanu
> * need to be fired as a result.
> */
> for (i = 0; i < DLM_HASH_BUCKETS; i++) {
> - bucket = &(dlm->lockres_hash[i]);
> + bucket = dlm_lockres_hash(dlm, i);
> hlist_for_each_entry(res, iter, bucket, hash_node) {
> /* always prune any $RECOVERY entries for dead nodes,
> * otherwise hangs can occur during later recovery */
> diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c
> --- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c.orig 2009-04-22 11:00:47.000000000 +0800
> +++ ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c 2009-04-22 11:33:26.000000000 +0800
> @@ -49,6 +49,34 @@
> #define MLOG_MASK_PREFIX (ML_DLM|ML_DLM_DOMAIN)
> #include "cluster/masklog.h"
>
> +static void dlm_free_pagevec(void **vec, int pages)
> +{
> + while (pages--)
> + free_page((unsigned long)vec[pages]);
> + kfree(vec);
> +}
> +
> +static void **dlm_alloc_pagevec(int pages)
> +{
> + void **vec = kmalloc(pages * sizeof(void *), GFP_KERNEL);
> + int i;
> +
> + if (!vec)
> + return NULL;
> +
> + for (i = 0; i < pages; i++) {
> + vec[i] = (void *)__get_free_page(GFP_KERNEL);
> + if (!vec[i])
> + goto out_free;
> + }
> + mlog(0, "Allocated DLM hash pagevec; %d pages (%lu expected), %Zd "
> + "buckets per page\n", pages, DLM_HASH_PAGES, DLM_BUCKETS_PER_PAGE);
> + return vec;
> +out_free:
> + dlm_free_pagevec(vec, i);
> + return NULL;
> +}
> +
> /*
> * ocfs2 node maps are array of long int, which limits to send them freely
> * across the wire due to endianness issues. To workaround this, we convert
> @@ -127,7 +155,7 @@ void __dlm_insert_lockres(struct dlm_ctx
>
> q = &res->lockname;
> q->hash = full_name_hash(q->name, q->len);
> - bucket = &(dlm->lockres_hash[q->hash % DLM_HASH_BUCKETS]);
> + bucket = dlm_lockres_hash(dlm, q->hash);
>
> /* get a reference for our hashtable */
> dlm_lockres_get(res);
> @@ -151,7 +179,7 @@ struct dlm_lock_resource * __dlm_lookup_
>
> hash = full_name_hash(name, len);
>
> - bucket = &(dlm->lockres_hash[hash % DLM_HASH_BUCKETS]);
> + bucket = dlm_lockres_hash(dlm, hash);
>
> /* check for pre-existing lock */
> hlist_for_each(iter, bucket) {
> @@ -394,7 +422,7 @@ static int dlm_migrate_all_locks(struct
> for (i = 0; i < DLM_HASH_BUCKETS; i++) {
> redo_bucket:
> n = 0;
> - bucket = &dlm->lockres_hash[i];
> + bucket = dlm_lockres_hash(dlm, i);
> iter = bucket->first;
> while (iter) {
> n++;
> @@ -1356,7 +1384,8 @@ static struct dlm_ctxt *dlm_alloc_ctxt(c
> goto leave;
> }
>
> - dlm->lockres_hash = (struct hlist_head *) __get_free_page(GFP_KERNEL);
> + dlm->lockres_hash = (struct hlist_head **)
> + dlm_alloc_pagevec(DLM_HASH_PAGES);
> if (!dlm->lockres_hash) {
> mlog_errno(-ENOMEM);
> kfree(dlm->name);
> @@ -1366,7 +1395,7 @@ static struct dlm_ctxt *dlm_alloc_ctxt(c
> }
>
> for (i=0; i<DLM_HASH_BUCKETS; i++)
> - INIT_HLIST_HEAD(&dlm->lockres_hash[i]);
> + INIT_HLIST_HEAD(dlm_lockres_hash(dlm, i));
>
> strcpy(dlm->name, domain);
> dlm->key = key;
> diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmcommon.h.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmcommon.h
> --- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmcommon.h.orig 2009-04-22 10:59:51.000000000 +0800
> +++ ./svnocfs2-1.2/fs/ocfs2/dlm/dlmcommon.h 2009-04-22 11:26:06.000000000 +0800
> @@ -37,7 +37,14 @@
> #define DLM_THREAD_SHUFFLE_INTERVAL 5 // flush everything every 5 passes
> #define DLM_THREAD_MS 200 // flush at least every 200 ms
>
> -#define DLM_HASH_BUCKETS (PAGE_SIZE / sizeof(struct hlist_head))
> +#define DLM_HASH_SIZE_DEFAULT (1 << 17)
> +#if DLM_HASH_SIZE_DEFAULT < PAGE_SIZE
> +# define DLM_HASH_PAGES 1
> +#else
> +# define DLM_HASH_PAGES (DLM_HASH_SIZE_DEFAULT / PAGE_SIZE)
> +#endif
> +#define DLM_BUCKETS_PER_PAGE (PAGE_SIZE / sizeof(struct hlist_head))
> +#define DLM_HASH_BUCKETS (DLM_HASH_PAGES * DLM_BUCKETS_PER_PAGE)
>
> enum dlm_ast_type {
> DLM_AST = 0,
> @@ -86,7 +93,7 @@ enum dlm_ctxt_state {
> struct dlm_ctxt
> {
> struct list_head list;
> - struct hlist_head *lockres_hash;
> + struct hlist_head **lockres_hash;
> struct list_head dirty_list;
> struct list_head purge_list;
> struct list_head pending_asts;
> @@ -136,6 +143,13 @@ struct dlm_ctxt
> struct list_head dlm_eviction_callbacks;
> };
>
> +static inline
> +struct hlist_head *dlm_lockres_hash(struct dlm_ctxt *dlm, unsigned i)
> +{
> + return dlm->lockres_hash[(i / DLM_BUCKETS_PER_PAGE) % DLM_HASH_PAGES]
> + + (i % DLM_BUCKETS_PER_PAGE);
> +}
> +
> /* these keventd work queue items are for less-frequently
> * called functions that cannot be directly called from the
> * net message handlers for some reason, usually because
>
> _______________________________________________
> Ocfs2-devel mailing list
> Ocfs2-devel at oss.oracle.com
> http://oss.oracle.com/mailman/listinfo/ocfs2-devel
>
^ permalink raw reply [flat|nested] 3+ messages in thread
* [Ocfs2-devel] [PATCH 1/1] OCFS2: fasten dlm_lock_resource hash_table lookups
2009-04-23 2:27 ` Sunil Mushran
@ 2009-04-23 11:06 ` Wengang Wang
0 siblings, 0 replies; 3+ messages in thread
From: Wengang Wang @ 2009-04-23 11:06 UTC (permalink / raw)
To: ocfs2-devel
Hi Sunil,
Sunil Mushran wrote:
> Wengang,
>
> When we backport, we like to add bread crumbs so that we can track back
> to the original patch. This helps us in cases where we, say, forget to
> include a patch. Also, we want to retain original authorship.
>
> For example:
> http://oss.oracle.com/git/?p=ocfs2-1.4.git;a=commitdiff;h=174f54db9af7c67a1afaa947aed934b9c80f588c
>
>
> So don't merge the patches. Add the mainline commit at the top of the
> patch. As 1.2 is svn, ensure the authors name is included in an
> Authored-by
> tag.
>
Ok, I will follow that for later backports.
> Has it been tested in 1.2?
>
No. Since I think this part(hash_table
construction/insertions/lookups/destruction) is independent and no
changes from the original patch,
I didn't test it --only made sure it pass compilation.
well, you want to test the functionility(correctness) or the performance?
thanks,
wengang.
> Thanks
> Sunil
>
> wengang wang wrote:
>> #backporting the 3 patches at
>> http://kernel.us.oracle.com/~smushran/srini/ to 1.2.
>>
>> enlarge hash_table capacity to fasten hash_table lookups.
>>
>
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2009-04-23 11:06 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2009-04-22 6:04 [Ocfs2-devel] [PATCH 1/1] OCFS2: fasten dlm_lock_resource hash_table lookups wengang wang
2009-04-23 2:27 ` Sunil Mushran
2009-04-23 11:06 ` Wengang Wang
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.