From: Dave Chinner <david@fromorbit.com>
To: xfs@oss.sgi.com
Subject: [PATCH 3/5] repair: phase 6 is trivially parallelisable
Date: Thu, 12 Dec 2013 18:22:23 +1100 [thread overview]
Message-ID: <1386832945-19763-4-git-send-email-david@fromorbit.com> (raw)
In-Reply-To: <1386832945-19763-1-git-send-email-david@fromorbit.com>
From: Dave Chinner <dchinner@redhat.com>
Phase 6 is currently single threaded, but it iterates AGs one at a
time. When there are hundreds of AGs that need scanning, this takes
a long time. Given that all the objects that the AG traversal works
on are per-ag, we can simply parallelise this into a strided AG
processing like phase 3 and 4.
Unpatched: 8m40s
patched: 1m10s (7 threads)
Big win!
Signed-off-by: Dave Chinner <dchinner@redhat.com>
---
repair/phase6.c | 56 +++++++++++++++++++++++++++++++++++++++++++++++---------
1 file changed, 47 insertions(+), 9 deletions(-)
diff --git a/repair/phase6.c b/repair/phase6.c
index d2d4a44..d82f900 100644
--- a/repair/phase6.c
+++ b/repair/phase6.c
@@ -51,6 +51,7 @@ typedef struct dotdot_update {
static dotdot_update_t *dotdot_update_list;
static int dotdot_update;
+static pthread_mutex_t dotdot_lock;
static void
add_dotdot_update(
@@ -64,12 +65,14 @@ add_dotdot_update(
do_error(_("malloc failed add_dotdot_update (%zu bytes)\n"),
sizeof(dotdot_update_t));
+ pthread_mutex_lock(&dotdot_lock);
dir->next = dotdot_update_list;
dir->irec = irec;
dir->agno = agno;
dir->ino_offset = ino_offset;
dotdot_update_list = dir;
+ pthread_mutex_unlock(&dotdot_lock);
}
/*
@@ -2918,34 +2921,68 @@ update_missing_dotdot_entries(
* these entries parents were updated, rebuild them again
* set dotdot_update flag so processing routines do not count links
*/
+ pthread_mutex_lock(&dotdot_lock);
dotdot_update = 1;
while (dotdot_update_list) {
dir = dotdot_update_list;
dotdot_update_list = dir->next;
+ dir->next = NULL;
+ pthread_mutex_unlock(&dotdot_lock);
+
process_dir_inode(mp, dir->agno, dir->irec, dir->ino_offset);
free(dir);
+
+ pthread_mutex_lock(&dotdot_lock);
}
+ pthread_mutex_unlock(&dotdot_lock);
}
static void
traverse_ags(
- xfs_mount_t *mp)
+ xfs_mount_t *mp)
{
- int i;
- work_queue_t queue;
+ int i, j;
+ xfs_agnumber_t agno;
+ work_queue_t *queues;
prefetch_args_t *pf_args[2];
/*
* we always do prefetch for phase 6 as it will fill in the gaps
* not read during phase 3 prefetch.
*/
- queue.mp = mp;
- pf_args[0] = start_inode_prefetch(0, 1, NULL);
- for (i = 0; i < glob_agcount; i++) {
- pf_args[(~i) & 1] = start_inode_prefetch(i + 1, 1,
- pf_args[i & 1]);
- traverse_function(&queue, i, pf_args[i & 1]);
+ if (!ag_stride) {
+ work_queue_t queue;
+
+ queue.mp = mp;
+ pf_args[0] = start_inode_prefetch(0, 1, NULL);
+ for (i = 0; i < glob_agcount; i++) {
+ pf_args[(~i) & 1] = start_inode_prefetch(i + 1, 1,
+ pf_args[i & 1]);
+ traverse_function(&queue, i, pf_args[i & 1]);
+ }
+ return;
}
+
+ /*
+ * create one worker thread for each segment of the volume
+ */
+ queues = malloc(thread_count * sizeof(work_queue_t));
+ for (i = 0, agno = 0; i < thread_count; i++) {
+ create_work_queue(&queues[i], mp, 1);
+ pf_args[0] = NULL;
+ for (j = 0; j < ag_stride && agno < glob_agcount; j++, agno++) {
+ pf_args[0] = start_inode_prefetch(agno, 1, pf_args[0]);
+ queue_work(&queues[i], traverse_function, agno,
+ pf_args[0]);
+ }
+ }
+
+ /*
+ * wait for workers to complete
+ */
+ for (i = 0; i < thread_count; i++)
+ destroy_work_queue(&queues[i]);
+ free(queues);
}
void
@@ -2957,6 +2994,7 @@ phase6(xfs_mount_t *mp)
memset(&zerocr, 0, sizeof(struct cred));
memset(&zerofsx, 0, sizeof(struct fsxattr));
orphanage_ino = 0;
+ pthread_mutex_init(&dotdot_lock, NULL);
do_log(_("Phase 6 - check inode connectivity...\n"));
--
1.8.4.rc3
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
next prev parent reply other threads:[~2013-12-12 7:22 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-12-12 7:22 [PATCH 0/5] xfs_repair: scalability inmprovements Dave Chinner
2013-12-12 7:22 ` [PATCH 1/5] repair: translation lookups limit scalability Dave Chinner
2013-12-12 18:26 ` Christoph Hellwig
2013-12-12 18:58 ` Brian Foster
2013-12-12 7:22 ` [PATCH 2/5] repair: per AG locks contend for cachelines Dave Chinner
2013-12-12 18:27 ` Christoph Hellwig
2013-12-12 18:58 ` Brian Foster
2013-12-12 20:46 ` Dave Chinner
2013-12-12 7:22 ` Dave Chinner [this message]
2013-12-12 18:43 ` [PATCH 3/5] repair: phase 6 is trivially parallelisable Christoph Hellwig
2013-12-12 20:53 ` Dave Chinner
2013-12-12 18:59 ` Brian Foster
2013-12-12 7:22 ` [PATCH 4/5] libxfs: buffer cache hashing is suboptimal Dave Chinner
2013-12-12 18:28 ` Christoph Hellwig
2013-12-12 18:59 ` Brian Foster
2013-12-12 20:56 ` Dave Chinner
2013-12-13 14:23 ` Brian Foster
2013-12-12 7:22 ` [PATCH 5/5] repair: limit auto-striding concurrency apprpriately Dave Chinner
2013-12-12 18:29 ` Christoph Hellwig
2013-12-12 21:00 ` Dave Chinner
2013-12-12 18:59 ` Brian Foster
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1386832945-19763-4-git-send-email-david@fromorbit.com \
--to=david@fromorbit.com \
--cc=xfs@oss.sgi.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox