From: Mike Christie <michaelc@cs.wisc.edu>
To: sekharan@us.ibm.com
Cc: andmike@us.ibm.com, linux-scsi@vger.kernel.org,
asson_ronald@emc.com, James.Bottomley@HansenPartnership.com,
device-mapper development <dm-devel@redhat.com>,
Benoit_Arthur@emc.com, jens.axboe@oracle.com, agk@redhat.com
Subject: Re: [PATCH 4/7] scsi_dh: add EMC Clariion device handler
Date: Tue, 22 Apr 2008 16:50:17 -0500 [thread overview]
Message-ID: <480E5D99.7050300@cs.wisc.edu> (raw)
In-Reply-To: <480E5D08.3000905@cs.wisc.edu>
Mike Christie wrote:
> Chandra Seetharaman wrote:
>> On Thu, 2008-04-17 at 12:14 -0500, Mike Christie wrote:
>>> Chandra Seetharaman wrote:
>>>> On Wed, 2008-04-16 at 11:29 -0500, Mike Christie wrote:
>>>>> Chandra Seetharaman wrote:
>>>>>> +
>>>>>> +static int send_cmd(struct scsi_device *sdev, int cmd)
>>>>>> +{
>>>>>> + struct request *rq = get_req(sdev, cmd);
>>>>>> +
>>>>>> + if (!rq)
>>>>>> + return SCSI_DH_RES_TEMP_UNAVAIL;
>>>>>> +
>>>>>> + return blk_execute_rq(sdev->request_queue, NULL, rq, 1);
>>>>>> +}
>>>>>> +
>>>>> My only concerns are:
>>>>>
>>>>> 1. EMC and HP need to send a command to every device to transition
>>>>> them. Because we do blk_execute_rq from the dm multipath workqueue
>>>>> we can now only failover/failback for a couple devices at a time.
>>>>>
>>>>> I am not sure if this is a big deal, because this the error handler
>>>>> path so it is going to be slower than the normal path. But it seems
>>>>> like
>>>> Yes. But...
>>>>
>>>> pg_init() due to failover/failback will be sent only when I/O is
>>>> sent/resent to a multipath device, isn't it ? and we don't expect I/Os
>>>> to be sent to all the devices at the same time (all the time), do we ?
>>>>
>>> I am not sure what you mean by all the time, because I am talking about
>>
>> What I meant was that we do not expect I/Os to be sent to all the
>> devices at all the times (pg_init will be sent only when I/Os fails on a
>> path, right ?).
>>
>> Sorry for not being clear.
>>
>
> No problem.
>
>
>>> failover times above. And for failover I think I said yes in the
>>> previous mail. For EMC we are currently sending failover commands to
>>> all the devices at the same time, because EMC does not do the
>>> controller failover RDAC does.
>>
>> RDAC doesn't do controller failover. It also does per lun failover.
>>
>
> Oh yeah, I forgot.
>
>
>>>> So, as you pointed, is it a big deal ? :)
>>>>
>>> In the previous mail I specifically said users might care, because
>>> they are picky about failover times, real 3m39.728s
>> user 0m4.135s
>> sys 0m14.536s
>>
>>> so the answer is to your question is what I said before, maybe :) I
>>> said I am not sure, because I do not have any numbers for the
>>> failover times.
>>
>> Since RDAC also does the failover per device (as is the case with EMC),
>> I ran tests on about 49 luns. I ran disktest on all the disks at the
>
> Thanks.
>
>> same time and disabled/enabled the port to the preferred path to
>> generate failover and failback.
>>
>> Let me know what do you think.
>>
>> Here are the results:
>> Tests run in an idle system. With 49 luns and the following script:
>> ******************************************************
>> for i in `ls -1 /dev/mapper/mpath*`
>> do
>> disktest $i -L 4000 -t 100 -P X &
>> sleep 1
>> done
>>
>> wait
>> ******************************************************
>> Simple Run:
>>
>> with patchset: 2.6.25-mm1:
>> real 3m30.122s real 3m29.746s
>> user 0m4.069s user 0m4.099s
>> sys 0m14.876s sys 0m14.535s
>> -----------------------------------------------
>
> Is this just a boot up test or a test just running IO but no
> failback/failover?
>
>>
>> Failover Run:
>>
>> with patchset: 2.6.25-mm1:
>> real 5m18.875s real 5m31.741s
>> user 0m4.069s user 0m3.883s
>> sys 0m14.838s sys 0m13.822s
>
> Ehh, I have no idea if this is good or bad. Does it mean it is talking
> 13 more seconds to complete?
>
Oops, I read that wrong. With the new code it is 13 seconds faster. I
have no concerns about that.
next prev parent reply other threads:[~2008-04-22 21:50 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
2008-04-16 1:18 [PATCH 0/7] scsi_dh: Move hardware handlers from dm to SCSI Chandra Seetharaman
2008-04-16 1:18 ` [PATCH 1/7] scsi_dh: add skeleton for SCSI Device Handlers Chandra Seetharaman
2008-04-16 1:18 ` [PATCH 2/7] scsi_dh: add lsi rdac device handler Chandra Seetharaman
2008-04-16 1:18 ` [PATCH 3/7] scsi_dh: add hp sw " Chandra Seetharaman
2008-04-16 1:18 ` [PATCH 4/7] scsi_dh: add EMC Clariion " Chandra Seetharaman
2008-04-16 16:29 ` Mike Christie
2008-04-16 23:59 ` [dm-devel] " Chandra Seetharaman
2008-04-17 17:14 ` Mike Christie
2008-04-22 21:09 ` [dm-devel] " Chandra Seetharaman
2008-04-22 21:47 ` Mike Christie
2008-04-22 21:50 ` Mike Christie [this message]
2008-04-22 23:45 ` [dm-devel] " Chandra Seetharaman
2008-04-16 1:18 ` [PATCH 5/7] scsi_dh: Use SCSI device handler in dm-multipath Chandra Seetharaman
2008-04-16 1:18 ` [PATCH 6/7] scsi_dh: Remove hardware handlers from dm Chandra Seetharaman
2008-04-16 1:19 ` [PATCH 7/7] scsi_dh: Remove hardware handler infrastructure " Chandra Seetharaman
-- strict thread matches above, loose matches on Subject: below --
2008-04-17 22:22 [PATCH 0/7] scsi_dh: Move hardware handlers from dm to SCSI Chandra Seetharaman
2008-04-17 22:23 ` [PATCH 4/7] scsi_dh: add EMC Clariion device handler Chandra Seetharaman
2008-04-17 21:18 [PATCH 0/7] scsi_dh: Move hardware handlers from dm to SCSI Chandra Seetharaman
2008-04-17 21:19 ` [PATCH 4/7] scsi_dh: add EMC Clariion device handler Chandra Seetharaman
2008-04-01 22:51 [PATCH 0/7] scsi_dh: Move hardware handlers from dm to SCSI Chandra Seetharaman
2008-04-01 22:51 ` [PATCH 4/7] scsi_dh: add EMC Clariion device handler Chandra Seetharaman
2008-03-11 1:33 [PATCH 0/7] scsi_dh: Move hardware handlers from dm to SCSI Chandra Seetharaman
2008-03-11 1:33 ` [PATCH 4/7] scsi_dh: add EMC Clariion device handler Chandra Seetharaman
2008-02-28 1:08 [PATCH 0/7] Move hardware handlers from dm layer to SCSI layer Chandra Seetharaman
2008-02-28 1:08 ` [PATCH 4/7] scsi_dh: add EMC Clariion device handler Chandra Seetharaman
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=480E5D99.7050300@cs.wisc.edu \
--to=michaelc@cs.wisc.edu \
--cc=Benoit_Arthur@emc.com \
--cc=James.Bottomley@HansenPartnership.com \
--cc=agk@redhat.com \
--cc=andmike@us.ibm.com \
--cc=asson_ronald@emc.com \
--cc=dm-devel@redhat.com \
--cc=jens.axboe@oracle.com \
--cc=linux-scsi@vger.kernel.org \
--cc=sekharan@us.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).