public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
* "xfs_log_force: error 5 returned." for drive that was removed.
@ 2016-04-17 14:33 Joe Wendt
  2016-04-18 18:54 ` Carlos Maiolino
  0 siblings, 1 reply; 4+ messages in thread
From: Joe Wendt @ 2016-04-17 14:33 UTC (permalink / raw)
  To: xfs


[-- Attachment #1.1: Type: text/plain, Size: 846 bytes --]

Hello! This may be a silly question or an interesting one...

We had a drive fail in a production server, which spawned this error in the
logs:
XFS (sde1): xfs_log_force: error 5 returned.

The dead array was lazy-unmounted, and the drive was hot-swapped, but when
the RAID array was rebuilt, it came online as /dev/sdk instead of /dev/sde.

Now /dev/sde1 doesn't exist in the system, but we still see this message
every 30 seconds. I'm assuming a reboot will clear out whatever is still
trying to access sde1, but I'm trying to avoid that if possible. Could
someone point me in the direction of what XFS might still be trying to do
with that device?

lsof hasn't given me any clues. I can't run xfs_repair on a volume that
isn't there. I haven't been able to find anything similar yet online. Any
help would be greatly appreciated!

Thanks,
Joe

[-- Attachment #1.2: Type: text/html, Size: 1040 bytes --]

[-- Attachment #2: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: "xfs_log_force: error 5 returned." for drive that was removed.
  2016-04-17 14:33 "xfs_log_force: error 5 returned." for drive that was removed Joe Wendt
@ 2016-04-18 18:54 ` Carlos Maiolino
  2016-05-04 15:48   ` Joe Wendt
  0 siblings, 1 reply; 4+ messages in thread
From: Carlos Maiolino @ 2016-04-18 18:54 UTC (permalink / raw)
  To: xfs

On Sun, Apr 17, 2016 at 09:33:27AM -0500, Joe Wendt wrote:
>    Hello! This may be a silly question or an interesting one...
>    We had a drive fail in a production server, which spawned this error in
>    the logs:
>    XFS (sde1): xfs_log_force: error 5 returned.
>    The dead array was lazy-unmounted, and the drive was hot-swapped, but
>    when the RAID array was rebuilt, it came online as /dev/sdk instead of
>    /dev/sde.
>    Now /dev/sde1 doesn't exist in the system, but we still see this
>    message every 30 seconds. I'm assuming a reboot will clear out whatever
>    is still trying to access sde1, but I'm trying to avoid that if
>    possible. Could someone point me in the direction of what XFS might
>    still be trying to do with that device?
>    lsof hasn't given me any clues. I can't run xfs_repair on a volume that
>    isn't there. I haven't been able to find anything similar yet online.
>    Any help would be greatly appreciated!
>    Thanks,
>    Joe

I believe this is the same problem being discussed in this thread:

XFS hung task in xfs_ail_push_all_sync() when unmounting FS after disk
failure/recovery.

Can you get a stack dump of the system (sysrq-t) and post it in some pastebin?


> _______________________________________________
> xfs mailing list
> xfs@oss.sgi.com
> http://oss.sgi.com/mailman/listinfo/xfs


-- 
Carlos

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: "xfs_log_force: error 5 returned." for drive that was removed.
  2016-04-18 18:54 ` Carlos Maiolino
@ 2016-05-04 15:48   ` Joe Wendt
  2016-05-05 10:05     ` Carlos Maiolino
  0 siblings, 1 reply; 4+ messages in thread
From: Joe Wendt @ 2016-05-04 15:48 UTC (permalink / raw)
  To: xfs


[-- Attachment #1.1: Type: text/plain, Size: 1923 bytes --]

Thanks for the reply, and I'm sorry for the delay. Another admin rebooted
the server before I had a chance to collect more info. I'll take a look at
the other thread in case it comes up again. I think we'll avoid the
lazy-unmount in the future though.

Thanks again!
-Joe

On Mon, Apr 18, 2016 at 1:54 PM, Carlos Maiolino <cmaiolino@redhat.com>
wrote:

> On Sun, Apr 17, 2016 at 09:33:27AM -0500, Joe Wendt wrote:
> >    Hello! This may be a silly question or an interesting one...
> >    We had a drive fail in a production server, which spawned this error
> in
> >    the logs:
> >    XFS (sde1): xfs_log_force: error 5 returned.
> >    The dead array was lazy-unmounted, and the drive was hot-swapped, but
> >    when the RAID array was rebuilt, it came online as /dev/sdk instead of
> >    /dev/sde.
> >    Now /dev/sde1 doesn't exist in the system, but we still see this
> >    message every 30 seconds. I'm assuming a reboot will clear out
> whatever
> >    is still trying to access sde1, but I'm trying to avoid that if
> >    possible. Could someone point me in the direction of what XFS might
> >    still be trying to do with that device?
> >    lsof hasn't given me any clues. I can't run xfs_repair on a volume
> that
> >    isn't there. I haven't been able to find anything similar yet online.
> >    Any help would be greatly appreciated!
> >    Thanks,
> >    Joe
>
> I believe this is the same problem being discussed in this thread:
>
> XFS hung task in xfs_ail_push_all_sync() when unmounting FS after disk
> failure/recovery.
>
> Can you get a stack dump of the system (sysrq-t) and post it in some
> pastebin?
>
>
> > _______________________________________________
> > xfs mailing list
> > xfs@oss.sgi.com
> > http://oss.sgi.com/mailman/listinfo/xfs
>
>
> --
> Carlos
>
> _______________________________________________
> xfs mailing list
> xfs@oss.sgi.com
> http://oss.sgi.com/mailman/listinfo/xfs
>

[-- Attachment #1.2: Type: text/html, Size: 2861 bytes --]

[-- Attachment #2: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: "xfs_log_force: error 5 returned." for drive that was removed.
  2016-05-04 15:48   ` Joe Wendt
@ 2016-05-05 10:05     ` Carlos Maiolino
  0 siblings, 0 replies; 4+ messages in thread
From: Carlos Maiolino @ 2016-05-05 10:05 UTC (permalink / raw)
  To: xfs

On Wed, May 04, 2016 at 10:48:13AM -0500, Joe Wendt wrote:
>    Thanks for the reply, and I'm sorry for the delay. Another admin
>    rebooted the server before I had a chance to collect more info. I'll
>    take a look at the other thread in case it comes up again. I think
>    we'll avoid the lazy-unmount in the future though.
>    Thanks again!
>    -Joe
> 

It certainly looks the same problem, which should be fixed by the patchset we
are working on, to add a configurable behavior for different kinds of errors.

>    On Mon, Apr 18, 2016 at 1:54 PM, Carlos Maiolino
>    <[1]cmaiolino@redhat.com> wrote:
> 
>    On Sun, Apr 17, 2016 at 09:33:27AM -0500, Joe Wendt wrote:
>    >    Hello! This may be a silly question or an interesting one...
>    >    We had a drive fail in a production server, which spawned this
>    error in
>    >    the logs:
>    >    XFS (sde1): xfs_log_force: error 5 returned.
>    >    The dead array was lazy-unmounted, and the drive was hot-swapped,
>    but
>    >    when the RAID array was rebuilt, it came online as /dev/sdk
>    instead of
>    >    /dev/sde.
>    >    Now /dev/sde1 doesn't exist in the system, but we still see this
>    >    message every 30 seconds. I'm assuming a reboot will clear out
>    whatever
>    >    is still trying to access sde1, but I'm trying to avoid that if
>    >    possible. Could someone point me in the direction of what XFS
>    might
>    >    still be trying to do with that device?
>    >    lsof hasn't given me any clues. I can't run xfs_repair on a volume
>    that
>    >    isn't there. I haven't been able to find anything similar yet
>    online.
>    >    Any help would be greatly appreciated!
>    >    Thanks,
>    >    Joe
> 
>      I believe this is the same problem being discussed in this thread:
>      XFS hung task in xfs_ail_push_all_sync() when unmounting FS after
>      disk
>      failure/recovery.
>      Can you get a stack dump of the system (sysrq-t) and post it in some
>      pastebin?
>      > _______________________________________________
>      > xfs mailing list
>      > [2]xfs@oss.sgi.com
>      > [3]http://oss.sgi.com/mailman/listinfo/xfs
>      --
>      Carlos
>      _______________________________________________
>      xfs mailing list
>      [4]xfs@oss.sgi.com
>      [5]http://oss.sgi.com/mailman/listinfo/xfs
> 
> References
> 
>    1. mailto:cmaiolino@redhat.com
>    2. mailto:xfs@oss.sgi.com
>    3. http://oss.sgi.com/mailman/listinfo/xfs
>    4. mailto:xfs@oss.sgi.com
>    5. http://oss.sgi.com/mailman/listinfo/xfs

> _______________________________________________
> xfs mailing list
> xfs@oss.sgi.com
> http://oss.sgi.com/mailman/listinfo/xfs


-- 
Carlos

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2016-05-05 10:05 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2016-04-17 14:33 "xfs_log_force: error 5 returned." for drive that was removed Joe Wendt
2016-04-18 18:54 ` Carlos Maiolino
2016-05-04 15:48   ` Joe Wendt
2016-05-05 10:05     ` Carlos Maiolino

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox