gluster-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Gluster-devel] difficult bug in 2.5 mainline


From: Amar S. Tumballi
Subject: Re: [Gluster-devel] difficult bug in 2.5 mainline
Date: Mon, 2 Jul 2007 05:25:09 +0530

Hi Harris,
With the latest patch this bug is fixed. Also, i hope it should fix the
problem of 'rm -rf' too.. please confirm.

i am looking into other strange bug reported by you.

-bulde

On 7/2/07, Harris Landgarten <address@hidden> wrote:

Disabling posix-locks changes the problem

The client crashes along with the lock-server brick

Here is the bt from the client:

#0  unify_bg_cbk (frame=0xe080168, cookie=0xe1109c8, this=0x8057730,
op_ret=0, op_errno=13) at unify.c:83
83          callcnt = --local->call_count;
(gdb) bt
#0  unify_bg_cbk (frame=0xe080168, cookie=0xe1109c8, this=0x8057730,
op_ret=0, op_errno=13) at unify.c:83
#1  0xb75b96e5 in client_unlink_cbk (frame=0xe1109c8, args=0x8059248) at
client-protocol.c:2969
#2  0xb75beff5 in notify (this=0x8057730, event=2, data=0x8095338) at
client-protocol.c:4184
#3  0xb7f73827 in transport_notify (this=0x0, event=235405672) at
transport.c:152
#4  0xb7f74299 in sys_epoll_iteration (ctx=0xbfb96248) at epoll.c:54
#5  0xb7f738fd in poll_iteration (ctx=0xbfb96248) at transport.c:260
#6  0x0804a170 in main (argc=6, argv=0xbfb96324) at glusterfs.c:341
(gdb) print local
$1 = (unify_local_t *) 0x0

Harris

----- Original Message -----
From: "Harris Landgarten" <address@hidden>
To: "gluster-devel" <address@hidden>
Sent: Sunday, July 1, 2007 10:56:05 AM (GMT-0500) America/New_York
Subject: [Gluster-devel] difficult bug in 2.5 mainline

I am trying to track down a bug that is causing hangs in 2.5-patch-249 and
all previous.

This happens during a full Zimbra backup of certain accounts to
/mnt/glusterfs/backups

The first stage of the backup copies indexes and primary storage to
/mnt/glusterfs/backups/tmp
All of this data resides in local storage and the writing to gluster is
successful.

The next stage copies secondary storage to /mnt/glusterfs/backups/tmp
This fails in the following way:

Brick1 hangs with no errors
Brick2 hangs with no errors
Zimbra client hangs with no errors
second client loses connectivity

The second client bails after 2 min but cannot connect
The Zimbra client never bails

I then restart the bricks

After both bricks are restarted, the second client reconnects and a hung
df -h completes

Zimbra client stays in a hung unconnected start

ls -l /mnt/glusterfs hangs

Only way is reset is

kill -9 pidof glusterfs
umount /mnt/glusterfs

glusterfs

Post mortem examination of /mnt/glusterfs/backups/tmp shows that a few
files have the written from the secondary storage volume. I this can over
15,000 files should have been written.

Note: this only happen with large email boxed with some large >10M files.

Note: with patch-247 the zimbra client would seqfault. With 249 it just
hangs in unrecoverable state.


Harris


_______________________________________________
Gluster-devel mailing list
address@hidden
http://lists.nongnu.org/mailman/listinfo/gluster-devel



_______________________________________________
Gluster-devel mailing list
address@hidden
http://lists.nongnu.org/mailman/listinfo/gluster-devel




--
Amar Tumballi
http://amar.80x25.org
[bulde on #gluster/irc.gnu.org]


reply via email to

[Prev in Thread] Current Thread [Next in Thread]