gluster-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Gluster-devel] Crash with mainline-2.5 patch 240 and dbench


From: Anand Avati
Subject: Re: [Gluster-devel] Crash with mainline-2.5 patch 240 and dbench
Date: Thu, 28 Jun 2007 10:34:09 +0530

Dale,
 We too got this bug recently. this fix is on the way.

avati

2007/6/28, Dale Dude <address@hidden>:

Running dbench produces this backtrace:

#0  0xb75b5dff in unify_lookup_cbk (frame=0x86f9bc0, cookie=0x8055e08,
this=0x80560b8, op_ret=0, op_errno=146211160, inode=0x8745fe0,
buf=0x87ed060) at unify.c:310
#1  0xb75c2249 in client_lookup_cbk (frame=0x8ec7598, args=0x8a004c0) at
client-protocol.c:3797
#2  0xb75c36f5 in notify (this=0x8055e08, event=2, data=0x808dc70) at
client-protocol.c:4184
#3  0xb7f36ae2 in transport_notify (this=0x8055e08, event=1) at
transport.c:152
#4  0xb7f3725e in sys_epoll_iteration (ctx=0x8b70158) at epoll.c:54
#5  0xb7f36c8d in poll_iteration (ctx=0x8b70158) at transport.c:260
#6  0x0804a2e9 in main (argc=3, argv=0xbf95b494) at glusterfs.c:341

==========================
glusterfs.log:
2007-06-27 15:51:40 C [fuse-bridge.c:455:fuse_fd_cbk] glusterfs-fuse:
open() got EINTR
2007-06-27 16:04:17 C [common-utils.c:205:gf_print_trace]
debug-backtrace: Got signal (11), printing backtrace
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace: /lib/libglusterfs.so.0(gf_print_trace+0x2d) [0xb7f355f9]
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace: [0xffffe420]
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace: /lib/glusterfs/1.3.0-pre5/xlator/protocol/client.so
[0xb75c2249]
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace:
/lib/glusterfs/1.3.0-pre5/xlator/protocol/client.so(notify+0x855)
[0xb75c36f5]
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace: /lib/libglusterfs.so.0(transport_notify+0x37)
[0xb7f36ae2]
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace: /lib/libglusterfs.so.0(sys_epoll_iteration+0xd7)
[0xb7f3725e]
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace: /lib/libglusterfs.so.0(poll_iteration+0x1d) [0xb7f36c8d]
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace: [glusterfs] [0x804a2e9]
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace: /lib/tls/i686/cmov/libc.so.6(__libc_start_main+0xd2)
[0xb7de7ea2]
2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
debug-backtrace: [glusterfs] [0x8049c31]

=============================
glusterfs-client.vol
volume server1
         type protocol/client
         option transport-type tcp/client     # for TCP/IP transport
         option remote-host 127.0.0.1     # IP address of the remote brick
         option remote-subvolume volumenamespace
end-volume

volume server1vol1
         type protocol/client
         option transport-type tcp/client     # for TCP/IP transport
         option remote-host 127.0.0.1     # IP address of the remote brick
         option remote-subvolume clusterfs1
end-volume

###################

volume bricks
  type cluster/unify
  option namespace server1
  option readdir-force-success on  # ignore failed mounts
  subvolumes server1vol1

  option scheduler rr
  option rr.limits.min-free-disk 5 #%
end-volume

volume writebehind   #writebehind improves write performance a lot
  type performance/write-behind
  option aggregate-size 131072 # in bytes
  subvolumes bricks
end-volume

#volume statprefetch
  #type performance/stat-prefetch
  #option cache-seconds 20
  #subvolumes writebehind
#end-volume

================================
glusterfs-server.vol:
volume clusterfs1
  type storage/posix
  option directory /volume1
end-volume

#volume clusterfs1
   #type performance/io-threads
   #option thread-count 8
   #subvolumes volume1
#end-volume

#######

volume volumenamespace
  type storage/posix
  option directory /volume.namespace
end-volume

###

volume clusterfs
  type protocol/server
  option transport-type tcp/server
  subvolumes clusterfs1 volumenamespace
  option auth.ip.clusterfs1.allow *
  option auth.ip.volumenamespace.allow *
end-volume


_______________________________________________
Gluster-devel mailing list
address@hidden
http://lists.nongnu.org/mailman/listinfo/gluster-devel




--
Anand V. Avati


reply via email to

[Prev in Thread] Current Thread [Next in Thread]