gluster-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Gluster-devel] Whole failure when one glusterfsd brought down.


From: Dale Dude
Subject: [Gluster-devel] Whole failure when one glusterfsd brought down.
Date: Wed, 21 Mar 2007 16:14:39 -0400
User-agent: Thunderbird 2.0pre (Windows/20070320)

Using code pulled March 21st @ 1am EST. Kernel 2.6.20-10 on Ubuntu Feisty 32bit.

I have 3 machines serving with glusterfsd and mounted the cluster from the first server. If I kill one of the glusterfsd's on any of the machines, the mount point becomes broken with the 'Transport' error below. Also, glusterfs will produce this error even if I unmount and remount with the 1 glusterfsd server still down. Is this expected results or shouldn't the mount continue to work even though one of the servers has "died"?

ls: reading directory local/: Transport endpoint is not connected

glusterfs.log produces this:
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed [Mar 21 06:25:17] [ERROR/tcp-client.c:284/tcp_connect()] tcp/client:non-blocking connect() returned: 111 (Connection refused) [Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed [Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed [Mar 21 06:25:17] [ERROR/tcp-client.c:284/tcp_connect()] tcp/client:non-blocking connect() returned: 111 (Connection refused) [Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed [Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed


======================================
glusterfs-server.vol used by all the servers:
(ignore my bad volume naming, was just testing)

volume brick
 type storage/posix                   # POSIX FS translator
 option directory /home/export        # Export this directory
end-volume

volume iothreads
  type performance/io-threads
  option thread-count 8
  subvolumes brick
end-volume

volume server
 type protocol/server
 option transport-type tcp/server     # For TCP/IP transport
 subvolumes iothreads
 option auth.ip.brick.allow * # Allow access to "brick" volume
end-volume


======================================
glusterfs-client.vol used on server1:
(ignore my bad volume naming, was just testing)

volume client1
 type protocol/client
 option transport-type tcp/client     # for TCP/IP transport
 option remote-host 1.1.1.1     # IP address of the remote brick
 option remote-subvolume brick        # name of the remote volume
end-volume

volume client2
 type protocol/client
 option transport-type tcp/client     # for TCP/IP transport
 option remote-host 2.2.2.2     # IP address of the remote brick
 option remote-subvolume brick        # name of the remote volume
end-volume

volume client3
 type protocol/client
 option transport-type tcp/client     # for TCP/IP transport
 option remote-host 3.3.3.3     # IP address of the remote brick
 option remote-subvolume brick        # name of the remote volume
end-volume

volume bricks
 type cluster/unify
 subvolumes client1 client2 client3
 option scheduler alu
option alu.limits.min-free-disk 60GB # Stop creating files when free-space lt 60GB
 option alu.limits.max-open-files 10000
option alu.order disk-usage:read-usage:write-usage:open-files-usage:disk-speed-usage option alu.disk-usage.entry-threshold 2GB # Units in KB, MB and GB are allowed option alu.disk-usage.exit-threshold 60MB # Units in KB, MB and GB are allowed
 option alu.open-files-usage.entry-threshold 1024
 option alu.open-files-usage.exit-threshold 32
 option alu.stat-refresh.interval 10sec
end-volume

volume writebehind   #writebehind improves write performance a lot
 type performance/write-behind
 option aggregate-size 131072 # in bytes
 subvolumes bricks
end-volume




reply via email to

[Prev in Thread] Current Thread [Next in Thread]