[Top][All Lists]
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Gluster-devel] performance seems extremely bad
From: |
Anand Avati |
Subject: |
Re: [Gluster-devel] performance seems extremely bad |
Date: |
Fri, 20 Jul 2007 02:10:28 +0530 |
Dale,
it is tough to relate the performance numbers provided by dbench to actual
real-life application's "performance" over a filesystem. dbench usually
reports very low numbers on filesystem implementations via fuse, due to the
marginally increased latency for every meta data operation. But for real
life applications the perfrmance is almost as good as a local disk and
sometimes better for very heavy I/O. Can you compare the dbench numbers (on
the latest TLA patchset) with, say, NFS or other similar network
filesystems?
thanks,
avati
2007/7/5, Dale Dude <address@hidden>:
Kernel 2.6.15. mainline-2.5 patch 275. fuse 2.6.5
Tested with: dbench -t 10 10. Is performance supposed to be this bad?
Glusterfs /volumes: Throughput 15.8983 MB/sec 10 procs
Bypass glusterfs direct to /volume1: Throughput 65.0482 MB/sec 10 procs
Bypass glusterfs direct to /volume2: Throughput 66.5139 MB/sec 10 procs
=============
client.vol:
volume server1
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 127.0.0.1 # IP address of the remote brick
option remote-subvolume volumenamespace
end-volume
volume server1vol1
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 127.0.0.1 # IP address of the remote brick
option remote-subvolume clusterfs1
end-volume
volume server1vol2
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 127.0.0.1 # IP address of the remote brick
option remote-subvolume clusterfs2
end-volume
volume bricks
type cluster/unify
option namespace server1
option readdir-force-success on # ignore failed mounts
subvolumes server1vol1 server1vol2
option scheduler rr
option rr.limits.min-free-disk 5 #%
end-volume
volume writebehind #writebehind improves write performance a lot
type performance/write-behind
option aggregate-size 131072 # in bytes
subvolumes bricks
end-volume
volume readahead
type performance/read-ahead
option page-size 65536 # unit in bytes
option page-count 16 # cache per file = (page-count x page-size)
subvolumes writebehind
end-volume
volume iothreads
type performance/io-threads
option thread-count 32
subvolumes readahead
end-volume
==============================
server.vol:
volume volume1
type storage/posix
option directory /volume1
end-volume
#volume posixlocks1
#type features/posix-locks
#option mandatory on # enables mandatory locking on all files
#subvolumes volume1
#end-volume
volume clusterfs1
type performance/io-threads
option thread-count 16
subvolumes volume1
end-volume
#######
volume volume2
type storage/posix
option directory /volume2
end-volume
#volume posixlocks2
#type features/posix-locks
#option mandatory on # enables mandatory locking on all files
#subvolumes volume2
#end-volume
volume clusterfs2
type performance/io-threads
option thread-count 16
subvolumes volume2
end-volume
#######
volume volumenamespace
type storage/posix
option directory /volume.namespace
end-volume
###
volume clusterfs
type protocol/server
option transport-type tcp/server
subvolumes clusterfs1 clusterfs2 volumenamespace
option auth.ip.clusterfs1.allow *
option auth.ip.clusterfs2.allow *
option auth.ip.volumenamespace.allow *
end-volume
_______________________________________________
Gluster-devel mailing list
address@hidden
http://lists.nongnu.org/mailman/listinfo/gluster-devel
--
Anand V. Avati