[Gluster-users] rm -rf errors

Sacerdoti, Federico Federico.Sacerdoti at DEShawResearch.com
Mon May 11 16:49:35 UTC 2009


Thanks. I updated to 2.0.0 but the daemons will not start and give a
very generic error that does not help
 
2009-05-11 12:41:56 E [glusterfsd.c:483:_xlator_graph_init] drdan0199:
validating translator failed
2009-05-11 12:41:56 E [glusterfsd.c:1145:main] glusterfs: translator
initialization failed.  exiting
 
Can you see something wrong in the volume file? This works fine for
2.0.0rc4
 
--START--
volume storage
  type storage/posix
  option directory /scratch/glusterfs/export
end-volume
 
# Required for AFR (file replication) module
volume locks
  type features/locks
  subvolumes storage
end-volume
 
volume brick
  type performance/io-threads
#option thread-count 1
  option thread-count 8 
  subvolumes locks
end-volume
 
volume server
  type protocol/server
  subvolumes brick
  option transport-type tcp
  option auth.addr.brick.allow 10.232.*
end-volume
 
volume drdan0191
  type protocol/client
  option transport-type tcp
  option remote-host drdan0191.en.desres.deshaw.com
  option remote-subvolume brick
end-volume
 
volume drdan0192
  type protocol/client
  option transport-type tcp
  option remote-host drdan0192.en.desres.deshaw.com
  option remote-subvolume brick
end-volume

[...]
 
volume nufa
  type cluster/nufa
  option local-volume-name `hostname -s`
  #subvolumes replicate1 replicate2 replicate3 replicate4 replicate5
  subvolumes drdan0191 drdan0192 drdan0193 drdan0194 drdan0195 drdan0196
drdan0197 drdan0198 drdan0199 drdan0200
end-volume
 
# This, from https://savannah.nongnu.org/bugs/?24972, does the 
# filesystem mounting at server start time. Like an /etc/fstab entry
volume fuse
  type mount/fuse
  option direct-io-mode 1
  option entry-timeout 1
  #option attr-timeout 1 (not recognized in 2.0)
  option mountpoint /mnt/glusterfs
  subvolumes nufa
end-volume 
--END--
 
Thanks,
fds
 

________________________________

From: Liam Slusser [mailto:lslusser at gmail.com] 
Sent: Thursday, May 07, 2009 1:51 PM
To: Sacerdoti, Federico
Cc: gluster-users at gluster.org
Subject: Re: [Gluster-users] rm -rf errors


You should try upgrading to the 2.0.0 release and try again.  They fixed
all sorts of bugs. 

liam


On Thu, May 7, 2009 at 8:21 AM, Sacerdoti, Federico
<Federico.Sacerdoti at deshawresearch.com> wrote:


	Hello,
	
	I am evaluating glusterfs and have seen some strange behavior
with
	remove. I have gluster/2.0.0rc4 setup on 10 linux nodes
connected with
	GigE. The config is Nufa/fuse with one storage brick per server,
as seen
	in the attached nufa.vol config file, which I use for both
clients and
	servers.
	
	My experiment is to launch 10 parallel writers, each of whom
writes
	32GiB worth of data in small files (2MB) to a shared
gluster-fuse
	mounted filesystem. The files are named uniquely per client, so
each
	file is only written once. This worked well, and I am seeing
performance
	close to that of native disk, even with 8-writers per node.
	
	However when I do a parallel "rm -rf writedir/" on the 10 nodes,
where
	writedir is the directory written in by the parallel writers
described
	above, I see strange effects. There are 69,000 UNLINK errors in
the
	glusterfsd.log of one server, in the form shown below. This
alone is not
	surprising as the operation is ocurring in parallel. However the
remove
	took much longer than expected, 92min, and more surprisingly the
rm
	command exited 0 but files remained in the writedir!
	
	I ran rm -rf writedir from a single client, and it too exited 0
but left
	the writedir non-empty. Is this expected?
	
	Thanks,
	Federico
	
	--From glusterfsd.log--
	2009-05-04 11:35:15 E [fuse-bridge.c:964:fuse_unlink_cbk]
	glusterfs-fuse: 5764889: UNLINK() /write.2MB.runid1.p1/5 => -1
(No such
	file or directory)
	2009-05-04 11:35:15 E [dht-common.c:1294:dht_err_cbk] nufa:
subvolume
	drdan0192 returned -1 (No such file or directory)
	2009-05-04 11:35:15 E [fuse-bridge.c:964:fuse_unlink_cbk]
	glusterfs-fuse: 5764894: UNLINK() /write.2MB.runid1.p1/51 => -1
(No such
	file or directory)
	--end--
	 <<nufa.vol>>
	
	_______________________________________________
	Gluster-users mailing list
	Gluster-users at gluster.org
	http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
	
	


-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20090511/57af954a/attachment.html>


More information about the Gluster-users mailing list