DragonFly users List (threaded) for 2008-06
DragonFly BSD
DragonFly users List (threaded) for 2008-06
[Date Prev][Date Next]  [Thread Prev][Thread Next]  [Date Index][Thread Index]

HAMMER on a VINUM volume


From: Antonio Huete Jimenez <tuxillo@xxxxxxxxxxxxxxxxx>
Date: Tue, 10 Jun 2008 00:39:47 +0200

Hi,

In my way of setting up a master machine for the diskless environment I need, I just created a vinum volume called datapool built from 4 hard disks (3 IDE UDMA 100 and one 80MB/s SCSI) where I'll store all data for diskless stations.

As HAMMER write performance was just improved I wanted to do a very basic check and here are the results.

UFS
---------------------------------------------------
master# mount /dev/vinum/datapool /datapool/

master# mount | grep datapool
/dev/vinum/datapool on /datapool (ufs, local, soft-updates)

master# df -h | grep datapool
/dev/vinum/datapool   141G   2.0K   130G     0%    /datapool

master# cd /datapool

master# dd if=/dev/zero of=./test01.dat count=1024 bs=1024K
1024+0 records in
1024+0 records out
1073741824 bytes transferred in 20.514075 secs (52341713 bytes/sec)

master# dd if=/dev/zero of=./test02.dat count=100 bs=1024K
100+0 records in
100+0 records out
104857600 bytes transferred in 1.891579 secs (55433902 bytes/sec)

master# dd if=/dev/zero of=./test03.dat count=4096 bs=1024K
4096+0 records in
4096+0 records out
4294967296 bytes transferred in 83.502730 secs (51435053 bytes/sec)


HAMMER -----------------------------------------------------

master# newfs_hammer -L datapool /dev/vinum/datapool
Volume 0 DEVICE /dev/vinum/datapool size 143.11GB
initialize freemap volume 0
Initializing B-Tree blockmap
Initializing Large-Data blockmap
Initializing Small-Data blockmap
---------------------------------------------
1 volume total size 143.11GB
boot-area-size:       64.00MB
memory-log-size:       0.50GB
undo-buffer-size:    152.00MB
zone-limit:            4.00TB
total-pre-allocated: 232.00MB

master# mount_hammer /dev/vinum/datapool /datapool/

master# df -h | grep datapool
datapool      142G     0B   142G     0%    /datapool

master# mount | grep datapool
datapool on /datapool (hammer, local)

master# cd /datapool/

master# dd if=/dev/zero of=./test01.dat count=1024 bs=1024K
1024+0 records in
1024+0 records out
1073741824 bytes transferred in 32.417771 secs (33122013 bytes/sec)

master# dd if=/dev/zero of=./test02.dat count=100 bs=1024K
100+0 records in
100+0 records out
104857600 bytes transferred in 3.143593 secs (33355971 bytes/sec)

master# dd if=/dev/zero of=./test03.dat count=4096 bs=1024K
4096+0 records in
4096+0 records out
4294967296 bytes transferred in 136.746060 secs (31408344 bytes/sec)

One thing I noticed is that in a after formatting with UFS we lose 10GB that UFS reserves for its management, but I didn't noticed any loss in a hammer partition. I like that!

Yeah, yeah, I know this is a very basic check and there are a lot of factors that have influence in these results, but I'm not an expert but an user.


Regards, Antonio Huete



[Date Prev][Date Next]  [Thread Prev][Thread Next]  [Date Index][Thread Index]