Hello Team ZFSGuru,
I tried to find my post earlier this year, and the advice I got from Cipher.. Oddly enough, the forum allows only the view of the latest postings (one page), and apparently I cannot go beyond that, as there is no page turn to see older discussions.
Eventually I've been able to find what I was looking for, using google (it helped me find a pointer to that conversation, which now I've saved locally).
I'm using Firefox and Chromium browsers on Linux, to browse the site.
Thanks for clarifying this issue.
HG

Powered by ZFSguru version 0.3.1
Running official system image 10.2.003 featuring FreeBSD 10.2-RELEASE-p1 with ZFS v5000.
Running Root-on-ZFS distribution.
Hardware
Motherboard Asus (M5A99X EVO R2.0)
64-bit octo core AMD FX(tm)-8320 Eight-Core Processor - Running at 3500 MHz ( scales from 1400 MHz to 3500 MHz ) -doing some rsync copies.
16 GiB physical memory, 15.4 GiB usable ECC RAM.
11 normal harddrives - in fact 2 ssds partitioned and used as advised by Cipher, and, in the interim, one 6 disk pool (6x3TB) in raidz2 configuration.
The reminder of the disks are for the time being just hanging off the system, not in use.
The pool is running on a Megaraid (SAS9240-8i) flushed in IT mode with the latest FW from AVAGO/LSI.
Gigabit networking - using the port on the MB; I also have a second card which is not in use at themoment.

I have a performance question.

I'm running a rsync copy from a raiserfs disk (mounted ro in zfsguru) to the raidz2 pool. the amount of data is roughly 2.7TB and has not finished after about 30 hours of running. But it is getting closer.
The raiserfs disk is connected to the mb sata, as is also the ssd root mirror. The raidz2 pool is on the LSI controller.

I read on the internet reported copies of 3TB in 7-8 hours, which is fantastic... Is there anything in particular I should look into relative to performance improvement options, with this setup?

I installed bonnie++ from fdbs repository and the ran it against the pool. I don't think it is too spectacular. Any comments and /or thoughts?

You can use gstat to determine disk I/O utilisation. For example with this command:

gstat -f gpt

This should give you an idea whether the disks are the bottleneck. It is likely reiserfs code is simply not that fast. This applies to other filesystems as well on BSD. Only UFS and ZFS are really usable on BSD/ZFSguru.