Nice thread you have going here, I am trying to get acceptable IPoIB results, but they aren't looking so good yet. I'll post some results and I have a couple of questions.
All machines are Dell R620's with E5-2660 cpu's and 20Gb cards. I tried connecting the test machines directly with a cable, but I get the same results in our 9024 switch. I went through the Mellanox optimizations in the above PDF.
Test #1: Ubuntu 12.04, Qlogic 7280 HCA's:
iperf results:
1 thread: [ 3] 0.0-10.0 sec 2.75 GBytes 2.37 Gbits/sec
2 threads: [SUM] 0.0-10.0 sec 2.57 GBytes 2.21 Gbits/sec
4 threads: [SUM] 0.0-10.0 sec 2.65 GBytes 2.28 Gbits/sec
Performance degrades rapidly past P4. This speed is around 20% what I would expect. Here is a quick RDMA test:
# ib_rdma_bw -n 20000 10.166.1.81
.....
18907: Bandwidth peak (#3751 to #13407): 1867.03 MB/sec
18907: Bandwidth average: 1773.97 MB/sec
Test #2: el6, Mellanox MT25208 HCA's:
iperf results:
1 thread: [ 3] 0.0-10.0 sec 2.95 GBytes 2.53 Gbits/sec
2 threads: [SUM] 0.0-10.2 sec 3.52 GBytes 2.97 Gbits/sec
4 threads: [SUM] 0.0-10.5 sec 3.53 GBytes 2.90 Gbits/sec
8 threads: [SUM] 0.0-10.2 sec 5.07 GBytes 4.28 Gbits/sec
12 threads: [SUM] 0.0-10.2 sec 4.81 GBytes 4.06 Gbits/sec
16 threads: [SUM] 0.0-10.2 sec 6.12 GBytes 5.17 Gbits/sec
24 threads: [SUM] 0.0-10.0 sec 5.98 GBytes 5.13 Gbits/sec
Looks like these are more scalable, but interestingly RDMA performance isn't as good:
14046: Bandwidth peak (#0 to #13202): 1492.46 MB/sec
14046: Bandwidth average: 1492.43 MB/sec
OK time for questions:
1) I will try Ubuntu on the Mellanox cards as the next test. Are there any other suggestions to try? I also just got a couple of 10 GbE cards too, that should be fun.
2) Is there anything like netcat or iperf for rdma? If not, in my spare time maybe I'll try to hack an RDMA option into netcat.
2) NFS-RDMA on el6 causes a weird kernel panic on the server if clients disconnect. Any idea how to deal with this one?
Thanks,
Mark