[OmniOS-discuss] slowness in aggr - bonded 10G

Lee Damon nomad at ee.washington.edu
Mon Sep 17 16:16:58 UTC 2018


I have four file servers all of which have at least two (some three)
aggregates made of two 10GB lines.  Three of the four servers are getting
pretty good throughput as reported by iperf:

(hvfs1 has iperf -s, other hosts iperf -c)
[ ID] Interval           Transfer     Bandwidth
[  5]   0.00-10.00  sec  8.04 GBytes  6.91 Gbits/sec                  sender
[  5]   0.00-10.00  sec  8.04 GBytes  6.91 Gbits/sec
receiver

but one, fs2, is getting roughly half that speed when it is running iperf
-c:

[ ID] Interval           Transfer     Bandwidth
[  4]   0.00-10.00  sec  3.81 GBytes  3.27 Gbits/sec                  sender
[  4]   0.00-10.00  sec  3.81 GBytes  3.27 Gbits/sec
receiver
(with occasional bumps to 4.05 Gb/s)

Even stranger, when I run iperf -c on fs2 and iperf -s on hvfs1 I get a
better (though still not as good) result bouncing between 4.6 Gb/s and 5.2
Gb/s.

It doesn't matter if I test via aggr0 or aggr_nas0, I get roughly the same
numbers. It also doesn't make any difference if aggr0 is dual-i40e or
dual-ixgbe devices.

Tests were done with hvfs1 acting as the iperf -s server via aggr_front0
(the equivalent to aggr0 on the other hosts) or aggr_nas0 as appropriate.
All hosts are directly connected to the same Juniper switch. (aggr info is
postpended at the end of this email.)

One theory I had was that fs2 wasn't actually using the full throughput of
the bond so I physically removed one of the two cables. Sure enough, the
bandwidth reported by iperf remained around 3.2Gb/s. I tested this with
both aggr0 and aggr_nas0 with the same result. There are times where it
gets closer to the 6Gb of the other hosts so it is clearly getting more
than just one link's worth but not often.

Unlike the other servers, fs2 is running 151026. The rest are all on
151022. I didn't see any references to aggr in the release notes for 151024
or 151026 but I'm not going to claim exhaustive search.

Aggregates were created with the device equivalent of

   dladm create-aggr -L active -l i40e0 -l i40e1 aggr0

My question is, what blatantly obvious thing did I miss in creating the
aggregates on fs2?

thanks,
nomad


aggr info:

fs1 (omnios-r151022-5e982daae6):
LINK            POLICY   ADDRPOLICY           LACPACTIVITY  LACPTIMER
 FLAGS
aggr0           L4       auto                 passive       short
 -----  (not used for this test)
aggr_nas0       L4       auto                 active        short
 -----
aggr_front0     L4       auto                 active        short
 -----

LINK        CLASS     MTU    STATE    BRIDGE     OVER
e1000g1     phys      1500   up       --         --
e1000g0     phys      1500   up       --         --
aggr0       aggr      1500   up       --         e1000g0,e1000g1
ixgbe0      phys      1500   up       --         --
ixgbe1      phys      1500   up       --         --
aggr_nas0   aggr      1500   up       --         ixgbe0,ixgbe1
ixgbe2      phys      1500   up       --         --
ixgbe3      phys      1500   up       --         --
aggr_front0 aggr      1500   up       --         ixgbe2,ixgbe3


fs2 (omnios-r151026-51c7d6fd75):
LINK            POLICY   ADDRPOLICY           LACPACTIVITY  LACPTIMER
 FLAGS
aggr_nas0       L4       auto                 active        short
 -----
aggr_net10      L4       auto                 active        short
 -----  (not used for this test)
aggr0           L4       auto                 active        short
 -----

LINK        CLASS     MTU    STATE    BRIDGE     OVER
i40e2       phys      1500   up       --         --
igb0        phys      1500   up       --         --
i40e0       phys      1500   up       --         --
i40e3       phys      1500   up       --         --
igb1        phys      1500   up       --         --
i40e1       phys      1500   up       --         --
ixgbe0      phys      1500   unknown  --         --
ixgbe1      phys      1500   unknown  --         --
aggr_nas0   aggr      1500   up       --         i40e2,i40e3
aggr_net10  aggr      1500   up       --         igb0,igb1
aggr0       aggr      1500   up       --         i40e0,i40e1


hvfs1 (omnios-r151022-5e982daae6):
LINK            POLICY   ADDRPOLICY           LACPACTIVITY  LACPTIMER
 FLAGS
aggr0           L4       auto                 active        short
 -----  (not used for this test)
aggr_front0     L4       auto                 active        short
 -----
aggr_nas0       L4       auto                 active        short
 -----

LINK        CLASS     MTU    STATE    BRIDGE     OVER
ixgbe0      phys      1500   up       --         --
igb0        phys      1500   up       --         --
igb1        phys      1500   up       --         --
igb2        phys      1500   unknown  --         --
oce0        phys      1500   up       --         --
igb3        phys      1500   unknown  --         --
ixgbe1      phys      1500   up       --         --
oce1        phys      1500   up       --         --
aggr0       aggr      1500   up       --         igb0,igb1
aggr_front0 aggr      1500   up       --         ixgbe0,ixgbe1
aggr_nas0   aggr      1500   up       --         oce0,oce1


hvfs2 (omnios-r151022-89f6242508):
LINK            POLICY   ADDRPOLICY           LACPACTIVITY  LACPTIMER
 FLAGS
aggr_nas0       L4       auto                 active        short
 -----
aggr0           L4       auto                 active        short
 -----

LINK        CLASS     MTU    STATE    BRIDGE     OVER
ixgbe2      phys      1500   up       --         --
ixgbe0      phys      1500   up       --         --
ixgbe3      phys      1500   up       --         --
ixgbe1      phys      1500   up       --         --
aggr_nas0   aggr      1500   up       --         ixgbe0,ixgbe1
aggr0       aggr      1500   up       --         ixgbe3,ixgbe2
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://omniosce.org/ml-archive/attachments/20180917/a03e8cef/attachment-0001.html>


More information about the OmniOS-discuss mailing list