You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users-cn@cloudstack.apache.org by 虚无缥缈 <83...@qq.com> on 2014/07/18 07:39:55 UTC
为什么我的双网卡绑定,总是一个网卡有流量呢。
A、B两台设备,都是双网卡绑定,mode=0,交换机上也做了端口组,交换机是cisco 2960,A上起了nfs,B上挂载nfs,然后进行dd测试,cacti上看总是一个端口流量,手动down掉一个网卡,另外一个网卡才有流量。
A:
[root@backup ~]# cat /proc/net/bonding/bond0
Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
Bonding Mode: load balancing (round-robin)
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0
Slave Interface: em1
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 1
Permanent HW addr:
Slave queue ID: 0
Slave Interface: em2
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 0
Permanent HW addr:
B:
root@storage-xen bak_201]# cat /proc/net/bonding/bond0
Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
Bonding Mode: load balancing (round-robin)
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0
Slave Interface: em1
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 0
Permanent HW addr:
Slave queue ID: 0
Slave Interface: em2
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 0
Permanent HW addr:
Slave queue ID: 0
交换机:
interface Port-channel1
switchport access vlan 150
switchport mode access
interface GigabitEthernet0/3
switchport access vlan 150
switchport mode access
channel-group 1 mode on
spanning-tree portfast
!
interface GigabitEthernet0/10
switchport access vlan 150
switchport mode access
channel-group 1 mode on
spanning-tree portfast
Re: 回复: 为什么我的双网卡绑定,总是一个网卡有流量呢。
Posted by Guo Star <gh...@gmail.com>.
你需要清楚理解mode=0的工作模式,一个连接的建立只能选择一条"通道"。如果你要模拟,最好测试apache benchmark来测试.
在 2014年10月7日 下午11:37,kenwayliu <ke...@gmail.com>写道:
> 业务压力不够,建议使用多客户端测试
>
>
>
> kenwayliu
>
> 发件人: 虚无缥缈
> 发送时间: 2014-07-21 14:44
> 收件人: users-cn
> 主题: 回复: 为什么我的双网卡绑定,总是一个网卡有流量呢。
> 正文已经贴了交换机配置,Cisco 2960s交换机。
>
>
>
>
> ------------------ 原始邮件 ------------------
> 发件人: "linuxbqj@gmail.com";<li...@gmail.com>;
> 发送时间: 2014年7月21日(星期一) 中午1:20
> 收件人: "users-cn"<us...@cloudstack.apache.org>;
>
> 主题: Re: 为什么我的双网卡绑定,总是一个网卡有流量呢。
>
>
>
> mode=0:平衡负载模式,有自动备援,但需要”Switch”支援及设定。
>
> http://support.huawei.com/ecommunity/bbs/10155553.html
>
> 2014-07-18 13:39 GMT+08:00 虚无缥缈 <83...@qq.com>:
> > A、B两台设备,都是双网卡绑定,mode=0,交换机上也做了端口组,交换机是cisco
> 2960,A上起了nfs,B上挂载nfs,然后进行dd测试,cacti上看总是一个端口流量,手动down掉一个网卡,另外一个网卡才有流量。
> >
> > A:
> > [root@backup ~]# cat /proc/net/bonding/bond0
> > Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
> >
> >
> > Bonding Mode: load balancing (round-robin)
> > MII Status: up
> > MII Polling Interval (ms): 100
> > Up Delay (ms): 0
> > Down Delay (ms): 0
> >
> >
> > Slave Interface: em1
> > MII Status: up
> > Speed: 1000 Mbps
> > Duplex: full
> > Link Failure Count: 1
> > Permanent HW addr:
> > Slave queue ID: 0
> >
> >
> > Slave Interface: em2
> > MII Status: up
> > Speed: 1000 Mbps
> > Duplex: full
> > Link Failure Count: 0
> > Permanent HW addr:
> >
> >
> >
> > B:
> > root@storage-xen bak_201]# cat /proc/net/bonding/bond0
> > Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
> >
> >
> > Bonding Mode: load balancing (round-robin)
> > MII Status: up
> > MII Polling Interval (ms): 100
> > Up Delay (ms): 0
> > Down Delay (ms): 0
> >
> >
> > Slave Interface: em1
> > MII Status: up
> > Speed: 1000 Mbps
> > Duplex: full
> > Link Failure Count: 0
> > Permanent HW addr:
> > Slave queue ID: 0
> >
> >
> > Slave Interface: em2
> > MII Status: up
> > Speed: 1000 Mbps
> > Duplex: full
> > Link Failure Count: 0
> > Permanent HW addr:
> > Slave queue ID: 0
> >
> >
> >
> > 交换机:
> > interface Port-channel1
> > switchport access vlan 150
> > switchport mode access
> >
> >
> >
> > interface GigabitEthernet0/3
> > switchport access vlan 150
> > switchport mode access
> > channel-group 1 mode on
> > spanning-tree portfast
> > !
> >
> > interface GigabitEthernet0/10
> > switchport access vlan 150
> > switchport mode access
> > channel-group 1 mode on
> > spanning-tree portfast
>
>
>
> --
> 白清杰 (Born Bai)
>
> Mail: linuxbqj@gmail.com
>
> 敬畏耶和华是智慧的开端
>
Re: 回复: 为什么我的双网卡绑定,总是一个网卡有流量呢。
Posted by kenwayliu <ke...@gmail.com>.
业务压力不够,建议使用多客户端测试
kenwayliu
发件人: 虚无缥缈
发送时间: 2014-07-21 14:44
收件人: users-cn
主题: 回复: 为什么我的双网卡绑定,总是一个网卡有流量呢。
正文已经贴了交换机配置,Cisco 2960s交换机。
------------------ 原始邮件 ------------------
发件人: "linuxbqj@gmail.com";<li...@gmail.com>;
发送时间: 2014年7月21日(星期一) 中午1:20
收件人: "users-cn"<us...@cloudstack.apache.org>;
主题: Re: 为什么我的双网卡绑定,总是一个网卡有流量呢。
mode=0:平衡负载模式,有自动备援,但需要”Switch”支援及设定。
http://support.huawei.com/ecommunity/bbs/10155553.html
2014-07-18 13:39 GMT+08:00 虚无缥缈 <83...@qq.com>:
> A、B两台设备,都是双网卡绑定,mode=0,交换机上也做了端口组,交换机是cisco 2960,A上起了nfs,B上挂载nfs,然后进行dd测试,cacti上看总是一个端口流量,手动down掉一个网卡,另外一个网卡才有流量。
>
> A:
> [root@backup ~]# cat /proc/net/bonding/bond0
> Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
>
>
> Bonding Mode: load balancing (round-robin)
> MII Status: up
> MII Polling Interval (ms): 100
> Up Delay (ms): 0
> Down Delay (ms): 0
>
>
> Slave Interface: em1
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 1
> Permanent HW addr:
> Slave queue ID: 0
>
>
> Slave Interface: em2
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr:
>
>
>
> B:
> root@storage-xen bak_201]# cat /proc/net/bonding/bond0
> Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
>
>
> Bonding Mode: load balancing (round-robin)
> MII Status: up
> MII Polling Interval (ms): 100
> Up Delay (ms): 0
> Down Delay (ms): 0
>
>
> Slave Interface: em1
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr:
> Slave queue ID: 0
>
>
> Slave Interface: em2
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr:
> Slave queue ID: 0
>
>
>
> 交换机:
> interface Port-channel1
> switchport access vlan 150
> switchport mode access
>
>
>
> interface GigabitEthernet0/3
> switchport access vlan 150
> switchport mode access
> channel-group 1 mode on
> spanning-tree portfast
> !
>
> interface GigabitEthernet0/10
> switchport access vlan 150
> switchport mode access
> channel-group 1 mode on
> spanning-tree portfast
--
白清杰 (Born Bai)
Mail: linuxbqj@gmail.com
敬畏耶和华是智慧的开端
回复: 为什么我的双网卡绑定,总是一个网卡有流量呢。
Posted by 虚无缥缈 <83...@qq.com>.
正文已经贴了交换机配置,Cisco 2960s交换机。
------------------ 原始邮件 ------------------
发件人: "linuxbqj@gmail.com";<li...@gmail.com>;
发送时间: 2014年7月21日(星期一) 中午1:20
收件人: "users-cn"<us...@cloudstack.apache.org>;
主题: Re: 为什么我的双网卡绑定,总是一个网卡有流量呢。
mode=0:平衡负载模式,有自动备援,但需要”Switch”支援及设定。
http://support.huawei.com/ecommunity/bbs/10155553.html
2014-07-18 13:39 GMT+08:00 虚无缥缈 <83...@qq.com>:
> A、B两台设备,都是双网卡绑定,mode=0,交换机上也做了端口组,交换机是cisco 2960,A上起了nfs,B上挂载nfs,然后进行dd测试,cacti上看总是一个端口流量,手动down掉一个网卡,另外一个网卡才有流量。
>
> A:
> [root@backup ~]# cat /proc/net/bonding/bond0
> Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
>
>
> Bonding Mode: load balancing (round-robin)
> MII Status: up
> MII Polling Interval (ms): 100
> Up Delay (ms): 0
> Down Delay (ms): 0
>
>
> Slave Interface: em1
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 1
> Permanent HW addr:
> Slave queue ID: 0
>
>
> Slave Interface: em2
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr:
>
>
>
> B:
> root@storage-xen bak_201]# cat /proc/net/bonding/bond0
> Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
>
>
> Bonding Mode: load balancing (round-robin)
> MII Status: up
> MII Polling Interval (ms): 100
> Up Delay (ms): 0
> Down Delay (ms): 0
>
>
> Slave Interface: em1
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr:
> Slave queue ID: 0
>
>
> Slave Interface: em2
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr:
> Slave queue ID: 0
>
>
>
> 交换机:
> interface Port-channel1
> switchport access vlan 150
> switchport mode access
>
>
>
> interface GigabitEthernet0/3
> switchport access vlan 150
> switchport mode access
> channel-group 1 mode on
> spanning-tree portfast
> !
>
> interface GigabitEthernet0/10
> switchport access vlan 150
> switchport mode access
> channel-group 1 mode on
> spanning-tree portfast
--
白清杰 (Born Bai)
Mail: linuxbqj@gmail.com
敬畏耶和华是智慧的开端
Re: 为什么我的双网卡绑定,总是一个网卡有流量呢。
Posted by "linuxbqj@gmail.com" <li...@gmail.com>.
mode=0:平衡负载模式,有自动备援,但需要”Switch”支援及设定。
http://support.huawei.com/ecommunity/bbs/10155553.html
2014-07-18 13:39 GMT+08:00 虚无缥缈 <83...@qq.com>:
> A、B两台设备,都是双网卡绑定,mode=0,交换机上也做了端口组,交换机是cisco 2960,A上起了nfs,B上挂载nfs,然后进行dd测试,cacti上看总是一个端口流量,手动down掉一个网卡,另外一个网卡才有流量。
>
> A:
> [root@backup ~]# cat /proc/net/bonding/bond0
> Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
>
>
> Bonding Mode: load balancing (round-robin)
> MII Status: up
> MII Polling Interval (ms): 100
> Up Delay (ms): 0
> Down Delay (ms): 0
>
>
> Slave Interface: em1
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 1
> Permanent HW addr:
> Slave queue ID: 0
>
>
> Slave Interface: em2
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr:
>
>
>
> B:
> root@storage-xen bak_201]# cat /proc/net/bonding/bond0
> Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)
>
>
> Bonding Mode: load balancing (round-robin)
> MII Status: up
> MII Polling Interval (ms): 100
> Up Delay (ms): 0
> Down Delay (ms): 0
>
>
> Slave Interface: em1
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr:
> Slave queue ID: 0
>
>
> Slave Interface: em2
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr:
> Slave queue ID: 0
>
>
>
> 交换机:
> interface Port-channel1
> switchport access vlan 150
> switchport mode access
>
>
>
> interface GigabitEthernet0/3
> switchport access vlan 150
> switchport mode access
> channel-group 1 mode on
> spanning-tree portfast
> !
>
> interface GigabitEthernet0/10
> switchport access vlan 150
> switchport mode access
> channel-group 1 mode on
> spanning-tree portfast
--
白清杰 (Born Bai)
Mail: linuxbqj@gmail.com
敬畏耶和华是智慧的开端
Re: 添加虚拟主机失败
Posted by "linuxbqj@gmail.com" <li...@gmail.com>.
日志比较少
可以再增加一些日志内容吗?
2014-07-18 14:32 GMT+08:00 18337171837@163.com <18...@163.com>:
> 各位,今天在cs后台添加主机失败
> 具体日志如下:
> 内存使用率27%,cpu使用率50%,要创建一个2G内存,2核心的cpu,从日志看应该内存和cpu是足够的,我看二级存储也是正常的,请教下这个是什么问题。
>
> 2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Hosts's actual total CPU: 48000 and CPU after applying overprovisioning: 96000
> 2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Free CPU: 47000 , Requested CPU: 2000
> 2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Free RAM: 101071503360 , Requested RAM: 2147483648
> 2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Host has enough CPU and RAM available
> 2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) STATS: Can alloc CPU from host: 1, used: 49000, reserved: 0, actual total: 48000, total with overprovisioning: 96000; requested cpu:2000,alloc_from_last_host?:false ,considerReservedCapacity?: true
> 2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) STATS: Can alloc MEM from host: 1, used: 34225520640, reserved: 0, total: 135297024000; requested mem: 2147483648,alloc_from_last_host?:false ,considerReservedCapacity?: true
> 2014-07-18 14:19:45,968 DEBUG [allocator.impl.FirstFitAllocator] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Found a suitable host, adding to list: 1
--
白清杰 (Born Bai)
Mail: linuxbqj@gmail.com
敬畏耶和华是智慧的开端
添加虚拟主机失败
Posted by "18337171837@163.com" <18...@163.com>.
各位,今天在cs后台添加主机失败
具体日志如下:
内存使用率27%,cpu使用率50%,要创建一个2G内存,2核心的cpu,从日志看应该内存和cpu是足够的,我看二级存储也是正常的,请教下这个是什么问题。
2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Hosts's actual total CPU: 48000 and CPU after applying overprovisioning: 96000
2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Free CPU: 47000 , Requested CPU: 2000
2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Free RAM: 101071503360 , Requested RAM: 2147483648
2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Host has enough CPU and RAM available
2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) STATS: Can alloc CPU from host: 1, used: 49000, reserved: 0, actual total: 48000, total with overprovisioning: 96000; requested cpu:2000,alloc_from_last_host?:false ,considerReservedCapacity?: true
2014-07-18 14:19:45,968 DEBUG [cloud.capacity.CapacityManagerImpl] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) STATS: Can alloc MEM from host: 1, used: 34225520640, reserved: 0, total: 135297024000; requested mem: 2147483648,alloc_from_last_host?:false ,considerReservedCapacity?: true
2014-07-18 14:19:45,968 DEBUG [allocator.impl.FirstFitAllocator] (Job-Executor-10:job-2119 = [ afcde3f0-cd36-4809-b379-9bd6861669f9 ] FirstFitRoutingAllocator) Found a suitable host, adding to list: 1