see also:http://www.vmware.com/files/pdf/techpaper/vmware-multipathing-configuration-software-iSCSI-port-binding.pdf

好文:http://bbs.vmsky.com/thread-35898-1-1.html

http://communities.vmware.com/thread/397882?start=0  通常iscsi不使用LACP或者NIC Teaming

Multipathing between a server and storage array provides the ability to load-balance between paths when all
paths are present and to handle failures of a path at any point between the server and the storage. Multipathing
is a de facto standard for most Fibre Channel SAN environments. In most software iSCSI environments,
multipathing is possible at the VMkernel network adapter level, but not the default configuration.
In a VMware vSphere? environment, the default iSCSI configuration for VMware? ESXi? servers creates only one
path from the software iSCSI adapter (vmhba) to each iSCSI target. To enable failover at the path level and to
load-balance I/O traffic between paths, the administrator must configure port binding to create multiple paths
between the software iSCSI adapters on ESXi servers and the storage array.
Without port binding, all iSCSI LUNs will be detected using a single path per target. By default, ESX will use
only one vmknic as egress port to connect to each target, and you will be unable to use path failover or to loadbalance
I/O between different paths to the iSCSI LUNs. This is true even if you have configured network adapter
teaming using more than one uplink for the VMkernel port group used for iSCSI. In case of simple network
adapter teaming, traffic will be redirected at the network layer to the second network adapter during
connectivity failure through the first network card, but failover at the path level will not be possible, nor will
load balancing between multiple paths.
Some of the user guides and documentation refer to vmknic-based software iSCSI multipathing as “port
binding” or simply as “software iSCSI multipathing.” This paper provides an overview of how to enable vmknicbased
software iSCSI multipathing, as well as the procedure by which to verify port binding configuration.

Configuring vmknic-Based iSCSI Multipathing
To enable vmknic-based software iSCSI multipathing, you must:
1.Create two VMkernel port groups and connect one uplink to each of them.
2.Bind each VMkernel network adapter to the software iSCSI adapter. Then run a rediscovery of iSCSI targets to
detect multiple paths to them.

Refer to the following sections to learn the network configuration and port binding procedure required to enable
vmknic-based multipathing for software iSCSI.

注意:可以在不同的vSwitch上建立vmkernel port group.但是如果vmkernel网络适配器是在同一个子网上,则必须建立在同一个vSwitch上。

确保每个vmkernel port group 只对应一个活动的物理网卡。

Multipathing for Software iSCSI的更多相关文章

  1. How to configure ESXi to boot via Software iSCSI?

    http://blogs.vmware.com/vsphere/2011/11/how-to-configure-esxi-to-boot-via-software-iscsi.html Introd ...

  2. iSCSI Network Designs: Part 5 – iSCSI Multipathing, Host Bus Adapters, High Availability and Redundancy

    iSCSI Network Designs: Part 5 – iSCSI Multipathing, Host Bus Adapters, High Availability and Redunda ...

  3. VCAP5-DCA Objective 1.3 – Configure and Manage Complex Multipathing and PSA Plug-ins

    http://virtuallyhyper.com/2012/10/vcap5-dca-objective-1-3-configure-and-manage-complex-multipathing- ...

  4. VCAP5-DCA – What’s new?

    see also: 韩国人的教材:http://ddii.pe.kr/ Section 1.1 – Implement and Manage complex storage Determine use ...

  5. day3:vcp考试

    Q41. An administrator creates a custom ESXi firewall rule using an XML file, however the rules do no ...

  6. RedHat 和 Mirantis OpenStack 产品的版本和功能汇总和对比(持续更新)

    Mirantis 和 Red Hat 作为 OpenStack 商业化产品领域的两大领军企业,在行业内有重要的地位.因此,研究其产品版本发布周期和所支持的功能,对制定 OpenStack 产品的版本和 ...

  7. xenserver+starwind架构布署

    主机 CPU 和主板均需支持 INTER-VT/ AMD-VT ,主板默认可能没开  进BISO开启  下载最新的 xenserver ,授权注册一下 轻松得到授权文件 (鄙视一下VMWARE,看人家 ...

  8. VMware vSphere服务器虚拟化实验六 vCenter Server 添加储存

                                                                          VMware vSphere服务器虚拟化实验六 vCente ...

  9. 虚拟化(四):vsphere高可用功能前提-共享存储搭建(使用微软提供的iscsi software target,也可以使用免费开源的openfiler)

    虚拟化(一):虚拟化及vmware产品介绍 虚拟化(二):虚拟化及vmware workstation产品使用 虚拟化(三):vsphere套件的安装注意及使用 虚拟化(四):vsphere高可用功能 ...

随机推荐

  1. ASIHttpRequest release 包无法发出请求

    ASIHttpRequest 在 Release 模式下,Optimize 后会导致发不出请求. 解决方案: 去掉这两个文件的 Optimization:ASIFormDataRequest.m AS ...

  2. 【Devops】【docker】【CI/CD】jenkins源码管理,添加SSH地址后报错+Jenkins构建报错:Please make sure you have the correct access rights and the repository exists.

    jenkins源码管理,添加SSH地址后报错: Could not read from remote repository. Please make sure you have the correct ...

  3. .Net中集合排序

    public class StockQuantity { public StockQuantity(string status, DateTime dateTime, int quantity) { ...

  4. ASP.NET MVC:some benefits of asp.net mvc

    Full control over HTML Full control over URLs Better separation of concerns Extensibility Testabilit ...

  5. Java获取URL链接的文件类型

    问题发生: Java从网络批量读取图片并保存至本网站服务器后再插入文章中 今天转入一篇文章 http://news.qq.com/a/20170605/045860.htm 发现图片未能成功上传 查看 ...

  6. 如何进行Logstash logstash-input-jdbc插件的离线安装

    我们单位的服务器位于隔离区,不允许链接互联网,因此整理了在ELK集群上离线安装Logstash的jdbc input插件的方法,供大家参考. 总体思路是需要一台中转的机器,这台机器需要能够访问互联网, ...

  7. .Net Core配置文件读取整理

    一 .配置文件说明 1.配置,主要是 指在程序中使用的一些特殊参数,并且大多数 仅在程序启动的之后指定不需要修改. 2.在以前.Net项目中配置文件主要指app.config或web.config,但 ...

  8. [转]PHP SOCKET编程

    FROM : http://blog.csdn.net/hguisu/article/details/7448528 1. 预备知识 一直以来很少看到有多少人使用php的socket模块来做一些事情, ...

  9. GraphX中Pregel单源点最短路径(转)

    原文链接:GraphX中Pregel单源点最短路径 GraphX中的单源点最短路径例子,使用的是类Pregel的方式. 核心部分是三个函数: 1.节点处理消息的函数  vprog: (VertexId ...

  10. Asp.Net验证控件浅析

    ASP.NET公有六种验证控件,分别如下: 控件名           功能描叙  RequiredFieldValidator(必须字段验证) 用于检查是否有输入值  CompareValidato ...