1 / 44

XenServer Storage Integration Deep Dive

XenServer Storage Integration Deep Dive. Agenda. XenServer 5.5 Storage Architecture. Multipathing. Vendor Integration. StorageLink. Platinum Edition. Enterprise Edition. Free Edition. Citrix XenServer & Essentials 5.5 Family. NEW. Stage Management. Lab Management.

len-hill
Download Presentation

XenServer Storage Integration Deep Dive

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. XenServerStorage Integration Deep Dive

  2. Agenda XenServer 5.5 Storage Architecture Multipathing Vendor Integration StorageLink

  3. Platinum Edition Enterprise Edition Free Edition Citrix XenServer & Essentials 5.5 Family NEW Stage Management Lab Management Provisioning Services (p+v) Essentials Provisioning Services (virtual) WorkloadBalancing NEW StorageLinkTM High Availability Performance Monitoring Workflow Studio Orchestration Live Migration (XenMotion) Active Directory Integration NEW Generic Storage Snapshotting 64bit, Windows, Linux Workloads XenServer XenCenter Management Shared Storage (iSCSI, FC, NFS) No socket restriction

  4. XenServer 5.5 Storage Architecture

  5. Expanded Backup Support

  6. Storage Technologies XenServer 5.5 iSCSI / FC XenServer 5.0 iSCSI / FC XenServer 5.0 / 5.5 NFS / EXT3 Storage Repository Storage Repository Storage Repository LUN LUN Filesystem VHD header VHD header LVM Logical Volume LVM Logical Volume .VHD file .VHD file LVM Logical Volume LVM Logical Volume LVM Volume Group LVM Volume Group VM virtual disk

  7. LVM XenServer 5.5 (LVHD) Replaces LVM for SRs Hosts VHD files directly on LVM volumes Best of both worlds Features of VHD Performance of LVM Adds Advanced Storage features Fast Cloning Snapshots Fast and simple upgrade Backwards compatible

  8. Multipathing

  9. Why using Multipathing? • Path redundancy to storage • Performance increase by load sharing algorithms • Many fibre channel environments by default have multiple paths LUN 1 Storage controller 1 LUN 1 FCHBA 1 LUN 1 LUN 1 FCHBA 2 Storage controller 2 LUN 1 FC Switches XenServer Storage Subsystem

  10. Enabling Multipathing • xe host-param-set other-config:multipathing=true uuid=host_uuid • xe host-param-set other-config:multipathhandle=dmp uuid=host_uuid • Note: Do not enable multipathing by other ways (e.g. CLI)!!!

  11. XenServer supports 2 multipathing technologies See details: http://support.citrix.com/article/ctx118791

  12. DMP vs RDAC MPP • Check if RDAC MPP is running • lsmod | grep mppVhba • „multipath –ll“ would show MD device as output (if DMP is active) • Use only 1 technology • When RDAP MPP is running use it • Otherwise use DMP

  13. MPP RDAC: Path check Lun #0 - WWN: 600a0b80001fdf0800001d9c49b0caa1 ---------------- LunObject: present DevState: OPTIMAL Controller 'A' Path -------------------- Path #1: LunPathDevice: present DevState: OPTIMAL Path #2: LunPathDevice: present DevState: OPTIMAL Controller 'B' Path -------------------- Path #1: LunPathDevice: present DevState: OPTIMAL Path #2: LunPathDevice: present DevState: OPTIMAL • mpputil

  14. DMP: Path check • Monitoring using XenCenter • Monitoring using CLI • Command: multipath -ll

  15. Multipathing & Software iSCSI

  16. iSCSI with Software Initiator • IP addressing to be done by XenServer Dom-0 • Multipathing also to be done by XenServer Dom-0 • Dom-0 IP configuration is essential Storage LAN Ctrl 1 XenServer Dom-0 LUN 1 LUN 1 IP LUN 1 IP IP NIC 1 LUN 1 XenServer Switches Storage Subsystem

  17. Best practice configuration: iSCSI storage with multipathing Subnet: 255.255.255.0 • Separation of subnets also on IP base Subnet 1 Storage LAN Adapter 1 Port 1 IP: 192.168.1.201 NIC 1 NIC 1 IP: 192.168.1.10 Port 2 IP: 192.168.2.201 NIC 2 NIC 2 IP: 192.168.2.10 Storage LAN Adapter 2 Port 1 IP: 192.168.1.202 LUN 1 Port 2 IP: 192.168.2.202 XenServer FC Switches Subnet 2 Storage Subsystem

  18. Not recommended configurations for multipathing and iSCSI: Both server NICs in same subnet Mixing of NIC teaming and multipathing Subnet 1 Subnet 1 NIC 1 NIC 1 Team IP: 192.168.1.10 NIC 1 IP: 192.168.1.10 NIC 2 NIC 2 NIC 2 IP: 192.168.1.11 XenServer XenServer

  19. Multipathing with Software InitiatorXenServer 5 • XenServer 5 supports Multipathing with iSCSI software initiator • Prerequisites are: • iSCSI target uses same IQN on all ports • iSCSI target ports operate in portal mode • Multipathing reliability has been enhanced massively in XenServer 5.5

  20. How to check if iSCSI target operates in portal mode? • Execute iscsiadm -m discovery --type sendtargets --portal <ip address of 1 target> • Output must show alls IPs of the target ports with identical IQN • Example:192.168.0.161:3260,1 iqn.strawberry:litchie192.168.0.204:3260,2 iqn.strawberry:litchie • When connecting to iSCSI target using XenCenter Storage Repository Wizard, also all target IPs should show up after Discovery

  21. NetApp Integration

  22. NetApp Storage • NetApp Storage supports Multipathing • For configuring NetApp storage and modification of multipath.conf see whitepaperhttp://support.citrix.com/article/CTX118842 • NetApp typically supports portal mode for iSCSI multipathing for iSCSI SW Initiator is supported • Especially for low-end NetApp storage (e.g. FAS2020) with limited LAN adapters special considerations take place

  23. NetApp low-end storage (iSCSI) • Often limited by NIC configuration • Example: 2 NICs per head • 1 aggregate / LUN is represented by 1 head at a time (other head for fault tolerance) • Thus: 2 NICs effectively can be used for storage connection • Typically Filer delivers non-block-based protocols (e.g. CIFS) which also require redundancy as well as block based protocols (e.g. iSCSI)

  24. Example FAS2020: Scenario 1nonetworkreduncancyforiSCSIand CIFSseparationofnetworks CIFS Network Controller 0(active) NIC 0 NIC 1 NIC 0 Controller 1(fault tolerance) NIC 0 NIC 1 NIC 1 iSCSI Network

  25. Example FAS2020: Scenario 2network redundancy for iSCSI and CIFSno separation of networks CIFS & iSCSI Network Controller 0 (active) NIC 0 vif / bond NIC Bond NIC 1 NIC 0 Controller 1(fault tolerance) NIC 0 NIC 1 vif / bond NIC 1

  26. Example FAS2020: Scenario 3network redundancy for iSCSI (multipathing) and CIFSseparation of networks NIC Bond Ctrl 1(active) CIFS VLAN CIFS VLAN NIC 0 iSCSI VLAN NIC 0 iSCSI VLAN NIC 1 Vif / bond NIC 1 Multipathing Controller 1(fault tolerance) Same configuration NIC 2 NIC 3

  27. Dell / Equalogic Integration

  28. Dell Equalogic Support • XenServer 5.5 includes Adapter (min. firmware 4.0.1 required) • Redundant path configuration does not depend on using adapter or not • All PS series are supported as running same OS • StorageLink Gateway support planned

  29. Dell / Equalogic • See whitepaper for Dell / Equalogic storagehttp://support.citrix.com/article/CTX118841 • Each Equalogic has two controllers • Only 1 controller is active • Uses „Group ID“ address on storage side(similar to bonding / teaming on server side) • Only connection over group ID, no direct connection to the iSCSI ports possible • Therefore multipathing cannot be used  bonding on XenServer side

  30. Datacore Integration

  31. Multipathing architecture with Datacore Subnet: 255.255.255.0 • Different IQNs for targets – no portal mode possible!! Subnet 1 Storage Controller 1 Port 1 IP: 192.168.1.201 NIC 1 NIC 1 IP: 192.168.1.10 IQN 1 NIC 2 NIC 2 IP: 192.168.2.10 Storage Controller 2 IQN 2 LUN 1 Port 2 IP: 192.168.2.202 XenServer Switches Subnet 2

  32. Datacore hints • Special attention for software iSCSI • Follow Datacore technical bulletin: TB15ftp://support.datacore.com/psp/tech_bulletins/TechBulletinsAll/TB15b_Citrix%20XenServer_config_501.pdf • Datacore in VM • O.k. when not using HA • Configuration possible, but take care about booting the whole environment • Take care when updating XenServer

  33. StorageLink

  34. Logical advancement of XenServer integrated storage adapters Netapp & Equalogic Storage adapter

  35. Citrix StorageLink Overview (XenServer) Data Path XenServer Control Path SAN Storage Cloud Guest iSCSI / FC Snap-in for XenServer StorageLink

  36. Leveraging the best of virtualization and storage

  37. StorageLink Overview XenServer Hyper-V Netapp EqualLogic DOM0 VSM Bridge VDS Parent Partition XenServer Hyper-V Data Path Data Path Virtual Storage Manager Control Path SAN/NAS

  38. StorageLink Gateway Overview SMI-S is the preferred method of integration as it requires no custom development work XenServer Dell Vendor-specific VSM Storage Adapters run in separate processes

  39. Storage Technologies XenServer 5.5 iSCSI / FC XenServer 5.5 iSCSI / FC + Storage Repository Storage Repository LUN LUN LUN VHD header VHD header LVM Logical Volume LVM Logical Volume LVM Volume Group VM virtual disk

  40. Snapshot types

  41. StorageLink: Microsoft look-and-feel

  42. Essentials: Example usage scenarioEffective creation of VMs from template VM clone 1x VM clone 3x VM Template 1. Copyof LUN 1. Copyof LUN 2. ModificationofZoning 2. ModificationofZoning 3. Creationof VM 3. Creationof VM Hugemanualeffort 4. Assignmentof LUN to VM 4. Assignmentof LUN to VM 1. Copyof LUN fewmouseclicks 2. ModificationofZoning 3. Creationof VM 4. Assignmentof LUN to VM 1. Copyof LUN • Effectiveness: • Fast-Cloningusing Storage Snapshots • Fullyautomatedstorageand SAN configuration • for FC andiSCSI 2. ModificationofZoning 3. Creationof VM LUN 4. Assignmentof LUN to VM

  43. StorageLink: Supported Storages • StorageLink HCLhttp://hcl.vmd.citrix.com/SLG-HCLHome.aspx

More Related