Module Lessons Lesson 1: Storage Concepts Lesson 2: Configure iSCSI Storage Lesson 3: Configure NAS/NFS Storage Lesson 4: Fibre Channel SAN Storage Lesson 5: VMware vSphere VMFS
Trang 1Configure and Manage Virtual Storage
Module 6
Trang 2You Are Here
Course Introduction
Introduction to Virtualization
Virtual Machines
VMware vCenter Server
Configure and Manage Virtual Networks
Configure and Manage Virtual Storage
Managing Virtual Machines
Data Protection
Access & Authentication Control
Resource Management and Monitoring
Trang 3Importance
Storage options give you the flexibility to set up your storage based
on your cost, performance, and manageability requirements Shared storage is useful for disaster recovery, high availability, and moving virtual machines between hosts
Trang 4Module Lessons
Lesson 1: Storage Concepts
Lesson 2: Configure iSCSI Storage
Lesson 3: Configure NAS/NFS Storage
Lesson 4: Fibre Channel SAN Storage
Lesson 5: VMware vSphere VMFS Datastores
Lesson 6: VMware vSphere Storage Appliance
Trang 5
Lesson 1:
Storage Concepts
Trang 6Learner Objectives
After this lesson, you should be able to do the following:
Describe VMware vSphere® storage technologies and datastores
Describe the storage device naming convention
Trang 7Direct Attached
File system
NAS
NFS VMware vSphere VMFS
VMware®
ESXi™
hosts
Trang 8Storage Protocol Overview
Storage
Protocol
Supports boot from SAN
Supports VMware vSphere®
vMotion®
Supports vSphere High Availability
Supports vSphere DRS
Supports Raw Device Mapping
Trang 9Datastore
A datastore is a logical storage unit that can use disk space on one physical device, or span several physical devices
Types of datastores:
Network File System (NFS)
Datastores are used to hold virtual machine files, templates, and ISO images
Trang 10
Uses a 1MB block size,
good for storing large virtual
disk files
Uses subblock addressing,
good for storing small files
Provides on-disk, block-level
locking
host host
VMFS Datastore
Trang 11NFS
NFS:
Is storage shared over the
network at the file system
level
Supports NFS version 3
Trang 12Storage Device Naming Conventions
Storage devices are identified in several ways:
SCSI ID – Unique SCSI identifier
Canonical name – The Network Address Authority (NAA) ID is a
unique LUN identifier, guaranteed to be persistent across reboots
• In addition to NAA IDs, devices can also be identified with mpx or t10
identifiers
Runtime name – Uses the convention vmhbaN:C:T:L This name is
not persistent through reboots
Trang 13Viewing Storage Maps
target
HBA
LUN
Trang 14Physical Storage Considerations
Discuss vSphere storage needs with your storage administration
team, such as:
LUN sizes
I/O bandwidth
Disk cache parameters
Zoning and masking
Identical LUN presentation to each ESXi host
Active-active or active-passive arrays
Export properties for NFS datastores
Trang 15Review of Learner Objectives
You should be able to do the following:
Describe vSphere storage technologies and datastores
Describe the storage device naming convention
Trang 16Lesson 2:
Configure iSCSI Storage
Trang 17Learner Objectives
After this lesson, you should be able to do the following:
Describe uses of IP storage with ESXi
Describe iSCSI components and addressing
Configure iSCSI initiators
Trang 18iSCSI Components
Trang 20iSCSI Initiators
Trang 21Configuring Software iSCSI
To configure the iSCSI software initiator:
1. Configure a VMkernel port for accessing IP storage
2. Enable the iSCSI software adapter
3. Configure the iSCSI IQN name and alias (if required)
4. Configure iSCSI software adapter properties, such as static/dynamic discovery addresses and iSCSI port binding
5. Configure iSCSI security (Challenge Handshake Authentication
Protocol (CHAP))
Trang 22ESXi Network Configuration for IP Storage
A VMkernel port must be created for ESXi to access software iSCSI
The same port can be used to access NAS/NFS storage
To optimize your vSphere networking setup:
Separate iSCSI networks from NAS/NFS networks
• Physical separation is preferred
• If that is not possible, use VLANs
Trang 23iSCSI Target-Discovery Methods
Two discovery methods are
supported:
Static
Dynamic (also known as
SendTargets)
The SendTargets response
returns iSCSI qualified name and
all available IP addresses
iSCSI target
192.168.36.101:3260
SendTargets request
SendTargets response
192.168.36.101:3260
Trang 24iSCSI Security: CHAP
iSCSI initiators use CHAP for
• Software iSCSI only
ESXi also supports per-target
CHAP authentication
Software iSCSI only
Different credentials for each
target
Software iSCSI properties > General tab
Target authenticates host
Host authenticates target
Trang 25Configuring Hardware iSCSI
To configure the iSCSI hardware initiator:
1. Install the iSCSI hardware adapter
For independent hardware iSCSI adapters
• Verify properly formatted IP address and IQN names
For dependent hardware iSCSI adapters
• Determine the name of the physical NIC associated with adapter so that port binding is properly configured
2. Modify the iSCSI name and configure the iSCSI alias
3. Configure iSCSI target addresses
4. Configure iSCSI security (CHAP)
Trang 26
Multipathing with iSCSI Storage
Use multiple NICs
Connect each NIC to a
separate VMkernel port
Associate VMkernel ports
with iSCSI initiator
Configure port binding in the Properties window of the iSCSI adapter
Trang 27Review of Learner Objectives
You should be able to do the following:
Describe uses of IP storage with ESXi
Describe iSCSI components and addressing
Configure iSCSI initiators
Trang 28Lesson 3:
Configure NAS/NFS Storage
Trang 29Learner Objectives
After this lesson, you should be able to do the following:
Describe NFS components and addressing
Create an NFS datastore
Trang 30NFS Components
directory to share with the ESXi host over the network
VMkernel port defined on virtual switch
ESXi host with
NIC mapped to
virtual switch
NAS device or a
server with storage
Trang 31Addressing and Access Control with NFS
192.168.81.72 VMkernel port configured with 192.168.81.33
Trang 32Configuring an NFS Datastore
Create a VMkernel port:
For better performance and security, separate it from the iSCSI
network
Provide the following information:
NFS server name (or IP address)
Folder on the NFS server, for example, /LUN1 and /LUN2
Whether to mount the NFS file system read-only:
• Default is to mount read/write
NFS datastore name
Trang 33Viewing IP Storage Information
Datastores view > Storage Views tab
Hosts and Clusters view > Configuration tab > Storage link
Trang 35Multipathing and NFS Storage
NIC teams with Route
single switch
or stacked switches
spanned, teamed switch ports
(feature may not be available on all switches)
vmnic0 vmnic1 allow VMkernel to make
Trang 36Lab 7
In this lab, you will configure access to iSCSI and NFS datastore
1. Add a VMkernel port group to a standard virtual switch
2. Configure the iSCSI software adapter
3. Configure access to NFS datastores
4. View iSCSI and NFS storage information
Trang 37Lab 8
In this lab, you will design a network configuration for an ESXi host, based on a set of requirements
1. Analyze the requirements
2. Design virtual switches and physical connections
Trang 38Review of Learner Objectives
You should be able to do the following:
Describe NFS components and addressing
Create an NFS datastore
Trang 39Lesson 4:
Fibre Channel SAN Storage
Trang 40Learner Objectives
After this lesson, you should be able to do the following:
Describe uses of Fibre Channel with ESXi
Describe Fibre Channel components and addressing
Access Fibre Channel storage
Trang 41Using Fibre Channel with ESXi
ESXi supports:
8Gb Fibre Channel
Fibre Channel over Ethernet (FCoE)
Trang 42Fibre Channel SAN Components
Trang 43Fibre Channel Addressing and Access Control
Trang 44Accessing Fibre Channel Storage
Install one or more supported Fibre Channel adapters in the ESXi
host
The Fibre Channel adapters are recognized by the host during the
boot sequence
Trang 45Viewing Fibre Channel Storage Information
The Storage Views tab provides
information about all SCSI adapters
and NAS mounts
Trang 46FCoE Adapters
Converged Network Adapter
Network driver
FC driver
ESXi host
Network driver
Software
FC
ESXi 5.0 host
NIC with FCoE support
Ethernet IP frames
to LAN devices
FC frames to FC storage arrays
Trang 47Configuring Software FCoE: Create VMkernel Port
Connect the VMkernel to physical
FCoE NICs that are installed on your
host
The VLAN ID and Priority class is
discovered during FCoE initialization
This class is not configured in
VMkernel port
vmnic2 NIC with
FCoE support
Trang 48Configuring Software FCoE: Activate Software FCoE Adapter
The second step is to add
the software FCoE
adapter
Select host -> Configuration tab -> Storage Adapters link -> Add
Trang 49Multipathing with Fibre Channel
Multipathing allows continued
access to SAN LUNs in the
event of hardware failure It
also provides load balancing
(optional)
Trang 50Multipathing with Software FCoE
VMkernel ports
Physical adapter: vmnic2
VMkernel label: FCoE-2
Trang 51Review of Learner Objectives
You should be able to do the following:
Describe uses of Fibre Channel with ESXi
Describe Fibre Channel components and addressing
Access Fibre Channel storage
Trang 52Lesson 5:
VMware vSphere VMFS Datastores
Trang 54Using a VMFS Datastore with ESXi
Use VMFS datastores whenever possible:
VMFS is optimized for storing and accessing large files
A VMFS can have a maximum volume size of 64TB
NFS datastores are great for storing virtual machines But some
functions are not supported
Use RDMs if your virtual machine:
• Is performing SAN snapshotting
• Is clustered to a physical machine
• Has large amounts of data that you do not want to convert into a virtual disk
Trang 55To create a VMFS datastore, use the Add Storage wizard:
Select the storage type, Disk/LUN
Select an available LUN
Specify a datastore name
Specify the datastore size: use full or partial LUN
Creating a VMFS Datastore
Trang 56Viewing VMFS Datastores
Click the Storage link in the Configuration tab
Trang 57Browsing Datastore Contents
Right-click the datastore in the host’s Summary tab or click the
Storage link in the Configuration tab
Trang 58Managing Overcommitted Datastores
An overcommitted datastore can occur when there are many
thin-provisioned virtual disks that use close to their maximum
allotted disk space
Actively monitor your datastore capacity:
Alarms assist through notifications:
• Datastore disk overallocation
• Virtual machine disk usage
Use reporting to view space usage
Actively manage your datastore capacity:
Increase datastore capacity when necessary
Use vSphere Storage vMotion to mitigate space usage issues on a
particular datastore
Trang 59Increasing the Size of a VMFS Datastore
Increase a VMFS
datastore’s size to give it
more space or possibly to
improve performance
Two ways to dynamically
increase the size of a VMFS
datastore:
Add an extent (LUN)
Expand the datastore
within its extent
Trang 60Comparing Methods for Increasing VMFS Datastore Size
Trang 61Before Increasing the Size of a VMFS Datastore
In general, before making any changes to your storage allocation:
Perform a rescan to ensure that your host sees the most current
storage
Quiesce I/O on all disks involved
Note the unique identifier (for example, the NAA ID) of the volume
that you want to expand
Trang 62Delete or Unmount a VMFS Datastore
Trang 63Multipathing Algorithms
Arrays provide different
features, some offer
active-active storage processors (SP)
and others offer active-passive
SPs
vSphere 5 offers native path
selection, load-balancing and
failover mechanisms
Third-party vendors can create
their own software to be
installed on your ESXi hosts
that will allow the ESXi host to
properly interact with the
storage arrays it uses
storage processors
switches
Trang 64Managing Multiple Storage Paths
To modify the number of storage paths to use, select the datastore
to modify and click that datastore’s Properties link
Click Manage Paths in the Properties window
Trang 65Configuring Storage Load Balancing
Path selection policies exist for:
Trang 66Lab 9
In this lab, you will work with VMFS datastores
1. Review your shared storage configuration
2. Change the name of a VMFS datastore
3. Create a VMFS datastore
4. Expand a VMFS datastore to consume unused space on a LUN
5. Remove a VMFS datastore
6. Extend a VMFS datastore
Ask your instructor which LUNs contain VMFS datastores that
should not be removed or reformatted
Trang 67Review of Learner Objectives
You should be able to do the following:
Create a VMFS datastore
Increase the size of a VMFS datastore
Delete a VMFS datastore
Trang 68Lesson 6:
VMware vSphere Storage Appliance
Trang 69Learner Objectives
After this lesson, you should be able to do the following:
Describe the architecture and requirements of the VMware vSphere
Storage Appliance cluster configuration
Discuss how a VSA cluster handles failures
Trang 70Provides Storage framework for vMotion, HA , and DRS
vSphere Storage Appliance: Introduction (1)
Creates Shared
Storage
Simple manageability
Abstraction From Underlying Hardware
Resilient to server failures
Highly available during disk (spindle) failure
Pools server disk capacity to form shared storage
Leverages vSphere Thin provisioning for space utilization
Enables storage scalability Installed, configure and managed via vCenter
Trang 71vSphere Storage Appliance – Introduction (2)
vSphere vSphere vSphere
Trang 72VSA Cluster Configuration Requirements
VSA cluster
front end
back end
ESXi host 10.15.20.150 10.15.20.201
24GB RAM,
8 hard disks, RAID controller, Gigabit Ethernet switches
The VSA
Manager is used
to install the VSA
cluster
Trang 73VSA Manager
Trang 76VSA Cluster Service
VSA Datastore 2 VSA
Datastore 1
Trang 77Review of Learner Objectives
You should be able to do the following:
Describe the architecture and requirements of the VMware vSphere
Storage Appliance cluster configuration
Discuss how a VSA cluster handles failures
Trang 78Key Points
Use VMFS datastores to hold virtual machine files
NFS datastores are useful as a repository for ISO images
Shared storage is integral to vSphere features like vMotion, vSphere High Availability, and DRS