Upload
vmug-it
View
37
Download
1
Embed Size (px)
Citation preview
+
From Zero to Colo - vCloud Director in my lab
With Mike Laverick (VMware)Blog: www.mikelaverick.comEmail: [email protected] Twitter: @mike_laverick
+Agenda
�The Home Lab Backstory - Long, Long ago in a galaxy called 2003….
�Former vSphere Setup
�CH-CH-CH Changes - vSphere5.1 Setup
� Compute
� Network
� Storage
�vCD Lesson Learned…
�My Lab To-Do List…
+The Home Lab Backstory - Long, Long ago in 2003…
� My first attempt with ESX 2.0/vCenter 1.0
� Location: Under my desk
� Girlfriend Impact: NIL
+The vCloud Suite: SDDC Era� Virtual Appliances where possible/necessary
� vCenter Server Appliance (VCSA)
� Feature Parity with Windows version
� Switch allowed me to completely reconfigure resources around vCloud/SDDC agenda
� Reduce “infrastructure VM” footprint
� Beware of plug-ins; Support for the web-client (e.g. NetApp VSC)
� vCloud Director Virtual Appliances (vCD-VA)
� Use built-in Oracle XE DB
� Dead easy to setup (No Packages, DB setup)
� Beware: No multi-cell, No migration
� Beware: Demo only; Labs; Training purposes…
� vShield Manager Virtual Appliance (Mandatory)
� vSphere Replication Appliance (VR)
� vSphere Data Protection Appliance (vDP)
+vSphere5/SRM5.0/View5.1 Era –� SRM 5.0 Period (2011)
� Hello 2x Dell Equallogics
� Hello 1x NS-120 & 1x NS-20
� Hello 2x NetApp 2040s
� Hello massive colocation bill!!!
� VMware Employee Period (2012) >>>>>>>>>>>
� HomeLab & ProLab Merge
� Goodbye EMC
� Goodbye 2xPDU
� Hello 24U of extra racks space
� Hello to 14 AMPs extra power!
� Location: Quality Colocation
� Costs: £870 GBP, $1,300 USD
� Girlfriend Impact: Married 2013,4th May
+Virtual Silos
� The VMware Cluster as the New Silo?
� Discrete Blocks of:
� Compute
� Network
� Storage
� Q. Why do we like silos?
� Q. Why do we hate silos?
+Compute Continued…
� One Site; Two Clusters
� “Infrastructure” Resource Pool – No Management Cluster �
� GOAL: Maximize Resource; Setup Tiered Clusters
� Decisions:
� Different CPU types forced DRS separation
� Gold Cluster = HP DL 385s
� WHY? = More memory & FC connected to SAS storage
� Silver Cluster = Lenovo TS200
� WHY? = Less RAM, Only 1GP pipe to either SAS/SATA on NFS/iSCSI
+Storage Continued…
� Destroyed Datastores!
� Except
� Infrastructure datastore
� Templates
� Software (iso, templates, miscs)
� Originally a much more complicated layout – 7 Tiers!!!
� 4 Tiers
� 1. Platinum (NetApp, FC, SAS, SnapMirror – 5min RPO)
� 2. Gold (Dell, iSCSI, SAS, Replication – 1hr RPO)
� 3. Silver (NetApp, NFS, SATA, Datastore Cluster, 3x300 GB, Thinly Provisioned)
� 4. Bronze (Dell, iSCSI, SATA, Thinly Provisioned)
� vSphere Replication for Replication from 3>4 or 4<3
+Storage Anxieties…� Many Organizational Tenants sharing the SAME datastore
� What about Site Recovery Manager?
� What about performance – Capacity management isn’t the issue
� With Array-based Replication (ABR)
� One Failover to rule them all?
� No per-vApp Failover
� No per-Organization failover
� Solutions?
� Platinum/Gold datastores per-Organization �
� vSphere Replication ☺
� VMware vVols ☺
+Network Continued…
� Goodbye Standard Switch
� Struggle to provide redundancy/separate with the “Combo Approach”
� Many of the Adv features of vCD require Distributed vSwitch
� Classical Approach:
� Two DvSwitches
� One for internal vSphere Networking (vMotion, IP Storage, FT, Management)
� One for Virtual DataCenter
� Backed by two VMNICs each…
+Network Anxieties…
� All my Provider vDCs share the SAME DvSwitch
� What about “Fat Finger Syndrome”?
� How realistic is that?
� Time to re-examine “Best Practices”
� Do best practices represents an ideal OR an ideal filtered through the limitations of a technology
� Provider vDCs in vCD 1.x – One Cluster, No Tiering of Storage
� Provider vDCs in vCD 5.x – Many clusters, Tiering of Storage
+Lesson Learned� When thinking about a Provider vDC
� All the resources matter
� Compute + Storage + Networking
� By far the easiest for me was compute
� But my “Gold” cluster has no FT Support �
� Prepare to make compromises/trade offs
� UNLESS all your hosts are the SAME
� VXLAN needs enabling on Distributed Switches via vSphere Client
� Prior to creating a Provider vDC
� Watch out with VMs already on the cluster – vCD ESX Agent
� Running existing “infrastructure” VMs on a cluster
� Stops the install of the vCD Agent
� Has to be done on per-ESX host basis (easy)
+More Lessons Learned…
� Get your VLANs sorted BEFORE you use them in vCD…
� Beware of Orphaned VLAN references in the vCD Databases
� http://kb.vmware.com/kb/2003988
+Work out your IP before you start!
�“Wrong”
� 192.168.3.x – “External Network”
� 172.168.x.x – “Organization Network”
� 10.x.x.x – “vApp Network”
�“Right”
� 10.x.x.x – “External Network”
� 172.168.x.x – “Organization Network”
� 192.168.1.x – “vApp Network”
�Keep it simple – whole ranges dedicated
+IP Ranges can be tricky to change� Even with vApps powered off – options unavailable
� Gateway Address
� Network Mask
� Resolution involves admin:
� Add new vApp Network
� Remap all VMs to new vApp Network
� Remove old vApp Network
+vApp Networks & Edge Gateway
� Every vApp Network you create:
� Creates a vCNS Edge Gateway
� Consumes resources
� Solution
� Create two vApps per Organization
� TypeA: One on the Organization Network
� TypeB: One on its own vApp Network
� Power off the Type B vApp to save resources
� Beware of static MAC/IP on Power Offs
+Establish a meaningful naming convention…
� I KNOW EVERYONE SAYS THIS, BUT IN A HOME LAB DON’T YOU CUT CORNERS SOMETIMES?
� <ORGNAME><NetworkType><Purposes>
� CORPHQ-OrgNetCorp-EdgeGateway
� CORPHQ-vAppNet-WebGateway
� Makes screengrabs, documentation & troubleshooting soooo much easier…
� Register Edge Gateway devices in DNS…
� Helps with SysLog – watch out for stale DNS Records…
+Your Lab?
� Nested ESX
� http://communities.vmware.com/community/vmtn/bestpractices/nested
� vTARDIS? - http://vinf.net/2011/10/03/vtardis-5-at-a-vmug-near-you/
� Workstation/Fusion – Out of the box nested ESX…
� AutoLab - http://www.labguides.com/autolab/
� Redundant Array of Inexpensive PCs (RAIPC)
� Community Hardware Page:
� http://communities.vmware.com/community/vmtn/bestpractices/crc
+vINCEPTION Levels
� vINCEPTION Level 0
� Physical ESX hosts, Virtual Everything Else (DC, vCenter, vCD)
� vINCEPTION Level 1
� vApp of vSphere running under Level 0
� Including vCD, vCNS Manager
� vINCEPTION Level 2
� vApps running under Level 1…
+Lab Future…� The DONE List
� Make my External Juniper Firewall work with vShield
� Need new servers?
� Dell?
+Follow my journey journal…
� Text Blogpost
� Follow my journey!
� Search for “Mike Laverick – My vCloud Journey Journal”
� Subscribe iTunes:
� http://tinyurl.com/audiowag (Audio)
� http://tinyurl.com/videowag (Video)
+
Questions (Welcomed) & (Attempts at) Answers
Blog: www.mikelaverick.com
Email: [email protected]
Twitter: @mike_laverick