Thursday, January 26, 2012

Where to find the SQL 64 bit Native Client for vCenter Installs

Microsoft® SQL Server® 2008 R2 Native Client  X64 Package (sqlncli.msi)

VMware Host Profiles giving errors about Path Selection Policy for naa and mpx devices

I am trying to use host profiles in a full NFS vCenter, no block level storage other than the local SSD disks for the ESXi 5 servers to boot from and some various emulated CDROM’s attached to some of the hosts. Because I don’t really care about the uniformity of my local storage I just want to ignore these errors and get a Green “Compliant” result for my cluster so I can fix the important issues. I tried going into Edit Profile and removing a lot of data from under Storage Configuration/PSA & NMP & iSCSI, but this led to more errors, such as below:

Failures Against Host Profile

Host state doesn’t match specification: device mpx.vmhba32:xx:xx:xx Path Selection Policy needs to be set to default for claiming SATP

Host state doesn’t match specification: device mpx.vmhba32:xx:xx:xx needs to be reset


To resolve the issue, you need to right click on the Actual storage profile under “Host Profiles” on the left hand side of the screen and choose “Enable/Disable Profile Configuration…”


Then unselect items from here such as PSA device configuration and whatever else causing you unnecessary grief.


After hitting OK, rescan your hosts and you should finally get your Green Light!

Credits go to the knowledge base and to Kevin Space.

Monday, January 23, 2012

Setup VLAN tagging for ESXi on a Dell Blade Server using Dell PowerConnect M8024-k chassis switches

When using ESX on a server it is good to have a lot of network cards, especially if using vCD (vCloud) .  This is a follow up to my previous post about dividing a single Dell NIC into multiple nics (partitioning)

Before you begin, you need to design your setup, for my example I want 9 VLAN’s.

1010 – mgmt (the vCenter & vCM & vCell, etc…)

1020 – guests (external facing)

1030 – vMotion

1040 – NFS (not using FC)

4010 – vcdni1

4020 – vcdni2

4030 – vcdni3

4040 – vcdni4

4050 – vcdni5

Step 1 Login and look around.

Login to Dell OpenManage, Open Switching/VLAN/ choose “VLAN Membership”.  Out of the box you will only see the default VLAN 1.  Assuming you are using multiple chassis with clusters that will span these chassis like I am, you will need Tagging to flow through your core switches into these.  On this default VLAN, you will see the “Lags” at the bottom left hand side, the Current is set to “F” which is forbidden, meaning this default VLAN will not pass through the trunk into the core switch and cross chassis, for the default VLAN 1, this is good, for the other VLANs, we will change it.


Step 2, Add VLAN’s. 

Under “VLAN Membership”, click “Add” near the top, type in your VLAN ID and name, then click Apply.  Repeat step to create all of your VLAN’s.


Step 3, Change VLAN Lag type to Tagged.

Click on Detail after you have created your VLANs.  Choose your VLAN under “Show VLAN”, Under Lags, change the “Static” box from “F” to “T” by clicking it a couple times.  Then click apply, repeat for all new VLAN’s you’ve created that you want to have flow through your core network.


Step 4, Change VLAN Port Settings to “Trunk”

**Warning** before you do this, you must have iDRAC console access to your blade or you may lose connectivity to it****

Click on “Port Settings” which is just below “VLAN Membership”.  You will now see the port Detail page.  For each Port Te1/0/1 through Te1/0/16 change the Port VLAN Mode from Access to Trunk, then Click Apply.  This will allow the blades to pass multiple VLAN networks to and from themselves.    Repeat this step 16 times.


Step 5, Modify ESXi to accommodate the new VLANs

Open a iDRAC session to your blade(s).   F2 to Login, Choose “Configure Management Network”, Choose “VLAN (optional)”, then type in your VLAN ID


Hit Enter, then ESC.  It will ask if you want to Apply changes and restart mangement network?  Say (Y)es.

NOTE ***In the dell Switch UI, make SURE to click the little floppy disk picture in the upper right to Save your work when your done or you’ll get to repeat it after your next power outage like I did***

You should be done, repeat these steps to get all your blades online and using VLAN’s.

Wednesday, January 18, 2012

Divide single Dell NIC into multiple NICs, going from 2 to 8 nics per blade Dell Blade Server

I was looking for the Dell equivolent to HP Flex-10’s FlexFabric Adapter.  In dell speak this is called “independent NIC partitioning“ or just “NIC partitioning” or NPAR. 

First let me give you some background into my new system.  It consists of M1000E blade cabinets, M710HD Blade Servers with Broadcom 57712-k 10GbE 2P nics, and PowerConnect M8024-k Cabinet Switches.  My goal is to turn two 10GB NIC’s into four NIC’s made up of two 1GB NIC’s and two 9GB NIC’s

Before you connect to iDRAC, If you want to use your mouse, you must set the Mouse Mode to USC\Diags (also don’t do this through remote desktop)  Make this change from the iDRAC GUI.  I always change the media to attach or I won’t be able to install ESX later.


Apply the setting, make sure to wait for the confirmation, otherwise it didn’t happen.

To make the actual NIC Partitioning changes, you must use the ""Dell Unified Server Configurator”, which being the noob I am, I tried to find for download, but apparently you access it during the blade server boot by pressing F10 to access the UEFI (System Services")


Then you will see something like this when it boots to UEFI


Select “Hardware Configuration”, you can do that with your mouse or the arrows on the keyboard. 


Then choose tab over to HII Advanced Configuration


You will need to do the following twice, once per nic.


Then Device Configuration Menu


Change Disabled to “Enabled”


After you hit back, you will see a new options, “NIC Partitioning Configuration Menu”, select it now.


Then Select “Global Bandwidth Allocation Menu”


Tab doesn’t work on the next page, some person decided to make tab only go back and forth between the top row and the back button.  However, luckily you can use your UP and Down arrows to move between rows.

Unfortunately it doesn’t appear you can Partition a NIC into less than 4, I only want 2, so using that methodology I will create one with 90 and one with 10, the other two I will give 1 to because I can’t give them 0.  You are allowed to over allocate the Maximum Bandwidth.  1%=100MB, so the find 2 NIC’s I will create will be 100MB adapters I won’t end up using.

Type in your number, then hit enter, then choose the next row.

HINT: if using the mouse and it gets a bit squirrly, Play with “ALT-C” and toggle Hide Local Cursor on and off to improve mouse response.


Choose Back, then Back, then Finish,  It will prompt you to Save, of course say yes, then repeat for your next adapter.


After you have done both NIC’s, Choose back, then Exit and Reboot.

Here comes the really fun part.  After installing and booting into ESX, it appears that only the 2nd NIC took the partitioning command as the first one is still at 25/25/25/25 and the second one is 10/90/1/1 correctly.  I made the changes again for the first NIC and rebooted again, and it works.  I verified this over 3 different new blades, if anyone knows the solution, I’d love to hear it.


Credits, I found the information for this article HERE

Missing LUNS from openfiler iscsi device on ESXi 5.0

My entire home vCenter cluster went down, I couldn’t get a response from anything so I rebooted the hosts and the storage, after reboots the two LUNS presented from iscsi were still there, but only one shown inside of ESX.  I rebooted the openfiler iscsi with the force check disk option, after reboot I was able to see both LUNS.  Thank God

Tuesday, January 17, 2012

See old entries in vSphere “Tasks & Events”

Go into PowerCLI
Export it out so that you can search it.
[vSphere PowerCLI] C:\Program Files (x86)\VMware\Infrastructure\vSphere PowerCLI
> Get-VIEvent > c:\event.txt

Tuesday, January 10, 2012

How to reboot a dell Server’s iDRAC

Login to iDRAC remotely, “help” is about the only command you can run besides racadm.  To reboot the iDRAC management card, type “racadm racreset”, wait about a minute and it should be up again.  You can run a constant ping on the iDRAC to watch it go down and back up.  Other racadm commands can be found here: