Perhaps I’m just a little slow, but this one took me a while to figure out. I know there’s a lot of documentation floating around, but none of it seemed to work for my environment.
I was given a Dell PowerEdge 2950 server (2.5″ SAS backplane) with 8x 2.5″ 147GB 10KRPM SAS drives. It only had 16GB RAM so I upgraded to 32GB. Set the RAID to 10 (588GB raw) and away I went. I was also given a Cisco 2950 switch to play around with.
First I installed vmware esxi 4.1 (I know 5.0 is already out, but I already had an older server running esxi41 and wanted to keep them roughly the same). I then realized the free hypervisor didn’t include many features I was used to (Live migrations is huge in my book). For shame.
It was then that a fellow worker wanted to install Server 2008 R2 with hyper-v. For shame.
I like free, and I like a lot of the features of KVM, so I went with Proxmox 1.9.
I plugged eth0 into port 21 (vlan 40) for management
I plugged eth1 into port 5 (trunk) for trunking
Here’s my proxmox networking config (/etc/network/interfaces)
* as this didn’t like my tabbing, here’s a link to the txt file*
# LOCALHOST
auto lo
iface lo inet loopback# INTERFACES
iface eth0 inet manual
iface eth1 inet manual# PRIMARY BOND
auto bond0
iface bond0 inet manual
slaves eth0 eth1
bond_miimon 100
bond_mode 4# SECONDARY BONDS
auto bond0.40
iface bond0.40 inet manual
vlan-raw-device bond0auto bond0.10
iface bond0.10 inet manual
vlan-raw-device bond0auto bond0.20
iface bond0.20 inet manual
vlan-raw-device bond0auto bond0.30
iface bond0.30 inet manual
vlan-raw-device bond0# BRIDGES
auto vmbr0
iface vmbr0 inet static
address 10.4.0.216
netmask 255.255.255.0
gateway 10.4.0.5
bridge_ports bond0.40
bridge_stp off
bridge_fd 0auto vmbr1
iface vmbr1 inet manual
bridge_ports bond0.10
bridge_stp off
bridge_fd 0auto vmbr2
iface vmbr2 inet manual
bridge_ports bond0.20
bridge_stp off
bridge_fd 0auto vmbr3
iface vmbr3 inet manual
bridge_ports bond0.30
bridge_stp off
bridge_fd 0
And here’s my Cisco switch configuration (just for the ports used obviously):
interface FastEthernet0/5
description proxmox VM
switchport trunk allowed vlan 10,20,30,40
switchport mode trunk
spanning-tree portfastinterface FastEthernet0/21
description proxmox Management
switchport access vlan 40
spanning-tree portfast
Next I’ll be testing some iscsi server connectivity.
Well I went ahead and ran some initial tests:
POC server:
2.7GHz AMD x64, 4GB RAM, 160GB SATA 7200, 10/100/1000
CPU BOGOMIPS: 10772.71
REGEX/SECOND: 418377
HD SIZE: 36.67 GB (/dev/mapper/pve-root)
BUFFERED READS: 98.46 MB/sec
AVERAGE SEEK TIME: 11.16 ms
FSYNCS/SECOND: 573.28
Dell 2950 Production:
Dual 3.16GHz Xeon (X5460), 32GB RAM, 8X 147GB SAS 10000, Dual 10/100/1000
CPU BOGOMIPS: 50539.82
REGEX/SECOND: 1036781
HD SIZE: 94.49 GB (/dev/mapper/pve-root)
BUFFERED READS: 282.70 MB/sec
AVERAGE SEEK TIME: 5.22 ms
FSYNCS/SECOND: 1998.46
Still setting up the test iSCSI