DHCP problems after Ubiquiti EdgeRouter firmware upgrade

In my homelab I use a Ubiquiti EdgeRouter Lite 3-port and UniFi AC Access Points for some time now. After upgrading the Ubiquiti EdgeRouter to the latest firmware (EdgeOS 1.10.1) my WIFI devices where unable to get an IP address. I have different VLANs defined on the EdgeRouter for the WIFI networks. Each VLAN has it’s own DHCP scope configured.

In the EdgeRouter GUI I didn’t find any clue why the WIFI devices didn’t get an IP address anymore, so I opened a SSH session to the EdgeRouter and start troubleshooting. First I tried to start the DHCP service by using this command.

sudo service dhcpd start

The following error is displayed:

[….] Cannot start the DHCP server because configuration file /opt/vyatta/etc/d [FAILconf is absent. … failed!

The DHCP service cannot be started, that’s the problem why the WIFI devices didn’t get an IP address anymore. Then i looked in the following log files:

  • cat /var/log/messages
  • cat /var/log/vyatta/vyatta-commit.log

In the vyatta-commit.log the following error is displayed under the [service dhcp-server] section:

[ service dhcp-server ]
Static DHCP lease IP ‘’ under static mapping ‘Chromecast’
under shared network name ‘WIFI’ is already is in by static-mapping ”.
DHCP server configuration commit aborted due to error(s).

In the DHCP scope for the WIFI VLAN there was a static IP mapping called “Chromecast”. I removing the “Chromecast” static IP mapping in the GUI of the EdgeRouter. In the SSH session tried to start the DHCP service by using the following command.

sudo service dhcpd start

Starting DHCP server daemon…

The DHCP service is started.  In the vyatta-commit.log no new errors are displayed and the WIFI devices were able to get an IP address.  Removing the “Chromecast” static mapping cleared the duplicate static IP error.

The mobile SDDC and EUC lab environment

With my  company I work for (Ictivity), we decided to develop a mobile Software Defined Data Center (SDDC) and End User Computing (EUC) lab environment. This mobile lab environment will be used to demo the VMware SDDC and End User Computing (EUC) stack with integration of third party solutions.  One of the reasons to use a physical lab environment instead of cloud services was flexibility and  having no dependencies.

The past moths I’ve got asked what components we used to build this lab environment. So here is an quick overview. The environment logically looks like the picture below:

Demo Environment

This environment contains three physical hosts with VMware ESXi installed and one switch. One ESXi host function as management host. On this management host the following software bits are installed:

  • vSphere 6
  • VSAN Witness
  • NSX Manager
  • Fortigate VMX
  • vRealize components
  • The End User Computing stack such as Horizon View, App Volumes, User Environment Manager and Identity Manager
  • Veeam

The other 2 ESXi hosts function as demo cluster. On this 2-node cluster the following software bits are installed:

  • vSphere 6
  • Virtual SAN (VSAN) All Flash (AF) configuration
  • NSX integration
  • Windows 10l
  • Windows Server 2012 R2

A laptop is used to connect to the lab environment.

What component are used? 

Some highlights of this lab are:

  • 4U rackmount flightcase
  • Mini-ITX motherboard
  • Intel Xeon D-1541 single socket System-on-Chip 8 core processor
  • 2 x 10 GbE Ethernet adapters
  • Only SSD is used
  • IPMI port


The case is a robust custom made 19″ 4U rackmount flightcase with a removable front and back. It has two wheels so you can carry it easily around. This case contains 3 servers and one switch.Here is a picture of the case including alle the hosts and the switch.

IMG_0622Indeling Flightcase


The flightcase contains three SuperMicro SYS-5018D-FN4T 1U Rackmount hosts with the following hardware specifications:

  • Chassis: SuperMicro 19″ 1U with a 200W Gold level power supply. Optimized for Mini-ITX (SuperChassis SC505-203B)
  • Motherboard: Super X10SDV-8C-TLN4F Mini-ITX board
  • Processor: 1 x Intel Xeon D-1541 single socket System-on-Chip. This processor contains 8 cores with 16 threads (hypertreading)
  • Memory: 4x DDR4 DIMM sockets (maximum 128 GB, 4 x 32GB DDR4 ECC  memory)
  • LAN: 2 x 10GbE and 2 x 1 GbE and 1 x IPMI LAN port
  • Expansion slots: 1 x PCIe 3.0 x16 slot and a M.2 PCIe 3.0 x4
  • Video: Aspeed AST2400
  • USB: 2x USB 3.0 and 4x USB 2.0

Management host

  • Memory: 4 x 32GB = 128 GB
  • SSD: 2 x Samsung PM863 MZ-7LM1T9E – SSD Enterprise – 1.92 TB – intern – 2.5″ – SATA 6Gb
  • Disk: Seagate Enterprise 6 TB disk (for backup)
  • USB Stick: Sandisk Ultra Fit USB3 16 GB (for booting ESXi)

Demo hosts 

Each host contains the following hardware:

  • Memory: 2 x 16GB = 32 GB per server
  • SSD: 1 x Intel P3500 SSD 1.2 TB PCIe 3.0 x4 (NVMe) and Samsung 950 Pro V-Nand M.2 PCI-e SSD 512GB
  • USB Stick: Sandisk Ultra Fit USB3 16 GB (for booting ESXi)


  • Switch: Netgear ProSafe Plus XS708E 8 x 10 Gbps +SFP slot


  • 6 x UTP CAT6 0.50 cm cables
  • 1 x UTP CAT6 5m
  • 1 x UTP CAT6 10m


Processor host

The two Intel X552/X557-AT NICs are not recognized by ESXi 6.5 and lower versions by default. To enable the Intel X552/x557 2 x 10GbE NICs download the Intel driver on the VMware website (link). Extract the ZIP file and install the offline bundle by using the following command:

esxcli software vib install -d /vmfs/volumes/datastore/driver/ixgbe-4.4.1-2159203-offline_bundle-3848596.zip

With this mobile SSDC lab environment we archived the following benefits:

  • Mobile and easy to carry around
  • Flexibility to install the latest VMware SDDC and 3e party software
  • No dependency
  • Enough horsepower
  • Low noise and power consumption
  • Remote accessible from our datacenter
  • IPMI and KVM support


Home lab extension with an Intel NUC 6th generation

For my home lab I bought a 6th generation Intel NUC. The Intel NUC has following specifications:

  • Intel NUC6i3SYH
  • Intel i3-6100u (Skylake) 2.3 GHz Dual Core, 3 MB cache, 15W TDP
  • 2 memory slots for DDR4-2133 SODIMM memory, maximum is 32 GB memory
  • Intel HD Graphics 520 GPU
  • Intel I219-V Gigabit network adapter and Intel Wireless-AC 8260 WIFI adapter
  • Option to install a 2.5″ HDD/SDD and a M.2 SSD card (2242 or 2280)
  • 4 USB 3.0 ports (2 in the front and 2 on the rear)
  • SD card reader (SDXC cards)
  • Case and a 19V AC-DC adapter

IMG_9206 IMG_9209

The Intel NUC will be used as management server for my Software Defined DataCenter (SDDC) home lab environment. The Intel NUC will host VMs such as:

  • Domain Controller + DNS
  • vCenter Server Appliance
  • Virtual SAN witness appliance
  • Veeam backup
  • Etc.

The VMs are stored on a Synology NAS. The Intel NUC will use a NFS connection to the Synology NAS.  The NUC will not have any disks. It will boot ESXi from the USB stick


The 6th generation Intel NUC leaves two choices for choosing a CPU:

  • Intel I3 Skylake available on the NUC6i3SYH model
  • Intel I5 Skylake available on the NUC6i5SYH model

Both CPUs have 2 cores and support hypertreading. The table below gives a quick comparison between both processors:


For this configuration the Intel NUC with the I3-6100u processor is sufficient and saves 100 euro. The I3 has 2 cores and hypertreading, so 4 logical processors are displayed in the hypervisor.


Other advanced technologies such as VT-x, VT-d, EPT are fully supported.


The Intel NUC has 2 memory slots and support up to 32 GB DDR4 2133 MHz SODIMM memory. I added  2 Crucial 16 GB DDR4-2133(CT16G4SFD8213) modules which makes a total of 32 GB memory.

IMG_9186 IMG_9190

I use the same memory as suggested by the blog “virten.net” link.

Network card

The Intel NUC has an Intel I219-V Gigabit network adapter and a wireless network card. Only the Intel I219-V can be used with VMware ESXi.


The NUC has a M.2 (PCIe Gen 3 x4) slot and a Intel AHCI SATA-600 controller. It is possible to install a 2.5″ SDD or harddisk in the drive cage.


The VMs are on a Synology NAS. So the NUC will not have any disks other than a USB drive for booting VMware ESXi.

VMware ESXi

An USB 3 stick is used to boot VMWare ESXi. On the USB stick is VMware ESXi 6.0 U1b (VMware-VMvisor-Installer-201601001 3380124.x86_64) installed. For creating a USB stick with ESXi 6 you can use the blogpost here. Only step 1 till 3 are needed.

There is no need to add extra drivers to the ESXI image because the network and storage adapter are recognized by default.

LAN Storage

Passthrough is supported by the CPU and motherboard.


Nesting such as VMware in VMware and Hyper-V in VMware is possible. Below is an screenshot of a Hyper-V Server with a VM hosted on ESXi.


Power consumption

The average power consumption of the NUC is between 20 and 30 watt with a couple of VMs active.


 Component Amount Total
Intel NUC NUC6i3SYH 1 € 299,00
Crucial 16 GB DDR4-2133 2 € 235,80
 USB3 Stick 16 GB 1  € 10,00
Total € 544,80


The 6th generation Intel NUC is an great and easy option for creating a small ESXi home lab. I use the Intel NUC as management server with a couple of VMs. Another use case is creating a 2/3- node hybrid Virtual SAN (VSAN) cluster. Put a Samsung 950 PRO in the M.2 slot for caching and a 2.5″ HDD as capacity tier. Easy.

Pros and cons


  • All in-one-package including a motherboard, processor, enclosure and power adapter.
  • Supports up to 32 GB of memory
  • Easy to install
  • Small Form Factor
  • Low noise & power consumption


  • The hardware is not on the VMware HCL
  • Need a converter to connect to a DVI or VGA monitor
  • Only 2 cores available
  • No expansion possibilities such as adding an extra netwerk card
  • No remote management