This is somewhat of a longshot, but I am wondering if anyone might be able to explain the following:
I have a VMware Host Server running Ubuntu Server 8.04 LTS, and VMware Server 2.0. The User CPU usage on the host keeps climbing, until I reboot the server when it goes back down to almost nothing - and starts to climb again. This is ever since I swapped a Server 2003 VM for a Server 2008 one (upgraded my DC, and migrated to 2008). I can't find any problem with the Windows server though, or any abnormally high CPU usage on the VM itself.
I have been keeping up-to-date on the host, so I've been through about 3 kernel upgrades, numerous recompiles of VMware, and a whole new VMware Server version when the last one was released not too long ago. I just can't figure this one out.
Any suggestions would be very well appreciated, I'm now just looking for things to try!
I can't answer the question, but I can add an extra little bit of anecdotal evidence. I have noticed this too, with a Linux hosts (Debian/Etch and Debian/Lenny).
After some mucking around when I first noticed the issue, I came to the conclusion that the issue is VMWare, not the guests themselves. Upon stopping all services in the VMs on a particular machine the CPU use remained high despite the OSs in the VMs doing nothing. On shutting down each of the four VMs the excessive host CPU use dropped by about 25% per VM (I didn't measure this by any scientific means, but certainly no one VM seems to be imposing most of the load). After restrting the VMs the CPU use remained where it used to be, even with the services in the VMs active, and the load has begun to slowly rise over time with no related rise in apparent useful activity.
In both the cases that I have noticed this happen, the host OS has been 32-bit Linux and the guest OSs have also been 32 bit Linux.
I have not seen it in all cases though. On my home server (64 bit Linux kernel with 32-bit user-land running one large and two small 32-bit Linux VMS and occasionally Windows VMS for testing) and the main dev/test host at work (64-bit Linux (both kernel and userland) host running mainly Windows VMs, some 32-bit and some 64) this aberrant behaviour doesn't seem to be present. All the above are running VMWare Server 2.
So to cut a long story short: it isn't just you, and it isn't just Windows based guests, but it doesn't seem to be a consistent problem (as many arrangements don't see it as do, in my limited experience). Though unfortunately I can't help any further than that as I've not had the time to look into the problem in more detail.
First, I've seen this with VMware Server, too - on both Windows and Linux. In my experience it was related to running both 64-bit and 32-bit VMs at the same time.
Though it may not be an option for you, I'd suggest ESXi - the lightweight, free edition of ESX.
I experience the same issue on a production server. Host OS is Debian Linux 64 bit. Guests are 4 Linux machines and 1 Windows XP. All guests are 32 bit. When i restart all VM services on the host, the CPU scheduling is fine and host CPU per VM is on par with real guest activity. However after some weeks the CPU scheduling increases until more or less the maximum available CPU time is consumed over the running VMs. The CPU usage on the host at that time is more or less 10x the initial CPU usage of the moment where the VMs just got started.
To me it seems a leak in the CPU scheduling on the host. Instaed of leaking RAM, it's leaking CPU cycles :-). How am I sure this doesnt happen on ESXi?
Thomasgg