Not 30x, more like 10-15x and windows vmsThat aside, does anyone else have a server with 30x RAM utilization v. CPU utilization?
Next year.I thought the real use for this stuff was supposed to be a layer between DRAM and NAND ssds that can be addressed as DRAM.
How convenient, my EXACT home lab/cluster config ratio per ESXi host (8 cores, 128GB memory, UP systems).Normal for general virtualization, unless you have some specific CPU heavy stuff (and we usually configure those hosts differently).
Quick rule for me is 16gb or more per core is a good ratio.
this is usually the case on our database vms. when we add more ram they can load more data to process and if the vcpu isn't sized appropriately with the increase in ram, we get that imbalance. that's why i always push them to test the resource increases with their codes on the UAT/load-test server before making any increase.@Patrick, did you ever get a definite answer on this?
Normally people will say to add more memory when suffering from performance problems. While that is usually good advice, I have been that they have seen times times when adding more memory without adding correpondingly more CPU will make the problem worse. I don't remember the details of the explanation unfortunately.
Does anybody know of a serious explanation of the difference adding more CPU vs more memory, and whether this is different according to hypervisor?