This is a follow up from a post I made 8 months ago: Why ATG’s Core Based Licensing is Stupid
With the new Westmere hex-core CPU’s out now, the problem has gotten worse. A mid-high or high end Westmere CPU presents as 12 cores. So what does this really mean?
I just ran the numbers, and basically a mid-high end single CPU server in 2008 (Xeon 5450) would cost me 4 ATG cores worth of licensing, and would handle X amount of traffic.
A mid-high end single CPU server in 2010 (Westmere 5650) would cost me 12 ATG cores worth of licensing, and will only handle X+35 to 70% traffic (based on published SPECint, SPECint_rate, and SPECfp scores for the CPUs).
So it’s a 300% increase in costs to handle 35 to 70% more traffic. Or just to provision with modern hardware. That’s crazy.