In Windows 7/Server 2008 and to a similar extent Windows 8/Server 2012, the kernel's internal time is set to an interval of 15.6ms. (Use clockres to confirm current 'tick rate' on Windows). Win8/Server2012 have a tickless kernel w/timer coalescing (info here) and the platform shows similar performance characteristics with C* to Windows 7 with a slight edge in performance to win8/server 2012 in my testing (the testing and results of which are outside the scope of this ticket).
Some arguments against lowering the system's internal timer to 1ms are here. These seem largely constrained to "it'll drain your battery" and "it'll prevent your processor from being as effective in sleep states". The 2nd is somewhat of a concern as we don't want Windows users to all of a sudden have increased CPU-usage bills from virtualized environments. In the comments, one individual mentions a VirtualBox VM spinning at 10-20% cpu just from changing that flag alone which seems mathematically unlikely, but is worth keeping an eye on and testing.
A Microsoft publication that largely reinforces the cautionary tale on power consumption can be found here.
With the cautionary tales on our radar, the impact on throughput and latency on the 2.2 branch on Windows is fairly dramatic. A couple of caveats on these #'s: I'm not completely saturating the system as the thread count is relatively low (keeping it consistent with other testing where it was saturating), and the read #'s from our 2012 test environment are not affected by this timer change while I see it on 3 other bare-metal installations. The testing environment is new and we haven't worked out the kinks yet, however the write / mixed illustrate the throughput and latency #'s I've mentioned above; for reads the cpu's are sitting idle at 1-5% used by stress and C* so something else clearly needs to be addressed there; I included them for completeness sake.
Some preliminary testing on OpenStack indicates kernel-space syscall saturation w/this patch that actually degrades performance, however the unpatched performance numbers in our OpenStack environment are low enough that I question their validity.
Opening this ticket w/attached branch to get it on the radar / conversation going, and I'm going to update this from being hard-coded to being a tunable in the .yaml.
Initial patch available here.