4K Overclocking Benchmarks Run: No fps Gain in Video Games Edit: May-5, 2017 I asked if anyone has any 4K Gaming benchmarks, testing if Overclocking has any effect at 4K Gaming, eventually I decided to test it myself and publish the results for others. To make my story short: There is absolutely no benefit Overclocking your CPU if you use your PC for Gaming in 4K or Gaming in less than 4K with 60Hz Monitor/TV. As you will see the single lower result is in CPU based Physics benchmarks, even API Overhead test has almost identical results. The reason I decided to do it is because I got tired of heat, its hot 9 out of 12 months over here so getting "free" heating is unnecessary, also I bet that my Overclocked PC with 1080ti, WC kit, 4 HDDS and 4 SSDs electricity costs are not cheap either and for nothing since I only game from time to time. My Next benchmark run will be testing if there is any Performance degradation under Windows 10 Pro when running with Hyper-Threading Enabled. Test System: Mobo: Asus Sabertooth X99 (Newest Bios:3505) RAM: 32GB 2666Mhz Enabled through XMP GPU: Gigabyte Aorus GTX1080 Ti Games and Benchmarks Installed to separate SSD Windows 10 Pro Creators Update, Game Mode Off, High Performance Power Plan. Before running the tests I used Display Driver Uninstaller v18.104.22.168 to clean up nvidia drivers and installed Driver v378.92 Display resolution and Benchmarks set at 3840x2160 GPU overclocked during both benchmark runs: Power Limit: 135%, Temp:90, Core: +60, Mem: +400, FAN: Locked @ 100% (Found out that keeping fan at 100% locked allows for cooler GPU and less/almost no throttling) During benchmarking the GPU was locked at 2000Mhz most of the time, sometimes doing 2025Mhz and sometimes 1985Mhz Intel 5820K @ 4.5Ghz, Hyper-Threading:OFF, CPU Cache @ 45, C-Staes Off, all Power Efficiency features Off in Bios. VS Intel 5820K @ Default Clocks, Hyper-Threading:OFF, CPU Cache @ 31, Processor Integrated VR Efficiency Mode: ON, C-States up to Including C6: ON, ASUS EPU Power Saving Mode: On, Enhanced Turbo Boost: ON (I enabled this to get proper all core boost while gaming.) Metro Last Light Redux Settings: 5820K @ 4.5Ghz = 64fps 5820K @ Default = 63fps Rise of the Tomb Raider DirectX 11 5820K @ 4.5Ghz = 63fps 5820K @ Default = 65fps DirectX 12 5820K @ 4.5Ghz = 62fps 5820K @ Default = 63fps Deus Ex: Mankind Divided DirectX 11 5820K @ 4.5Ghz = AVG:41, Min:33, Max:50 5820K @ Default = AVG:42, Min:33, Max:51 DirectX 12 5820K @ 4.5Ghz = AVG:41, Min:32, Max:50 5820K @ Default = AVG:41, Min:33, Max:50 Superposition Benchmark 5820K @ 4.5Ghz = Results: 9793, Avg:73, Min:57, Max:96 5820K @ Default = Results: 9959 , Avg:74, Min:59, Max:100 Fire Strike Ultra (Default 4K Test) 5820K @ 4.5Ghz = Graphics: 7253, Physics: 12788, Combined: 3943 Graphics Test 1: 41.99FPS, Graphics Test 2: 26.32FPS, Physics Test: 40.60FPS, Combined test: 18.34FPS 5820K @ Default = Graphics: 7579, Physics: 1840, Combined: 1955 Graphics Test 1: 42.91FPS , Graphics Test 2: 26.75FPS , Physics Test: 5.84FPS, Combined test: 9.10FPS API Overhead Feature Test (Default 4K Test) 5820K @ 4.5Ghz = DirectX 11 single-thread: 1421011, DirectX 11 multi-thread: 1314870, DirectX 12: 6592312, Vulkan: 25879380 5820K @ Default = DirectX 11 single-thread: 1344875, DirectX 11 multi-thread: 1128938, DirectX 12: 5713722, Vulkan: 24059458 Edit: Based on these Benchmarks from Techpowerup, even at FHD and 2K, Modern CPUs no matter its speed or amount of cores (if you above 4) has no effect on game FPS.