I knew it wasn't my card holding things back already because of using MSI afterburner to show GPU and CPU usage stats but I wasn't sure just how much a new CPU and RAM would help.
I previously had an i5-3550 and 8G of DDR3 1600c9(xmp profile). I now have an R5-3600 and 16G of DDR4 3200c16. Card is about the same as a 10603G.
Things I can notice that are different:
I no longer microfreeze and drop ~10fps whenever an in-game overlay message pops up on screen.
Everything just feels MUCH smoother and movement is more fluid even in games where I never went under 60 fps before.
I use the same Sata SSD that I had before but it is loading much faster now at all points of all the games that I was struggling to hold 60fps at before.
The games are now using about 1G more of system ram than before. Even though every game I own uses under 5gb apart from Assassins Creed:Origins. I suppose this might have to do with windows and its reserving of ram.
The speaker sound is a huge improvement and no longer sounds as tinny. Not as good as the USB headphones but nice anyway.
What's not different:
-Ultra/High shadows, Reflections, Particle effects and Volumetric fog are what tanks my framerate and that is GPU. I can generally go Ultra textures/detail and geometry as well as draw distance but this has always been the case. Luckily, I think Ultra shadows look unrealistic and the fog just makes things hard to see for little benefit. I'm good with all this.
- Life tends to throw the unexpected at you so the decision might be out of my hands anyway.
PS: the fucking memory is RGB(low stock so I had no choice) and since I'm using a stripped down beta bios for compatibility, I don't think I can stop the Las Vegas light show for now.
Don't get me started on the cryengine conetracing. I have deep intimate knowledge of that codebase and holy shit it's bad. I rewrote parts of it to try and remove the hacky parts, then handed it off to a co-worker who rewrote most of it, and in non-demo scenarios it is just plain terrible