Question for the devs or knowledgeable users

I have been doing in depth benchmarking for various detail, effects and distance settings. My system specs: Ryzen 1600 3.8ghz, 16gb ram, 250gb OS ssd, 500gb games ssd, gtx 1080. I am finding that as you increase the quality of settings, GPU and CPU utilization decreases. Another thing I am seeing is that the game will not utilize more than 8gb system ram, 5gb video ram but will use 15gb of pagefile. I know the “ultra” settings are for “future hardware” yet this game will not fully utilize the vram of a GPU and prefers to stream textures from the pagefile. How can the devs justify the “future hardware” settings when the high/very high quality settings can’t even utilize current hardware properly? This game is terribly optimized, textures pop in rather than fade in/out and the engine bottle necks performance with texture streaming from the ssd.

What am I doing wrong? Which settings effect CPU, which ones GPU? How much can mods improve optimization?

1 Like

Here is an example of what I am experiencing.
Settings:
Resolution 1440p
borderless
HD Textures off
antialias off
Object quality Ultra
game effects low
lighting low
particles low
physics low
post process low
shader quality low
shadows ultra
textures ultra
water detail ultra
volumetric lighting low
vegetation detail ultra
object distance max
LOD distance max
vegetation distance max
motion blur off

If anyone would like, I can upload many more benchmark screen caps.