Originally Posted by ssateneth
All of these point to corruption of data in RAM. CPU issues would arise more likely with clock watchdog timeouts, WHEA errors or L1/L2 cache errors (found in hwinfo64), or in the case of unstable cache (though not always), system freezes.
Testing ram stability while maintaining high speeds can take a very long time. "stable enough" comes from situations where you only run short stress tests or stresses not good enough that leads to problems where your system is inexplicably crashing now where it wasn't before. Some people say "oh, i degraded my system somehow". No, it just wasn't stable to begin with. Unless you are shoving dangerous volts for long periods of time without appropriate cooling (I draw the line at 1.55 vcore, 1.7 vdimm, 1.35v sa and 1.35v io), I can't see any "degrading" happening unless the part was destined to die to begin with.
I prefer GSAT and HCI Memtest (At least 8 hours, sometimes 16 as even the most intermittent/rare error means a fail in my book. "stable enough" is not stable), and sometimes P95 with AVX and/or linpack stress tests if I want to test cooling (At least 8 hours). It'll take about 1.5 to 3 weeks to dial in a proper 24/7 overclock and requires you not using the PC at the same time.
Could go into the process of dialing in a stable overclock or troubleshooting the issues you are getting, but to do it properly requires a lot of detail. You'd more or less have to start at stock speeds / JEDEC non-XMP settings and go from there and change one thing at a time between each stress test. If you change a bunch of things and then you start getting errors, you won't know which setting exactly is causing the error and you're back to square one.
I never get any WHEA errors, none ever reported in HWinfo and also I am running this script which basically stops prime95 instantly when an error is found which also shows whether there are any WHEA errors in cmd console, never reported any errors though(will attach image).
I am suspecting that there might be something wrong with my motherboard because I have previously run memtest86 for 1 hour on each of the sticks and two combined together which resulted in no errors being found. Moreover, I just did 30minutes of prime95 stress test on each of the sticks installed in DIMMA2 and DIMMAB2 on separate occasions (ram stick 1 in dimma2 and then in dimmab2 and ram stick 2 in dimma2 and then in dimmab2). Also, did stress test both sticks for 20minutes in DIMMA2 and DIMMB2 just now. I would like to mention that I've done this before, at first it seems like both sticks work, but after a system restart, the system starts to crash again under load.
: After I installed both of the sticks into the motherboard, I ran prime95 for 1 whole hour with no crashes or errors, restarted my system and ran prime95 again and 2 minutes into the test, system froze
I noticed that when a single ram stick is installed and run under prime95, the CPU temperature sits at 66 and max being 69 whereas on the other side when both sticks are installed, the CPU temperature is 78 and at max being 79. Is that normal?
In terms of voltages applied to my CPU and RAM, I haven't ever applied more than 1.33 on CPU and 1.360 on RAM.
Also, there seems to be vdroop, at least yesterday(hasn't happened today as I set LLC to auto), I ran prime95 for around 40minutes and as soon as I stopped the workers, my system froze. Are there any ways to reduce it apart from LLC?
Originally Posted by BroadPwns
The problem that does not concern the OP and did concern people with heavy power hungry gpus, I believe SeaSonic already fixed that with new batch of PSUs.
1. Restoring factory default bios settings; if that fail
2. Formatting your system disk; if that fail
3. Changing motherboard
I have formatted my system disk before, did a clean install. I would love to try changing motherboard but I do not have any for my socket.