Overclock.net banner
121 - 140 of 303 Posts
@fpsflow

Can I just check a few things with you?

From my understanding of your previous explanations, is my thinking correct here:

Default total TDC is the RED CORE + ORANGE SOC + YELLOW USR ? (see screenshot labels)
This for the 552W Aqua Extreme bios would be 448A + 82A + 86A = 616A

My card is the 7900XTX Sapphire Pulse, same PCB as the Nitro+ for anyone following at home (just single bios vs dual bios).
According to TechPowerUp, that's 17 Phases at 70A, so 1190A for this card.

@fpsflow So I divide them to figure out my scale? 1190 / 616 = 1.93

I was using a 1.5x scaler, but then tried 2.0x. Now thinking it doesn't need to be anywhere near that high, 1190A is crazy.

How do I calculate the likely Wattage being drawn when making these adjustments?
(I have a wall power monitor coming from Amazon today anyway!!!)

Also trying to understand how to best use the GFX_VID offset?
  • Wondering what people think is a "safe" for various cards. 7900XTX uses 1150mV as default, is 1200mV pretty tame? How about 1250mV? I can't find any discussion online, even with people using EVC2.
  • If I apply GFX_VID +50mV with the script, does that mean 1150mV set through Adrenaline is allowing 1200mV at the GPU?
  • If so, do offsets in Adrenaline (or MCT) also have the script induced offset applied, so 1100mV in Adrenaline is 1150mV actual?

I'm trying to figure out what this does for the V/F curve and how to play this for maximum core clock when both power delivery and temps are not a problem in my case (Corsair AX1200i, extreme watercooling).

VRAM_VID at +145mV got me about 100Mhz with Fast Timings and seemingly stable. Higher seems to regress performance and/or introduce texture artifacts. Temps are still great.


Image


Below is a couple of HWINFO logs compared. In GREEN is an old run with my 550W bios set to 484W (to emulate a Nitro+), vs in RED my run last night with Scaler at 2.0x, GFX_VID +50mv and VRAM_VID +145mV. Adrenaline (MCT) was set to 115% Power Limit and 1090mV.

HWINFO output was not recalibrated (no adjustments made for 2x scaler).

Not sure what this is telling me exactly, but I'm sure @fpsflow will know :) Do I double the Amps shown here for the RED run? Again, how might I guess what the TBP was actually?

Image
 
Discussion starter · #123 · (Edited)
Default total TDC is the RED CORE + ORANGE SOC + YELLOW USR ? (see screenshot labels)
No, on RDNA3 there's a TDC limit and a TGP (Total Graphics Power) limit. TDC_GFX, in my 7800XT case it was around 220A if I remember right. And then there's the TGP limit at 280W I think.
TGP is made up out of powers of all 5 rails, VDDCR_GFX+VDDCR_SOC+VDDCR_URS+VDDCI_MEM+VDDIO.
In your case TDC limit is 448A on GFX. And TGP limit is 460W. So when you modify the current gain value you end up affecting both these limits at the same time, TDC and TGP.

With a max of 1190A for GFX current, and a TDC limit of 448A, you'd divide 1190A by 448A to get the scaling factor. So ~x2.65. This makes it so your new TDC_GFX limit goes from 448A to 1190A. This is the maximum possible power, considering your VRM's limit is 1190A. You move the software limit up to hardware limit using x2.65 value. This is the "balls out" value.

Now, when you modify TDC_GFX, that value will be wrongly reported by HWinfo, along with TGP. So ideally you use the same scaling factor for the other 4 rails (I think the script does this on all 5 rails), and then you multiply the HWinfo values for all 5 rails and for TGP with the same value you used in the script, to get back the real current/power values. Same goes for the HWinfo TDC limit sensors.
If I didn't make any mistake, you'd have to multiply by same factor in the script all these HWinfo sensors, to get back real values reporting:

Number Screenshot

Some of the rails are missing from various categories, but apply the multiplying factor to current sensors, current limits, each 5 rails power reporting + TGP (since it's made up out of all 5 of them), and GPU PPT should be TGP I think (seems close enough).
HWinfo has an option in Customize tab for each sensor to add a multiplier value, that's where you add the one you used in the script (x1.5 or 2.0 or whatever). This way you get real HWinfo data.


Regarding VID offsets, I didn't go over 1.495V on VRAM chips, seemed to risky (and fast timings would crash at high clocks, never used it).
You could experiment with the rest of small rails, I ended up using 20-25mV extra on VDDCI_MEM, about 10mV less on VDDCR_SOC and can't remember on USR rail but I tweaked that as well, 10mV up or down seemed to worked better in my case.

For GFX VID that depends on the load, but my general experience has been that VID offset gets applied on the standard curve. This seems to be different on RDNA4, but on RDNA3 I would always get the offset for the same running condtions. Higher/lower voltage and more/less heat.
With extra GFX voltage you can sustain higher clocks, that's the short of it. I went as high as +200mV on GFX, on my 7800XT, but compensated with heavy UV (since I was on stock cooling). Remember that VID always gets applied on RDNA3, UV isn't, that depends on running conditions and algorithm.
So you will peak at stock_voltage+VID, 1150mV+200mV=1350mV even without load and even with UV applied. Under load, UV will remove some voltage, but under light load/idle you will peak at max voltage, so take care with how much +VID you use. I don't know what is the max the die can take, VR limit is set at 1.5V or 1.56V max. I didn't have the guts to get anywhere near that, but I peaked at 1.35V few times.
Using +VID and -UV was a trick that allowed me to get some extra clocks in some cases, seems like a nice trick on RDNA3. Since VID/UV behave slightly differently you might be able to get something out of this, in some situations.
 
@fpsflow Thanks for the additional explanation.

With a max of 1190A for GFX current, and a TDC limit of 448A, you'd divide 1190A by 448A to get the scaling factor. So ~x2.65. This makes it so your new TDC_GFX limit goes from 448A to 1190A. This is the maximum possible power, considering your VRM's limit is 1190A. You move the software limit up to hardware limit using x2.65 value. This is the "balls out" value.
So was doing a little benchmarking with a x1.90 scaling factor (before I read you post just now), and it was scaling nicely with steady performance increases in Steel Nomad.

All of a sudden when I pushed the core clock one more bracket up (3300-3400 to 3350-3450) the PSU protection tripped as soon as the load came on (Corsair AX1200i). Still unsure if it was OCP or OVP.
I reset power and tried a few more times (because YOLO) and could replicate the issue regularly. Dropping clocks back down would pass the benchmark without shutting the PSU down.

I realised this was clearly this was pulling a LOT of power to trip a 1200W PSU. Though my temps had been very undercontrol due watercooling so I didn't realise HOW MUCH power it was pulling until later.

Remembering that the AX1200i can output sensor data, I managed to find my Corsair iLink dongle in a drawer and hooked the PSU up through a USB header, then regrettably installed Corsair iCUE and immediately disabled all services except the one needed to talk to HWiNFO 64.

Although the polling rate on the PSU is terrible, at least HWiNFO is now able to see and log some key PSU metrics.... time to experiment.

Here is a Furmark 2 running the 2160p benchmark.

First run in GREEN is the 7900XTX running with no unlocked power through the script, but it does have OC settings applied in MCT. It's a locked 550W as expected.

The second run in RED is the 7900XTX running with a 1.5x scaling factor from the script (no VID adjustments). This is showing 550W too... but have a look at the rest of the power draw!!!! HOLY CRAP!!!!

Click image for full size!



Yes, that is an average power draw increase of 389W from the wall to the PSU and 364W from the PSU to PC (93.5% efficiency, not bad) and then presumably all that extra power into the GPU which was previously already pulling 550W.

The PSU 12V current readout also showed an increase of 27.55A, or ~330W.

Remember, the only thing that changed here was using the script to set a 1.5x power scaling factor.

Umm. 900W+ 7900XTX anyone?

Below you can see the aftermath in HWiNFO. I intentionally disabled monitoring of a lot of the PSU sensors because they poll so slow - stupid Corsair software interfaces not playing nice with others.
The HWiNFO sensors for all 5 rails and for TGP have been scaled by 1.5x to match the script scaling.

Image


You can see my overkill watercooling (MO-RA 420 with average water temp 16-20C) managed to keep the hotspot under 100C, so actually the card didn't throttle at all.
It actually saw a 17.8% performance increase in this basic Furmark (2) 4K benchmark, though I never really know why Furmark clock speeds are so low compared to normal titles.

@fpsflow Many questions here. :ROFLMAO:
  • How did 1.5x scaling result in such a HUGE actually power increase?
  • Have I just mathed wrong here?
  • If I had a 1500W PSU I assume this GPU would keep pulling power over 1000W until it thermal throttled?
  • What to try next? ;)
I guess I could repeat this experiment at 1.1x, 1.2x, 1.3x, 1.4x etc just to see what the measured PSU power increase is each time.

Summary: Those running the 550W Aqua 7900XTX bios - be careful with this (awesome) script !!!
 
Discussion starter · #125 · (Edited)
@fpsflow Thanks for the additional explanation.



So was doing a little benchmarking with a x1.90 scaling factor (before I read you post just now), and it was scaling nicely with steady performance increases in Steel Nomad.

All of a sudden when I pushed the core clock one more bracket up (3300-3400 to 3350-3450) the PSU protection tripped as soon as the load came on (Corsair AX1200i). Still unsure if it was OCP or OVP.
I reset power and tried a few more times (because YOLO) and could replicate the issue regularly. Dropping clocks back down would pass the benchmark without shutting the PSU down.

I realised this was clearly this was pulling a LOT of power to trip a 1200W PSU. Though my temps had been very undercontrol due watercooling so I didn't realise HOW MUCH power it was pulling until later.

Remembering that the AX1200i can output sensor data, I managed to find my Corsair iLink dongle in a drawer and hooked the PSU up through a USB header, then regrettably installed Corsair iCUE and immediately disabled all services except the one needed to talk to HWiNFO 64.

Although the polling rate on the PSU is terrible, at least HWiNFO is now able to see and log some key PSU metrics.... time to experiment.

Here is a Furmark 2 running the 2160p benchmark.

First run in GREEN is the 7900XTX running with no unlocked power through the script, but it does have OC settings applied in MCT. It's a locked 550W as expected.

The second run in RED is the 7900XTX running with a 1.5x scaling factor from the script (no VID adjustments). This is showing 550W too... but have a look at the rest of the power draw!!!! HOLY CRAP!!!!

Click image for full size!



Yes, that is an average power draw increase of 389W from the wall to the PSU and 364W from the PSU to PC (93.5% efficiency, not bad) and then presumably all that extra power into the GPU which was previously already pulling 550W.

The PSU 12V current readout also showed an increase of 27.55A, or ~330W.

Remember, the only thing that changed here was using the script to set a 1.5x power scaling factor.

Umm. 900W+ 7900XTX anyone?

Below you can see the aftermath in HWiNFO. I intentionally disabled monitoring of a lot of the PSU sensors because they poll so slow - stupid Corsair software interfaces not playing nice with others.
The HWiNFO sensors for all 5 rails and for TGP have been scaled by 1.5x to match the script scaling.

Image


You can see my overkill watercooling (MO-RA 420 with average water temp 16-20C) managed to keep the hotspot under 100C, so actually the card didn't throttle at all.
It actually saw a 17.8% performance increase in this basic Furmark (2) 4K benchmark, though I never really know why Furmark clock speeds are so low compared to normal titles.

@fpsflow Many questions here. :ROFLMAO:
  • How did 1.5x scaling result in such a HUGE actually power increase?
  • Have I just mathed wrong here?
  • If I had a 1500W PSU I assume this GPU would keep pulling power over 1000W until it thermal throttled?
  • What to try next? ;)
I guess I could repeat this experiment at 1.1x, 1.2x, 1.3x, 1.4x etc just to see what the measured PSU power increase is each time.

Summary: Those running the 550W Aqua 7900XTX bios - be careful with this (awesome) script !!!
Using x1.5 means a max power limit of 825W. But remember, this is Pout (out of VRMs) not Pin (at connector). Means it's without any other losses. Add the VRM losses to get power at GPU connector, then add the rest of PC power draw on top of that (CPU/RAM/chipset/SSDs/water pump/fans etc), and then add mains to 12V losses on top of that and you get whole power supply load.
Using x1.9 factor would increase TGP (total VRMs output for all 5 rails) limit to 1045W, makes sense it would trip your protections. You're over 1100W-1150W at GPU connectors, not considering the rest of PC + efficiency losses in power supply.

TBP (Total Board Power) should reflect the power draw at connector, but that is an estimated value, it's TGP + some onboard smaller rails/fans/RGB + efficiency losses. Using the script messes with both TGP and TBP, you can correct TGP but you cannot correct TBP, since it's comprised out of wrongly reported TGP + other stuff. So the multiplier value doesn't work on it. But you can subtract uncorrected TGP from TBP, that difference should be "accurate" to whatever degree it is stock (edit: or not, if it's a percent of TGP. depends on how it's calculated internally), and add that difference on top of corrected TGP, to get a sense of TBP.
 
Umm. 900W+ 7900XTX anyone?
I thought it was common knowledge that an unshackled Navi 31 would easily be the most power-hungry GPU ever made on ambient cooling?
 
  • Rep+
Reactions: NattyKathy
Using x1.5 means a max power limit of 825W. But remember, this is Pout (out of VRMs) not Pin (at connector). Means it's without any other losses. Add the VRM losses to get power at GPU connector, then add the rest of PC power draw on top of that (CPU/RAM/chipset/SSDs/water pump/fans etc), and then add mains to 12V losses on top of that and you get whole power supply load.
Using x1.9 factor would increase TGP (total VRMs output for all 5 rails) limit to 1045W, makes sense it would trip your protections. You're over 1100W-1150W at GPU connectors, not considering the rest of PC + efficiency losses in power supply.

TBP (Total Board Power) should reflect the power draw at connector, but that is an estimated value, it's TGP + some onboard smaller rails/fans/RGB + efficiency losses. Using the script messes with both TGP and TBP, you can correct TGP but you cannot correct TBP, since it's comprised out of wrongly reported TGP + other stuff. So the multiplier value doesn't work on it. But you can subtract uncorrected TGP from TBP, that difference should be "accurate" to whatever degree it is stock (edit: or not, if it's a percent of TGP. depends on how it's calculated internally), and add that difference on top of corrected TGP, to get a sense of TBP.
Thanks again, that's helpful to better understand the relationship here.

What I hadn't appreciated earlier was that the scaling factor applied directly to TGP, I guess that makes sense when I think about it, but I was focused on the core TDC relationship vs phases/rating of the MOSFETs and didn't think or realise that TGP (which was already at 460 with my bios at 115% PL) would also directly scale.

I also didn't know TGP was measured but TBP was an estimate. I've always used TBP as a reference but will start ignoring that and referencing TGP going forward.

The good news is I think I can keep it at 1.5 now and just use PL between 100% and 115% to regulate TDC and therefore TGP.

I thought it was common knowledge that an unshackled Navi 31 would easily be the most power-hungry GPU ever made on ambient cooling?
I mean that's what I was hoping for, hadn't counted on it to being this easy to reach crazy XOC levels !! :ROFLMAO:
 
Discussion starter · #129 · (Edited)
What I hadn't appreciated earlier was that the scaling factor applied directly to TGP, I guess that makes sense when I think about it, but I was focused on the core TDC relationship vs phases/rating of the MOSFETs and didn't think or realise that TGP (which was already at 460 with my bios at 115% PL) would also directly scale.
Yeah, GFX TDC affects GFX power which is included in TGP.
RDNA4 measures TBP at connector, and applies the limit on this value. RDNA3 measures TGP and applies the power limit on it, and only estimates TBP.
 
Outstanding work. Got two 7900 GRE coming. Hopefully the Samsung idle power issue causing black screens won't be a problem.

I had an asrock taichi xtx that did vram non fast timing 2840 (samsung) without any mods.
 
  • Rep+
Reactions: fpsflow
感谢您的分享。我只在AMD Software里+10%功耗,使我的9070XT的TBP limit保持在334W。然后将GFX电压+60mv,soc电压-10mv,VRAM-65mv 达到了timespy最高显卡分数32449。我注意到在linux里增加GFX电压的结果貌似和在AMD Software里减少核心电压的效果一样。原本想要通过测试的话,我在AMD Software里最多只能降低60mv的核心电压,但是现在额外在linux里+60mv GFX后依旧能通过测试。
 
感谢您的分享。我只在AMD Software里+10%功耗,使我的9070XT的TBP limit保持在334W。然后将GFX电压+60mv,soc电压-10mv,VRAM-65mv 达到了timespy最高显卡分数32449。我注意到在linux里增加GFX电压的结果貌似和在AMD Software里减少核心电压的效果一样。原本想要通过测试的话,我在AMD Software里最多只能降低60mv的核心电压,但是现在额外在linux里+60mv GFX后依旧能通过测试。
我的显存颗粒品牌为Hynix,我认为1.35V的电压似乎有点过于冗余了,所以我将VRAM电压减少了65mv,它现在工作在1.285v,这使得在334W的TPB limit下,核心能够分到更多的功耗,对于提高核心的boost频率很有帮助
 
Discussion starter · #136 · (Edited)
我的显存颗粒品牌为Hynix,我认为1.35V的电压似乎有点过于冗余了,所以我将VRAM电压减少了65mv,它现在工作在1.285v,这使得在334W的TPB limit下,核心能够分到更多的功耗,对于提高核心的boost频率很有帮助
VRAM current use is around 20-25A. At 65mV less, you save about 1.6W. I don't think that matters for anything, not even temperatures. That 1.6W is spread across 8 memory chips, so it means 0.2W less on each memory chip. But lower clocks may make for much lower current, haven't tested this. Still, you're looking at 5-10W saved, at most, that's my best guess.
 
VRAM current use is around 20-25A. At 65mV less, you save about 1.6W. I don't think that matters for anything, not even temperatures. That 1.6W is spread across 8 memory chips, so it means 0.2W less on each memory chip. But lower clocks may make for much lower current, haven't tested this. Still, you're looking at 5-10W saved, at most, that's my best guess.
将显存电压由1.35v减少到1.285v后,显卡分数由32400提升到了32500,确实只有一点点作用。
 
121 - 140 of 303 Posts