You are not logged in.

#1 2025-06-20 07:10:17

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Random GPU lag

So I've been noticing random lag spikes, specifically related with my nvidia dedicated GPU.
I've tried to monitor what's going on with it by running:

nvidia-smi -l 1

It goes like this for a bit of time:

Fri Jun 20 12:05:23 2025       
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 570.153.02             Driver Version: 570.153.02     CUDA Version: 12.8     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 2050        Off |   00000000:01:00.0 Off |                  N/A |
| N/A   51C    P8              2W /   55W |      15MiB /   4096MiB |      0%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+
                                                                                         
+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI              PID   Type   Process name                        GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|    0   N/A  N/A            1196      G   /usr/lib/Xorg                             4MiB |
+-----------------------------------------------------------------------------------------+

And then at the exact time a lagspike happens I get this error:

Unexpected NVML event
Error occurred while processing the event: Unknown Error

I have no clue what is causing this, but it has been getting progressively worse over time.
Sometimes it just doesn't happen at all for days. My power profile is set to Performance.
Here are more details about my system:

theairblow@theairblow
---------------------
OS: Arch Linux x86_64
Host: ASUS TUF Gaming A17 FA706NF_FA706NF (1.0)
Kernel: Linux 6.14.7-arch2-1
Uptime: 25 mins
Packages: 2048 (pacman)
Shell: zsh 5.9
Display (B173HAN04.9 ): 1920x1080 @ 144 Hz in 17"
DE: KDE Plasma 6.4.0
WM: KWin (X11)
WM Theme: Breeze
Theme: Breeze (Dark) [Qt], Breeze-Dark [GTK2], Breeze [GTK3]
Icons: Papirus-Dark [Qt], Papirus-Dark [GTK2/3/4]
Font: Noto Sans (10pt) [Qt], Noto Sans (10pt) [GTK2/3/4]
Cursor: breeze (24px)
Terminal: konsole 25.4.2
CPU: AMD Ryzen 5 7535HS (12) @ 4.60 GHz
GPU 1: NVIDIA GeForce RTX 2050 [Discrete]
GPU 2: AMD Radeon 680M [Integrated]
Memory: 5.56 GiB / 14.38 GiB (39%)
Swap: 5.67 MiB / 15.98 GiB (0%)
Disk (/): 420.99 GiB / 502.88 GiB (84%) - ext4
Disk (/mnt/storage): 231.34 GiB / 468.38 GiB (49%) - ext4
Local IP (wlan0): 192.168.0.187/24
Battery (A32-K55): 100% [AC Connected]
Locale: en_US.UTF-8

Offline

#2 2025-06-20 08:10:32

d_fajardo
Member
Registered: 2017-07-28
Posts: 1,687

Re: Random GPU lag

Which NVIDIA package did you install? The GPU fan is not registering in your smi. My hunch is GPU memory issue.
You can test your GPU memory:
https://aur.archlinux.org/packages/cuda_memtest

Offline

#3 2025-06-20 08:22:20

seth
Member
From: Won't reply 2 private help req
Registered: 2012-09-03
Posts: 75,081

Re: Random GPU lag

Battery (A32-K55): 100% [AC Connected]

Charger events?

If you drain the battery a bit, does the next "unexpected event" hit around when it's hitting 100% again?

Online

#4 2025-06-20 08:37:47

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

d_fajardo wrote:

Which NVIDIA package did you install? The GPU fan is not registering in your smi. My hunch is GPU memory issue.
You can test your GPU memory:
https://aur.archlinux.org/packages/cuda_memtest

I'm using nvidia-open-dkms, my GPU is from the Ampere family according to Wikipedia (which is newer than Turing)

cuda_memtest fails to compile due to missing nvcc. I tried to add /opt/cuda/bin/ into PATH, but it just resulted in an even weirder error

  #$ cudafe++ --c++17 --gnu_version=150101 --display_error_number
  --orig_src_file_name "CMakeCUDACompilerId.cu" --orig_src_path_name
  "/home/theairblow/.cache/yay/cuda_memtest/src/cuda_memtest/build/CMakeFiles/4.0.3-dirty/CompilerIdCUDA/CMakeCUDACompilerId.cu"
  --allow_managed --m64 --parse_templates --gen_c_file_name
  "tmp/CMakeCUDACompilerId.cudafe1.cpp" --stub_file_name
  "CMakeCUDACompilerId.cudafe1.stub.c" --gen_module_id_file
  --module_id_file_name "tmp/CMakeCUDACompilerId.module_id"
  "tmp/CMakeCUDACompilerId.cpp4.ii"

  /usr/include/c++/15.1.1/type_traits(554): error: type name is not allowed
        : public __bool_constant<__is_pointer(_Tp)>
                                              ^
  /usr/include/c++/15.1.1/type_traits(554): error: identifier "__is_pointer"
  is undefined
        : public __bool_constant<__is_pointer(_Tp)>
                                 ^
  /usr/include/c++/15.1.1/type_traits(876): error: type name is not allowed
        : public __bool_constant<__is_volatile(_Tp)>
                                               ^
  /usr/include/c++/15.1.1/type_traits(876): error: identifier "__is_volatile"
  is undefined
        : public __bool_constant<__is_volatile(_Tp)>

... and so on

P.S. Found the /etc/profile.d/cuda.sh script, trying it again. Look like I'm supposed to log out first.

Last edited by TheAirBlow (2025-06-20 08:40:59)

Offline

#5 2025-06-20 08:41:51

d_fajardo
Member
Registered: 2017-07-28
Posts: 1,687

Re: Random GPU lag

Is this the stock kernel? If it is why are you using dkms?

cuda_memtest fails to compile due to missing nvcc

You might need to install cuda.

Offline

#6 2025-06-20 08:43:49

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

d_fajardo wrote:

Which NVIDIA package did you install? The GPU fan is not registering in your smi. My hunch is GPU memory issue.
You can test your GPU memory:
https://aur.archlinux.org/packages/cuda_memtest

$ cuda_memtest
[06/20/2025 13:41:09][theairblow][0]:Running cuda memtest, version 1.2.3
[06/20/2025 13:41:09][theairblow][0]:NVRM version: NVIDIA UNIX Open Kernel Module for x86_64  575.64  Release Build  (root@theairblow)  
[06/20/2025 13:41:09][theairblow][0]:num_gpus=1
[06/20/2025 13:41:09][theairblow][0]:Device name=NVIDIA GeForce RTX 2050, global memory size=3951624192, serial=unknown (NVML runtime error)
[06/20/2025 13:41:09][theairblow][0]:Attached to device 0 successfully.
[06/20/2025 13:41:09][theairblow][0]:WARNING: driver reported at least 3854565376 bytes are free but largest possible allocation is 3850371072 bytes.
[06/20/2025 13:41:09][theairblow][0]:Allocated 3672 MB
[06/20/2025 13:41:09][theairblow][0]:Test0 [Walking 1 bit]
[06/20/2025 13:41:10][theairblow][0]:Test0 finished in 1.4 seconds
[06/20/2025 13:41:10][theairblow][0]:Test1 [Own address test]
[06/20/2025 13:41:11][theairblow][0]:Test1 finished in 0.4 seconds
[06/20/2025 13:41:11][theairblow][0]:Test2 [Moving inversions, ones&zeros]
[06/20/2025 13:41:13][theairblow][0]:Test2 finished in 2.1 seconds
[06/20/2025 13:41:13][theairblow][0]:Test3 [Moving inversions, 8 bit pat]
[06/20/2025 13:41:15][theairblow][0]:Test3 finished in 2.1 seconds
[06/20/2025 13:41:15][theairblow][0]:Test4 [Moving inversions, random pattern]
[06/20/2025 13:41:16][theairblow][0]:Test4 finished in 1.1 seconds
[06/20/2025 13:41:16][theairblow][0]:Test5 [Block move, 64 moves]
[06/20/2025 13:41:17][theairblow][0]:Test5 finished in 0.6 seconds
[06/20/2025 13:41:17][theairblow][0]:Test6 [Moving inversions, 32 bit pat]
[06/20/2025 13:42:31][theairblow][0]:Test6 finished in 74.6 seconds
[06/20/2025 13:42:31][theairblow][0]:Test7 [Random number sequence]
[06/20/2025 13:42:33][theairblow][0]:Test7 finished in 1.7 seconds
[06/20/2025 13:42:33][theairblow][0]:Test8 [Modulo 20, random pattern]
[06/20/2025 13:42:33][theairblow][0]:test8[mod test]: p1=0x144be586, p2=0xebb41a79
[06/20/2025 13:42:38][theairblow][0]:Test8 finished in 5.5 seconds

$ cuda_memtest --stress --num_passes 1 --num_iterations 100
[06/20/2025 13:43:02][theairblow][0]:Running cuda memtest, version 1.2.3
[06/20/2025 13:43:02][theairblow][0]:NVRM version: NVIDIA UNIX Open Kernel Module for x86_64  575.64  Release Build  (root@theairblow)  
[06/20/2025 13:43:02][theairblow][0]:num_gpus=1
[06/20/2025 13:43:02][theairblow][0]:Device name=NVIDIA GeForce RTX 2050, global memory size=3951624192, serial=unknown (NVML runtime error)
[06/20/2025 13:43:02][theairblow][0]:Attached to device 0 successfully.
[06/20/2025 13:43:02][theairblow][0]:WARNING: driver reported at least 3854565376 bytes are free but largest possible allocation is 3850371072 bytes.
[06/20/2025 13:43:02][theairblow][0]:Allocated 3672 MB
[06/20/2025 13:43:02][theairblow][0]:Test10 [Memory stress test]
[06/20/2025 13:43:02][theairblow][0]:Test10 with pattern=0x4bbd97c54fc61f62
[06/20/2025 13:43:10][theairblow][0]:Test10 finished in 7.4 seconds

No errors were reported.

Offline

#7 2025-06-20 08:45:07

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

d_fajardo wrote:

Is this the stock kernel? If it is why are you using dkms?

cuda_memtest fails to compile due to missing nvcc

You might need to install cuda.

It makes a script in /etc/profile.d, which runs once in the login shell.
The build failed because I was supposed to log out first.

Offline

#8 2025-06-20 08:46:13

seth
Member
From: Won't reply 2 private help req
Registered: 2012-09-03
Posts: 75,081

Re: Random GPU lag

d_fajardo wrote:

The GPU fan is not registering in your smi.

Notebook with integrated cooling system where GPU and CPU share a heat-pipe?

Online

#9 2025-06-20 08:59:43

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

seth wrote:
d_fajardo wrote:

The GPU fan is not registering in your smi.

Notebook with integrated cooling system where GPU and CPU share a heat-pipe?

asus-isa-000a
Adapter: ISA adapter
cpu_fan:     3600 RPM
gpu_fan:     3600 RPM

Offline

#10 2025-06-20 09:01:39

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

seth wrote:

Battery (A32-K55): 100% [AC Connected]

Charger events?

If you drain the battery a bit, does the next "unexpected event" hit around when it's hitting 100% again?

Drained it to 70% with stress and then plugged it back in.
The same error and lag spike happened shortly after.

Offline

#11 2025-06-20 10:13:32

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

Observations:
1. Error if I disconnect or plug back in the charger
2. Error after staying on 98% for a bit because I configured for it to stay at approx. 95%
3. Error a few seconds after I configured it back to charge to 100%
4. Error purely at random with the associated lagspike

It is crazy inconsistent and I don't think it's possible to draw a meaningful conclusion.

Offline

#12 2025-06-20 10:34:36

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

seth wrote:

Battery (A32-K55): 100% [AC Connected]

Charger events?

If you drain the battery a bit, does the next "unexpected event" hit around when it's hitting 100% again?

Yup. It just got to 100% and a bunch of lagspikes occured.

Offline

#13 2025-06-20 13:25:25

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

Just noticed a bunch of errors in my kernel logs:

[   31.147051] NVRM: testIfDsmSubFunctionEnabled: GPS ACPI DSM called before _acpiDsmSupportedFuncCacheInit subfunction = 10.
[   31.147055] NVRM: testIfDsmSubFunctionEnabled: GPS ACPI DSM called before _acpiDsmSupportedFuncCacheInit subfunction = 11.
[   31.186151] [drm] Initialized nvidia-drm 0.0.0 for 0000:01:00.0 on minor 0
[   31.186777] nvidia 0000:01:00.0: [drm] Cannot find any crtc or sizes
[   43.023003] Bluetooth: RFCOMM TTY layer initialized
[   43.023028] Bluetooth: RFCOMM socket layer initialized
[   43.023041] Bluetooth: RFCOMM ver 1.11
[   43.225384] nvme nvme0: using unchecked data buffer
[   43.277498] block nvme0n1: No UUID available providing old NGUID
[   46.101946] warning: `kdeconnectd' uses wireless extensions which will stop working for Wi-Fi 7 hardware; use nl80211
[   46.270629] wlan0: authenticate with cc:28:aa:c2:db:24 (local address=b0:47:e9:4a:38:aa)
[   46.271079] wlan0: send auth to cc:28:aa:c2:db:24 (try 1/3)
[   46.309516] wlan0: authenticated
[   46.310405] wlan0: associate with cc:28:aa:c2:db:24 (try 1/3)
[   46.317715] wlan0: RX AssocResp from cc:28:aa:c2:db:24 (capab=0x11 status=0 aid=1)
[   46.326436] wlan0: associated
[   88.055756] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[   88.055773] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Physical Layer, (Receiver ID)
[   88.055778] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000001/0000a000
[   88.055783] nvidia 0000:01:00.0:    [ 0] RxErr                  (First)
[   88.055824] pcieport 0000:00:01.1: PME: Spurious native interrupt!
[   88.055835] pcieport 0000:00:01.1: PME: Spurious native interrupt!
[   88.055977] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[   88.055993] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[   88.055997] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[   88.056001] nvidia 0000:01:00.0:    [ 6] BadTLP                
[   88.056188] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[   88.056203] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[   88.056207] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[   88.056212] nvidia 0000:01:00.0:    [ 6] BadTLP                
[   88.056400] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[   88.056414] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[   88.056418] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[   88.056423] nvidia 0000:01:00.0:    [ 6] BadTLP                
[   88.056734] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[   88.056748] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[   88.056752] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[   88.056757] nvidia 0000:01:00.0:    [ 6] BadTLP                
[   88.056951] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[   88.056965] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[   88.056970] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[   88.056974] nvidia 0000:01:00.0:    [ 6] BadTLP                
[   88.057164] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[   88.057179] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Physical Layer, (Receiver ID)
[   88.057183] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000001/0000a000
[   88.057188] nvidia 0000:01:00.0:    [ 0] RxErr                  (First)
[   88.057367] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[   88.057378] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[   88.057380] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[   88.057382] nvidia 0000:01:00.0:    [ 6] BadTLP                
[   88.057558] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[   88.057569] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[   88.057571] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[   88.057573] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  122.202687] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  122.202700] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  122.202703] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  122.202705] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  122.202737] pcieport 0000:00:01.1: PME: Spurious native interrupt!
[  122.202903] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  122.202911] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  122.202913] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  122.202915] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  122.203092] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  122.203102] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  122.203104] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  122.203105] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  122.203280] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  122.203290] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  122.203292] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  122.203294] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  122.203467] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  122.203477] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  122.203478] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  122.203480] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  122.203676] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  122.203685] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  122.203686] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  122.203688] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  122.203862] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  122.203872] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  122.203874] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  122.203876] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  129.999325] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  129.999344] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Physical Layer, (Receiver ID)
[  129.999349] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000001/0000a000
[  129.999354] nvidia 0000:01:00.0:    [ 0] RxErr                  (First)
[  129.999548] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  129.999563] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  129.999567] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  129.999572] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  129.999760] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  129.999777] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  129.999783] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  129.999788] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  129.999982] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  129.999996] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  130.000001] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  130.000005] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  130.000485] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  130.000500] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  130.000505] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  130.000509] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  130.000697] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  130.000712] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  130.000716] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  130.000720] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  130.000908] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  130.000924] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  130.000928] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  130.000932] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  213.523527] input: EDIFIER WH950NB (AVRCP) as /devices/virtual/input/input30
[  275.161781] tun: Universal TUN/TAP device driver, 1.6
[  275.934661] amn0: entered promiscuous mode
[  636.197978] input: EDIFIER WH950NB (AVRCP) as /devices/virtual/input/input31
[  907.692071] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[  907.692087] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[  907.692094] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[  907.692101] nvidia 0000:01:00.0:    [ 6] BadTLP                
[  907.692125] pcieport 0000:00:01.1: PME: Spurious native interrupt!
[  907.692146] pcieport 0000:00:01.1: PME: Spurious native interrupt!
[ 1259.489331] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[ 1259.489340] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[ 1259.489342] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[ 1259.489344] nvidia 0000:01:00.0:    [ 6] BadTLP                
[ 1259.489356] pcieport 0000:00:01.1: PME: Spurious native interrupt!
[ 1259.489360] pcieport 0000:00:01.1: PME: Spurious native interrupt!
[ 1259.493695] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[ 1259.493701] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[ 1259.493703] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[ 1259.493705] nvidia 0000:01:00.0:    [ 6] BadTLP                
[ 1259.507513] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[ 1259.507524] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[ 1259.507527] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[ 1259.507530] nvidia 0000:01:00.0:    [ 6] BadTLP                
[ 1272.176828] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[ 1272.176839] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[ 1272.176844] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[ 1272.176847] nvidia 0000:01:00.0:    [ 6] BadTLP                
[ 1272.279609] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[ 1272.279627] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[ 1272.279634] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[ 1272.279640] nvidia 0000:01:00.0:    [ 6] BadTLP                
[ 1272.597438] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[ 1272.597455] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[ 1272.597462] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[ 1272.597468] nvidia 0000:01:00.0:    [ 6] BadTLP                
[ 1272.892708] pcieport 0000:00:01.1: AER: Correctable error message received from 0000:01:00.0
[ 1272.892721] nvidia 0000:01:00.0: PCIe Bus Error: severity=Correctable, type=Data Link Layer, (Receiver ID)
[ 1272.892725] nvidia 0000:01:00.0:   device [10de:25ad] error status/mask=00000040/0000a000
[ 1272.892729] nvidia 0000:01:00.0:    [ 6] BadTLP                
[ 1871.637817] NVRM: _threadNodeCheckTimeout: _threadNodeCheckTimeout: currentTime: f43adfc7b44c0 >= f43accb207e00
[ 1871.637831] NVRM: _threadNodeCheckTimeout: _threadNodeCheckTimeout: Timeout was set to: 4000 msecs!
[ 1871.638197] NVRM: _threadNodeCheckTimeout: _threadNodeCheckTimeout: currentTime: f43adfc7b44c0 >= f43accb207e00
[ 1871.638205] NVRM: _threadNodeCheckTimeout: _threadNodeCheckTimeout: Timeout was set to: 4000 msecs!
[ 1871.638445] NVRM: _threadNodeCheckTimeout: _threadNodeCheckTimeout: currentTime: f43adfc7b44c0 >= f43accb207e00
[ 1871.638452] NVRM: _threadNodeCheckTimeout: _threadNodeCheckTimeout: Timeout was set to: 4000 msecs!
[ 1871.638467] NVRM: _threadNodeCheckTimeout: _threadNodeCheckTimeout: currentTime: f43adfc7b44c0 >= f43accb207e00
[ 1871.638473] NVRM: _threadNodeCheckTimeout: _threadNodeCheckTimeout: Timeout was set to: 4000 msecs!
[ 1871.638481] NVRM: kflcnWaitForHalt_TU102: Timeout waiting for Falcon to halt
[ 1871.638490] NVRM: s_executeBooterUcode_TU102: failed to execute Booter: status 0x65, mailbox 0x7654f000
[ 1871.638496] NVRM: kgspExecuteBooterLoad_TU102: failed to execute Booter Load: 0x65
[ 1871.638502] NVRM: kgspBootstrap_TU102: failed to execute Booter Load (ucode for initial boot): 0x65
[ 1871.638508] NVRM: gpuPowerManagementResume: GSP boot failed at resume (bootMode 0x1): 0x65
[ 1871.640182] nvidia 0000:01:00.0: AMD-Vi: Event logged [IO_PAGE_FAULT domain=0x0011 address=0x472a1000 flags=0x0000]
[ 1871.640215] nvidia 0000:01:00.0: AMD-Vi: Event logged [IO_PAGE_FAULT domain=0x0011 address=0x472a1a00 flags=0x0000]
[ 1872.432821] ACPI: button: The lid device is not compliant to SW_LID.
[ 1877.688524] NVRM: _kgspLogXid119: ********************************* GSP Timeout **********************************
[ 1877.688532] NVRM: _kgspLogXid119: Note: Please also check logs above.
[ 1877.688543] NVRM: GPU at PCI:0000:01:00: GPU-c34b60d5-4a6e-7dc5-1146-ffb629959268
[ 1877.688546] NVRM: Xid (PCI:0000:01:00): 119, pid=17092, name=kworker/5:3, Timeout after 6s of waiting for RPC response from GPU0 GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080205b 0x4).
[ 1877.688574] NVRM: GPU0 GSP RPC buffer contains function 76 (GSP_RM_CONTROL) and data 0x000000002080205b 0x0000000000000004.
[ 1877.688577] NVRM: GPU0 RPC history (CPU -> GSP):
[ 1877.688579] NVRM:     entry function                   data0              data1              ts_start           ts_end             duration actively_polling
[ 1877.688581] NVRM:      0    76   GSP_RM_CONTROL        0x000000002080205b 0x0000000000000004 0x00063800c5242ef4 0x0000000000000000          y
[ 1877.688586] NVRM:     -1    47   UNLOADING_GUEST_DRIVE 0x0000000000000000 0x0000000000000000 0x00063800c40bd19c 0x00063800c4100017 274043us  
[ 1877.688591] NVRM:     -2    10   FREE                  0x00000000c1e00083 0x0000000000000000 0x00063800c40bc711 0x00063800c40bd139   2600us  
[ 1877.688595] NVRM:     -3    10   FREE                  0x000000000000000a 0x0000000000000000 0x00063800c40bc3e4 0x00063800c40bc70b    807us  
[ 1877.688598] NVRM:     -4    10   FREE                  0x000000000000000b 0x0000000000000000 0x00063800c40bbab0 0x00063800c40bc2c4   2068us  
[ 1877.688602] NVRM:     -5    10   FREE                  0x0000000000000006 0x0000000000000000 0x00063800c40bb5a9 0x00063800c40bba7d   1236us  
[ 1877.688605] NVRM:     -6    10   FREE                  0x0000000000000002 0x0000000000000000 0x00063800c40ba9aa 0x00063800c40bb544   2970us  
[ 1877.688608] NVRM:     -7    10   FREE                  0x0000000000000005 0x0000000000000000 0x00063800c40ba242 0x00063800c40ba998   1878us  
[ 1877.688611] NVRM: GPU0 RPC event history (CPU <- GSP):
[ 1877.688614] NVRM:     entry function                   data0              data1              ts_start           ts_end             duration during_incomplete_rpc
[ 1877.688616] NVRM:      0    4108 UCODE_LIBOS_PRINT     0x0000000000000000 0x0000000000000000 0x00063800c40c5dab 0x00063800c40c5dac      1us  
[ 1877.688620] NVRM:     -1    4128 GSP_POST_NOCAT_RECORD 0x0000000000000002 0x0000000000000028 0x00063800c40c1e69 0x00063800c40c1e6c      3us  
[ 1877.688624] NVRM:     -2    4111 PERF_BRIDGELESS_INFO_ 0x0000000000000000 0x0000000000000000 0x00063800c40c1d00 0x00063800c40c1d00           
[ 1877.688627] NVRM:     -3    4099 POST_EVENT            0x0000000000000021 0x0000000000000100 0x00063800c380ba24 0x00063800c380ba32     14us  
[ 1877.688631] NVRM:     -4    4099 POST_EVENT            0x0000000000000021 0x0000000000000020 0x00063800c378db1b 0x00063800c378db27     12us  
[ 1877.688635] NVRM:     -5    4099 POST_EVENT            0x0000000000000021 0x0000000000000001 0x00063800c2d313b7 0x00063800c2d313bf      8us  
[ 1877.688638] NVRM:     -6    4108 UCODE_LIBOS_PRINT     0x0000000000000000 0x0000000000000000 0x00063800c2d14fb7 0x00063800c2d14fb7           
[ 1877.688641] NVRM:     -7    4108 UCODE_LIBOS_PRINT     0x0000000000000000 0x0000000000000000 0x00063800c2d14e7d 0x00063800c2d14e7e      1us  
[ 1877.688648] CPU: 5 UID: 0 PID: 17092 Comm: kworker/5:3 Tainted: G           OE       6.15.2-arch1-1 #1 PREEMPT(full)  806378c57c3c21a60e39b7d20019ada706b7af8b
[ 1877.688654] Tainted: [O]=OOT_MODULE, [E]=UNSIGNED_MODULE
[ 1877.688655] Hardware name: ASUSTeK COMPUTER INC. ASUS TUF Gaming A17 FA706NF_FA706NF/FA706NF, BIOS FA706NF.308 01/08/2025
[ 1877.688657] Workqueue: kacpi_notify acpi_os_execute_deferred
[ 1877.688665] Call Trace:
[ 1877.688668]  <TASK>
[ 1877.688671]  dump_stack_lvl+0x5d/0x80
[ 1877.688681]  _kgspRpcRecvPoll+0x5b5/0x800 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.688913]  _issueRpcAndWait+0xc2/0x920 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.689107]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.689112]  ? osGetCurrentThread+0x26/0x60 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.689372]  ? _rmGpuLockIsOwner+0x24/0x90 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.689619]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.689623]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.689628]  rpcRmApiControl_GSP+0x274/0x960 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.689791]  ? update_curr+0x33/0x2a0
[ 1877.689806]  subdeviceCtrlCmdPerfSetPowerstate_KERNEL+0xaa/0x1c0 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.689995]  resControl_IMPL+0x1a5/0x1b0 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.690134]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.690141]  serverControl+0x48d/0x590 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.690273]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.690282]  _rmapiRmControl+0x598/0x980 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.690473]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.690495]  rmapiControlWithSecInfo+0x79/0x140 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.690713]  rmapiControl+0x24/0x40 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.690899]  RmPowerSourceChangeEvent+0x58/0x70 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.691189]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.691195]  RmPowerManagement+0x1c3/0x1cc [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.691474]  ? rm_transition_dynamic_power+0x53/0x13d [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.691758]  RmGcxPowerManagement+0x21d/0x3a0 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.692004]  ? _rmGpuLockIsOwner+0x24/0x90 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.692268]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.692273]  rm_transition_dynamic_power+0x8a/0x13d [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.692532]  ? __pci_find_next_ht_cap+0x59/0xe0
[ 1877.692543]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.692548]  ? __pfx_pci_pm_runtime_resume+0x10/0x10
[ 1877.692552]  ? __pfx_pci_pm_runtime_resume+0x10/0x10
[ 1877.692556]  nv_pmops_runtime_resume+0x65/0xf0 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.692676]  __rpm_callback+0x48/0x1f0
[ 1877.692682]  ? os_alloc_mem+0x104/0x120 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.692806]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.692810]  ? __pfx_pci_pm_runtime_resume+0x10/0x10
[ 1877.692815]  rpm_callback+0x6d/0x80
[ 1877.692819]  rpm_resume+0x4af/0x6d0
[ 1877.692823]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.692826]  ? _portMemAllocNonPagedUntracked+0x2c/0x40 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.693013]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.693020]  __pm_runtime_resume+0x52/0x90
[ 1877.693027]  pci_device_shutdown+0x1b/0x60
[ 1877.693033]  nv_indicate_not_idle+0x2e/0x40 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.693180]  os_ref_dynamic_power+0x146/0x220 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.693481]  rm_power_source_change_event+0x5b/0x189 [nvidia 227502a4f29567f98a4afecb2d3280c636b9e445]
[ 1877.693734]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1877.693740]  ? acpi_evaluate_object+0x195/0x450
[ 1877.693753]  acpi_ev_notify_dispatch+0x4e/0x70
[ 1877.693759]  acpi_os_execute_deferred+0x1a/0x30
[ 1877.693764]  process_one_work+0x193/0x350
[ 1877.693772]  worker_thread+0x2d7/0x410
[ 1877.693778]  ? __pfx_worker_thread+0x10/0x10
[ 1877.693781]  kthread+0xfc/0x240
[ 1877.693786]  ? __pfx_kthread+0x10/0x10
[ 1877.693789]  ret_from_fork+0x34/0x50
[ 1877.693795]  ? __pfx_kthread+0x10/0x10
[ 1877.693798]  ret_from_fork_asm+0x1a/0x30
[ 1877.693808]  </TASK>
[ 1877.693841] NVRM: _kgspLogXid119: ********************************************************************************
[ 1877.693846] NVRM: _issueRpcAndWait: rpcRecvPoll timedout for fn 76!
[ 1877.693850] NVRM: subdeviceCtrlCmdPerfSetPowerstate_KERNEL: NV2080_CTRL_CMD_PERF_SET_POWERSTATE RPC failed
[ 1883.694291] NVRM: Xid (PCI:0000:01:00): 119, pid=17092, name=kworker/5:3, Timeout after 6s of waiting for RPC response from GPU0 GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a81 0x4).
[ 1883.694306] NVRM: _issueRpcAndWait: rpcRecvPoll timedout for fn 76!
[ 1883.694308] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 1889.695140] NVRM: Xid (PCI:0000:01:00): 119, pid=17092, name=kworker/5:3, Timeout after 6s of waiting for RPC response from GPU0 GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080205b 0x4).
[ 1889.695252] NVRM: nvAssertFailedNoLog: Assertion failed: Back to back GSP RPC timeout detected! GPU marked for reset @ kernel_gsp.c:2314
[ 1889.695321] NVRM: _issueRpcAndWait: rpcRecvPoll timedout for fn 76!
[ 1889.695344] NVRM: subdeviceCtrlCmdPerfSetPowerstate_KERNEL: NV2080_CTRL_CMD_PERF_SET_POWERSTATE RPC failed
[ 1895.695822] NVRM: Rate limiting GSP RPC error prints for GPU at PCI:0000:01:00 (printing 1 of every 30).  The GPU likely needs to be reset.
[ 1895.695829] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 1895.696011] NVRM: rm_power_source_change_event: rm_power_source_change_event: Failed to handle Power Source change event, status=0x65
[ 1901.696588] NVRM: nvCheckOkFailedNoLog: Check failed: Call timed out [NV_ERR_TIMEOUT] (0x00000065) returned from pRmApi->Control(pRmApi, pGpu->hInternalClient, pGpu->hInternalSubdevice, NV2080_CTRL_CMD_INTERNAL_LOG_OOB_XID, &params, sizeof(params)) @ gpu.c:6468
[ 1901.696595] NVRM: Xid (PCI:0000:01:00): 154, GPU recovery action changed from 0x0 (None) to 0x1 (GPU Reset Required)
[ 1907.697356] NVRM: RmCheckForGcxSupportOnCurrentState: NVRM, Failed to get GCx pre-requisite, status=0x65
[ 1913.698116] NVRM: nvAssertFailedNoLog: Assertion failed: 0 @ kern_perf_pwr.c:191
[ 1919.698876] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 1919.699063] NVRM: RmHandleDNotifierEvent: RmHandleDNotifierEvent: Failed to handle ACPI D-Notifier event, status=0x65
[ 1925.699651] NVRM: RmCheckForGcxSupportOnCurrentState: NVRM, Failed to get GCx pre-requisite, status=0x65
[ 1931.700396] NVRM: subdeviceCtrlCmdPerfSetPowerstate_KERNEL: NV2080_CTRL_CMD_PERF_SET_POWERSTATE RPC failed
[ 1937.701151] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 1937.701341] NVRM: rm_power_source_change_event: rm_power_source_change_event: Failed to handle Power Source change event, status=0x65
[ 1943.701911] NVRM: RmCheckForGcxSupportOnCurrentState: NVRM, Failed to get GCx pre-requisite, status=0x65
[ 1949.702665] NVRM: nvAssertFailedNoLog: Assertion failed: 0 @ kern_perf_pwr.c:191
[ 1955.703414] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 1955.703602] NVRM: RmHandleDNotifierEvent: RmHandleDNotifierEvent: Failed to handle ACPI D-Notifier event, status=0x65
[ 1961.704173] NVRM: RmCheckForGcxSupportOnCurrentState: NVRM, Failed to get GCx pre-requisite, status=0x65
[ 1967.704918] NVRM: subdeviceCtrlCmdPerfSetPowerstate_KERNEL: NV2080_CTRL_CMD_PERF_SET_POWERSTATE RPC failed
[ 1973.705670] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 1973.705876] NVRM: rm_power_source_change_event: rm_power_source_change_event: Failed to handle Power Source change event, status=0x65
[ 1979.706424] NVRM: RmCheckForGcxSupportOnCurrentState: NVRM, Failed to get GCx pre-requisite, status=0x65
[ 1985.707174] NVRM: nvAssertFailedNoLog: Assertion failed: 0 @ kern_perf_pwr.c:191
[ 1991.707912] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 1991.708108] NVRM: RmHandleDNotifierEvent: RmHandleDNotifierEvent: Failed to handle ACPI D-Notifier event, status=0x65
[ 1997.708661] NVRM: RmCheckForGcxSupportOnCurrentState: NVRM, Failed to get GCx pre-requisite, status=0x65
[ 2003.709409] NVRM: subdeviceCtrlCmdPerfSetPowerstate_KERNEL: NV2080_CTRL_CMD_PERF_SET_POWERSTATE RPC failed
[ 2009.710143] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 2009.710328] NVRM: rm_power_source_change_event: rm_power_source_change_event: Failed to handle Power Source change event, status=0x65
[ 2015.710894] NVRM: subdeviceCtrlCmdPerfSetPowerstate_KERNEL: NV2080_CTRL_CMD_PERF_SET_POWERSTATE RPC failed
[ 2021.711637] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 2021.711824] NVRM: rm_power_source_change_event: rm_power_source_change_event: Failed to handle Power Source change event, status=0x65
[ 2027.712378] NVRM: nvAssertFailedNoLog: Assertion failed: 0 @ kern_perf_pwr.c:191
[ 2033.713106] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 2033.713292] NVRM: RmHandleDNotifierEvent: RmHandleDNotifierEvent: Failed to handle ACPI D-Notifier event, status=0x65
[ 2039.713881] NVRM: subdeviceCtrlCmdPerfSetPowerstate_KERNEL: NV2080_CTRL_CMD_PERF_SET_POWERSTATE RPC failed
[ 2045.714583] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 2045.714778] NVRM: rm_power_source_change_event: rm_power_source_change_event: Failed to handle Power Source change event, status=0x65
[ 2051.715323] NVRM: Xid (PCI:0000:01:00): 119, pid=934, name=nv_queue, Timeout after 6s of waiting for RPC response from GPU0 GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a7d7 0x2).
[ 2051.715335] NVRM: _issueRpcAndWait: rpcRecvPoll timedout for fn 76!
[ 2051.715342] NVRM: RmCheckForGcxSupportOnCurrentState: NVRM, Failed to get GCx pre-requisite, status=0x65
[ 2057.716132] NVRM: subdeviceCtrlCmdPerfSetPowerstate_KERNEL: NV2080_CTRL_CMD_PERF_SET_POWERSTATE RPC failed
[ 2063.716796] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 2063.717082] NVRM: rm_power_source_change_event: rm_power_source_change_event: Failed to handle Power Source change event, status=0x65
[ 2069.717533] NVRM: RmCheckForGcxSupportOnCurrentState: NVRM, Failed to get GCx pre-requisite, status=0x65
[ 2075.718272] NVRM: subdeviceCtrlCmdPerfSetPowerstate_KERNEL: NV2080_CTRL_CMD_PERF_SET_POWERSTATE RPC failed
[ 2081.718990] NVRM: _kperfSendPostPowerStateCallback: Error getting Aux Power State:0x65
[ 2081.719177] NVRM: rm_power_source_change_event: rm_power_source_change_event: Failed to handle Power Source change event, status=0x65
[ 2087.719728] NVRM: RmCheckForGcxSupportOnCurrentState: NVRM, Failed to get GCx pre-requisite, status=0x65

nvidia-smi -l 1 straight up just refuses to run a this point. Freezes and no response to SIGINT.

P.S. This seems to have been caused by me removing pcie_aspm=off while I was testing out kernel arguments.

Last edited by TheAirBlow (2025-06-20 13:36:40)

Offline

#14 2025-06-20 15:37:24

seth
Member
From: Won't reply 2 private help req
Registered: 2012-09-03
Posts: 75,081

Online

#15 2025-06-20 16:43:02

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

Nothing is logged when I plug/unplug my charger. That one was caused because I removed pcie_aspm=off.
I assumed NVreg_EnableGpuFirmware=0 was available in the open-source driver, but re-reading the wiki page changed my mind.
Switched to the binary blob, I'll continue monitoring this issue.

Last edited by TheAirBlow (2025-06-20 16:43:41)

Offline

#16 2025-06-20 16:44:12

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

The same random freezes still happen.

Offline

#17 2025-06-20 17:20:04

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

[ 1230.024499] INFO: task kworker/4:0:33 blocked for more than 122 seconds.
[ 1230.024507]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.024509] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.024511] task:kworker/4:0     state:D stack:0     pid:33    tgid:33    ppid:2      task_flags:0x4208060 flags:0x00004000
[ 1230.024516] Workqueue: kacpi_notify acpi_os_execute_deferred
[ 1230.024523] Call Trace:
[ 1230.024525]  <TASK>
[ 1230.024529]  __schedule+0x402/0x1320
[ 1230.024535]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.024539]  ? kmem_cache_alloc_noprof+0x12d/0x410
[ 1230.024544]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.024546]  ? acpi_ut_allocate_object_desc_dbg+0x5a/0x110
[ 1230.024550]  schedule+0x27/0xd0
[ 1230.024552]  schedule_preempt_disabled+0x15/0x30
[ 1230.024554]  rwsem_down_write_slowpath+0x1f4/0x6e0
[ 1230.024557]  ? acpi_ut_update_ref_count.part.0+0xdd/0x8f0
[ 1230.024561]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.024564]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.024567]  down_write+0x5a/0x60
[ 1230.024570]  os_acquire_rwlock_write+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.024813]  _nv051520rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.025111]  _nv053004rm+0x28c/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.025310]  _nv000839rm+0x27/0x70 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.025512]  rm_acpi_notify+0xf1/0x280 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.025811]  acpi_ev_notify_dispatch+0x4e/0x70
[ 1230.025817]  acpi_os_execute_deferred+0x1a/0x30
[ 1230.025821]  process_one_work+0x193/0x350
[ 1230.025826]  worker_thread+0x2d7/0x410
[ 1230.025831]  ? __pfx_worker_thread+0x10/0x10
[ 1230.025834]  kthread+0xfc/0x240
[ 1230.025838]  ? __pfx_kthread+0x10/0x10
[ 1230.025841]  ret_from_fork+0x34/0x50
[ 1230.025845]  ? __pfx_kthread+0x10/0x10
[ 1230.025848]  ret_from_fork_asm+0x1a/0x30
[ 1230.025856]  </TASK>
[ 1230.025866] INFO: task kworker/1:1:105 blocked for more than 122 seconds.
[ 1230.025870]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.025872] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.025875] task:kworker/1:1     state:D stack:0     pid:105   tgid:105   ppid:2      task_flags:0x4208060 flags:0x00004000
[ 1230.025879] Workqueue: kacpi_notify acpi_os_execute_deferred
[ 1230.025883] Call Trace:
[ 1230.025885]  <TASK>
[ 1230.025887]  __schedule+0x402/0x1320
[ 1230.025893]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.025896]  ? kmem_cache_alloc_noprof+0x12d/0x410
[ 1230.025901]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.025903]  ? acpi_ut_allocate_object_desc_dbg+0x5a/0x110
[ 1230.025909]  schedule+0x27/0xd0
[ 1230.025912]  schedule_preempt_disabled+0x15/0x30
[ 1230.025915]  rwsem_down_write_slowpath+0x1f4/0x6e0
[ 1230.025918]  ? asm_sysvec_apic_timer_interrupt+0x1a/0x20
[ 1230.025925]  down_write+0x5a/0x60
[ 1230.025928]  os_acquire_rwlock_write+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.026192]  _nv051520rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.026582]  _nv053004rm+0x28c/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.026777]  _nv000839rm+0x27/0x70 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.026967]  rm_acpi_notify+0xf1/0x280 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.027157]  acpi_ev_notify_dispatch+0x4e/0x70
[ 1230.027159]  acpi_os_execute_deferred+0x1a/0x30
[ 1230.027162]  process_one_work+0x193/0x350
[ 1230.027165]  worker_thread+0x2d7/0x410
[ 1230.027168]  ? __pfx_worker_thread+0x10/0x10
[ 1230.027170]  kthread+0xfc/0x240
[ 1230.027173]  ? __pfx_kthread+0x10/0x10
[ 1230.027175]  ret_from_fork+0x34/0x50
[ 1230.027177]  ? __pfx_kthread+0x10/0x10
[ 1230.027179]  ret_from_fork_asm+0x1a/0x30
[ 1230.027184]  </TASK>
[ 1230.027187] INFO: task kworker/1:2:121 blocked for more than 122 seconds.
[ 1230.027190]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.027191] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.027193] task:kworker/1:2     state:D stack:0     pid:121   tgid:121   ppid:2      task_flags:0x4208060 flags:0x00004000
[ 1230.027196] Workqueue: kacpi_notify acpi_os_execute_deferred
[ 1230.027199] Call Trace:
[ 1230.027200]  <TASK>
[ 1230.027202]  __schedule+0x402/0x1320
[ 1230.027205]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.027207]  ? kmem_cache_alloc_noprof+0x12d/0x410
[ 1230.027210]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.027212]  ? acpi_ut_allocate_object_desc_dbg+0x5a/0x110
[ 1230.027216]  schedule+0x27/0xd0
[ 1230.027218]  schedule_preempt_disabled+0x15/0x30
[ 1230.027220]  rwsem_down_write_slowpath+0x1f4/0x6e0
[ 1230.027222]  ? acpi_ut_update_ref_count.part.0+0xdd/0x8f0
[ 1230.027225]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.027228]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.027231]  down_write+0x5a/0x60
[ 1230.027233]  os_acquire_rwlock_write+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.027383]  _nv051520rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.027687]  _nv053004rm+0x28c/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.028001]  _nv000839rm+0x27/0x70 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.028257]  rm_acpi_notify+0xf1/0x280 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.028447]  acpi_ev_notify_dispatch+0x4e/0x70
[ 1230.028450]  acpi_os_execute_deferred+0x1a/0x30
[ 1230.028453]  process_one_work+0x193/0x350
[ 1230.028457]  worker_thread+0x2d7/0x410
[ 1230.028460]  ? __pfx_worker_thread+0x10/0x10
[ 1230.028463]  kthread+0xfc/0x240
[ 1230.028465]  ? __pfx_kthread+0x10/0x10
[ 1230.028467]  ret_from_fork+0x34/0x50
[ 1230.028469]  ? __pfx_kthread+0x10/0x10
[ 1230.028472]  ret_from_fork_asm+0x1a/0x30
[ 1230.028477]  </TASK>
[ 1230.028486] INFO: task kworker/5:1:137 blocked for more than 122 seconds.
[ 1230.028489]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.028491] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.028492] task:kworker/5:1     state:D stack:0     pid:137   tgid:137   ppid:2      task_flags:0x4208060 flags:0x00004000
[ 1230.028498] Workqueue: kacpi_notify acpi_os_execute_deferred
[ 1230.028501] Call Trace:
[ 1230.028504]  <TASK>
[ 1230.028506]  __schedule+0x402/0x1320
[ 1230.028510]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.028513]  ? kmem_cache_alloc_noprof+0x12d/0x410
[ 1230.028516]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.028519]  ? acpi_ut_allocate_object_desc_dbg+0x5a/0x110
[ 1230.028523]  schedule+0x27/0xd0
[ 1230.028538]  schedule_preempt_disabled+0x15/0x30
[ 1230.028547]  rwsem_down_write_slowpath+0x1f4/0x6e0
[ 1230.028559]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.028567]  down_write+0x5a/0x60
[ 1230.028575]  os_acquire_rwlock_write+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.028838]  _nv051520rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.029138]  _nv053004rm+0x28c/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.029319]  _nv000839rm+0x27/0x70 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.029516]  rm_acpi_notify+0xf1/0x280 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.029706]  acpi_ev_notify_dispatch+0x4e/0x70
[ 1230.029709]  acpi_os_execute_deferred+0x1a/0x30
[ 1230.029711]  process_one_work+0x193/0x350
[ 1230.029714]  worker_thread+0x2d7/0x410
[ 1230.029717]  ? __pfx_worker_thread+0x10/0x10
[ 1230.029720]  kthread+0xfc/0x240
[ 1230.029722]  ? __pfx_kthread+0x10/0x10
[ 1230.029724]  ret_from_fork+0x34/0x50
[ 1230.029726]  ? __pfx_kthread+0x10/0x10
[ 1230.029728]  ret_from_fork_asm+0x1a/0x30
[ 1230.029733]  </TASK>
[ 1230.029738] INFO: task kworker/11:1:187 blocked for more than 122 seconds.
[ 1230.029740]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.029742] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.029744] task:kworker/11:1    state:D stack:0     pid:187   tgid:187   ppid:2      task_flags:0x4208060 flags:0x00004000
[ 1230.029747] Workqueue: kacpi_notify acpi_os_execute_deferred
[ 1230.029751] Call Trace:
[ 1230.029752]  <TASK>
[ 1230.029754]  __schedule+0x402/0x1320
[ 1230.029757]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.029759]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.029763]  schedule+0x27/0xd0
[ 1230.029765]  schedule_preempt_disabled+0x15/0x30
[ 1230.029767]  rwsem_down_write_slowpath+0x1f4/0x6e0
[ 1230.029771]  down_write+0x5a/0x60
[ 1230.029774]  os_acquire_rwlock_write+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.029924]  _nv051520rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.030219]  _nv053004rm+0x28c/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.030399]  rm_power_source_change_event+0x3c/0x184 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.030593]  ? acpi_ut_release_mutex+0xef/0x1b0
[ 1230.030599]  acpi_ev_notify_dispatch+0x4e/0x70
[ 1230.030601]  acpi_os_execute_deferred+0x1a/0x30
[ 1230.030604]  process_one_work+0x193/0x350
[ 1230.030607]  worker_thread+0x2d7/0x410
[ 1230.030610]  ? __pfx_worker_thread+0x10/0x10
[ 1230.030612]  kthread+0xfc/0x240
[ 1230.030614]  ? __pfx_kthread+0x10/0x10
[ 1230.030617]  ret_from_fork+0x34/0x50
[ 1230.030619]  ? __pfx_kthread+0x10/0x10
[ 1230.030621]  ret_from_fork_asm+0x1a/0x30
[ 1230.030626]  </TASK>
[ 1230.030628] INFO: task kworker/8:2:218 blocked for more than 122 seconds.
[ 1230.030630]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.030632] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.030633] task:kworker/8:2     state:D stack:0     pid:218   tgid:218   ppid:2      task_flags:0x4208060 flags:0x00004000
[ 1230.030637] Workqueue: kacpi_notify acpi_os_execute_deferred
[ 1230.030639] Call Trace:
[ 1230.030641]  <TASK>
[ 1230.030642]  __schedule+0x402/0x1320
[ 1230.030645]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.030647]  ? kmem_cache_alloc_noprof+0x12d/0x410
[ 1230.030650]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.030652]  ? acpi_ut_allocate_object_desc_dbg+0x5a/0x110
[ 1230.030656]  schedule+0x27/0xd0
[ 1230.030658]  schedule_preempt_disabled+0x15/0x30
[ 1230.030660]  rwsem_down_write_slowpath+0x1f4/0x6e0
[ 1230.030662]  ? acpi_ut_update_ref_count.part.0+0xdd/0x8f0
[ 1230.030665]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.030668]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.030671]  down_write+0x5a/0x60
[ 1230.030673]  os_acquire_rwlock_write+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.030823]  _nv051520rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.031141]  _nv053004rm+0x28c/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.031322]  _nv000839rm+0x27/0x70 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.031521]  rm_acpi_notify+0xf1/0x280 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.031713]  acpi_ev_notify_dispatch+0x4e/0x70
[ 1230.031716]  acpi_os_execute_deferred+0x1a/0x30
[ 1230.031719]  process_one_work+0x193/0x350
[ 1230.031723]  worker_thread+0x2d7/0x410
[ 1230.031726]  ? __pfx_worker_thread+0x10/0x10
[ 1230.031729]  kthread+0xfc/0x240
[ 1230.031731]  ? __pfx_kthread+0x10/0x10
[ 1230.031733]  ret_from_fork+0x34/0x50
[ 1230.031736]  ? __pfx_kthread+0x10/0x10
[ 1230.031738]  ret_from_fork_asm+0x1a/0x30
[ 1230.031743]  </TASK>
[ 1230.031747] INFO: task nv_queue:262 blocked for more than 122 seconds.
[ 1230.031750]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.031751] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.031753] task:nv_queue        state:D stack:0     pid:262   tgid:262   ppid:2      task_flags:0x208040 flags:0x00004000
[ 1230.031756] Call Trace:
[ 1230.031757]  <TASK>
[ 1230.031759]  __schedule+0x402/0x1320
[ 1230.031763]  ? os_execute_work_item+0x40/0x90 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.031914]  schedule+0x27/0xd0
[ 1230.031916]  schedule_preempt_disabled+0x15/0x30
[ 1230.031918]  rwsem_down_write_slowpath+0x1f4/0x6e0
[ 1230.031920]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.031925]  down_write+0x5a/0x60
[ 1230.031928]  os_acquire_rwlock_write+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.032086]  _nv051520rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.032531]  _nv053004rm+0x28c/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.032847]  _nv059758rm+0x63/0x230 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.033018]  ? __pfx__main_loop+0x10/0x10 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.033169]  rm_execute_work_item+0x66/0x1f0 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.033364]  os_execute_work_item+0x68/0x90 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.033523]  _main_loop+0x93/0x150 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.033703]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.033707]  ? __pfx__main_loop+0x10/0x10 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.033853]  kthread+0xfc/0x240
[ 1230.033857]  ? __pfx_kthread+0x10/0x10
[ 1230.033859]  ret_from_fork+0x34/0x50
[ 1230.033862]  ? __pfx_kthread+0x10/0x10
[ 1230.033864]  ret_from_fork_asm+0x1a/0x30
[ 1230.033869]  </TASK>
[ 1230.033873] INFO: task kworker/10:2:328 blocked for more than 122 seconds.
[ 1230.033876]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.033878] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.033880] task:kworker/10:2    state:D stack:0     pid:328   tgid:328   ppid:2      task_flags:0x4208160 flags:0x00004000
[ 1230.033884] Workqueue: kacpi_notify acpi_os_execute_deferred
[ 1230.033887] Call Trace:
[ 1230.033888]  <TASK>
[ 1230.033890]  __schedule+0x402/0x1320
[ 1230.033894]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.033896]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.033900]  schedule+0x27/0xd0
[ 1230.033902]  schedule_preempt_disabled+0x15/0x30
[ 1230.033903]  rwsem_down_write_slowpath+0x1f4/0x6e0
[ 1230.033908]  down_write+0x5a/0x60
[ 1230.033911]  os_acquire_rwlock_write+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.034059]  _nv051520rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.034352]  _nv053004rm+0x28c/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.034539]  rm_power_source_change_event+0x3c/0x184 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.034849]  ? acpi_ut_release_mutex+0xef/0x1b0
[ 1230.034857]  acpi_ev_notify_dispatch+0x4e/0x70
[ 1230.034861]  acpi_os_execute_deferred+0x1a/0x30
[ 1230.034865]  process_one_work+0x193/0x350
[ 1230.034870]  worker_thread+0x2d7/0x410
[ 1230.034874]  ? __pfx_worker_thread+0x10/0x10
[ 1230.034877]  kthread+0xfc/0x240
[ 1230.034881]  ? __pfx_kthread+0x10/0x10
[ 1230.034884]  ret_from_fork+0x34/0x50
[ 1230.034888]  ? __pfx_kthread+0x10/0x10
[ 1230.034890]  ret_from_fork_asm+0x1a/0x30
[ 1230.034897]  </TASK>
[ 1230.034906] INFO: task kworker/0:2:670 blocked for more than 122 seconds.
[ 1230.034910]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.034912] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.034914] task:kworker/0:2     state:D stack:0     pid:670   tgid:670   ppid:2      task_flags:0x4208060 flags:0x00004000
[ 1230.034919] Workqueue: kacpi_notify acpi_os_execute_deferred
[ 1230.034922] Call Trace:
[ 1230.034924]  <TASK>
[ 1230.034927]  __schedule+0x402/0x1320
[ 1230.034930]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.034934]  ? kmem_cache_alloc_noprof+0x12d/0x410
[ 1230.034939]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.034941]  ? acpi_ut_allocate_object_desc_dbg+0x5a/0x110
[ 1230.034946]  schedule+0x27/0xd0
[ 1230.034949]  schedule_preempt_disabled+0x15/0x30
[ 1230.034952]  rwsem_down_write_slowpath+0x1f4/0x6e0
[ 1230.034955]  ? acpi_ut_update_ref_count.part.0+0xdd/0x8f0
[ 1230.034959]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.034963]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.034967]  down_write+0x5a/0x60
[ 1230.034970]  os_acquire_rwlock_write+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.035237]  _nv051520rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.035554]  _nv053004rm+0x28c/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.035742]  _nv000839rm+0x27/0x70 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.035932]  rm_acpi_notify+0xf1/0x280 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.036121]  acpi_ev_notify_dispatch+0x4e/0x70
[ 1230.036123]  acpi_os_execute_deferred+0x1a/0x30
[ 1230.036126]  process_one_work+0x193/0x350
[ 1230.036129]  worker_thread+0x2d7/0x410
[ 1230.036132]  ? __pfx_worker_thread+0x10/0x10
[ 1230.036134]  kthread+0xfc/0x240
[ 1230.036137]  ? __pfx_kthread+0x10/0x10
[ 1230.036139]  ret_from_fork+0x34/0x50
[ 1230.036141]  ? __pfx_kthread+0x10/0x10
[ 1230.036143]  ret_from_fork_asm+0x1a/0x30
[ 1230.036148]  </TASK>
[ 1230.036157] INFO: task nvidia-powerd:1053 blocked for more than 122 seconds.
[ 1230.036160]       Tainted: P           OE       6.15.2-arch1-1 #1
[ 1230.036161] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[ 1230.036162] task:nvidia-powerd   state:D stack:0     pid:1053  tgid:1053  ppid:1      task_flags:0x400100 flags:0x00000002
[ 1230.036165] Call Trace:
[ 1230.036167]  <TASK>
[ 1230.036168]  __schedule+0x402/0x1320
[ 1230.036173]  schedule+0x27/0xd0
[ 1230.036175]  schedule_preempt_disabled+0x15/0x30
[ 1230.036177]  rwsem_down_read_slowpath+0x25f/0x4d0
[ 1230.036180]  ? os_memcpy_to_user+0x56/0x80 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.036331]  down_read+0x48/0xb0
[ 1230.036334]  os_acquire_rwlock_read+0x2b/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.036485]  _nv051518rm+0x10/0x40 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.036894]  _nv053004rm+0x2cc/0x360 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.037086]  _nv053023rm+0x15f/0x1e0 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.037266]  _nv015696rm+0x345/0x680 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.037442]  _nv052978rm+0x69/0xd0 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.037663]  _nv013970rm+0x86/0xa0 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.037848]  _nv000792rm+0x7e9/0xe30 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.038041]  rm_ioctl+0x7f/0x400 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.038232]  nvidia_unlocked_ioctl+0x52a/0x8e0 [nvidia a60a6f8bfc4051b9fdde010080cb3f1d258a8f56]
[ 1230.038381]  __x64_sys_ioctl+0x97/0xc0
[ 1230.038385]  do_syscall_64+0x7b/0x810
[ 1230.038388]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.038390]  ? switch_fpu_return+0x4e/0xd0
[ 1230.038393]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.038395]  ? arch_exit_to_user_mode_prepare.isra.0+0x7c/0x90
[ 1230.038398]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.038400]  ? syscall_exit_to_user_mode+0x37/0x1c0
[ 1230.038403]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.038404]  ? do_syscall_64+0x87/0x810
[ 1230.038406]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.038408]  ? arch_exit_to_user_mode_prepare.isra.0+0x7c/0x90
[ 1230.038410]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.038411]  ? syscall_exit_to_user_mode+0x37/0x1c0
[ 1230.038414]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.038415]  ? do_syscall_64+0x87/0x810
[ 1230.038417]  ? srso_alias_return_thunk+0x5/0xfbef5
[ 1230.038419]  ? irqentry_exit_to_user_mode+0x2c/0x1b0
[ 1230.038421]  entry_SYSCALL_64_after_hwframe+0x76/0x7e
[ 1230.038424] RIP: 0033:0x7f1392ff9ecd
[ 1230.038450] RSP: 002b:00007ffd71d52640 EFLAGS: 00000246 ORIG_RAX: 0000000000000010
[ 1230.038453] RAX: ffffffffffffffda RBX: 0000000000000020 RCX: 00007f1392ff9ecd
[ 1230.038454] RDX: 00007ffd71d527a0 RSI: 00000000c020462a RDI: 0000000000000005
[ 1230.038455] RBP: 00007ffd71d52690 R08: 00007ffd71d527a0 R09: 00007ffd71d527bc
[ 1230.038457] R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000005
[ 1230.038458] R13: 00000000c020462a R14: 000000000000002a R15: 00007ffd71d527bc
[ 1230.038462]  </TASK>
[ 1230.038463] Future hung task reports are suppressed, see sysctl kernel.hung_task_warnings

Now nvidia-smi just straight up freezes. It also made my system lag for ~5 seconds.

Last edited by TheAirBlow (2025-06-20 17:20:33)

Offline

#18 2025-06-20 17:54:45

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

Observation: it seems to run just fine when I unplug my charger and put it on Performance profile. No random freezes happen at all in this state.

Offline

#19 2025-06-20 23:21:53

seth
Member
From: Won't reply 2 private help req
Registered: 2012-09-03
Posts: 75,081

Re: Random GPU lag

seth wrote:

Battery (A32-K55): 100% [AC Connected]

Charger events?

Please post your complete system journal for the boot:

sudo journalctl -b | curl -F 'file=@-' 0x0.st

There were threads with some daemon removing the GPU from the bus in response, causing the the still loaded driver to talk to the void…

Online

#20 2025-06-21 08:30:20

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

seth wrote:
seth wrote:

Battery (A32-K55): 100% [AC Connected]

Charger events?

Please post your complete system journal for the boot:

sudo journalctl -b | curl -F 'file=@-' 0x0.st

There were threads with some daemon removing the GPU from the bus in response, causing the the still loaded driver to talk to the void…

https://0x0.st/8IWP.txt

Nothing is logged at the same time as the lagspike happens.
I just suspect it's my charger disconnecting/connecting too fast, because it looks like there's something up with the port?

Offline

#21 2025-06-21 14:06:34

seth
Member
From: Won't reply 2 private help req
Registered: 2012-09-03
Posts: 75,081

Re: Random GPU lag

The kind of errors as in #17 isn't in there either.
Can you catch any charger events in https://wiki.archlinux.org/title/Acpid ?
Does anything show up when you explicitly pull and plug the charger?

Jun 21 13:13:03 theairblow supergfxd[894]: [DEBUG supergfxctl::pci_device] Laptop does not have a hotplug dgpu: DgpuNotFound
Jun 21 13:13:04 theairblow supergfxd[894]: [DEBUG supergfxctl] Did CommandArgs { inner: ["start", "nvidia-powerd.service"] }
Jun 21 13:13:04 theairblow supergfxd[894]: [INFO  supergfxctl::controller] reload: Reloaded gfx mode: Hybrid

If you're running in hybrid, in normal prime mode, why do you run that daemon at all?
Try to disable it and see whether the system stabilizes…

Online

#22 2025-06-21 14:19:05

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

seth wrote:

The kind of errors as in #17 isn't in there either.
Can you catch any charger events in https://wiki.archlinux.org/title/Acpid ?
Does anything show up when you explicitly pull and plug the charger?

Jun 21 13:13:03 theairblow supergfxd[894]: [DEBUG supergfxctl::pci_device] Laptop does not have a hotplug dgpu: DgpuNotFound
Jun 21 13:13:04 theairblow supergfxd[894]: [DEBUG supergfxctl] Did CommandArgs { inner: ["start", "nvidia-powerd.service"] }
Jun 21 13:13:04 theairblow supergfxd[894]: [INFO  supergfxctl::controller] reload: Reloaded gfx mode: Hybrid

If you're running in hybrid, in normal prime mode, why do you run that daemon at all?
Try to disable it and see whether the system stabilizes…

I disabled supergfxd along with nvidia-powerd.

Events when plugged in:

Jun 21 19:18:56 theairblow root[66620]: ACPI action undefined: ACPI0003:00
Jun 21 19:18:57 theairblow root[66624]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 19:18:57 theairblow root[66626]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 19:18:57 theairblow root[66628]: ACPI action undefined: PNP0C0A:00
Jun 21 19:18:57 theairblow root[66630]: ACPI action undefined: PNP0C0A:00
Jun 21 19:18:57 theairblow root[66632]: ACPI action undefined: ACPI0003:00
Jun 21 19:18:57 theairblow root[66634]: ACPI action undefined: PNP0C0A:00
Jun 21 19:18:57 theairblow root[66636]: ACPI action undefined: PNP0C0A:00

Events when plugged out:

Jun 21 19:19:49 theairblow root[66706]: ACPI action undefined: ACPI0003:00
Jun 21 19:19:49 theairblow root[66710]: ACPI action undefined: ACPI0003:00
Jun 21 19:19:49 theairblow root[66721]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 19:19:49 theairblow root[66723]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 19:19:49 theairblow root[66725]: ACPI action undefined: PNP0C0A:00
Jun 21 19:19:49 theairblow root[66727]: ACPI action undefined: ACPI0003:00

Last edited by TheAirBlow (2025-06-21 14:20:01)

Offline

#23 2025-06-21 15:07:40

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

At the same time as a lag spike happens I see just metric tons of events like this:

Jun 21 19:39:56 theairblow root[68751]: ACPI action undefined: ACPI0003:00
Jun 21 19:39:56 theairblow root[68755]: ACPI action undefined: ACPI0003:00
Jun 21 19:39:56 theairblow root[68767]: ACPI action undefined: ACPI0003:00
Jun 21 19:57:18 theairblow root[70118]: ACPI group/action undefined: button/up / UP
Jun 21 19:57:21 theairblow root[70185]: ACPI group/action undefined: button/up / UP
Jun 21 20:03:47 theairblow root[70999]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:47 theairblow root[71002]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:47 theairblow root[71009]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:03:47 theairblow root[71011]: ACPI action undefined: PNP0C0A:00
Jun 21 20:03:47 theairblow root[71015]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:47 theairblow root[71023]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:47 theairblow root[71039]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:03:47 theairblow root[71041]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:03:47 theairblow root[71043]: ACPI action undefined: PNP0C0A:00
Jun 21 20:03:47 theairblow root[71045]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:48 theairblow root[71048]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:48 theairblow root[71050]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:48 theairblow root[71052]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:48 theairblow root[71054]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:48 theairblow root[71056]: ACPI action undefined: PNP0C0A:00
Jun 21 20:03:49 theairblow root[71059]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:49 theairblow root[71064]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:49 theairblow root[71066]: ACPI action undefined: PNP0C0A:00
Jun 21 20:03:49 theairblow root[71068]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:03:49 theairblow root[71070]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:03:49 theairblow root[71072]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:49 theairblow root[71079]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:49 theairblow root[71081]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:50 theairblow root[71083]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:50 theairblow root[71089]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:03:50 theairblow root[71091]: ACPI action undefined: PNP0C0A:00
Jun 21 20:03:51 theairblow root[71096]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:51 theairblow root[71100]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:51 theairblow root[71102]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:52 theairblow root[71104]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:52 theairblow root[71108]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:52 theairblow root[71110]: ACPI action undefined: ACPI0003:00
Jun 21 20:03:54 theairblow root[71120]: ACPI action undefined: PNP0C0A:00
Jun 21 20:03:56 theairblow root[71125]: ACPI action undefined: PNP0C0A:00
Jun 21 20:03:56 theairblow root[71127]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:00 theairblow root[71134]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:00 theairblow root[71142]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:00 theairblow root[71144]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:00 theairblow root[71146]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:00 theairblow root[71148]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:00 theairblow root[71150]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:00 theairblow root[71152]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:00 theairblow root[71154]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:00 theairblow root[71156]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:01 theairblow root[71158]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:01 theairblow root[71160]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:01 theairblow root[71164]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:01 theairblow root[71168]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:01 theairblow root[71170]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:01 theairblow root[71176]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:01 theairblow root[71179]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:01 theairblow root[71183]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:01 theairblow root[71185]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:01 theairblow root[71189]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:01 theairblow root[71191]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:01 theairblow root[71195]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:02 theairblow root[71197]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:02 theairblow root[71201]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:02 theairblow root[71203]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:02 theairblow root[71210]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:02 theairblow root[71212]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:02 theairblow root[71216]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:02 theairblow root[71218]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:02 theairblow root[71220]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:02 theairblow root[71224]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:02 theairblow root[71226]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:02 theairblow root[71231]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:02 theairblow root[71233]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:02 theairblow root[71235]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71237]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71240]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71242]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71244]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71250]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71252]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71254]: ACPI group/action undefined: button/up / UP
Jun 21 20:04:03 theairblow root[71259]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71261]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71263]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71267]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71276]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:03 theairblow root[71278]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71280]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71282]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71284]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:04 theairblow root[71288]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:04 theairblow root[71290]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71292]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71294]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71296]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71300]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71302]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71304]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71306]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71308]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71311]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:04 theairblow root[71313]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:05 theairblow root[71315]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:05 theairblow root[71317]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:05 theairblow root[71321]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71324]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71328]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71330]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71332]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71334]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71336]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71338]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:05 theairblow root[71340]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:05 theairblow root[71342]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:05 theairblow root[71344]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71346]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71348]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71350]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:05 theairblow root[71352]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71354]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71356]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71358]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71360]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71362]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71365]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71367]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:06 theairblow root[71369]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:06 theairblow root[71371]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:06 theairblow root[71373]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71375]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71377]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71379]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71381]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71383]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:06 theairblow root[71385]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71387]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:07 theairblow root[71389]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:07 theairblow root[71391]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:07 theairblow root[71393]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71395]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71399]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71401]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71403]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71405]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71409]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:07 theairblow root[71411]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:07 theairblow root[71413]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:07 theairblow root[71417]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71419]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71421]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71423]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:07 theairblow root[71425]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71427]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71429]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71431]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71434]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:08 theairblow root[71436]: ACPI group/action undefined: wmi / PNP0C14:01
Jun 21 20:04:08 theairblow root[71438]: ACPI action undefined: PNP0C0A:00
Jun 21 20:04:08 theairblow root[71440]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71442]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71444]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71446]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71448]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71450]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71452]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71454]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71456]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71458]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:08 theairblow root[71460]: ACPI action undefined: ACPI0003:00
Jun 21 20:04:10 theairblow root[71465]: ACPI action undefined: PNP0C0A:00

So it is indeed charger events that are causing this.

Offline

#24 2025-06-21 21:48:14

seth
Member
From: Won't reply 2 private help req
Registered: 2012-09-03
Posts: 75,081

Re: Random GPU lag

You could use the handler.sh script to maybe figure what these events are.
Is the charger plug possibly loose?

Online

#25 2025-06-22 04:33:47

TheAirBlow
Member
Registered: 2022-07-05
Posts: 75

Re: Random GPU lag

seth wrote:

You could use the handler.sh script to maybe figure what these events are.
Is the charger plug possibly loose?

I assume it is. Sometimes I plug it in and I hear a concoction of plug in/plug out sounds from Plasma.

Offline

Board footer

Powered by FluxBB