![]() |
儒雅的企鹅
1 年前 |
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement . We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed #1648 docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed #1648 zyr-NULL opened this issuewhen i use docker to create container, i get this error
docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.
Steps to reproduce the issue
when i executed the following command
sudo docker run --rm --gpus all nvidia/cuda:11.0.3-base-ubuntu20.04 nvidia-smi
i get the following error
docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy' nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.
but when i executed this following command, it has the expected output
sudo docker run hello-world
here is some Information
Some nvidia-container information
gpu-server@gpu-server:~$ nvidia-container-cli -k -d /dev/tty info
-- WARNING, the following logs are for debugging purposes only --
I0621 07:07:51.735875 4789 nvc.c:376] initializing library context (version=1.10.0, build=395fd41701117121f1fd04ada01e1d7e006a37ae)
I0621 07:07:51.735941 4789 nvc.c:350] using root /
I0621 07:07:51.735947 4789 nvc.c:351] using ldcache /etc/ld.so.cache
I0621 07:07:51.735963 4789 nvc.c:352] using unprivileged user 1000:1000
I0621 07:07:51.735984 4789 nvc.c:393] attempting to load dxcore to see if we are running under Windows Subsystem for Linux (WSL)
I0621 07:07:51.736064 4789 nvc.c:395] dxcore initialization failed, continuing assuming a non-WSL environment
W0621 07:07:51.739205 4791 nvc.c:273] failed to set inheritable capabilities
W0621 07:07:51.739329 4791 nvc.c:274] skipping kernel modules load due to failure
I0621 07:07:51.739807 4793 rpc.c:71] starting driver rpc service
W0621 07:08:16.774958 4789 rpc.c:121] terminating driver rpc service (forced)
I0621 07:08:20.481845 4789 rpc.c:135] driver rpc service terminated with signal 15
nvidia-container-cli: initialization error: driver rpc error: timed out
I0621 07:08:20.481972 4789 nvc.c:434] shutting down library context
Kernel version
Linux gpu-server 4.15.0-187-generic #198-Ubuntu SMP Tue Jun 14 03:23:51 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
Driver information
==============NVSMI LOG==============
Timestamp : Tue Jun 21 07:13:57 2022
Driver Version : 515.48.07
CUDA Version : 11.7
Attached GPUs : 4
GPU 00000000:01:00.0
Product Name : NVIDIA A100-SXM4-40GB
Product Brand : NVIDIA
Product Architecture : Ampere
Display Mode : Disabled
Display Active : Disabled
Persistence Mode : Disabled
MIG Mode
Current : Disabled
Pending : Disabled
Accounting Mode : Disabled
Accounting Mode Buffer Size : 4000
Driver Model
Current : N/A
Pending : N/A
Serial Number : 1561221014674
GPU UUID : GPU-b67da01e-feba-d839-62c5-2773d4e963f0
Minor Number : 0
VBIOS Version : 92.00.19.00.13
MultiGPU Board : No
Board ID : 0x100
GPU Part Number : 692-2G506-0202-002
Module ID : 3
Inforom Version
Image Version : G506.0202.00.02
OEM Object : 2.0
ECC Object : 6.16
Power Management Object : N/A
GPU Operation Mode
Current : N/A
Pending : N/A
GSP Firmware Version : 515.48.07
GPU Virtualization Mode
Virtualization Mode : None
Host VGPU Mode : N/A
IBMNPU
Relaxed Ordering Mode : N/A
Bus : 0x01
Device : 0x00
Domain : 0x0000
Device Id : 0x20B010DE
Bus Id : 00000000:01:00.0
Sub System Id : 0x144E10DE
GPU Link Info
PCIe Generation
Max : 4
Current : 4
Link Width
Max : 16x
Current : 16x
Bridge Chip
Type : N/A
Firmware : N/A
Replays Since Reset : 0
Replay Number Rollovers : 0
Tx Throughput : 0 KB/s
Rx Throughput : 0 KB/s
Fan Speed : N/A
Performance State : P0
Clocks Throttle Reasons
Idle : Active
Applications Clocks Setting : Not Active
SW Power Cap : Not Active
HW Slowdown : Not Active
HW Thermal Slowdown : Not Active
HW Power Brake Slowdown : Not Active
Sync Boost : Not Active
SW Thermal Slowdown : Not Active
Display Clock Setting : Not Active
FB Memory Usage
Total : 40960 MiB
Reserved : 571 MiB
Used : 0 MiB
Free : 40388 MiB
BAR1 Memory Usage
Total : 65536 MiB
Used : 1 MiB
Free : 65535 MiB
Compute Mode : Default
Utilization
Gpu : 0 %
Memory : 0 %
Encoder : 0 %
Decoder : 0 %
Encoder Stats
Active Sessions : 0
Average FPS : 0
Average Latency : 0
FBC Stats
Active Sessions : 0
Average FPS : 0
Average Latency : 0
Ecc Mode
Current : Enabled
Pending : Enabled
ECC Errors
Volatile
SRAM Correctable : 0
SRAM Uncorrectable : 0
DRAM Correctable : 0
DRAM Uncorrectable : 0
Aggregate
SRAM Correctable : 0
SRAM Uncorrectable : 0
DRAM Correctable : 0
DRAM Uncorrectable : 0
Retired Pages
Single Bit ECC : N/A
Double Bit ECC : N/A
Pending Page Blacklist : N/A
Remapped Rows
Correctable Error : 0
Uncorrectable Error : 0
Pending : No
Remapping Failure Occurred : No
Bank Remap Availability Histogram
Max : 640 bank(s)
High : 0 bank(s)
Partial : 0 bank(s)
Low : 0 bank(s)
None : 0 bank(s)
Temperature
GPU Current Temp : 32 C
GPU Shutdown Temp : 92 C
GPU Slowdown Temp : 89 C
GPU Max Operating Temp : 85 C
GPU Target Temperature : N/A
Memory Current Temp : 34 C
Memory Max Operating Temp : 95 C
Power Readings
Power Management : Supported
Power Draw : 54.92 W
Power Limit : 400.00 W
Default Power Limit : 400.00 W
Enforced Power Limit : 400.00 W
Min Power Limit : 100.00 W
Max Power Limit : 400.00 W
Clocks
Graphics : 1080 MHz
SM : 1080 MHz
Memory : 1215 MHz
Video : 975 MHz
Applications Clocks
Graphics : 1095 MHz
Memory : 1215 MHz
Default Applications Clocks
Graphics : 1095 MHz
Memory : 1215 MHz
Max Clocks
Graphics : 1410 MHz
SM : 1410 MHz
Memory : 1215 MHz
Video : 1290 MHz
Max Customer Boost Clocks
Graphics : 1410 MHz
Clock Policy
Auto Boost : N/A
Auto Boost Default : N/A
Voltage
Graphics : 731.250 mV
Processes : None
GPU 00000000:41:00.0
Product Name : NVIDIA A100-SXM4-40GB
Product Brand : NVIDIA
Product Architecture : Ampere
Display Mode : Disabled
Display Active : Disabled
Persistence Mode : Disabled
MIG Mode
Current : Disabled
Pending : Disabled
Accounting Mode : Disabled
Accounting Mode Buffer Size : 4000
Driver Model
Current : N/A
Pending : N/A
Serial Number : 1561221014888
GPU UUID : GPU-6ca82e47-c63a-1bea-38ad-d3af9e1dc26b
Minor Number : 1
VBIOS Version : 92.00.19.00.13
MultiGPU Board : No
Board ID : 0x4100
GPU Part Number : 692-2G506-0202-002
Module ID : 1
Inforom Version
Image Version : G506.0202.00.02
OEM Object : 2.0
ECC Object : 6.16
Power Management Object : N/A
GPU Operation Mode
Current : N/A
Pending : N/A
GSP Firmware Version : 515.48.07
GPU Virtualization Mode
Virtualization Mode : None
Host VGPU Mode : N/A
IBMNPU
Relaxed Ordering Mode : N/A
Bus : 0x41
Device : 0x00
Domain : 0x0000
Device Id : 0x20B010DE
Bus Id : 00000000:41:00.0
Sub System Id : 0x144E10DE
GPU Link Info
PCIe Generation
Max : 4
Current : 4
Link Width
Max : 16x
Current : 16x
Bridge Chip
Type : N/A
Firmware : N/A
Replays Since Reset : 0
Replay Number Rollovers : 0
Tx Throughput : 0 KB/s
Rx Throughput : 0 KB/s
Fan Speed : N/A
Performance State : P0
Clocks Throttle Reasons
Idle : Active
Applications Clocks Setting : Not Active
SW Power Cap : Not Active
HW Slowdown : Not Active
HW Thermal Slowdown : Not Active
HW Power Brake Slowdown : Not Active
Sync Boost : Not Active
SW Thermal Slowdown : Not Active
Display Clock Setting : Not Active
FB Memory Usage
Total : 40960 MiB
Reserved : 571 MiB
Used : 0 MiB
Free : 40388 MiB
BAR1 Memory Usage
Total : 65536 MiB
Used : 1 MiB
Free : 65535 MiB
Compute Mode : Default
Utilization
Gpu : 0 %
Memory : 0 %
Encoder : 0 %
Decoder : 0 %
Encoder Stats
Active Sessions : 0
Average FPS : 0
Average Latency : 0
FBC Stats
Active Sessions : 0
Average FPS : 0
Average Latency : 0
Ecc Mode
Current : Enabled
Pending : Enabled
ECC Errors
Volatile
SRAM Correctable : 0
SRAM Uncorrectable : 0
DRAM Correctable : 0
DRAM Uncorrectable : 0
Aggregate
SRAM Correctable : 0
SRAM Uncorrectable : 0
DRAM Correctable : 0
DRAM Uncorrectable : 0
Retired Pages
Single Bit ECC : N/A
Double Bit ECC : N/A
Pending Page Blacklist : N/A
Remapped Rows
Correctable Error : 0
Uncorrectable Error : 0
Pending : No
Remapping Failure Occurred : No
Bank Remap Availability Histogram
Max : 640 bank(s)
High : 0 bank(s)
Partial : 0 bank(s)
Low : 0 bank(s)
None : 0 bank(s)
Temperature
GPU Current Temp : 30 C
GPU Shutdown Temp : 92 C
GPU Slowdown Temp : 89 C
GPU Max Operating Temp : 85 C
GPU Target Temperature : N/A
Memory Current Temp : 40 C
Memory Max Operating Temp : 95 C
Power Readings
Power Management : Supported
Power Draw : 57.45 W
Power Limit : 400.00 W
Default Power Limit : 400.00 W
Enforced Power Limit : 400.00 W
Min Power Limit : 100.00 W
Max Power Limit : 400.00 W
Clocks
Graphics : 915 MHz
SM : 915 MHz
Memory : 1215 MHz
Video : 780 MHz
Applications Clocks
Graphics : 1095 MHz
Memory : 1215 MHz
Default Applications Clocks
Graphics : 1095 MHz
Memory : 1215 MHz
Max Clocks
Graphics : 1410 MHz
SM : 1410 MHz
Memory : 1215 MHz
Video : 1290 MHz
Max Customer Boost Clocks
Graphics : 1410 MHz
Clock Policy
Auto Boost : N/A
Auto Boost Default : N/A
Voltage
Graphics : 700.000 mV
Processes : None
GPU 00000000:81:00.0
Product Name : NVIDIA A100-SXM4-40GB
Product Brand : NVIDIA
Product Architecture : Ampere
Display Mode : Disabled
Display Active : Disabled
Persistence Mode : Disabled
MIG Mode
Current : Disabled
Pending : Disabled
Accounting Mode : Disabled
Accounting Mode Buffer Size : 4000
Driver Model
Current : N/A
Pending : N/A
Serial Number : 1561221015040
GPU UUID : GPU-7e4b55d2-75fc-8ab5-e212-09e69e84704b
Minor Number : 2
VBIOS Version : 92.00.19.00.13
MultiGPU Board : No
Board ID : 0x8100
GPU Part Number : 692-2G506-0202-002
Module ID : 2
Inforom Version
Image Version : G506.0202.00.02
OEM Object : 2.0
ECC Object : 6.16
Power Management Object : N/A
GPU Operation Mode
Current : N/A
Pending : N/A
GSP Firmware Version : 515.48.07
GPU Virtualization Mode
Virtualization Mode : None
Host VGPU Mode : N/A
IBMNPU
Relaxed Ordering Mode : N/A
Bus : 0x81
Device : 0x00
Domain : 0x0000
Device Id : 0x20B010DE
Bus Id : 00000000:81:00.0
Sub System Id : 0x144E10DE
GPU Link Info
PCIe Generation
Max : 4
Current : 4
Link Width
Max : 16x
Current : 16x
Bridge Chip
Type : N/A
Firmware : N/A
Replays Since Reset : 0
Replay Number Rollovers : 0
Tx Throughput : 0 KB/s
Rx Throughput : 0 KB/s
Fan Speed : N/A
Performance State : P0
Clocks Throttle Reasons
Idle : Active
Applications Clocks Setting : Not Active
SW Power Cap : Not Active
HW Slowdown : Not Active
HW Thermal Slowdown : Not Active
HW Power Brake Slowdown : Not Active
Sync Boost : Not Active
SW Thermal Slowdown : Not Active
Display Clock Setting : Not Active
FB Memory Usage
Total : 40960 MiB
Reserved : 571 MiB
Used : 0 MiB
Free : 40388 MiB
BAR1 Memory Usage
Total : 65536 MiB
Used : 1 MiB
Free : 65535 MiB
Compute Mode : Default
Utilization
Gpu : 0 %
Memory : 0 %
Encoder : 0 %
Decoder : 0 %
Encoder Stats
Active Sessions : 0
Average FPS : 0
Average Latency : 0
FBC Stats
Active Sessions : 0
Average FPS : 0
Average Latency : 0
Ecc Mode
Current : Enabled
Pending : Enabled
ECC Errors
Volatile
SRAM Correctable : 0
SRAM Uncorrectable : 0
DRAM Correctable : 0
DRAM Uncorrectable : 0
Aggregate
SRAM Correctable : 0
SRAM Uncorrectable : 0
DRAM Correctable : 0
DRAM Uncorrectable : 0
Retired Pages
Single Bit ECC : N/A
Double Bit ECC : N/A
Pending Page Blacklist : N/A
Remapped Rows
Correctable Error : 0
Uncorrectable Error : 0
Pending : No
Remapping Failure Occurred : No
Bank Remap Availability Histogram
Max : 640 bank(s)
High : 0 bank(s)
Partial : 0 bank(s)
Low : 0 bank(s)
None : 0 bank(s)
Temperature
GPU Current Temp : 32 C
GPU Shutdown Temp : 92 C
GPU Slowdown Temp : 89 C
GPU Max Operating Temp : 85 C
GPU Target Temperature : N/A
Memory Current Temp : 33 C
Memory Max Operating Temp : 95 C
Power Readings
Power Management : Supported
Power Draw : 54.65 W
Power Limit : 400.00 W
Default Power Limit : 400.00 W
Enforced Power Limit : 400.00 W
Min Power Limit : 100.00 W
Max Power Limit : 400.00 W
Clocks
Graphics : 1080 MHz
SM : 1080 MHz
Memory : 1215 MHz
Video : 975 MHz
Applications Clocks
Graphics : 1095 MHz
Memory : 1215 MHz
Default Applications Clocks
Graphics : 1095 MHz
Memory : 1215 MHz
Max Clocks
Graphics : 1410 MHz
SM : 1410 MHz
Memory : 1215 MHz
Video : 1290 MHz
Max Customer Boost Clocks
Graphics : 1410 MHz
Clock Policy
Auto Boost : N/A
Auto Boost Default : N/A
Voltage
Graphics : 712.500 mV
Processes : None
GPU 00000000:C1:00.0
Product Name : NVIDIA A100-SXM4-40GB
Product Brand : NVIDIA
Product Architecture : Ampere
Display Mode : Disabled
Display Active : Disabled
Persistence Mode : Disabled
MIG Mode
Current : Disabled
Pending : Disabled
Accounting Mode : Disabled
Accounting Mode Buffer Size : 4000
Driver Model
Current : N/A
Pending : N/A
Serial Number : 1561221014695
GPU UUID : GPU-66ba085a-5496-d204-d4da-aa9f112d3fd8
Minor Number : 3
VBIOS Version : 92.00.19.00.13
MultiGPU Board : No
Board ID : 0xc100
GPU Part Number : 692-2G506-0202-002
Module ID : 0
Inforom Version
Image Version : G506.0202.00.02
OEM Object : 2.0
ECC Object : 6.16
Power Management Object : N/A
GPU Operation Mode
Current : N/A
Pending : N/A
GSP Firmware Version : 515.48.07
GPU Virtualization Mode
Virtualization Mode : None
Host VGPU Mode : N/A
IBMNPU
Relaxed Ordering Mode : N/A
Bus : 0xC1
Device : 0x00
Domain : 0x0000
Device Id : 0x20B010DE
Bus Id : 00000000:C1:00.0
Sub System Id : 0x144E10DE
GPU Link Info
PCIe Generation
Max : 4
Current : 4
Link Width
Max : 16x
Current : 16x
Bridge Chip
Type : N/A
Firmware : N/A
Replays Since Reset : 0
Replay Number Rollovers : 0
Tx Throughput : 0 KB/s
Rx Throughput : 0 KB/s
Fan Speed : N/A
Performance State : P0
Clocks Throttle Reasons
Idle : Active
Applications Clocks Setting : Not Active
SW Power Cap : Not Active
HW Slowdown : Not Active
HW Thermal Slowdown : Not Active
HW Power Brake Slowdown : Not Active
Sync Boost : Not Active
SW Thermal Slowdown : Not Active
Display Clock Setting : Not Active
FB Memory Usage
Total : 40960 MiB
Reserved : 571 MiB
Used : 0 MiB
Free : 40388 MiB
BAR1 Memory Usage
Total : 65536 MiB
Used : 1 MiB
Free : 65535 MiB
Compute Mode : Default
Utilization
Gpu : 0 %
Memory : 0 %
Encoder : 0 %
Decoder : 0 %
Encoder Stats
Active Sessions : 0
Average FPS : 0
Average Latency : 0
FBC Stats
Active Sessions : 0
Average FPS : 0
Average Latency : 0
Ecc Mode
Current : Enabled
Pending : Enabled
ECC Errors
Volatile
SRAM Correctable : 0
SRAM Uncorrectable : 0
DRAM Correctable : 0
DRAM Uncorrectable : 0
Aggregate
SRAM Correctable : 0
SRAM Uncorrectable : 0
DRAM Correctable : 0
DRAM Uncorrectable : 0
Retired Pages
Single Bit ECC : N/A
Double Bit ECC : N/A
Pending Page Blacklist : N/A
Remapped Rows
Correctable Error : 0
Uncorrectable Error : 0
Pending : No
Remapping Failure Occurred : No
Bank Remap Availability Histogram
Max : 640 bank(s)
High : 0 bank(s)
Partial : 0 bank(s)
Low : 0 bank(s)
None : 0 bank(s)
Temperature
GPU Current Temp : 30 C
GPU Shutdown Temp : 92 C
GPU Slowdown Temp : 89 C
GPU Max Operating Temp : 85 C
GPU Target Temperature : N/A
Memory Current Temp : 38 C
Memory Max Operating Temp : 95 C
Power Readings
Power Management : Supported
Power Draw : 59.01 W
Power Limit : 400.00 W
Default Power Limit : 400.00 W
Enforced Power Limit : 400.00 W
Min Power Limit : 100.00 W
Max Power Limit : 400.00 W
Clocks
Graphics : 1080 MHz
SM : 1080 MHz
Memory : 1215 MHz
Video : 975 MHz
Applications Clocks
Graphics : 1095 MHz
Memory : 1215 MHz
Default Applications Clocks
Graphics : 1095 MHz
Memory : 1215 MHz
Max Clocks
Graphics : 1410 MHz
SM : 1410 MHz
Memory : 1215 MHz
Video : 1290 MHz
Max Customer Boost Clocks
Graphics : 1410 MHz
Clock Policy
Auto Boost : N/A
Auto Boost Default : N/A
Voltage
Graphics : 737.500 mV
Processes : None
docker version
Client: Docker Engine - Community
Version: 20.10.17
API version: 1.41
Go version: go1.17.11
Git commit: 100c701
Built: Mon Jun 6 23:02:56 2022
OS/Arch: linux/amd64
Context: default
Experimental: true
Server: Docker Engine - Community
Engine:
Version: 20.10.11
API version: 1.41 (minimum version 1.12)
Go version: go1.16.9
Git commit: 847da18
Built: Thu Nov 18 00:35:16 2021
OS/Arch: linux/amd64
Experimental: false
containerd:
Version: 1.6.6
GitCommit: 10c12954828e7c7c9b6e0ea9b0c02b01407d3ae1
nvidia:
Version: 1.1.2
GitCommit: v1.1.2-0-ga916309
docker-init:
Version: 0.19.0
GitCommit: de40ad0
NVIDIA container library version
cli-version: 1.10.0
lib-version: 1.10.0
build date: 2022-06-13T10:39+00:00
build revision: 395fd41701117121f1fd04ada01e1d7e006a37ae
build compiler: x86_64-linux-gnu-gcc-7 7.5.0
build platform: x86_64
build flags: -D_GNU_SOURCE -D_FORTIFY_SOURCE=2 -DNDEBUG -std=gnu11 -O2 -g -fdata-sections -ffunction-sections -fplan9-extensions -fstack-protector -fno-strict-aliasing -fvisibility=hidden -Wall -Wextra -Wcast-align -Wpointer-arith -Wmissing-prototypes -Wnonnull -Wwrite-strings -Wlogical-op -Wformat=2 -Wmissing-format-attribute -Winit-self -Wshadow -Wstrict-prototypes -Wunreachable-code -Wconversion -Wsign-conversion -Wno-unknown-warning-option -Wno-format-extra-args -Wno-gnu-alignof-expression -Wl,-zrelro -Wl,-znow -Wl,-zdefs -Wl,--gc-sections
NVIDIA packages version
Desired=Unknown/Install/Remove/Purge/Hold
| Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
|/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
||/ Name Version Architecture Description
+++-====================================-=======================-=======================-=============================================================================
ii libnvidia-container-tools 1.10.0-1 amd64 NVIDIA container runtime library (command-line tools)
ii libnvidia-container1:amd64 1.10.0-1 amd64 NVIDIA container runtime library
un nvidia-container-runtime <none> <none> (no description available)
un nvidia-container-runtime-hook <none> <none> (no description available)
ii nvidia-container-toolkit 1.10.0-1 amd64 NVIDIA container runtime hook
un nvidia-docker <none> <none> (no description available)
ii nvidia-docker2 2.7.0-1 all nvidia-docker CLI wrapper
@elezar i make nvidia-container-toolkit=1.9.0-1
,but output remain the same.
Details are as follows
Desired=Unknown/Install/Remove/Purge/Hold
| Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
|/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
||/ Name Version Architecture Description
+++-====================================-=======================-=======================-=============================================================================
ii libnvidia-container-tools 1.10.0-1 amd64 NVIDIA container runtime library (command-line tools)
ii libnvidia-container1:amd64 1.10.0-1 amd64 NVIDIA container runtime library
un nvidia-container-runtime <none> <none> (no description available)
un nvidia-container-runtime-hook <none> <none> (no description available)
ii nvidia-container-toolkit 1.9.0-1 amd64 NVIDIA container runtime hook
un nvidia-docker <none> <none> (no description available)
ii nvidia-docker2 2.7.0-1 all nvidia-docker CLI wrapper
gpu-server@gpu-server:~$ sudo docker run --rm --gpus all nvidia/cuda:11.0.3-base-ubuntu20.04 nvidia-smi
docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.
@elezar this the same error
gpu-server@gpu-server:~$ dpkg -l '*nvidia*'
Desired=Unknown/Install/Remove/Purge/Hold
| Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
|/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
||/ Name Version Architecture Description
+++-====================================-=======================-=======================-=============================================================================
ii libnvidia-container-tools 1.9.0-1 amd64 NVIDIA container runtime library (command-line tools)
ii libnvidia-container1:amd64 1.9.0-1 amd64 NVIDIA container runtime library
un nvidia-container-runtime <none> <none> (no description available)
un nvidia-container-runtime-hook <none> <none> (no description available)
ii nvidia-container-toolkit 1.9.0-1 amd64 NVIDIA container runtime hook
un nvidia-docker <none> <none> (no description available)
ii nvidia-docker2 2.7.0-1 all nvidia-docker CLI wrapper
gpu-server@gpu-server:~$ systemctl restart docker
==== AUTHENTICATING FOR org.freedesktop.systemd1.manage-units ===
Authentication is required to restart 'docker.service'.
Authenticating as: gpu-server
Password:
==== AUTHENTICATION COMPLETE ===
gpu-server@gpu-server:~$ sudo docker run --rm --gpus all nvidia/cuda:11.0.3-base-ubuntu20.04 nvidia-smi
docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.
What happens when you run nvidia-smi
directly on the host?
Also, this warning in the log seems suspicious:
W0621 07:07:51.739205 4791 nvc.c:273] failed to set inheritable capabilities
Is there something unconventional about how you are running docker on your host? For example, rootless, in a snap sandbox, etc.
@klueska
when i run this nvidia-smi on the host, it will two minutes for expected results
gpu-server@gpu-server:~$ nvidia-smi
Tue Jun 21 13:47:51 2022
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 515.48.07 Driver Version: 515.48.07 CUDA Version: 11.7 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|===============================+======================+======================|
| 0 NVIDIA A100-SXM... Off | 00000000:01:00.0 Off | 0 |
| N/A 32C P0 52W / 400W | 0MiB / 40960MiB | 0% Default |
| | | Disabled |
+-------------------------------+----------------------+----------------------+
| 1 NVIDIA A100-SXM... Off | 00000000:41:00.0 Off | 0 |
| N/A 30C P0 53W / 400W | 0MiB / 40960MiB | 0% Default |
| | | Disabled |
+-------------------------------+----------------------+----------------------+
| 2 NVIDIA A100-SXM... Off | 00000000:81:00.0 Off | 0 |
| N/A 31C P0 52W / 400W | 0MiB / 40960MiB | 0% Default |
| | | Disabled |
+-------------------------------+----------------------+----------------------+
| 3 NVIDIA A100-SXM... Off | 00000000:C1:00.0 Off | 0 |
| N/A 30C P0 52W / 400W | 0MiB / 40960MiB | 0% Default |
| | | Disabled |
+-------------------------------+----------------------+----------------------+
+-----------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=============================================================================|
| No running processes found |
+-----------------------------------------------------------------------------+
i run docker in root model and i used sudo gpasswd -a gpu-server docker
, then run docker ,in addition i did not installed docker with snap.
If it’s taking that long to get a result from nvidia-smi on the host then I could understand why the RPC might time out in the nvidia-container-cli when trying to get a result from the driver RPC call.
Im assuming this is happening because (1) you are not running the nvidia-persistenced daemon and (2) your GPUs are not in persistence mode.
Both (1) and (2) achieve the same thing, but (1) is the preferred method to keep the GPU driver alive even when no clients are attached.
Try enabling one of these methods and report back.
XksA-me, zkytony, z-q-y, alkavan, juliusHuelsmann, suamin, fahnub, PPPPPYJ, FruiteePro, and krevas reacted with thumbs up emoji
alkavan, Cyebukayire, maykonpacheco, PPPPPYJ, FranLucchini, and krevas reacted with heart emoji
alkavan and maykonpacheco reacted with rocket emoji
All reactions
@klueska Yeah dude thanks =)
That's exactly it, running nvidia-persistenced
as root
without any additional parameters solved it.
You can confirm with nvidia-smi
that the GPU units load faster.
docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed
#1710
If it’s taking that long to get a result from nvidia-smi on the host then I could understand why the RPC might time out in the nvidia-container-cli when trying to get a result from the driver RPC call.
Im assuming this is happening because (1) you are not running the nvidia-persistenced daemon and (2) your GPUs are not in persistence mode.
Both (1) and (2) achieve the same thing, but (1) is the preferred method to keep the GPU driver alive even when no clients are attached.
Try enabling one of these methods and report back.
I tried using nvidia-smi -i 0 -pm ENABLED
but not luck. I'm having the same error as @zyr-NULL. I then tried using:
# nvidia-persistenced --user aurelien
nvidia-persistenced failed to initialize. Check syslog for more details.
Here are the logs:
# cat /var/log/syslog
Feb 6 14:15:46 pop-os nvidia-persistenced: device 0000:01:00.0 - persistence mode disabled.
Feb 6 14:15:46 pop-os nvidia-persistenced: device 0000:01:00.0 - NUMA memory offlined.
Feb 6 14:16:06 pop-os nvidia-persistenced: Failed to change ownership of /var/run/nvidia-persistenced: Operation not permitted
Feb 6 14:16:06 pop-os nvidia-persistenced: Shutdown (16491)
Feb 6 14:16:38 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb 6 14:16:38 pop-os nvidia-persistenced: Shutdown (16516)
Feb 6 14:16:46 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb 6 14:16:48 pop-os rtkit-daemon[1589]: message repeated 5 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb 6 14:17:01 pop-os nvidia-persistenced: device 0000:01:00.0 - persistence mode enabled.
Feb 6 14:17:01 pop-os nvidia-persistenced: device 0000:01:00.0 - NUMA memory onlined.
Feb 6 14:17:01 pop-os CRON[16652]: (root) CMD ( cd / && run-parts --report /etc/cron.hourly)
Feb 6 14:17:03 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb 6 14:17:03 pop-os nvidia-persistenced: Shutdown (16656)
Feb 6 14:18:03 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb 6 14:18:04 pop-os rtkit-daemon[1589]: message repeated 3 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb 6 14:18:33 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb 6 14:18:33 pop-os nvidia-persistenced: Shutdown (16799)
Feb 6 14:18:38 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb 6 14:18:53 pop-os rtkit-daemon[1589]: message repeated 7 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb 6 14:19:17 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb 6 14:19:17 pop-os nvidia-persistenced: Shutdown (16994)
Feb 6 14:20:04 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb 6 14:20:07 pop-os rtkit-daemon[1589]: message repeated 5 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb 6 14:20:56 pop-os systemd[1]: Starting Cleanup of Temporary Directories...
Feb 6 14:20:56 pop-os systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully.
Feb 6 14:20:56 pop-os systemd[1]: Finished Cleanup of Temporary Directories.
Feb 6 14:21:03 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb 6 14:21:07 pop-os rtkit-daemon[1589]: message repeated 3 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb 6 14:21:15 pop-os systemd[1]: Reloading.
Feb 6 14:23:27 pop-os chronyd[1007]: NTS-KE session with 52.10.183.132:4460 (oregon.time.system76.com) timed out
Feb 6 14:23:33 pop-os chronyd[1007]: NTS-KE session with 18.228.202.30:4460 (brazil.time.system76.com) timed out
Feb 6 14:23:33 pop-os chronyd[1007]: NTS-KE session with 15.237.97.214:4460 (paris.time.system76.com) timed out
Feb 6 14:23:34 pop-os chronyd[1007]: NTS-KE session with 3.134.129.152:4460 (ohio.time.system76.com) timed out
Feb 6 14:23:36 pop-os chronyd[1007]: NTS-KE session with 3.220.42.39:4460 (virginia.time.system76.com) timed out
Feb 6 14:24:21 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb 6 14:24:21 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb 6 14:25:04 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb 6 14:25:04 pop-os nvidia-persistenced: Shutdown (17372)
Any ideas @klueska? 🥲
Here's the error by the way:
$ sudo docker run --rm --gpus all nvidia/cuda:11.6.2-base-ubuntu20.04 nvidia-smi
docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: container error: cgroup subsystem devices not found: unknown.
I met this problem too. My GPU is A30 and GPU driver is 525.85.12.
State: Terminated
Reason: StartError
Message: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: detection error: nvml error: unknown error: unknown
[Tue Feb 21 21:10:41 2023] NVRM: Xid (PCI:0000:22:00): 119, pid=3660230, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
@fighterhit this may be related to changes in the GSP firmware paths and should be addressed in the v1.13.0-rc.1 release. Would you be able to give it a try?
Hi @elezar , yes I can, but do I need to restart the node or the containers on the node? Because it may need to be deployed in our production environment.
Or does this problem exist in previous versions, I can also accept lower versions :).
Updating the nvidia-container-toolkit
should not require a restart. The components of the NVIDIA container stack are only invoked when containers are created and would not affect running containers.
If this is the firmware issue that I think it is, then a lower version would not work.
What is the output of
ls -al /firmware/nvidia/525.85*/*
If there is a single gsp.bin
file there, then a v1.13
release is not required and any NVIDIA Container Toolkit version after v1.8.1
should work. If there are multiple gsp_*.bin
files there then v1.13.0-rc.1
would be required to address this.
Updating the nvidia-container-toolkit
should not require a restart. The components of the NVIDIA container stack are only invoked when containers are created and would not affect running containers.
If this is the firmware issue that I think it is, then a lower version would not work.
What is the output of
ls -al /firmware/nvidia/525.85*/*
If there is a single gsp.bin
file there, then a v1.13
release is not required and any NVIDIA Container Toolkit version after v1.8.1
should work. If there are multiple gsp_*.bin
files there then v1.13.0-rc.1
would be required to address this.
@elezar The output is:
ls: cannot access '/firmware/nvidia/525.85*/*': No such file or directory
It seems that there is no such /firmware
directory on our all GPU nodes. FYI, our driver is installed by NVIDIA-Linux-x86_64-525.85.12.run
.
Updating the nvidia-container-toolkit
should not require a restart. The components of the NVIDIA container stack are only invoked when containers are created and would not affect running containers.
If this is the firmware issue that I think it is, then a lower version would not work.
What is the output of
ls -al /firmware/nvidia/525.85*/*
If there is a single gsp.bin
file there, then a v1.13
release is not required and any NVIDIA Container Toolkit version after v1.8.1
should work. If there are multiple gsp_*.bin
files there then v1.13.0-rc.1
would be required to address this.
@elezar The output is:
ls: cannot access '/firmware/nvidia/525.85*/*': No such file or directory
It seems that there is no such /firmware
directory on our all GPU nodes. FYI, our driver is installed by NVIDIA-Linux-x86_64-525.85.12.run
.
Hi @elezar , do you have any more suggestions? Thanks!
The path is /lib/firmware/…
Thanks @klueska 😅 , there are indeed two gsp_*.bin
files in this directory.
-r--r--r-- 1 root root 25M Feb 18 11:34 /lib/firmware/nvidia/525.85.12/gsp_ad10x.bin
-r--r--r-- 1 root root 36M Feb 18 11:34 /lib/firmware/nvidia/525.85.12/gsp_tu10x.bin
The new RC adds support for detecting multiple GSP firmware files, which is required for container support to work correctly on the 525 driver.
The persistenced issue is still relevant but this is a new one related to the latest NVIDIA driver.
Thanks @klueska , how can I install this version v1.13.0-rc.1
? I searched by using apt-cache show nvidia-container-toolkit
but there seems to be no package for this version.
The persistenced issue is still relevant but this is a new one related to the latest NVIDIA driver.
Does the persistenced issue only appear on the latest driver (525.85.12)? Can I solve it by downgrading to a certain driver version?
Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.
That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.
Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.
Instructions here:
https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit
Thank you for your help @klueska. What confuses me is that in the past, our GPU clusters (1080Ti, 2080Ti, 3090, A30, A100) have not turned on the persistent mode and there is no such problem, only A30 nodes (with driver 525.85.12) have this problem, none of the other types of GPU nodes(1080Ti, 2080Ti, 3090 with driver 525.78.01) have this problem.
Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.
That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.
Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.
Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit
Hi @klueska @elezar , when I configure to use the experimental repo I get the following error and can't install the v1.13.0-rc.1
, my system distribution is debian11.
root@xxx:/etc/apt/sources.list.d# distribution=$(. /etc/os-release;echo $ID$VERSION_ID) \
&& curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \
&& curl -s -L https://nvidia.github.io/libnvidia-container/experimental/$distribution/libnvidia-container.list | \
sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
# Unsupported distribution!
# Check https://nvidia.github.io/libnvidia-container
Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.
That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.
Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.
Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit
Hi @klueska @elezar , when I configure to use the experimental repo I get the following error and can't install the v1.13.0-rc.1
, my system distribution is debian11.
root@xxx:/etc/apt/sources.list.d# distribution=$(. /etc/os-release;echo $ID$VERSION_ID) \
&& curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \
&& curl -s -L https://nvidia.github.io/libnvidia-container/experimental/$distribution/libnvidia-container.list | \
sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
# Unsupported distribution!
# Check https://nvidia.github.io/libnvidia-container
@fighterhit what distribution are you using? Please ensure that the distribution
variable is set to ubuntu18.04 and try again.
Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.
That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.
Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.
Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit
Hi @klueska @elezar , when I configure to use the experimental repo I get the following error and can't install the v1.13.0-rc.1
, my system distribution is debian11.
root@xxx:/etc/apt/sources.list.d# distribution=$(. /etc/os-release;echo $ID$VERSION_ID) \
&& curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \
&& curl -s -L https://nvidia.github.io/libnvidia-container/experimental/$distribution/libnvidia-container.list | \
sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
# Unsupported distribution!
# Check https://nvidia.github.io/libnvidia-container
@fighterhit what distribution are you using? Please ensure that the distribution
variable is set to ubuntu18.04 and try again.
@elezar My distribution is Debian GNU/Linux 11 (bullseye)
.
Hi @klueska @elezar , I test the v1.13.0-rc.1
but it still report some errors:
Warning Failed 12s (x3 over 64s) kubelet Error: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: ldcache error: open failed: /sbin/ldconfig.real: no such file or directory: unknown
Warning BackOff 11s (x3 over 44s) kubelet Back-off restarting failed container
Maybe /sbin/ldconfig.real
should be /sbin/ldconfig
?
For debian, yes, it should be @/sbin/ldconfig
(without the .real
), as seen here:
https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15
This file gets installed under/etc/nvidia-container-runtime/config.toml
The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?
For debian, yes, it should be @/sbin/ldconfig
(without the .real
), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15
This file gets installed under/etc/nvidia-container-runtime/config.toml
The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?
Yes @klueska , I failed to installed the latest version using the experimental repo(#1648 (comment)), so I followed @elezar advice and set the distro to ubuntu18.04
. Can I manually modify /etc/nvidia-container-runtime/config.toml
to make it work? But I'm not sure if this will have any other effects.
For debian, yes, it should be @/sbin/ldconfig
(without the .real
), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15
This file gets installed under/etc/nvidia-container-runtime/config.toml
The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?
Yes @klueska , I failed to installed the latest version using the experimental repo(#1648 (comment)), so I followed @elezar advice and set the distro to ubuntu18.04
. Can I manually modify /etc/nvidia-container-runtime/config.toml
to make it work? But I'm not sure if this will have any other effects.
Sorry about that. I was making an assumption about the distribution you are using. You can install the ubuntu18.04
package and then update the /etc/nvidia-container-runtime/config.toml
value after the fact. This should have no other effects.
For debian, yes, it should be @/sbin/ldconfig
(without the .real
), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15
This file gets installed under/etc/nvidia-container-runtime/config.toml
The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?
Yes @klueska , I failed to installed the latest version using the experimental repo(#1648 (comment)), so I followed @elezar advice and set the distro to ubuntu18.04
. Can I manually modify /etc/nvidia-container-runtime/config.toml
to make it work? But I'm not sure if this will have any other effects.
Sorry about that. I was making an assumption about the distribution you are using. You can install the ubuntu18.04
package and then update the /etc/nvidia-container-runtime/config.toml
value after the fact. This should have no other effects.
Thanks @elezar ! I have tried it and it works fine. It may be better if debian11
could be supported to use the experimental repo.
@fighterhit I have just double-checked our repo configuration and the issue is that for debian10
(to which debian11
redirects) the repo only had a libnvidia-container-experimental.list
and no libnvidia-container.list
file which is referred to by the instructions.
I have created a link / redirect for this now and the official instructions should work as expected. (It may take about 30 minutes for the changes to reflect in the repo though).
@fighterhit I have just double-checked our repo configuration and the issue is that for debian10
(to which debian11
redirects) the repo only had a libnvidia-container-experimental.list
and no libnvidia-container.list
file which is referred to by the instructions.
I have created a link / redirect for this now and the official instructions should work as expected. (It may take about 30 minutes for the changes to reflect in the repo though).
Thanks for your confirmation, it works now. @elezar
I am trying to install NVIDIA container toolkit on Amazon Linux 2, created a new ec2 instance and followed the instructions in this page and running into below issue. I get the same error when i tried this on a Ubuntu instance as well.
nvidia-container-cli: initialization error: load library failed: libnvidia-ml.so.1: cannot open shared object file: no such file or directory: unknown.
ERRO[0000] error waiting for container: context canceled
@elezar @klueska please advise how to fix this issue, appreciate your inputs. Thanks
Hi @elezar @klueska , unfortunately I have used the latest toolkit but this problem reappeared, I think this may be related to the driver, and I asked the driver community for help but got no more reply (https://github.com/NVIDIA/open-gpu-kernel-modules/issues/446), https://forums.developer.nvidia.com/t/timeout-waiting-for-rpc-from-gsp/244789 could you communicate with the driver team about this issue? Because some users in the community also encountered the same problem.Thanks!
[Mon Mar 6 21:09:45 2023] NVRM: GPU at PCI:0000:61:00: GPU-81f5d81e-7906-c145-3def-82e281d7b260
[Mon Mar 6 21:09:45 2023] NVRM: GPU Board Serial Number: 1322621149674
[Mon Mar 6 21:09:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977720, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a00b0 0x0).
[Mon Mar 6 21:09:45 2023] CPU: 15 PID: 977720 Comm: nvidia-smi Tainted: P OE 5.10.0-20-amd64 #1 Debian 5.10.158-2
[Mon Mar 6 21:09:45 2023] Hardware name: Inspur NF5468A5/YZMB-02382-101, BIOS 4.02.12 01/28/2022
[Mon Mar 6 21:09:45 2023] Call Trace:
[Mon Mar 6 21:09:45 2023] dump_stack+0x6b/0x83
[Mon Mar 6 21:09:45 2023] _nv011231rm+0x39d/0x470 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv011168rm+0x62/0x2e0 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv040022rm+0xdb/0x140 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv041148rm+0x2ce/0x3a0 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv015451rm+0x788/0x800 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv039541rm+0xac/0xe0 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv041150rm+0xac/0x140 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv041149rm+0x37a/0x4d0 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv039443rm+0xc9/0x150 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv039444rm+0x42/0x70 [nvidia]
[Mon Mar 6 21:09:45 2023] ? rm_cleanup_file_private+0x128/0x180 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv000554rm+0x49/0x60 [nvidia]
[Mon Mar 6 21:09:45 2023] ? _nv000694rm+0x7fb/0xc80 [nvidia]
[Mon Mar 6 21:09:45 2023] ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar 6 21:09:45 2023] ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar 6 21:09:45 2023] ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar 6 21:09:45 2023] ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar 6 21:09:45 2023] ? do_syscall_64+0x33/0x80
[Mon Mar 6 21:09:45 2023] ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar 6 21:10:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977720, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0050 0x0).
[Mon Mar 6 21:11:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977720, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0040 0x0).
[Mon Mar 6 21:12:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801348 0x410).
[Mon Mar 6 21:12:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x0 0x6c).
[Mon Mar 6 21:13:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x80 0x38).
[Mon Mar 6 21:14:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x2080 0x4).
[Mon Mar 6 21:15:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801348 0x410).
[Mon Mar 6 21:15:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x0 0x6c).
[Mon Mar 6 21:16:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x80 0x38).
[Mon Mar 6 21:17:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=905052, name=python, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801702 0x4).
[Mon Mar 6 21:18:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x2080 0x4).
[Mon Mar 6 21:18:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800110 0x84).
[Mon Mar 6 21:19:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801348 0x410).
[Mon Mar 6 21:20:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:21:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x0 0x6c).
[Mon Mar 6 21:21:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x80 0x38).
[Mon Mar 6 21:22:08 2023] INFO: task python:976158 blocked for more than 120 seconds.
[Mon Mar 6 21:22:08 2023] Tainted: P OE 5.10.0-20-amd64 #1 Debian 5.10.158-2
[Mon Mar 6 21:22:08 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[Mon Mar 6 21:22:08 2023] task:python state:D stack: 0 pid:976158 ppid:835062 flags:0x00000000
[Mon Mar 6 21:22:08 2023] Call Trace:
[Mon Mar 6 21:22:08 2023] __schedule+0x282/0x880
[Mon Mar 6 21:22:08 2023] schedule+0x46/0xb0
[Mon Mar 6 21:22:08 2023] rwsem_down_write_slowpath+0x246/0x4d0
[Mon Mar 6 21:22:08 2023] os_acquire_rwlock_write+0x31/0x40 [nvidia]
[Mon Mar 6 21:22:08 2023] _nv038505rm+0xc/0x30 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv039453rm+0x18d/0x1d0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv041182rm+0x45/0xd0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv041127rm+0x142/0x2b0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv039415rm+0x15a/0x2e0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv039416rm+0x5b/0x90 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv039416rm+0x31/0x90 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv012688rm+0x1d/0x30 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv039431rm+0xb0/0xb0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv012710rm+0x54/0x70 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv011426rm+0xc4/0x120 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv000659rm+0x63/0x70 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv000582rm+0x2c/0x40 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv000694rm+0x86c/0xc80 [nvidia]
[Mon Mar 6 21:22:08 2023] ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar 6 21:22:08 2023] ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar 6 21:22:08 2023] ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar 6 21:22:08 2023] ? do_syscall_64+0x33/0x80
[Mon Mar 6 21:22:08 2023] ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar 6 21:22:08 2023] INFO: task nvidia-smi:993195 blocked for more than 120 seconds.
[Mon Mar 6 21:22:08 2023] Tainted: P OE 5.10.0-20-amd64 #1 Debian 5.10.158-2
[Mon Mar 6 21:22:08 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[Mon Mar 6 21:22:08 2023] task:nvidia-smi state:D stack: 0 pid:993195 ppid: 58577 flags:0x00000004
[Mon Mar 6 21:22:08 2023] Call Trace:
[Mon Mar 6 21:22:08 2023] __schedule+0x282/0x880
[Mon Mar 6 21:22:08 2023] schedule+0x46/0xb0
[Mon Mar 6 21:22:08 2023] rwsem_down_write_slowpath+0x246/0x4d0
[Mon Mar 6 21:22:08 2023] os_acquire_rwlock_write+0x31/0x40 [nvidia]
[Mon Mar 6 21:22:08 2023] _nv038505rm+0xc/0x30 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv039453rm+0x18d/0x1d0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv041182rm+0x45/0xd0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv041127rm+0x142/0x2b0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv039415rm+0x15a/0x2e0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv039416rm+0x5b/0x90 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv039416rm+0x31/0x90 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv000560rm+0x59/0x70 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv000560rm+0x33/0x70 [nvidia]
[Mon Mar 6 21:22:08 2023] ? _nv000694rm+0x4ae/0xc80 [nvidia]
[Mon Mar 6 21:22:08 2023] ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar 6 21:22:08 2023] ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar 6 21:22:08 2023] ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar 6 21:22:08 2023] ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar 6 21:22:08 2023] ? do_syscall_64+0x33/0x80
[Mon Mar 6 21:22:08 2023] ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar 6 21:22:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800157 0x0).
[Mon Mar 6 21:23:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:24:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:24:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:25:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:26:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080852e 0x208).
[Mon Mar 6 21:27:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20808513 0x598).
[Mon Mar 6 21:27:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a068 0x4).
[Mon Mar 6 21:28:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a618 0x181c).
[Mon Mar 6 21:29:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a612 0xd98).
[Mon Mar 6 21:30:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20809009 0x8).
[Mon Mar 6 21:30:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar 6 21:31:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar 6 21:32:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x2080 0x4).
[Mon Mar 6 21:33:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar 6 21:33:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800110 0x84).
[Mon Mar 6 21:34:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:35:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x208f 0x0).
[Mon Mar 6 21:36:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x208f1105 0x8).
[Mon Mar 6 21:36:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800157 0x0).
[Mon Mar 6 21:37:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=905052, name=python, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801702 0x4).
[Mon Mar 6 21:38:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:39:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:39:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a00b0 0x0).
[Mon Mar 6 21:40:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:41:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0050 0x0).
[Mon Mar 6 21:42:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:42:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080852e 0x208).
[Mon Mar 6 21:43:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20808513 0x598).
[Mon Mar 6 21:44:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a068 0x4).
[Mon Mar 6 21:45:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a618 0x181c).
[Mon Mar 6 21:45:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0040 0x0).
[Mon Mar 6 21:46:18 2023] INFO: task python:976158 blocked for more than 120 seconds.
[Mon Mar 6 21:46:18 2023] Tainted: P OE 5.10.0-20-amd64 #1 Debian 5.10.158-2
[Mon Mar 6 21:46:18 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[Mon Mar 6 21:46:18 2023] task:python state:D stack: 0 pid:976158 ppid:835062 flags:0x00000000
[Mon Mar 6 21:46:18 2023] Call Trace:
[Mon Mar 6 21:46:18 2023] __schedule+0x282/0x880
[Mon Mar 6 21:46:18 2023] schedule+0x46/0xb0
[Mon Mar 6 21:46:18 2023] rwsem_down_write_slowpath+0x246/0x4d0
[Mon Mar 6 21:46:18 2023] os_acquire_rwlock_write+0x31/0x40 [nvidia]
[Mon Mar 6 21:46:18 2023] _nv038505rm+0xc/0x30 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv039453rm+0x18d/0x1d0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv041182rm+0x45/0xd0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv041127rm+0x142/0x2b0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv039415rm+0x15a/0x2e0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv039416rm+0x5b/0x90 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv039416rm+0x31/0x90 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv012688rm+0x1d/0x30 [nvidia]
[Mon Mar 6 21:46:18 2023] ? task_numa_fault+0x2a3/0xb70
[Mon Mar 6 21:46:18 2023] ? _nv039431rm+0xb0/0xb0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv012710rm+0x54/0x70 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv011426rm+0xc4/0x120 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv000659rm+0x63/0x70 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv000582rm+0x2c/0x40 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv000694rm+0x86c/0xc80 [nvidia]
[Mon Mar 6 21:46:18 2023] ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar 6 21:46:18 2023] ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar 6 21:46:18 2023] ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar 6 21:46:18 2023] ? do_syscall_64+0x33/0x80
[Mon Mar 6 21:46:18 2023] ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar 6 21:46:18 2023] INFO: task nvidia-smi:980026 blocked for more than 120 seconds.
[Mon Mar 6 21:46:18 2023] Tainted: P OE 5.10.0-20-amd64 #1 Debian 5.10.158-2
[Mon Mar 6 21:46:18 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[Mon Mar 6 21:46:18 2023] task:nvidia-smi state:D stack: 0 pid:980026 ppid:980025 flags:0x00000000
[Mon Mar 6 21:46:18 2023] Call Trace:
[Mon Mar 6 21:46:18 2023] __schedule+0x282/0x880
[Mon Mar 6 21:46:18 2023] ? psi_task_change+0x88/0xd0
[Mon Mar 6 21:46:18 2023] schedule+0x46/0xb0
[Mon Mar 6 21:46:18 2023] rwsem_down_read_slowpath+0x18e/0x500
[Mon Mar 6 21:46:18 2023] os_acquire_rwlock_read+0x31/0x40 [nvidia]
[Mon Mar 6 21:46:18 2023] _nv038503rm+0xc/0x30 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv039453rm+0x64/0x1d0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv041182rm+0x45/0xd0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv041133rm+0xfd/0x2b0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv012728rm+0x59a/0x690 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv039431rm+0x53/0xb0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv011404rm+0x52/0xa0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? _nv000694rm+0x5ae/0xc80 [nvidia]
[Mon Mar 6 21:46:18 2023] ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar 6 21:46:18 2023] ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar 6 21:46:18 2023] ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar 6 21:46:18 2023] ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar 6 21:46:18 2023] ? do_syscall_64+0x33/0x80
[Mon Mar 6 21:46:18 2023] ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar 6 21:46:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a612 0xd98).
[Mon Mar 6 21:47:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20809009 0x8).
[Mon Mar 6 21:48:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar 6 21:48:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar 6 21:49:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar 6 21:50:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800110 0x84).
[Mon Mar 6 21:51:02 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar 6 21:51:47 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800157 0x0).
[Mon Mar 6 21:52:32 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x208f 0x0).
[Mon Mar 6 21:53:17 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x208f1105 0x8).
[Mon Mar 6 21:54:02 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
I was getting similar errors trying to run nvidia-smi
after updating my Ubuntu 22.04 system.
I noticed that sudo systemctl status nvidia-persistenced
was not enabled, so I tried to enable it with sudo systemctl enable nvidia-persistenced
but that failed as well.
I then ran across the following forum post which led me to think that perhaps my drivers had somehow been disabled/corrupted by a recent system and/or kernel upgrade so I decided to reinstall (and upgrade) the nvidia driver using the PPA, and nvidia-smi
is working again. Hope this helps someone else...