相关文章推荐

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement . We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed #1648 docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed #1648 zyr-NULL opened this issue Jun 21, 2022 · 40 comments

Issue or feature description

when i use docker to create container, i get this error

docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.

Steps to reproduce the issue

  • when i executed the following command
    sudo docker run --rm --gpus all nvidia/cuda:11.0.3-base-ubuntu20.04 nvidia-smi
  • i get the following error
    docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy' nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.
  • but when i executed this following command, it has the expected output
    sudo docker run hello-world
  • here is some Information

  • Some nvidia-container information
  • gpu-server@gpu-server:~$ nvidia-container-cli -k -d /dev/tty info
    -- WARNING, the following logs are for debugging purposes only --
    I0621 07:07:51.735875 4789 nvc.c:376] initializing library context (version=1.10.0, build=395fd41701117121f1fd04ada01e1d7e006a37ae)
    I0621 07:07:51.735941 4789 nvc.c:350] using root /
    I0621 07:07:51.735947 4789 nvc.c:351] using ldcache /etc/ld.so.cache
    I0621 07:07:51.735963 4789 nvc.c:352] using unprivileged user 1000:1000
    I0621 07:07:51.735984 4789 nvc.c:393] attempting to load dxcore to see if we are running under Windows Subsystem for Linux (WSL)
    I0621 07:07:51.736064 4789 nvc.c:395] dxcore initialization failed, continuing assuming a non-WSL environment
    W0621 07:07:51.739205 4791 nvc.c:273] failed to set inheritable capabilities
    W0621 07:07:51.739329 4791 nvc.c:274] skipping kernel modules load due to failure
    I0621 07:07:51.739807 4793 rpc.c:71] starting driver rpc service
    W0621 07:08:16.774958 4789 rpc.c:121] terminating driver rpc service (forced)
    I0621 07:08:20.481845 4789 rpc.c:135] driver rpc service terminated with signal 15
    nvidia-container-cli: initialization error: driver rpc error: timed out
    I0621 07:08:20.481972 4789 nvc.c:434] shutting down library context
    

    Kernel version
    Linux gpu-server 4.15.0-187-generic #198-Ubuntu SMP Tue Jun 14 03:23:51 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux

    Driver information

    ==============NVSMI LOG==============
    Timestamp                                 : Tue Jun 21 07:13:57 2022
    Driver Version                            : 515.48.07
    CUDA Version                              : 11.7
    Attached GPUs                             : 4
    GPU 00000000:01:00.0
        Product Name                          : NVIDIA A100-SXM4-40GB
        Product Brand                         : NVIDIA
        Product Architecture                  : Ampere
        Display Mode                          : Disabled
        Display Active                        : Disabled
        Persistence Mode                      : Disabled
        MIG Mode
            Current                           : Disabled
            Pending                           : Disabled
        Accounting Mode                       : Disabled
        Accounting Mode Buffer Size           : 4000
        Driver Model
            Current                           : N/A
            Pending                           : N/A
        Serial Number                         : 1561221014674
        GPU UUID                              : GPU-b67da01e-feba-d839-62c5-2773d4e963f0
        Minor Number                          : 0
        VBIOS Version                         : 92.00.19.00.13
        MultiGPU Board                        : No
        Board ID                              : 0x100
        GPU Part Number                       : 692-2G506-0202-002
        Module ID                             : 3
        Inforom Version
            Image Version                     : G506.0202.00.02
            OEM Object                        : 2.0
            ECC Object                        : 6.16
            Power Management Object           : N/A
        GPU Operation Mode
            Current                           : N/A
            Pending                           : N/A
        GSP Firmware Version                  : 515.48.07
        GPU Virtualization Mode
            Virtualization Mode               : None
            Host VGPU Mode                    : N/A
        IBMNPU
            Relaxed Ordering Mode             : N/A
            Bus                               : 0x01
            Device                            : 0x00
            Domain                            : 0x0000
            Device Id                         : 0x20B010DE
            Bus Id                            : 00000000:01:00.0
            Sub System Id                     : 0x144E10DE
            GPU Link Info
                PCIe Generation
                    Max                       : 4
                    Current                   : 4
                Link Width
                    Max                       : 16x
                    Current                   : 16x
            Bridge Chip
                Type                          : N/A
                Firmware                      : N/A
            Replays Since Reset               : 0
            Replay Number Rollovers           : 0
            Tx Throughput                     : 0 KB/s
            Rx Throughput                     : 0 KB/s
        Fan Speed                             : N/A
        Performance State                     : P0
        Clocks Throttle Reasons
            Idle                              : Active
            Applications Clocks Setting       : Not Active
            SW Power Cap                      : Not Active
            HW Slowdown                       : Not Active
                HW Thermal Slowdown           : Not Active
                HW Power Brake Slowdown       : Not Active
            Sync Boost                        : Not Active
            SW Thermal Slowdown               : Not Active
            Display Clock Setting             : Not Active
        FB Memory Usage
            Total                             : 40960 MiB
            Reserved                          : 571 MiB
            Used                              : 0 MiB
            Free                              : 40388 MiB
        BAR1 Memory Usage
            Total                             : 65536 MiB
            Used                              : 1 MiB
            Free                              : 65535 MiB
        Compute Mode                          : Default
        Utilization
            Gpu                               : 0 %
            Memory                            : 0 %
            Encoder                           : 0 %
            Decoder                           : 0 %
        Encoder Stats
            Active Sessions                   : 0
            Average FPS                       : 0
            Average Latency                   : 0
        FBC Stats
            Active Sessions                   : 0
            Average FPS                       : 0
            Average Latency                   : 0
        Ecc Mode
            Current                           : Enabled
            Pending                           : Enabled
        ECC Errors
            Volatile
                SRAM Correctable              : 0
                SRAM Uncorrectable            : 0
                DRAM Correctable              : 0
                DRAM Uncorrectable            : 0
            Aggregate
                SRAM Correctable              : 0
                SRAM Uncorrectable            : 0
                DRAM Correctable              : 0
                DRAM Uncorrectable            : 0
        Retired Pages
            Single Bit ECC                    : N/A
            Double Bit ECC                    : N/A
            Pending Page Blacklist            : N/A
        Remapped Rows
            Correctable Error                 : 0
            Uncorrectable Error               : 0
            Pending                           : No
            Remapping Failure Occurred        : No
            Bank Remap Availability Histogram
                Max                           : 640 bank(s)
                High                          : 0 bank(s)
                Partial                       : 0 bank(s)
                Low                           : 0 bank(s)
                None                          : 0 bank(s)
        Temperature
            GPU Current Temp                  : 32 C
            GPU Shutdown Temp                 : 92 C
            GPU Slowdown Temp                 : 89 C
            GPU Max Operating Temp            : 85 C
            GPU Target Temperature            : N/A
            Memory Current Temp               : 34 C
            Memory Max Operating Temp         : 95 C
        Power Readings
            Power Management                  : Supported
            Power Draw                        : 54.92 W
            Power Limit                       : 400.00 W
            Default Power Limit               : 400.00 W
            Enforced Power Limit              : 400.00 W
            Min Power Limit                   : 100.00 W
            Max Power Limit                   : 400.00 W
        Clocks
            Graphics                          : 1080 MHz
            SM                                : 1080 MHz
            Memory                            : 1215 MHz
            Video                             : 975 MHz
        Applications Clocks
            Graphics                          : 1095 MHz
            Memory                            : 1215 MHz
        Default Applications Clocks
            Graphics                          : 1095 MHz
            Memory                            : 1215 MHz
        Max Clocks
            Graphics                          : 1410 MHz
            SM                                : 1410 MHz
            Memory                            : 1215 MHz
            Video                             : 1290 MHz
        Max Customer Boost Clocks
            Graphics                          : 1410 MHz
        Clock Policy
            Auto Boost                        : N/A
            Auto Boost Default                : N/A
        Voltage
            Graphics                          : 731.250 mV
        Processes                             : None
    GPU 00000000:41:00.0
        Product Name                          : NVIDIA A100-SXM4-40GB
        Product Brand                         : NVIDIA
        Product Architecture                  : Ampere
        Display Mode                          : Disabled
        Display Active                        : Disabled
        Persistence Mode                      : Disabled
        MIG Mode
            Current                           : Disabled
            Pending                           : Disabled
        Accounting Mode                       : Disabled
        Accounting Mode Buffer Size           : 4000
        Driver Model
            Current                           : N/A
            Pending                           : N/A
        Serial Number                         : 1561221014888
        GPU UUID                              : GPU-6ca82e47-c63a-1bea-38ad-d3af9e1dc26b
        Minor Number                          : 1
        VBIOS Version                         : 92.00.19.00.13
        MultiGPU Board                        : No
        Board ID                              : 0x4100
        GPU Part Number                       : 692-2G506-0202-002
        Module ID                             : 1
        Inforom Version
            Image Version                     : G506.0202.00.02
            OEM Object                        : 2.0
            ECC Object                        : 6.16
            Power Management Object           : N/A
        GPU Operation Mode
            Current                           : N/A
            Pending                           : N/A
        GSP Firmware Version                  : 515.48.07
        GPU Virtualization Mode
            Virtualization Mode               : None
            Host VGPU Mode                    : N/A
        IBMNPU
            Relaxed Ordering Mode             : N/A
            Bus                               : 0x41
            Device                            : 0x00
            Domain                            : 0x0000
            Device Id                         : 0x20B010DE
            Bus Id                            : 00000000:41:00.0
            Sub System Id                     : 0x144E10DE
            GPU Link Info
                PCIe Generation
                    Max                       : 4
                    Current                   : 4
                Link Width
                    Max                       : 16x
                    Current                   : 16x
            Bridge Chip
                Type                          : N/A
                Firmware                      : N/A
            Replays Since Reset               : 0
            Replay Number Rollovers           : 0
            Tx Throughput                     : 0 KB/s
            Rx Throughput                     : 0 KB/s
        Fan Speed                             : N/A
        Performance State                     : P0
        Clocks Throttle Reasons
            Idle                              : Active
            Applications Clocks Setting       : Not Active
            SW Power Cap                      : Not Active
            HW Slowdown                       : Not Active
                HW Thermal Slowdown           : Not Active
                HW Power Brake Slowdown       : Not Active
            Sync Boost                        : Not Active
            SW Thermal Slowdown               : Not Active
            Display Clock Setting             : Not Active
        FB Memory Usage
            Total                             : 40960 MiB
            Reserved                          : 571 MiB
            Used                              : 0 MiB
            Free                              : 40388 MiB
        BAR1 Memory Usage
            Total                             : 65536 MiB
            Used                              : 1 MiB
            Free                              : 65535 MiB
        Compute Mode                          : Default
        Utilization
            Gpu                               : 0 %
            Memory                            : 0 %
            Encoder                           : 0 %
            Decoder                           : 0 %
        Encoder Stats
            Active Sessions                   : 0
            Average FPS                       : 0
            Average Latency                   : 0
        FBC Stats
            Active Sessions                   : 0
            Average FPS                       : 0
            Average Latency                   : 0
        Ecc Mode
            Current                           : Enabled
            Pending                           : Enabled
        ECC Errors
            Volatile
                SRAM Correctable              : 0
                SRAM Uncorrectable            : 0
                DRAM Correctable              : 0
                DRAM Uncorrectable            : 0
            Aggregate
                SRAM Correctable              : 0
                SRAM Uncorrectable            : 0
                DRAM Correctable              : 0
                DRAM Uncorrectable            : 0
        Retired Pages
            Single Bit ECC                    : N/A
            Double Bit ECC                    : N/A
            Pending Page Blacklist            : N/A
        Remapped Rows
            Correctable Error                 : 0
            Uncorrectable Error               : 0
            Pending                           : No
            Remapping Failure Occurred        : No
            Bank Remap Availability Histogram
                Max                           : 640 bank(s)
                High                          : 0 bank(s)
                Partial                       : 0 bank(s)
                Low                           : 0 bank(s)
                None                          : 0 bank(s)
        Temperature
            GPU Current Temp                  : 30 C
            GPU Shutdown Temp                 : 92 C
            GPU Slowdown Temp                 : 89 C
            GPU Max Operating Temp            : 85 C
            GPU Target Temperature            : N/A
            Memory Current Temp               : 40 C
            Memory Max Operating Temp         : 95 C
        Power Readings
            Power Management                  : Supported
            Power Draw                        : 57.45 W
            Power Limit                       : 400.00 W
            Default Power Limit               : 400.00 W
            Enforced Power Limit              : 400.00 W
            Min Power Limit                   : 100.00 W
            Max Power Limit                   : 400.00 W
        Clocks
            Graphics                          : 915 MHz
            SM                                : 915 MHz
            Memory                            : 1215 MHz
            Video                             : 780 MHz
        Applications Clocks
            Graphics                          : 1095 MHz
            Memory                            : 1215 MHz
        Default Applications Clocks
            Graphics                          : 1095 MHz
            Memory                            : 1215 MHz
        Max Clocks
            Graphics                          : 1410 MHz
            SM                                : 1410 MHz
            Memory                            : 1215 MHz
            Video                             : 1290 MHz
        Max Customer Boost Clocks
            Graphics                          : 1410 MHz
        Clock Policy
            Auto Boost                        : N/A
            Auto Boost Default                : N/A
        Voltage
            Graphics                          : 700.000 mV
        Processes                             : None
    GPU 00000000:81:00.0
        Product Name                          : NVIDIA A100-SXM4-40GB
        Product Brand                         : NVIDIA
        Product Architecture                  : Ampere
        Display Mode                          : Disabled
        Display Active                        : Disabled
        Persistence Mode                      : Disabled
        MIG Mode
            Current                           : Disabled
            Pending                           : Disabled
        Accounting Mode                       : Disabled
        Accounting Mode Buffer Size           : 4000
        Driver Model
            Current                           : N/A
            Pending                           : N/A
        Serial Number                         : 1561221015040
        GPU UUID                              : GPU-7e4b55d2-75fc-8ab5-e212-09e69e84704b
        Minor Number                          : 2
        VBIOS Version                         : 92.00.19.00.13
        MultiGPU Board                        : No
        Board ID                              : 0x8100
        GPU Part Number                       : 692-2G506-0202-002
        Module ID                             : 2
        Inforom Version
            Image Version                     : G506.0202.00.02
            OEM Object                        : 2.0
            ECC Object                        : 6.16
            Power Management Object           : N/A
        GPU Operation Mode
            Current                           : N/A
            Pending                           : N/A
        GSP Firmware Version                  : 515.48.07
        GPU Virtualization Mode
            Virtualization Mode               : None
            Host VGPU Mode                    : N/A
        IBMNPU
            Relaxed Ordering Mode             : N/A
            Bus                               : 0x81
            Device                            : 0x00
            Domain                            : 0x0000
            Device Id                         : 0x20B010DE
            Bus Id                            : 00000000:81:00.0
            Sub System Id                     : 0x144E10DE
            GPU Link Info
                PCIe Generation
                    Max                       : 4
                    Current                   : 4
                Link Width
                    Max                       : 16x
                    Current                   : 16x
            Bridge Chip
                Type                          : N/A
                Firmware                      : N/A
            Replays Since Reset               : 0
            Replay Number Rollovers           : 0
            Tx Throughput                     : 0 KB/s
            Rx Throughput                     : 0 KB/s
        Fan Speed                             : N/A
        Performance State                     : P0
        Clocks Throttle Reasons
            Idle                              : Active
            Applications Clocks Setting       : Not Active
            SW Power Cap                      : Not Active
            HW Slowdown                       : Not Active
                HW Thermal Slowdown           : Not Active
                HW Power Brake Slowdown       : Not Active
            Sync Boost                        : Not Active
            SW Thermal Slowdown               : Not Active
            Display Clock Setting             : Not Active
        FB Memory Usage
            Total                             : 40960 MiB
            Reserved                          : 571 MiB
            Used                              : 0 MiB
            Free                              : 40388 MiB
        BAR1 Memory Usage
            Total                             : 65536 MiB
            Used                              : 1 MiB
            Free                              : 65535 MiB
        Compute Mode                          : Default
        Utilization
            Gpu                               : 0 %
            Memory                            : 0 %
            Encoder                           : 0 %
            Decoder                           : 0 %
        Encoder Stats
            Active Sessions                   : 0
            Average FPS                       : 0
            Average Latency                   : 0
        FBC Stats
            Active Sessions                   : 0
            Average FPS                       : 0
            Average Latency                   : 0
        Ecc Mode
            Current                           : Enabled
            Pending                           : Enabled
        ECC Errors
            Volatile
                SRAM Correctable              : 0
                SRAM Uncorrectable            : 0
                DRAM Correctable              : 0
                DRAM Uncorrectable            : 0
            Aggregate
                SRAM Correctable              : 0
                SRAM Uncorrectable            : 0
                DRAM Correctable              : 0
                DRAM Uncorrectable            : 0
        Retired Pages
            Single Bit ECC                    : N/A
            Double Bit ECC                    : N/A
            Pending Page Blacklist            : N/A
        Remapped Rows
            Correctable Error                 : 0
            Uncorrectable Error               : 0
            Pending                           : No
            Remapping Failure Occurred        : No
            Bank Remap Availability Histogram
                Max                           : 640 bank(s)
                High                          : 0 bank(s)
                Partial                       : 0 bank(s)
                Low                           : 0 bank(s)
                None                          : 0 bank(s)
        Temperature
            GPU Current Temp                  : 32 C
            GPU Shutdown Temp                 : 92 C
            GPU Slowdown Temp                 : 89 C
            GPU Max Operating Temp            : 85 C
            GPU Target Temperature            : N/A
            Memory Current Temp               : 33 C
            Memory Max Operating Temp         : 95 C
        Power Readings
            Power Management                  : Supported
            Power Draw                        : 54.65 W
            Power Limit                       : 400.00 W
            Default Power Limit               : 400.00 W
            Enforced Power Limit              : 400.00 W
            Min Power Limit                   : 100.00 W
            Max Power Limit                   : 400.00 W
        Clocks
            Graphics                          : 1080 MHz
            SM                                : 1080 MHz
            Memory                            : 1215 MHz
            Video                             : 975 MHz
        Applications Clocks
            Graphics                          : 1095 MHz
            Memory                            : 1215 MHz
        Default Applications Clocks
            Graphics                          : 1095 MHz
            Memory                            : 1215 MHz
        Max Clocks
            Graphics                          : 1410 MHz
            SM                                : 1410 MHz
            Memory                            : 1215 MHz
            Video                             : 1290 MHz
        Max Customer Boost Clocks
            Graphics                          : 1410 MHz
        Clock Policy
            Auto Boost                        : N/A
            Auto Boost Default                : N/A
        Voltage
            Graphics                          : 712.500 mV
        Processes                             : None
    GPU 00000000:C1:00.0
        Product Name                          : NVIDIA A100-SXM4-40GB
        Product Brand                         : NVIDIA
        Product Architecture                  : Ampere
        Display Mode                          : Disabled
        Display Active                        : Disabled
        Persistence Mode                      : Disabled
        MIG Mode
            Current                           : Disabled
            Pending                           : Disabled
        Accounting Mode                       : Disabled
        Accounting Mode Buffer Size           : 4000
        Driver Model
            Current                           : N/A
            Pending                           : N/A
        Serial Number                         : 1561221014695
        GPU UUID                              : GPU-66ba085a-5496-d204-d4da-aa9f112d3fd8
        Minor Number                          : 3
        VBIOS Version                         : 92.00.19.00.13
        MultiGPU Board                        : No
        Board ID                              : 0xc100
        GPU Part Number                       : 692-2G506-0202-002
        Module ID                             : 0
        Inforom Version
            Image Version                     : G506.0202.00.02
            OEM Object                        : 2.0
            ECC Object                        : 6.16
            Power Management Object           : N/A
        GPU Operation Mode
            Current                           : N/A
            Pending                           : N/A
        GSP Firmware Version                  : 515.48.07
        GPU Virtualization Mode
            Virtualization Mode               : None
            Host VGPU Mode                    : N/A
        IBMNPU
            Relaxed Ordering Mode             : N/A
            Bus                               : 0xC1
            Device                            : 0x00
            Domain                            : 0x0000
            Device Id                         : 0x20B010DE
            Bus Id                            : 00000000:C1:00.0
            Sub System Id                     : 0x144E10DE
            GPU Link Info
                PCIe Generation
                    Max                       : 4
                    Current                   : 4
                Link Width
                    Max                       : 16x
                    Current                   : 16x
            Bridge Chip
                Type                          : N/A
                Firmware                      : N/A
            Replays Since Reset               : 0
            Replay Number Rollovers           : 0
            Tx Throughput                     : 0 KB/s
            Rx Throughput                     : 0 KB/s
        Fan Speed                             : N/A
        Performance State                     : P0
        Clocks Throttle Reasons
            Idle                              : Active
            Applications Clocks Setting       : Not Active
            SW Power Cap                      : Not Active
            HW Slowdown                       : Not Active
                HW Thermal Slowdown           : Not Active
                HW Power Brake Slowdown       : Not Active
            Sync Boost                        : Not Active
            SW Thermal Slowdown               : Not Active
            Display Clock Setting             : Not Active
        FB Memory Usage
            Total                             : 40960 MiB
            Reserved                          : 571 MiB
            Used                              : 0 MiB
            Free                              : 40388 MiB
        BAR1 Memory Usage
            Total                             : 65536 MiB
            Used                              : 1 MiB
            Free                              : 65535 MiB
        Compute Mode                          : Default
        Utilization
            Gpu                               : 0 %
            Memory                            : 0 %
            Encoder                           : 0 %
            Decoder                           : 0 %
        Encoder Stats
            Active Sessions                   : 0
            Average FPS                       : 0
            Average Latency                   : 0
        FBC Stats
            Active Sessions                   : 0
            Average FPS                       : 0
            Average Latency                   : 0
        Ecc Mode
            Current                           : Enabled
            Pending                           : Enabled
        ECC Errors
            Volatile
                SRAM Correctable              : 0
                SRAM Uncorrectable            : 0
                DRAM Correctable              : 0
                DRAM Uncorrectable            : 0
            Aggregate
                SRAM Correctable              : 0
                SRAM Uncorrectable            : 0
                DRAM Correctable              : 0
                DRAM Uncorrectable            : 0
        Retired Pages
            Single Bit ECC                    : N/A
            Double Bit ECC                    : N/A
            Pending Page Blacklist            : N/A
        Remapped Rows
            Correctable Error                 : 0
            Uncorrectable Error               : 0
            Pending                           : No
            Remapping Failure Occurred        : No
            Bank Remap Availability Histogram
                Max                           : 640 bank(s)
                High                          : 0 bank(s)
                Partial                       : 0 bank(s)
                Low                           : 0 bank(s)
                None                          : 0 bank(s)
        Temperature
            GPU Current Temp                  : 30 C
            GPU Shutdown Temp                 : 92 C
            GPU Slowdown Temp                 : 89 C
            GPU Max Operating Temp            : 85 C
            GPU Target Temperature            : N/A
            Memory Current Temp               : 38 C
            Memory Max Operating Temp         : 95 C
        Power Readings
            Power Management                  : Supported
            Power Draw                        : 59.01 W
            Power Limit                       : 400.00 W
            Default Power Limit               : 400.00 W
            Enforced Power Limit              : 400.00 W
            Min Power Limit                   : 100.00 W
            Max Power Limit                   : 400.00 W
        Clocks
            Graphics                          : 1080 MHz
            SM                                : 1080 MHz
            Memory                            : 1215 MHz
            Video                             : 975 MHz
        Applications Clocks
            Graphics                          : 1095 MHz
            Memory                            : 1215 MHz
        Default Applications Clocks
            Graphics                          : 1095 MHz
            Memory                            : 1215 MHz
        Max Clocks
            Graphics                          : 1410 MHz
            SM                                : 1410 MHz
            Memory                            : 1215 MHz
            Video                             : 1290 MHz
        Max Customer Boost Clocks
            Graphics                          : 1410 MHz
        Clock Policy
            Auto Boost                        : N/A
            Auto Boost Default                : N/A
        Voltage
            Graphics                          : 737.500 mV
        Processes                             : None
    
  • docker version
  • Client: Docker Engine - Community
     Version:           20.10.17
     API version:       1.41
     Go version:        go1.17.11
     Git commit:        100c701
     Built:             Mon Jun  6 23:02:56 2022
     OS/Arch:           linux/amd64
     Context:           default
     Experimental:      true
    Server: Docker Engine - Community
     Engine:
      Version:          20.10.11
      API version:      1.41 (minimum version 1.12)
      Go version:       go1.16.9
      Git commit:       847da18
      Built:            Thu Nov 18 00:35:16 2021
      OS/Arch:          linux/amd64
      Experimental:     false
     containerd:
      Version:          1.6.6
      GitCommit:        10c12954828e7c7c9b6e0ea9b0c02b01407d3ae1
     nvidia:
      Version:          1.1.2
      GitCommit:        v1.1.2-0-ga916309
     docker-init:
      Version:          0.19.0
      GitCommit:        de40ad0
    
  • NVIDIA container library version
  • cli-version: 1.10.0
    lib-version: 1.10.0
    build date: 2022-06-13T10:39+00:00
    build revision: 395fd41701117121f1fd04ada01e1d7e006a37ae
    build compiler: x86_64-linux-gnu-gcc-7 7.5.0
    build platform: x86_64
    build flags: -D_GNU_SOURCE -D_FORTIFY_SOURCE=2 -DNDEBUG -std=gnu11 -O2 -g -fdata-sections -ffunction-sections -fplan9-extensions -fstack-protector -fno-strict-aliasing -fvisibility=hidden -Wall -Wextra -Wcast-align -Wpointer-arith -Wmissing-prototypes -Wnonnull -Wwrite-strings -Wlogical-op -Wformat=2 -Wmissing-format-attribute -Winit-self -Wshadow -Wstrict-prototypes -Wunreachable-code -Wconversion -Wsign-conversion -Wno-unknown-warning-option -Wno-format-extra-args -Wno-gnu-alignof-expression -Wl,-zrelro -Wl,-znow -Wl,-zdefs -Wl,--gc-sections
    
  • NVIDIA packages version
  • Desired=Unknown/Install/Remove/Purge/Hold
    | Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
    |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
    ||/ Name                                 Version                 Architecture            Description
    +++-====================================-=======================-=======================-=============================================================================
    ii  libnvidia-container-tools            1.10.0-1                amd64                   NVIDIA container runtime library (command-line tools)
    ii  libnvidia-container1:amd64           1.10.0-1                amd64                   NVIDIA container runtime library
    un  nvidia-container-runtime             <none>                  <none>                  (no description available)
    un  nvidia-container-runtime-hook        <none>                  <none>                  (no description available)
    ii  nvidia-container-toolkit             1.10.0-1                amd64                   NVIDIA container runtime hook
    un  nvidia-docker                        <none>                  <none>                  (no description available)
    ii  nvidia-docker2                       2.7.0-1                 all                     nvidia-docker CLI wrapper
              

    @elezar i make nvidia-container-toolkit=1.9.0-1,but output remain the same.

  • Details are as follows
  • Desired=Unknown/Install/Remove/Purge/Hold
    | Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
    |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
    ||/ Name                                 Version                 Architecture            Description
    +++-====================================-=======================-=======================-=============================================================================
    ii  libnvidia-container-tools            1.10.0-1                amd64                   NVIDIA container runtime library (command-line tools)
    ii  libnvidia-container1:amd64           1.10.0-1                amd64                   NVIDIA container runtime library
    un  nvidia-container-runtime             <none>                  <none>                  (no description available)
    un  nvidia-container-runtime-hook        <none>                  <none>                  (no description available)
    ii  nvidia-container-toolkit             1.9.0-1                 amd64                   NVIDIA container runtime hook
    un  nvidia-docker                        <none>                  <none>                  (no description available)
    ii  nvidia-docker2                       2.7.0-1                 all                     nvidia-docker CLI wrapper
    
    gpu-server@gpu-server:~$ sudo docker run --rm --gpus all nvidia/cuda:11.0.3-base-ubuntu20.04 nvidia-smi
    
    docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.
              

    @elezar this the same error

    gpu-server@gpu-server:~$ dpkg -l '*nvidia*'
    Desired=Unknown/Install/Remove/Purge/Hold
    | Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
    |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
    ||/ Name                                 Version                 Architecture            Description
    +++-====================================-=======================-=======================-=============================================================================
    ii  libnvidia-container-tools            1.9.0-1                 amd64                   NVIDIA container runtime library (command-line tools)
    ii  libnvidia-container1:amd64           1.9.0-1                 amd64                   NVIDIA container runtime library
    un  nvidia-container-runtime             <none>                  <none>                  (no description available)
    un  nvidia-container-runtime-hook        <none>                  <none>                  (no description available)
    ii  nvidia-container-toolkit             1.9.0-1                 amd64                   NVIDIA container runtime hook
    un  nvidia-docker                        <none>                  <none>                  (no description available)
    ii  nvidia-docker2                       2.7.0-1                 all                     nvidia-docker CLI wrapper
    gpu-server@gpu-server:~$ systemctl restart docker 
    ==== AUTHENTICATING FOR org.freedesktop.systemd1.manage-units ===
    Authentication is required to restart 'docker.service'.
    Authenticating as: gpu-server
    Password: 
    ==== AUTHENTICATION COMPLETE ===
    gpu-server@gpu-server:~$ sudo docker run --rm --gpus all nvidia/cuda:11.0.3-base-ubuntu20.04 nvidia-smi
    docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.
              

    What happens when you run nvidia-smi directly on the host?

    Also, this warning in the log seems suspicious:

    W0621 07:07:51.739205 4791 nvc.c:273] failed to set inheritable capabilities

    Is there something unconventional about how you are running docker on your host? For example, rootless, in a snap sandbox, etc.

    @klueska
    when i run this nvidia-smi on the host, it will two minutes for expected results

    gpu-server@gpu-server:~$ nvidia-smi
    Tue Jun 21 13:47:51 2022       
    +-----------------------------------------------------------------------------+
    | NVIDIA-SMI 515.48.07    Driver Version: 515.48.07    CUDA Version: 11.7     |
    |-------------------------------+----------------------+----------------------+
    | GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
    | Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
    |                               |                      |               MIG M. |
    |===============================+======================+======================|
    |   0  NVIDIA A100-SXM...  Off  | 00000000:01:00.0 Off |                    0 |
    | N/A   32C    P0    52W / 400W |      0MiB / 40960MiB |      0%      Default |
    |                               |                      |             Disabled |
    +-------------------------------+----------------------+----------------------+
    |   1  NVIDIA A100-SXM...  Off  | 00000000:41:00.0 Off |                    0 |
    | N/A   30C    P0    53W / 400W |      0MiB / 40960MiB |      0%      Default |
    |                               |                      |             Disabled |
    +-------------------------------+----------------------+----------------------+
    |   2  NVIDIA A100-SXM...  Off  | 00000000:81:00.0 Off |                    0 |
    | N/A   31C    P0    52W / 400W |      0MiB / 40960MiB |      0%      Default |
    |                               |                      |             Disabled |
    +-------------------------------+----------------------+----------------------+
    |   3  NVIDIA A100-SXM...  Off  | 00000000:C1:00.0 Off |                    0 |
    | N/A   30C    P0    52W / 400W |      0MiB / 40960MiB |      0%      Default |
    |                               |                      |             Disabled |
    +-------------------------------+----------------------+----------------------+
    +-----------------------------------------------------------------------------+
    | Processes:                                                                  |
    |  GPU   GI   CI        PID   Type   Process name                  GPU Memory |
    |        ID   ID                                                   Usage      |
    |=============================================================================|
    |  No running processes found                                                 |
    +-----------------------------------------------------------------------------+
    

    i run docker in root model and i used sudo gpasswd -a gpu-server docker, then run docker ,in addition i did not installed docker with snap.

    If it’s taking that long to get a result from nvidia-smi on the host then I could understand why the RPC might time out in the nvidia-container-cli when trying to get a result from the driver RPC call.

    Im assuming this is happening because (1) you are not running the nvidia-persistenced daemon and (2) your GPUs are not in persistence mode.

    Both (1) and (2) achieve the same thing, but (1) is the preferred method to keep the GPU driver alive even when no clients are attached.

    Try enabling one of these methods and report back.

    XksA-me, zkytony, z-q-y, alkavan, juliusHuelsmann, suamin, fahnub, PPPPPYJ, FruiteePro, and krevas reacted with thumbs up emoji alkavan, Cyebukayire, maykonpacheco, PPPPPYJ, FranLucchini, and krevas reacted with heart emoji alkavan and maykonpacheco reacted with rocket emoji All reactions

    @klueska Yeah dude thanks =)
    That's exactly it, running nvidia-persistenced as root without any additional parameters solved it.
    You can confirm with nvidia-smi that the GPU units load faster.

    docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed #1710

    If it’s taking that long to get a result from nvidia-smi on the host then I could understand why the RPC might time out in the nvidia-container-cli when trying to get a result from the driver RPC call.

    Im assuming this is happening because (1) you are not running the nvidia-persistenced daemon and (2) your GPUs are not in persistence mode.

    Both (1) and (2) achieve the same thing, but (1) is the preferred method to keep the GPU driver alive even when no clients are attached.

    Try enabling one of these methods and report back.

    I tried using nvidia-smi -i 0 -pm ENABLED but not luck. I'm having the same error as @zyr-NULL. I then tried using:

    # nvidia-persistenced --user aurelien
    nvidia-persistenced failed to initialize. Check syslog for more details.
    

    Here are the logs:

    # cat /var/log/syslog
    Feb  6 14:15:46 pop-os nvidia-persistenced: device 0000:01:00.0 - persistence mode disabled.
    Feb  6 14:15:46 pop-os nvidia-persistenced: device 0000:01:00.0 - NUMA memory offlined.
    Feb  6 14:16:06 pop-os nvidia-persistenced: Failed to change ownership of /var/run/nvidia-persistenced: Operation not permitted
    Feb  6 14:16:06 pop-os nvidia-persistenced: Shutdown (16491)
    Feb  6 14:16:38 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
    Feb  6 14:16:38 pop-os nvidia-persistenced: Shutdown (16516)
    Feb  6 14:16:46 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
    Feb  6 14:16:48 pop-os rtkit-daemon[1589]: message repeated 5 times: [ Supervising 8 threads of 5 processes of 1 users.]
    Feb  6 14:17:01 pop-os nvidia-persistenced: device 0000:01:00.0 - persistence mode enabled.
    Feb  6 14:17:01 pop-os nvidia-persistenced: device 0000:01:00.0 - NUMA memory onlined.
    Feb  6 14:17:01 pop-os CRON[16652]: (root) CMD (   cd / && run-parts --report /etc/cron.hourly)
    Feb  6 14:17:03 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
    Feb  6 14:17:03 pop-os nvidia-persistenced: Shutdown (16656)
    Feb  6 14:18:03 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
    Feb  6 14:18:04 pop-os rtkit-daemon[1589]: message repeated 3 times: [ Supervising 8 threads of 5 processes of 1 users.]
    Feb  6 14:18:33 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
    Feb  6 14:18:33 pop-os nvidia-persistenced: Shutdown (16799)
    Feb  6 14:18:38 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
    Feb  6 14:18:53 pop-os rtkit-daemon[1589]: message repeated 7 times: [ Supervising 8 threads of 5 processes of 1 users.]
    Feb  6 14:19:17 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
    Feb  6 14:19:17 pop-os nvidia-persistenced: Shutdown (16994)
    Feb  6 14:20:04 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
    Feb  6 14:20:07 pop-os rtkit-daemon[1589]: message repeated 5 times: [ Supervising 8 threads of 5 processes of 1 users.]
    Feb  6 14:20:56 pop-os systemd[1]: Starting Cleanup of Temporary Directories...
    Feb  6 14:20:56 pop-os systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully.
    Feb  6 14:20:56 pop-os systemd[1]: Finished Cleanup of Temporary Directories.
    Feb  6 14:21:03 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
    Feb  6 14:21:07 pop-os rtkit-daemon[1589]: message repeated 3 times: [ Supervising 8 threads of 5 processes of 1 users.]
    Feb  6 14:21:15 pop-os systemd[1]: Reloading.
    Feb  6 14:23:27 pop-os chronyd[1007]: NTS-KE session with 52.10.183.132:4460 (oregon.time.system76.com) timed out
    Feb  6 14:23:33 pop-os chronyd[1007]: NTS-KE session with 18.228.202.30:4460 (brazil.time.system76.com) timed out
    Feb  6 14:23:33 pop-os chronyd[1007]: NTS-KE session with 15.237.97.214:4460 (paris.time.system76.com) timed out
    Feb  6 14:23:34 pop-os chronyd[1007]: NTS-KE session with 3.134.129.152:4460 (ohio.time.system76.com) timed out
    Feb  6 14:23:36 pop-os chronyd[1007]: NTS-KE session with 3.220.42.39:4460 (virginia.time.system76.com) timed out
    Feb  6 14:24:21 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
    Feb  6 14:24:21 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
    Feb  6 14:25:04 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
    Feb  6 14:25:04 pop-os nvidia-persistenced: Shutdown (17372)
    

    Any ideas @klueska? 🥲

    Here's the error by the way:

    $ sudo docker run --rm --gpus all nvidia/cuda:11.6.2-base-ubuntu20.04 nvidia-smi
    docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
    nvidia-container-cli: container error: cgroup subsystem devices not found: unknown.
              

    I met this problem too. My GPU is A30 and GPU driver is 525.85.12.

     State:      Terminated
          Reason:   StartError
          Message:  failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
    nvidia-container-cli: detection error: nvml error: unknown error: unknown
    
    [Tue Feb 21 21:10:41 2023] NVRM: Xid (PCI:0000:22:00): 119, pid=3660230, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
    

    @fighterhit this may be related to changes in the GSP firmware paths and should be addressed in the v1.13.0-rc.1 release. Would you be able to give it a try?

    Hi @elezar , yes I can, but do I need to restart the node or the containers on the node? Because it may need to be deployed in our production environment.

    Or does this problem exist in previous versions, I can also accept lower versions :).

    Updating the nvidia-container-toolkit should not require a restart. The components of the NVIDIA container stack are only invoked when containers are created and would not affect running containers.

    If this is the firmware issue that I think it is, then a lower version would not work.

    What is the output of

    ls -al /firmware/nvidia/525.85*/*
    

    If there is a single gsp.bin file there, then a v1.13 release is not required and any NVIDIA Container Toolkit version after v1.8.1 should work. If there are multiple gsp_*.bin files there then v1.13.0-rc.1 would be required to address this.

    Updating the nvidia-container-toolkit should not require a restart. The components of the NVIDIA container stack are only invoked when containers are created and would not affect running containers.

    If this is the firmware issue that I think it is, then a lower version would not work.

    What is the output of

    ls -al /firmware/nvidia/525.85*/*
    

    If there is a single gsp.bin file there, then a v1.13 release is not required and any NVIDIA Container Toolkit version after v1.8.1 should work. If there are multiple gsp_*.bin files there then v1.13.0-rc.1 would be required to address this.

    @elezar The output is:

    ls: cannot access '/firmware/nvidia/525.85*/*': No such file or directory
    

    It seems that there is no such /firmware directory on our all GPU nodes. FYI, our driver is installed by NVIDIA-Linux-x86_64-525.85.12.run.

    Updating the nvidia-container-toolkit should not require a restart. The components of the NVIDIA container stack are only invoked when containers are created and would not affect running containers.
    If this is the firmware issue that I think it is, then a lower version would not work.
    What is the output of

    ls -al /firmware/nvidia/525.85*/*
    

    If there is a single gsp.bin file there, then a v1.13 release is not required and any NVIDIA Container Toolkit version after v1.8.1 should work. If there are multiple gsp_*.bin files there then v1.13.0-rc.1 would be required to address this.

    @elezar The output is:

    ls: cannot access '/firmware/nvidia/525.85*/*': No such file or directory
    

    It seems that there is no such /firmware directory on our all GPU nodes. FYI, our driver is installed by NVIDIA-Linux-x86_64-525.85.12.run.

    Hi @elezar , do you have any more suggestions? Thanks!

    The path is /lib/firmware/…

    Thanks @klueska 😅, there are indeed two gsp_*.bin files in this directory.

    -r--r--r-- 1 root root 25M Feb 18 11:34 /lib/firmware/nvidia/525.85.12/gsp_ad10x.bin
    -r--r--r-- 1 root root 36M Feb 18 11:34 /lib/firmware/nvidia/525.85.12/gsp_tu10x.bin
              

    The new RC adds support for detecting multiple GSP firmware files, which is required for container support to work correctly on the 525 driver.

    The persistenced issue is still relevant but this is a new one related to the latest NVIDIA driver.

    Thanks @klueska , how can I install this version v1.13.0-rc.1? I searched by using apt-cache show nvidia-container-toolkit but there seems to be no package for this version.

    The persistenced issue is still relevant but this is a new one related to the latest NVIDIA driver.

    Does the persistenced issue only appear on the latest driver (525.85.12)? Can I solve it by downgrading to a certain driver version?

    Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.

    That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.

    Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.

    Instructions here:
    https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit

    Thank you for your help @klueska. What confuses me is that in the past, our GPU clusters (1080Ti, 2080Ti, 3090, A30, A100) have not turned on the persistent mode and there is no such problem, only A30 nodes (with driver 525.85.12) have this problem, none of the other types of GPU nodes(1080Ti, 2080Ti, 3090 with driver 525.78.01) have this problem.

    Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.

    That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.

    Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.

    Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit

    Hi @klueska @elezar , when I configure to use the experimental repo I get the following error and can't install the v1.13.0-rc.1 , my system distribution is debian11.

    root@xxx:/etc/apt/sources.list.d# distribution=$(. /etc/os-release;echo $ID$VERSION_ID) \
          && curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \
          && curl -s -L https://nvidia.github.io/libnvidia-container/experimental/$distribution/libnvidia-container.list | \
             sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
             sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
    # Unsupported distribution!
    # Check https://nvidia.github.io/libnvidia-container
    

    Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.
    That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.
    Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.
    Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit

    Hi @klueska @elezar , when I configure to use the experimental repo I get the following error and can't install the v1.13.0-rc.1 , my system distribution is debian11.

    root@xxx:/etc/apt/sources.list.d# distribution=$(. /etc/os-release;echo $ID$VERSION_ID) \
          && curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \
          && curl -s -L https://nvidia.github.io/libnvidia-container/experimental/$distribution/libnvidia-container.list | \
             sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
             sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
    # Unsupported distribution!
    # Check https://nvidia.github.io/libnvidia-container
    

    @fighterhit what distribution are you using? Please ensure that the distribution variable is set to ubuntu18.04 and try again.

    Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.
    That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.
    Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.
    Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit

    Hi @klueska @elezar , when I configure to use the experimental repo I get the following error and can't install the v1.13.0-rc.1 , my system distribution is debian11.

    root@xxx:/etc/apt/sources.list.d# distribution=$(. /etc/os-release;echo $ID$VERSION_ID) \
          && curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \
          && curl -s -L https://nvidia.github.io/libnvidia-container/experimental/$distribution/libnvidia-container.list | \
             sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
             sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
    # Unsupported distribution!
    # Check https://nvidia.github.io/libnvidia-container
    

    @fighterhit what distribution are you using? Please ensure that the distribution variable is set to ubuntu18.04 and try again.

    @elezar My distribution is Debian GNU/Linux 11 (bullseye).

    Hi @klueska @elezar , I test the v1.13.0-rc.1 but it still report some errors:

      Warning  Failed     12s (x3 over 64s)  kubelet            Error: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
    nvidia-container-cli: ldcache error: open failed: /sbin/ldconfig.real: no such file or directory: unknown
      Warning  BackOff  11s (x3 over 44s)  kubelet  Back-off restarting failed container
    

    Maybe /sbin/ldconfig.real should be /sbin/ldconfig?

    For debian, yes, it should be @/sbin/ldconfig (without the .real), as seen here:
    https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15

    This file gets installed under/etc/nvidia-container-runtime/config.toml

    The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?

    For debian, yes, it should be @/sbin/ldconfig (without the .real), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15

    This file gets installed under/etc/nvidia-container-runtime/config.toml

    The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?

    Yes @klueska , I failed to installed the latest version using the experimental repo(#1648 (comment)), so I followed @elezar advice and set the distro to ubuntu18.04. Can I manually modify /etc/nvidia-container-runtime/config.toml to make it work? But I'm not sure if this will have any other effects.

    For debian, yes, it should be @/sbin/ldconfig (without the .real), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15
    This file gets installed under/etc/nvidia-container-runtime/config.toml
    The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?

    Yes @klueska , I failed to installed the latest version using the experimental repo(#1648 (comment)), so I followed @elezar advice and set the distro to ubuntu18.04. Can I manually modify /etc/nvidia-container-runtime/config.toml to make it work? But I'm not sure if this will have any other effects.

    Sorry about that. I was making an assumption about the distribution you are using. You can install the ubuntu18.04 package and then update the /etc/nvidia-container-runtime/config.toml value after the fact. This should have no other effects.

    For debian, yes, it should be @/sbin/ldconfig (without the .real), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15
    This file gets installed under/etc/nvidia-container-runtime/config.toml
    The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?

    Yes @klueska , I failed to installed the latest version using the experimental repo(#1648 (comment)), so I followed @elezar advice and set the distro to ubuntu18.04. Can I manually modify /etc/nvidia-container-runtime/config.toml to make it work? But I'm not sure if this will have any other effects.

    Sorry about that. I was making an assumption about the distribution you are using. You can install the ubuntu18.04 package and then update the /etc/nvidia-container-runtime/config.toml value after the fact. This should have no other effects.

    Thanks @elezar ! I have tried it and it works fine. It may be better if debian11 could be supported to use the experimental repo.

    @fighterhit I have just double-checked our repo configuration and the issue is that for debian10 (to which debian11 redirects) the repo only had a libnvidia-container-experimental.list and no libnvidia-container.list file which is referred to by the instructions.

    I have created a link / redirect for this now and the official instructions should work as expected. (It may take about 30 minutes for the changes to reflect in the repo though).

    @fighterhit I have just double-checked our repo configuration and the issue is that for debian10 (to which debian11 redirects) the repo only had a libnvidia-container-experimental.list and no libnvidia-container.list file which is referred to by the instructions.

    I have created a link / redirect for this now and the official instructions should work as expected. (It may take about 30 minutes for the changes to reflect in the repo though).

    Thanks for your confirmation, it works now. @elezar

    I am trying to install NVIDIA container toolkit on Amazon Linux 2, created a new ec2 instance and followed the instructions in this page and running into below issue. I get the same error when i tried this on a Ubuntu instance as well.

    nvidia-container-cli: initialization error: load library failed: libnvidia-ml.so.1: cannot open shared object file: no such file or directory: unknown.
    ERRO[0000] error waiting for container: context canceled 
    

    @elezar @klueska please advise how to fix this issue, appreciate your inputs. Thanks

    Hi @elezar @klueska , unfortunately I have used the latest toolkit but this problem reappeared, I think this may be related to the driver, and I asked the driver community for help but got no more reply (https://github.com/NVIDIA/open-gpu-kernel-modules/issues/446), https://forums.developer.nvidia.com/t/timeout-waiting-for-rpc-from-gsp/244789 could you communicate with the driver team about this issue? Because some users in the community also encountered the same problem.Thanks!

    [Mon Mar  6 21:09:45 2023] NVRM: GPU at PCI:0000:61:00: GPU-81f5d81e-7906-c145-3def-82e281d7b260
    [Mon Mar  6 21:09:45 2023] NVRM: GPU Board Serial Number: 1322621149674
    [Mon Mar  6 21:09:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977720, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a00b0 0x0).
    [Mon Mar  6 21:09:45 2023] CPU: 15 PID: 977720 Comm: nvidia-smi Tainted: P           OE     5.10.0-20-amd64 #1 Debian 5.10.158-2
    [Mon Mar  6 21:09:45 2023] Hardware name: Inspur NF5468A5/YZMB-02382-101, BIOS 4.02.12 01/28/2022
    [Mon Mar  6 21:09:45 2023] Call Trace:
    [Mon Mar  6 21:09:45 2023]  dump_stack+0x6b/0x83
    [Mon Mar  6 21:09:45 2023]  _nv011231rm+0x39d/0x470 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv011168rm+0x62/0x2e0 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv040022rm+0xdb/0x140 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv041148rm+0x2ce/0x3a0 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv015451rm+0x788/0x800 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv039541rm+0xac/0xe0 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv041150rm+0xac/0x140 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv041149rm+0x37a/0x4d0 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv039443rm+0xc9/0x150 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv039444rm+0x42/0x70 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? rm_cleanup_file_private+0x128/0x180 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv000554rm+0x49/0x60 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? _nv000694rm+0x7fb/0xc80 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
    [Mon Mar  6 21:09:45 2023]  ? __x64_sys_ioctl+0x8b/0xc0
    [Mon Mar  6 21:09:45 2023]  ? do_syscall_64+0x33/0x80
    [Mon Mar  6 21:09:45 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
    [Mon Mar  6 21:10:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977720, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0050 0x0).
    [Mon Mar  6 21:11:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977720, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0040 0x0).
    [Mon Mar  6 21:12:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801348 0x410).
    [Mon Mar  6 21:12:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x0 0x6c).
    [Mon Mar  6 21:13:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x80 0x38).
    [Mon Mar  6 21:14:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x2080 0x4).
    [Mon Mar  6 21:15:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801348 0x410).
    [Mon Mar  6 21:15:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x0 0x6c).
    [Mon Mar  6 21:16:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x80 0x38).
    [Mon Mar  6 21:17:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=905052, name=python, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801702 0x4).
    [Mon Mar  6 21:18:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x2080 0x4).
    [Mon Mar  6 21:18:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800110 0x84).
    [Mon Mar  6 21:19:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801348 0x410).
    [Mon Mar  6 21:20:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:21:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x0 0x6c).
    [Mon Mar  6 21:21:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x80 0x38).
    [Mon Mar  6 21:22:08 2023] INFO: task python:976158 blocked for more than 120 seconds.
    [Mon Mar  6 21:22:08 2023]       Tainted: P           OE     5.10.0-20-amd64 #1 Debian 5.10.158-2
    [Mon Mar  6 21:22:08 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
    [Mon Mar  6 21:22:08 2023] task:python          state:D stack:    0 pid:976158 ppid:835062 flags:0x00000000
    [Mon Mar  6 21:22:08 2023] Call Trace:
    [Mon Mar  6 21:22:08 2023]  __schedule+0x282/0x880
    [Mon Mar  6 21:22:08 2023]  schedule+0x46/0xb0
    [Mon Mar  6 21:22:08 2023]  rwsem_down_write_slowpath+0x246/0x4d0
    [Mon Mar  6 21:22:08 2023]  os_acquire_rwlock_write+0x31/0x40 [nvidia]
    [Mon Mar  6 21:22:08 2023]  _nv038505rm+0xc/0x30 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv039453rm+0x18d/0x1d0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv041182rm+0x45/0xd0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv041127rm+0x142/0x2b0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv039415rm+0x15a/0x2e0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv039416rm+0x5b/0x90 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv039416rm+0x31/0x90 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv012688rm+0x1d/0x30 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv039431rm+0xb0/0xb0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv012710rm+0x54/0x70 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv011426rm+0xc4/0x120 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv000659rm+0x63/0x70 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv000582rm+0x2c/0x40 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv000694rm+0x86c/0xc80 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? __x64_sys_ioctl+0x8b/0xc0
    [Mon Mar  6 21:22:08 2023]  ? do_syscall_64+0x33/0x80
    [Mon Mar  6 21:22:08 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
    [Mon Mar  6 21:22:08 2023] INFO: task nvidia-smi:993195 blocked for more than 120 seconds.
    [Mon Mar  6 21:22:08 2023]       Tainted: P           OE     5.10.0-20-amd64 #1 Debian 5.10.158-2
    [Mon Mar  6 21:22:08 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
    [Mon Mar  6 21:22:08 2023] task:nvidia-smi      state:D stack:    0 pid:993195 ppid: 58577 flags:0x00000004
    [Mon Mar  6 21:22:08 2023] Call Trace:
    [Mon Mar  6 21:22:08 2023]  __schedule+0x282/0x880
    [Mon Mar  6 21:22:08 2023]  schedule+0x46/0xb0
    [Mon Mar  6 21:22:08 2023]  rwsem_down_write_slowpath+0x246/0x4d0
    [Mon Mar  6 21:22:08 2023]  os_acquire_rwlock_write+0x31/0x40 [nvidia]
    [Mon Mar  6 21:22:08 2023]  _nv038505rm+0xc/0x30 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv039453rm+0x18d/0x1d0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv041182rm+0x45/0xd0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv041127rm+0x142/0x2b0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv039415rm+0x15a/0x2e0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv039416rm+0x5b/0x90 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv039416rm+0x31/0x90 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv000560rm+0x59/0x70 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv000560rm+0x33/0x70 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? _nv000694rm+0x4ae/0xc80 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
    [Mon Mar  6 21:22:08 2023]  ? __x64_sys_ioctl+0x8b/0xc0
    [Mon Mar  6 21:22:08 2023]  ? do_syscall_64+0x33/0x80
    [Mon Mar  6 21:22:08 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
    [Mon Mar  6 21:22:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800157 0x0).
    [Mon Mar  6 21:23:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:24:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:24:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:25:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:26:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080852e 0x208).
    [Mon Mar  6 21:27:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20808513 0x598).
    [Mon Mar  6 21:27:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a068 0x4).
    [Mon Mar  6 21:28:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a618 0x181c).
    [Mon Mar  6 21:29:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a612 0xd98).
    [Mon Mar  6 21:30:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20809009 0x8).
    [Mon Mar  6 21:30:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
    [Mon Mar  6 21:31:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
    [Mon Mar  6 21:32:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x2080 0x4).
    [Mon Mar  6 21:33:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
    [Mon Mar  6 21:33:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800110 0x84).
    [Mon Mar  6 21:34:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:35:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x208f 0x0).
    [Mon Mar  6 21:36:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x208f1105 0x8).
    [Mon Mar  6 21:36:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800157 0x0).
    [Mon Mar  6 21:37:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=905052, name=python, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801702 0x4).
    [Mon Mar  6 21:38:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:39:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:39:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a00b0 0x0).
    [Mon Mar  6 21:40:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:41:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0050 0x0).
    [Mon Mar  6 21:42:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:42:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080852e 0x208).
    [Mon Mar  6 21:43:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20808513 0x598).
    [Mon Mar  6 21:44:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a068 0x4).
    [Mon Mar  6 21:45:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a618 0x181c).
    [Mon Mar  6 21:45:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0040 0x0).
    [Mon Mar  6 21:46:18 2023] INFO: task python:976158 blocked for more than 120 seconds.
    [Mon Mar  6 21:46:18 2023]       Tainted: P           OE     5.10.0-20-amd64 #1 Debian 5.10.158-2
    [Mon Mar  6 21:46:18 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
    [Mon Mar  6 21:46:18 2023] task:python          state:D stack:    0 pid:976158 ppid:835062 flags:0x00000000
    [Mon Mar  6 21:46:18 2023] Call Trace:
    [Mon Mar  6 21:46:18 2023]  __schedule+0x282/0x880
    [Mon Mar  6 21:46:18 2023]  schedule+0x46/0xb0
    [Mon Mar  6 21:46:18 2023]  rwsem_down_write_slowpath+0x246/0x4d0
    [Mon Mar  6 21:46:18 2023]  os_acquire_rwlock_write+0x31/0x40 [nvidia]
    [Mon Mar  6 21:46:18 2023]  _nv038505rm+0xc/0x30 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv039453rm+0x18d/0x1d0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv041182rm+0x45/0xd0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv041127rm+0x142/0x2b0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv039415rm+0x15a/0x2e0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv039416rm+0x5b/0x90 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv039416rm+0x31/0x90 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv012688rm+0x1d/0x30 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? task_numa_fault+0x2a3/0xb70
    [Mon Mar  6 21:46:18 2023]  ? _nv039431rm+0xb0/0xb0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv012710rm+0x54/0x70 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv011426rm+0xc4/0x120 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv000659rm+0x63/0x70 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv000582rm+0x2c/0x40 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv000694rm+0x86c/0xc80 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? __x64_sys_ioctl+0x8b/0xc0
    [Mon Mar  6 21:46:18 2023]  ? do_syscall_64+0x33/0x80
    [Mon Mar  6 21:46:18 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
    [Mon Mar  6 21:46:18 2023] INFO: task nvidia-smi:980026 blocked for more than 120 seconds.
    [Mon Mar  6 21:46:18 2023]       Tainted: P           OE     5.10.0-20-amd64 #1 Debian 5.10.158-2
    [Mon Mar  6 21:46:18 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
    [Mon Mar  6 21:46:18 2023] task:nvidia-smi      state:D stack:    0 pid:980026 ppid:980025 flags:0x00000000
    [Mon Mar  6 21:46:18 2023] Call Trace:
    [Mon Mar  6 21:46:18 2023]  __schedule+0x282/0x880
    [Mon Mar  6 21:46:18 2023]  ? psi_task_change+0x88/0xd0
    [Mon Mar  6 21:46:18 2023]  schedule+0x46/0xb0
    [Mon Mar  6 21:46:18 2023]  rwsem_down_read_slowpath+0x18e/0x500
    [Mon Mar  6 21:46:18 2023]  os_acquire_rwlock_read+0x31/0x40 [nvidia]
    [Mon Mar  6 21:46:18 2023]  _nv038503rm+0xc/0x30 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv039453rm+0x64/0x1d0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv041182rm+0x45/0xd0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv041133rm+0xfd/0x2b0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv012728rm+0x59a/0x690 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv039431rm+0x53/0xb0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv011404rm+0x52/0xa0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? _nv000694rm+0x5ae/0xc80 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
    [Mon Mar  6 21:46:18 2023]  ? __x64_sys_ioctl+0x8b/0xc0
    [Mon Mar  6 21:46:18 2023]  ? do_syscall_64+0x33/0x80
    [Mon Mar  6 21:46:18 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
    [Mon Mar  6 21:46:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a612 0xd98).
    [Mon Mar  6 21:47:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20809009 0x8).
    [Mon Mar  6 21:48:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
    [Mon Mar  6 21:48:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
    [Mon Mar  6 21:49:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
    [Mon Mar  6 21:50:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800110 0x84).
    [Mon Mar  6 21:51:02 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
    [Mon Mar  6 21:51:47 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800157 0x0).
    [Mon Mar  6 21:52:32 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x208f 0x0).
    [Mon Mar  6 21:53:17 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x208f1105 0x8).
    [Mon Mar  6 21:54:02 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
              

    I was getting similar errors trying to run nvidia-smi after updating my Ubuntu 22.04 system.

    I noticed that sudo systemctl status nvidia-persistenced was not enabled, so I tried to enable it with sudo systemctl enable nvidia-persistenced but that failed as well.

    I then ran across the following forum post which led me to think that perhaps my drivers had somehow been disabled/corrupted by a recent system and/or kernel upgrade so I decided to reinstall (and upgrade) the nvidia driver using the PPA, and nvidia-smi is working again. Hope this helps someone else...

    https://forums.developer.nvidia.com/t/nvidia-smi-has-failed-because-it-couldnt-communicate-with-the-nvidia-driver-make-sure-that-the-latest-nvidia-driver-is-installed-and-running/197141/2

     
    推荐文章