You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The guest VM is running Ubuntu plucky with the following configuration:
Driver: NVIDIA UNIX Open Kernel Mode Setting Driver for x86_64 570.195.03 Release Build (dvs-builder@U22-I3-H04-03-1) Sat Sep 20 00:39:49 UTC 2025
Kernel: Linux version 6.14.0-35-generic (buildd@lcy02-amd64-078) (x86_64-linux-gnu-gcc-14 (Ubuntu 14.2.0-19ubuntu2) 14.2.0, GNU ld (GNU Binutils for Ubuntu) 2.44) Executing an Attestation of the GPU(x-nv-gpu-measurements-match) Failed #35-Ubuntu SMP PREEMPT_DYNAMIC Sat Oct 11 10:06:31 UTC 2025 (Ubuntu 6.14.0-35.35-generic 6.14.11)
The Kernel command line contains just console=ttyS0 pci=realloc,nocrs as documented on the deployment guide.
I'm running Ubuntu plucky because it ships with a more recent kernel version which supports the functionality needed by COCONUT-SVSM.
Output of dmesg | grep -i sev on the guest:
[ 2.319769] Memory Encryption Features active: AMD SEV SEV-ES SEV-SNP
[ 2.320704] SEV: Status: SEV SEV-ES SEV-SNP
[ 2.539717] SEV: APIC: wakeup_secondary_cpu() replaced with wakeup_cpu_via_vmgexit()
[ 3.677778] SEV: Using SNP CPUID table, 28 entries present.
[ 3.678703] SEV: SNP running at VMPL2.
[ 4.034162] SEV: SNP guest platform device initialized.
[ 6.156077] sev-guest sev-guest: Initialized SEV guest driver (using VMPCK2 communication key)
On the guest I've also removed the nouveau kernel module:
# modinfo nouveau
modinfo: ERROR: Module nouveau not found.
The times where the GPU is detected, I'm able to verify the GPU attestation on the guest and everything works fine:
Generating nonce in the local GPU Verifier ..
Number of GPUs available : 1
Fetching GPU 0 information from GPU driver.
All GPU Evidences fetched successfully
-----------------------------------
Verifying GPU: GPU-6a77de48-cf1d-2048-5be0-a63557bf0e6f
Driver version fetched : 570.195.03
VBIOS version fetched : 96.00.9f.00.01
Validating GPU certificate chains.
The firmware ID in the device certificate chain is matching with the one in the attestation report.
GPU attestation report certificate chain validation successful.
The certificate chain revocation status verification successful.
Authenticating attestation report
The nonce in the SPDM GET MEASUREMENT request message is matching with the generated nonce.
Driver version fetched from the attestation report : 570.195.03
VBIOS version fetched from the attestation report : 96.00.9f.00.01
Attestation report signature verification successful.
Attestation report verification successful.
Authenticating the RIMs.
Authenticating Driver RIM
Fetching the driver RIM from the RIM service.
RIM Schema validation passed.
driver RIM certificate chain verification successful.
The certificate chain revocation status verification successful.
driver RIM signature verification successful.
Driver RIM verification successful
Authenticating VBIOS RIM.
Fetching the VBIOS RIM from the RIM service.
RIM Schema validation passed.
vbios RIM certificate chain verification successful.
The certificate chain revocation status verification successful.
vbios RIM signature verification successful.
VBIOS RIM verification successful
Comparing measurements (runtime vs golden)
The runtime measurements are matching with the golden measurements.
GPU is in expected state.
GPU 0 with UUID GPU-6a77de48-cf1d-2048-5be0-a63557bf0e6f verified successfully.
Setting the GPU Ready State to READY
GPU Attestation is Successful.
Output of nvidia-smi conf-compute -q:
==============NVSMI CONF-COMPUTE LOG==============
CC State : ON
Multi-GPU Mode : None
CPU CC Capabilities : AMD SEV-SNP
GPU CC Capabilities : CC Capable
CC GPUs Ready State : Ready
I'm experiencing intermittent GPU detection failures in my VM, which is running the OS in live mode. This means every reboot starts from a clean OS image, yet I still frequently need multiple reboots for the GPU to be recognized. The dmesg output appears identical in both successful and failed detection attempts, with the exception of the previously reported stack trace.
Hello,
About half of the times when I boot a VM with an H100 attached with CC mode enabled I get the following error:
The guest VM is running Ubuntu plucky with the following configuration:
The Kernel command line contains just
console=ttyS0 pci=realloc,nocrsas documented on the deployment guide.I'm running Ubuntu plucky because it ships with a more recent kernel version which supports the functionality needed by COCONUT-SVSM.
Output of
dmesg | grep -i sevon the guest:On the guest I've also removed the nouveau kernel module:
The times where the GPU is detected, I'm able to verify the GPU attestation on the guest and everything works fine:
Output of
nvidia-smi conf-compute -q:The host is configured following the instructions at https://coconut-svsm.github.io/svsm/installation/INSTALL/.
I'm experiencing intermittent GPU detection failures in my VM, which is running the OS in live mode. This means every reboot starts from a clean OS image, yet I still frequently need multiple reboots for the GPU to be recognized. The dmesg output appears identical in both successful and failed detection attempts, with the exception of the previously reported stack trace.