This hasn’t happened to me yet but I was just thinking about it. Let’s say you have a server with an iGPU, and you use GPU passthrough to let VMs use the iGPU. And then one day the host’s ssh server breaks, maybe you did something stupid or there was a bad update. Are you fucked? How could you possibly recover, with no display and no SSH? The only thing I can think of is setting up serial access for emergencies like this, but I rarely hear about serial access nowadays so I wonder if there’s some other solution here.

  • berylenara@sh.itjust.worksOP
    link
    fedilink
    English
    arrow-up
    2
    ·
    13 days ago

    It only hijacks the GPU when I start the VM

    How did you do this? All the tutorials I read hijack the GPU at startup. Do you have to manually detach the GPU from the host before assigning it to the VM?

    • [email protected]@sh.itjust.works
      link
      fedilink
      arrow-up
      3
      ·
      13 days ago

      Interesting.
      I’m not doing anything special that wasn’t in one of the popular tutorials and I thought that’s how it was supposed to work, although it might very well be a “bug” how it behaves right now.

      I don’t know enough about this, but the drivers are blacklisted on the host at boot, yet the console is still displayed through the GPU’s HDMI at that time which might depend on the specific GPU (a vega64 in my case).

      The host doesn’t have a graphical desktop environment, just the shell.

      • berylenara@sh.itjust.worksOP
        link
        fedilink
        English
        arrow-up
        3
        ·
        12 days ago

        the drivers are blacklisted on the host at boot

        This is the problem I was alluding to, though I’m surprised you are still able to see the console despite the driver being blacklisted. I have heard of people using scripts to manually detach the GPU and attach it to a VM, but sounds like you don’t need that, which is interesting