• Tattorack@lemmy.world
    link
    fedilink
    English
    arrow-up
    15
    arrow-down
    2
    ·
    4 months ago

    Wouldn’t just one GPU be enough to run the Sphere, or a I getting something wrong?

    I remember hearing about that it’s not exactly high resolution, each “pixel” being a bunch of pretty large lamps.

    • ilinamorato@lemmy.world
      link
      fedilink
      English
      arrow-up
      17
      ·
      edit-2
      4 months ago

      Wikipedia says it’s 16,000x16,000 (which is way less than I thought). The way the math works, that’s 16x as big as a 4k monitor, so 16 GPUs would make sense. And there’s a screen inside and one outside, so double that. But I also can’t figure out why it needs five times that. Redundancy? Poor optimization? I dunno.

      • Tattorack@lemmy.world
        link
        fedilink
        English
        arrow-up
        12
        ·
        4 months ago

        But wouldn’t that be only necessary if it needed to render real-time graphics at such a scale? If I’m correct, all its doing is playing back videos.

        • ilinamorato@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          ·
          4 months ago

          I think it’s doing some non-trivial amount of rendering, since it’s often syncing graphics with music played live.

        • stormeuh@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 months ago

          Even if it’s just playing back videos, it still should compensate for the distortion of the spherical display. That’s a “simple” 3d transformation, but with the amount of pixels, coordinating between the GPUs and some redundancy, it doesn’t seem like an excessive amount of computing power. The whole thing is still an impressive excess though…

        • ilinamorato@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          4 months ago

          Someone elsewhere in the thread suggested it might be a marketing thing on Nvidia’s part, and that makes a lot of sense.

      • Mark@lemmy.ca
        link
        fedilink
        English
        arrow-up
        5
        ·
        4 months ago

        I work for a digital display company, and it is definitely redundancy. There will be at least two redundant display systems that go to the modules separately so they can switch between them to solve issues. If a component fails on one side they just switch to the other.

    • umbraroze@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      1
      ·
      4 months ago

      The way I think it, it’s possible a really small number of GPUs would be enough to render the framebuffer, you’d just need an army of low-power graphics units to receive the data and render it on screens.

      Having a high-power GPU for every screen is definitely a loss unless the render job is distributed really well and there’s also people around to admire the results at the distance where the pixel differences no longer matter. Which is to say, not here.