“We developed a deep neural network that maps the phase and amplitude of WiFi signals to UV coordinates within 24 human regions. The results of the study reveal that our model can estimate the dense pose of multiple subjects, with comparable performance to image-based approaches, by utilizing WiFi signals as the only input.”

  • Xavier@lemmy.ca
    link
    fedilink
    English
    arrow-up
    18
    arrow-down
    2
    ·
    10 months ago

    Henceforth, the building code shall make mandatory that every room be perfectly grounded Faraday cages (/s).

    Still, imagine lethal drones integrated with that technology (of course, they already have infrared, maybe even some adequate wavelength of X-rays).

    Nevertheless, pretty cool to see how far we can take preexisting technology with the help of some deep learning layers.

    • Maggoty@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      10 months ago

      Here’s what they’re putting in the goggles that Infantrymen wear now.

      I don’t care to guess what the drones are packing.

      • Shurimal@kbin.social
        link
        fedilink
        arrow-up
        4
        arrow-down
        1
        ·
        10 months ago

        What we know about drones is that they have cameras that can discern individuals from 10 km altitude.

        What we suspect is that US has Hubble-sized spy satellites that can do almost the same. There were a lot of classified military STS missions.

        What is theoretically possible is that US drones and spy sats can function as very large arrays (we do this with astronomical telescopes already) to dramatically increase spatial resolution.

        • Maggoty@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          10 months ago

          I’d believe it. When I was in the infantry 20 years ago we could see you 3km away with the optics mounted on our machineguns. And several kilometers with cameras mounted on towers. I don’t know how far they went but it was at least 5km because we were directing mortar fire with them and that’s about the range of the mortar system we were using.

        • SeducingCamel@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          ·
          10 months ago

          Oh I wonder if that’s how the Pic was taken that trump tweeted out of that rocket launch site, people didn’t think it was physically possible for a satellite to have that resolution

          • Shurimal@kbin.social
            link
            fedilink
            arrow-up
            1
            ·
            10 months ago

            It all comes down to the size of the mirror/lense—the bigger, the better. Up to a point. The biggest problem is air currents and different air densities refracting light and distorting the image. That’s what these laser beams are for on photos taken of astronomical observatories—they give reference light spot that can be used to calibrate adaptive optics to current atmospheric conditions reducing distortion.