• AppleTea@lemmy.zip
      link
      fedilink
      English
      arrow-up
      30
      arrow-down
      1
      ·
      5 hours ago

      whaaaat surely BYU, the school that claimed to have done cold fusion, is an upstanding pillar of academic research

    • Gork@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      14
      ·
      5 hours ago

      This would be the perfect use case for that fancy Apple VR headset they released a year or two so. Since it has built-in eye tracking, it would be easy to set up a test in a controlled environment where participants navigate it while looking around.

      • bleistift2@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        4
        ·
        4 hours ago

        Navigating that scene in real life (or even simulated) would make the data orders of magnitude more annoying to interpret. On a static image you can just overlay all eye movements and produce a heatmap. But for a subject that’s actually (or virtually) moving, none of the data would coincide and you’d have to manually find out which focus points were actually equal.

    • III@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      5 hours ago

      I feel like utilizing eye tracking would be used if they were to study this concept more deeply. That data would be more complicated to sift through given how much data and how many variables might come into play. Definitely more telling but also harder to analyze.