Artificial Intelligence has the potential to disrupt so many different dimensions of our society that the White House Office of Science & Technology Policy recently announced a series of four public workshops to look at some of the possible impacts of AI. The first of these workshops happened at the University of Washington on Tuesday, May 24th, and I was there to cover how some of these discussions may impact the virtual reality community.

The first AI public workshop was focused on law and policy, and I had a chance to talk to three different people about their perspectives on AI. I interviewed the White House Deputy U.S. Chief Technology Officer Edward Felten about how these workshops came about, and the government’s plan for addressing the issue.

I also talked with workshop attendee Sheila Dean, who is a privacy advocate about the implications of AI algorithms making judgments about identified individuals, as well as Ned Finkle, who is the Vice President of External Affairs at NVIDIA about the role of high-end GPUs in the AI revolution.

LISTEN TO THE VOICES OF VR PODCAST

There are a number of take-aways from this event that are relevant to the VR community.

First of all, there are going to be a number of different privacy issues, like the biometric data that could be collected from virtual reality technologies including eye tracking, attention, heart rate, emotional states, body language, and even EMG muscle data or EEG brainwaves. There are a number of companies now using machine learning techniques in order to analyze and make sense of these raw data streams. Storing this type of biometric data and what it means could have some privacy implications. For example, Conor Russomanno warned me that EEG data could have a unique fingerprint and so there could be implications of storing anonymized brainwave data because it could still get tracked back to you.

SEE ALSO
Road to VR’s 2023 Game of the Year Awards

I also discussed tracking user behavior and data with Linden Lab’s Ebbe Altberg, where we talked about the potential dangers of having companies ban users based upon observed behavior. Will there be AI approaches that either grant or deny access to virtual spaces based upon an aggregation of behavioral data or community feedback?

oculus-social-alpha-touched-up
See Also: Facebook-like ‘Code of Conduct’ Governs Voice Communication in Oculus Social Alpha

Sheila Dean was concerned that she didn’t hear a lot of voices that were advocating for privacy rights of users in the context of some these AI-driven tracking solutions. She sees that we’re in the middle of a battle where our privacy awareness and rights are eroding, and that users need to be aware of what’s at stake when AI neural nets start to flag us as targets within these databases. She says that consumers need to advocate for data access, privacy notice consent, privacy controls, and for people to be more aware of their privacy rights. We have the right to ask companies and the government to send us a copy of the data that they have about us because we still own all of our data.

Sheila also had a strong reaction to Oren Etzioni’s presentation. Etzioni is the CEO of the Allen Institute for Artificial Intelligence, and he had a rather optimistic take on AI and the risks. He had a slide that labeled SkyNet as a ‘Hollywood Myth’, though Sheila notes that SkyNet is a very real NSA program. She cites an article by the Intercept that there’s an actual NSA program called SKYNET that uses AI technologies to identify terrorist targets.

SEE ALSO
The First $100 You Should Spend on Meta Quest Games

At the same time, SkyNet is kind of seen as the ‘Hitler’ of AI discussions, and we could probably adapt Godwin’s Law to say, “As an online discussion [about AI] grows longer, the probability of a comparison involving [SkyNet] approaches 1.”

There have been a lot of overblown fears about AI seeded by dystopian sci-fi dramas coming out of Hollywood. These fears have the potential to prevent AI from contributing to the public good in many ways, from saving lives to making us smarter.

Microsoft Research’s Kate Crawford sees that discussions that jump straight to ‘SkyNet’ can make practical, nuanced discussions difficult. She was advocating for stronger ethics within the computer science community, as well as a more interdisciplinary approach to encompass many different perspectives with the AI as possible.

In Alex McDowell’s presentation at Unity’s VR/AR Vision Summit, he argued that VR represents a return to valuing multiple perspectives. Stories used to be transmitted through many generations through oral traditions where tribes would add and adapt the story based upon their own recent personal stories and experiences.

Alex says that the advent of print, film, and TV, marked a shift where we started to see canonical versions of stories that were told primarily from a singular perspective. But VR has the potential to show us the vulnerability of the first-person perspective, and as a result put more emphasis on ensuring that our machine learning approaches include a diversity of perspectives across many different domains.

SEE ALSO
The First $100 You Should Spend on Meta Quest Games

Right now AI is very narrow and focused on specific applications, but moving towards artificial general intelligence means that we’re going to have to discover some of the underlying principles that are transferable to building up a common sense framework for intelligence. Artificial general intelligence is one of the unsolved and hard problems in AI, and so far no one knows how to do this yet. But it’s likely that it’s going to require cross-disciplinary collaboration, holistic thinking, and other ingredients that have yet to be discovered.

Continue Reading on Page 2

1
2
Newsletter graphic

This article may contain affiliate links. If you click an affiliate link and buy a product we may receive a small commission which helps support the publication. More information.


  • Lots of interesting points there, but as a Glass explorer, I’m now hyper-sensitive to how people throw around the “Privacy” topic and I think it’s a bit abused here. Privacy: “the state or condition of being free from being observed or disturbed by other people” and “the state of being free from public attention.” Since one cannot be unobserved in public, “privacy” cannot be an issue in public. That means that every time someone tries to make issues about privacy that don’t relate to a “private” (lacking public visibility) context, it makes no sense and it seems that MUCH of the “privacy” talk these days is guilty of just that. As it applies to AI and biometrics, this is true. In your home, this kind of observation would violate your privacy IF it was uninvited. Outside your home (be it permanent or temporary) and perhaps a bathroom, there is little to no expectation of privacy though.

    • Bruno

      Privacy is a right of everyone. You want it even in public. A micro-example: you wear clothes for privacy (beside other reasons, of course)

  • DonGateley

    For novelist (and all around smart guy) Neil Stephenson’s take on what a more benign Skynet derivative could be, see the reticulum in his novel Anathem.

  • Michael Cordes

    Imao the real danger of skynet is not the domination by the machines themself – that is certainly a contradictory and crazy idea and furthermore is hiding the real danger: the domination of man; e.g. a single or a certain group of man with the machines as their tool