Artificial Intelligence has the potential to disrupt so many different dimensions of our society that the White House Office of Science & Technology Policy recently announced a series of four public workshops to look at some of the possible impacts of AI. The first of these workshops happened at the University of Washington on Tuesday, May 24th, and I was there to cover how some of these discussions may impact the virtual reality community.
The first AI public workshop was focused on law and policy, and I had a chance to talk to three different people about their perspectives on AI. I interviewed the White House Deputy U.S. Chief Technology Officer Edward Felten about how these workshops came about, and the government’s plan for addressing the issue.
I also talked with workshop attendee Sheila Dean, who is a privacy advocate about the implications of AI algorithms making judgments about identified individuals, as well as Ned Finkle, who is the Vice President of External Affairs at NVIDIA about the role of high-end GPUs in the AI revolution.
LISTEN TO THE VOICES OF VR PODCAST
There are a number of take-aways from this event that are relevant to the VR community.
First of all, there are going to be a number of different privacy issues, like the biometric data that could be collected from virtual reality technologies including eye tracking, attention, heart rate, emotional states, body language, and even EMG muscle data or EEG brainwaves. There are a number of companies now using machine learning techniques in order to analyze and make sense of these raw data streams. Storing this type of biometric data and what it means could have some privacy implications. For example, Conor Russomanno warned me that EEG data could have a unique fingerprint and so there could be implications of storing anonymized brainwave data because it could still get tracked back to you.
I also discussed tracking user behavior and data with Linden Lab’s Ebbe Altberg, where we talked about the potential dangers of having companies ban users based upon observed behavior. Will there be AI approaches that either grant or deny access to virtual spaces based upon an aggregation of behavioral data or community feedback?
Sheila Dean was concerned that she didn’t hear a lot of voices that were advocating for privacy rights of users in the context of some these AI-driven tracking solutions. She sees that we’re in the middle of a battle where our privacy awareness and rights are eroding, and that users need to be aware of what’s at stake when AI neural nets start to flag us as targets within these databases. She says that consumers need to advocate for data access, privacy notice consent, privacy controls, and for people to be more aware of their privacy rights. We have the right to ask companies and the government to send us a copy of the data that they have about us because we still own all of our data.
Sheila also had a strong reaction to Oren Etzioni’s presentation. Etzioni is the CEO of the Allen Institute for Artificial Intelligence, and he had a rather optimistic take on AI and the risks. He had a slide that labeled SkyNet as a ‘Hollywood Myth’, though Sheila notes that SkyNet is a very real NSA program. She cites an article by the Intercept that there’s an actual NSA program called SKYNET that uses AI technologies to identify terrorist targets.
At the same time, SkyNet is kind of seen as the ‘Hitler’ of AI discussions, and we could probably adapt Godwin’s Law to say, “As an online discussion [about AI] grows longer, the probability of a comparison involving [SkyNet] approaches 1.”
There have been a lot of overblown fears about AI seeded by dystopian sci-fi dramas coming out of Hollywood. These fears have the potential to prevent AI from contributing to the public good in many ways, from saving lives to making us smarter.
Microsoft Research’s Kate Crawford sees that discussions that jump straight to ‘SkyNet’ can make practical, nuanced discussions difficult. She was advocating for stronger ethics within the computer science community, as well as a more interdisciplinary approach to encompass many different perspectives with the AI as possible.
In Alex McDowell’s presentation at Unity’s VR/AR Vision Summit, he argued that VR represents a return to valuing multiple perspectives. Stories used to be transmitted through many generations through oral traditions where tribes would add and adapt the story based upon their own recent personal stories and experiences.
Alex says that the advent of print, film, and TV, marked a shift where we started to see canonical versions of stories that were told primarily from a singular perspective. But VR has the potential to show us the vulnerability of the first-person perspective, and as a result put more emphasis on ensuring that our machine learning approaches include a diversity of perspectives across many different domains.
Right now AI is very narrow and focused on specific applications, but moving towards artificial general intelligence means that we’re going to have to discover some of the underlying principles that are transferable to building up a common sense framework for intelligence. Artificial general intelligence is one of the unsolved and hard problems in AI, and so far no one knows how to do this yet. But it’s likely that it’s going to require cross-disciplinary collaboration, holistic thinking, and other ingredients that have yet to be discovered.
Continue Reading on Page 2
Page: 1 2