Facebook has announced the latest version of its successful standalone virtual reality (VR) headset, the Oculus Quest 2. The new device packs more computing power and a sharper screen than its predecessor, and is also US$100 cheaper.
The Oculus Quest 2 is the latest step in Facebook’s long-term strategy of making VR more accessible and popular. Facebook recently brought all its VR work under the umbrella of Facebook Reality Labs, it has announced new applications like the Infinite Office VR workplace, and will also require a Facebook login for future Oculus devices.
The compulsory link to Facebook has many consumers concerned, considering the social media giant’s chequered history with privacy and data. VR and its cousin, augmented reality (AR), are perhaps the most data-extractive digital sensors we’re likely to invite into our homes in the next decade.
Why does Facebook make virtual reality headsets?
Facebook acquired VR company Oculus in 2014 for an estimated US$2.3 billion. But where Oculus originally aimed at gamers, Facebook boss Mark Zuckerberg wants VR for social media.
At the same event last year, Zuckerberg said Facebook sees VR as a pathway to a new kind of “social computing platform” using the enhanced feeling of “presence” that VR affords. For Facebook, the introduction of VR-based computing will be like the leap from text-based command line interfaces to the graphical user interfaces we use today.
This may well be right. VR affords a strong feeling of embodied presence that offers new possibilities for entertainment, training, learning and connecting with others at a distance.
But if the VR future is the one Facebook is “working in the lab” on, it will function via the company’s existing social computing platform and business model of extracting data to deliver targeted advertisements.
Virtual reality collects real data
A VR headset collects data about the user, but also about the outside world. This is one of the key ethical issues of emerging “mixed reality” technologies.
As American VR researcher Jeremy Bailenson has written:
…commercial VR systems typically track body movements 90 times per second to display the scene appropriately, and high-end systems record 18 types of movements across the head and hands. Consequently, spending 20 minutes in a VR simulation leaves just under 2 million unique recordings of body language.
The way you move your body can be used to identify you, like a fingerprint, so everything you do in VR could be traced back to your individual identity.
Facebook’s Oculus Quest headsets also use outward-facing cameras to track and map their surroundings.
In late 2019 Facebook said they “don’t collect and store images or 3D maps of your environment on our servers today”. Note the word today, which tech journalist Ben Lang notes makes clear the company is not ruling out anything in the future.
Virtual reality leads to augmented reality
Facebook wants to collect this data to facilitate its plans for augmented reality (AR).
Where VR takes a user to a fully virtual environment, AR combines virtual elements with our real surroundings.
Last year Facebook unveiled the Live Maps application, a vision of an expansive surveillance apparatus presumably powered by AR glasses and data collected from Oculus Insight. Live Maps will provide many minor conveniences for Facebook users, like letting you know you’ve left your keys on the coffee table.
Now Facebook have announced their first steps towards making this a reality: Project Aria. This will involve people wearing glasses-like sensors around Seattle and the San Francisco Bay area, to collect the data to build what Wired co-founder Kevin Kelly calls “the mirrorworld”, the next big tech platform.
People are rightly concerned about the ethical implications of this kind of data extraction. Alongside Project Aria, Facebook launched its Responsible Innovation Principles page, and they’re already quick to emphasise that faces and license plates will be blurred in this data collection.
As we have argued elsewhere, framing questions about VR and AR surveillance in terms of individual privacy suits companies like Facebook very well. That’s because their previous failings are actually in the (un)ethical use of data (as in the case of Cambridge Analytica) and their asymmetric platform power.
We need more than just ‘tech ethics’
Many emerging technologies encounter what is known as the Collingridge problem: it is hard to predict the various impacts of a technology until it is extensively developed and widely used, but by then it is almost impossible to control or change.
We see this playing out right now, in efforts to regulate Google and Facebook’s power over news media.
As David Watts argues, big tech designs its own rules of ethics to avoid scrutiny and accountability:
Feelgood, high-level data ethics principles are not fit for the purpose of regulating big tech … The harms linked to big tech can only be addressed by proper regulation.
What might regulation of Facebook’s VR look like? Germany offers one such response – their antitrust regulations have resulted in Facebook withdrawing the headset from sale. We can only hope the technology doesn’t become too entrenched to be changed, or challenged.
But regulation has not always stopped Facebook in the past, who paid out US$550 million to settle a lawsuit for breaching biometric privacy laws. In the multi-billion dollar world of big-tech, it’s all a cost of doing business.
Another question we might ask ourselves is whether Facebook’s virtual-reality future and others like it really need to exist. Maybe there are other ways to avoid forgetting your keys.