The Parliamentary Office of Science and Technology has published new
impartial research looking into the use, misuse and collection of
biometric data.
Biometric data allows the unique identification of a person.
Facial, voice or fingerprint data is commonly used to unlock
personal devices, speed up passport checks or for online
banking.
Research indicates that people prefer to use biometric
verification to log in to devices over using a password or
authenticator app. As such, the ‘passwordless verification'
market is projected to be worth £46billion by 2032.
This new paper, co-authored by Vincent Saverat, PhD student at
Imperial College London, and Dr Simon Brawley from the
Parliamentary Office of Science and Technology, was based on
interviews with experts and brings together findings from over
170 sources to provide a unique overview of the how biometric
technology is advancing and where the risks might lie.
Highlights
- AI has made new uses of biometric data possible. It can be
used to classify individuals into different demographic
categories and to infer a person's emotional or psychological
state.
- Some systems can determine the characteristics of users, such
as age, and some systems even claim to be able to determine
sexuality, although studies relating to sexuality have been
widely criticised as ‘pseudo science'.
- Collection of biometric data is regulated differently in the
UK, EU and US.
- Successive governments have planned to legislate further
around the collection and use of biometric data. The previous
government's Data Protection and Digital Information Bill fell at
the general election, and the new government has proposed a
Digital Information and Smart Data Bill to ‘ensure [public] data
is well protected'. The Bill has not yet been
published.
- There have been several high-profile cases of legal action
involving biometric data. US firm Clearview AI, which provides a
‘face search' function drawing on a database of more than 40
billion images, was fined €20million for breaching EU data
regulations. In 2019 HMRC was found to have collected the voice
records of 5 million users without consent as part of a new voice
verification system. It was subsequently required to delete them
under GDPR regulations.
- Facial recognition is particularly controversial due to
privacy concerns and due to concerns about bias. UK police forces
use facial recognition software to biometrically identify
suspects, for example in dense crowds. However, there is evidence
to suggest multiple facial recognition systems give false
positive identifications more frequently when the individual
being identified is from a Black or Asian background. False
positives are also more common in women, children and older
people. This is likely because AI algorithms within the programme
were trained on datasets where certain groups were
under-represented.
Professor Lord Robert Winston, Vice Chair of the POST
board, said: “This fascinating POST research clearly
demonstrates that biometric technology is developing faster than
our understanding of its implications. Many of the technologies
outlined here would have been unthinkable a few years ago. If we
want to protect our most valued asset - what makes us unique, our
very identities - we must ensure that we keep abreast of how this
data is being used and ensure that it is effectively regulated.
This research will be invaluable in helping parliament play its
role in the debate.”
Vincent Saverat, POST fellow, said “Writing
this paper alongside the Parliamentary Office of Science and
Technology has given me the opportunity to ensure that clear and
impartial evidence on this quickly developing subject is shared
with the people charged with regulating it. Technological
progress can happen fast, and often brings with it unexpected
technical and ethical considerations. Drawing together this
research will help clarify some of issues at the core of new
biometric data uses."