BIOMETRIC SOLUTIONS TO IDENTITY-RELATED CRIME EVIDENCE VERSUS POLICY Russell G Smith Australian Institute of Criminology Conference paper presented at: Delivering crime prevention : making the evidence work Carlton Crest Hotel, Sydney 21-22 November 2005 This conference was organised by the Australian Institute of Criminology in conjunction with the Crime Prevention Division of the New South Wales Attorney General’s Department. http://www.aic.gov.au/conferences/2005-cp/ Note: copyright for these papers/speeches rests with the authors and the AIC accepts no liability for the material contained within them. These papers/speeches are not peer reviewed and in particular, speeches may not necessarily meet the standards of the AIC's formal publication series. Their presence on the AIC website should not be constituted as an endorsement of the views contained within them. Australian Institute of Criminology and Crime Prevention Division, Attorney-General’s Department, New South Wales Delivering Crime Prevention: Making the Evidence Work Carlton Crest Hotel, Sydney 21 November 2005 “Biometric Solutions to Identity-related Crime: Evidence versus Policy Russell G. Smith Introduction Identifying people with certainty is both a time-consuming and costly activity for public and private sector organisations. Each year in Australia, government agencies need to identify and record approximately half a million new Australian residents (including births, permanent new arrivals and long-term visitors), 2.5 million enrolment forms need to be processed by the Australian Electoral Commission, the Australian Taxation Office issues almost half a million new tax file numbers; Centrelink grants 2.8 million new claims for benefits, and the Department of Foreign Affairs and Trade issues over one million travel documents. On each occasion, evidence of identity is required. Millions of people every day also log-on to computer networks, for work, to withdraw cash from Automated Teller Machines, or to use the Internet for recreation or business. In 2005, the Reserve Bank reported that there were 772 million ATM withdrawals and 1,176 million EFTPOS transactions in Australia, each of which required a customer to enter a PIN (Australian Payments Clearing Association 2005). There are also millions of occasions each year in which individuals have to be identified for access to buildings, travel purposes such as air, train, bus and tram travel, collection of highway tolls, and use of retail customer cards in shops and businesses. In order to identify oneself, it is usual to produce or disclose something that you have (such as a card or other token), something that you know (such as a password or PIN), something related to who you are (such as a fingerprint or other biometric), or something indicating where you are located (such as an address or phone number). Of course, there are others, such as the use of a person’s name, and a variety of behavioural and psychological characteristics that can be used to identify people. Depending upon the degree of certainty with which one needs to establish a person’s identity, one or more of these various methods may be relied upon. Often only one method of information will be used, although so-called 'two-factor authentication' is now becoming the standard. Each, however, has its own vulnerabilities and risks, which are able to be exploited by those who want to act illegally. Identity-related Fraud Risks of Knowledge-based and Token-based Systems Recently, however, reliance on simple knowledge-based and token-based systems have created many opportunities for fraud. In the news media, reports regularly appear of credit card numbers and other personal information being taken from databases and misused. In the United States in May 2005, for example, the processor of payment card data, CardSystems Solutions Inc, had its database breached and credit card account 2 information including magnetic stripe data and cardholder names relating to over 40 million accounts were stolen. Over 130,000 Australian cardholders were affected as a result (Krim and Barbaro 2005). Cases continue to come before the courts of fraud facilitated through the misuse of computer passwords, and fraudulent identities have also been involved in the commission of a wide range of other criminal activities including obtaining finance dishonestly, opening bank accounts in false names, money laundering, car re-birthing, credit card skimming, obtaining family allowance benefits, obtaining security guard licences and shooters’ licences, avoiding driving demerit points and producing TOEFL Certificates for immigrants. It has recently been estimated that identity-related fraud cost $1.1 billion in Australia alone in 2002 (Cuganesan & Lacey 2003). The precise way in which identity-related fraud is perpetrated, and the ease with which it can be carried out, will depend on the type of authentication system that is in place. In particular, it will depend on the nature of the data which are associated with individuals, and how easy they are to obtain and/or replicate. For example, to defeat knowledge-based systems it is necessary for an offender to ascertain and use the piece of information that the user knows (password). This information may be learnt directly from the user (who may share their password with friends or co-workers, or be tricked into revealing it through ‘social engineering’), may be guessed or ‘cracked’ through the use of computer technology, or may be obtained through practices such as ‘shoulder-surfing’ (where an individual watches a person entering their password or PIN into a machine) or ‘dumpsterdiving’ (where an individual searches through a person’s rubbish for relevant information). Different challenges are faced by those who seek to defraud a token-based system. As such systems rely on the association of people with a physical object, circumventing them will generally require an individual to acquire the required token. This can be done either by stealing or purchasing a legitimately manufactured token, or forging a copy. The ease with which such objects can be counterfeited will depend on the nature of the object, and any document security features (such as holographic images) that have been incorporated. Although the use of such security features may make it more difficult to defraud tokenbased systems, with advances in computer technology it is usually possible for a determined identity thief to bypass even the most secure systems and counterfeit documents containing the appropriate security features (Smith 1999). Even if documents cannot be successfully counterfeited, it may still be possible to buy or steal them. Biometric Authentication Due to the vulnerabilities of knowledge-based and token-based systems, a number of organisations and government agencies are moving to ‘biometrics’. The United Kingdom Biometrics Working Group (2002, p.4) define biometrics as ‘the automated means of recognising a living person through the measurement of distinguishing physiological or behavioural traits’. In other words, biometric systems are based on who a person is, rather than what a person has or what he or she knows. Whether by fingerprint, voiceprint, iris 2 3 pattern or a number of other characteristics, it is possible to measure individuals’ personal attributes to help identify them. In Australia, for example, on 24 October 2005, a biometrically-enabled Passport was first made available in which the personal information currently recorded on the passport is now kept on a computer chip embedded in the centre pages of the Passport. Already some 2,500 e-Passports have been issued, and trials are being conducted involving airline staff and some others which enable them to use facial recognition technology in conjunction with the e-Passport to proceed through customs controls at airports (Nash 2005). Biometric systems entail two processes: enrolment and matching. In the enrolment phase, an individual’s biometric characteristic (such as a fingerprint) is acquired for the first time. The image acquired will usually be converted into a ‘template’, against which subsequent comparisons are made. In the matching phase, an individual’s biometric characteristic is captured again. This ‘live template’ is compared against previously enrolled data, seeking a match. The key question for biometric systems operating in verification mode is ‘are you who you claim to be?’. Other issues arise in the contexts of using biometrics for identification and ‘watch list’ checking, such as for border control and immigration purposes. Evaluating Biometric Systems The ways in which biometric systems can be evaluated differ depending upon the particular type of biometric technology to be tested, as well as the purpose for which it is being used - whether for identification or surveillance. No single test has been developed which can accurately measure all issues across different biometric devices, in a uniform way. Because of this multiplicity of ways in which biometric systems can be evaluated, it is not readily possible to determine which is the ‘best’ biometric, because some systems will perform well on one measure, but will be outperformed on others. The choice of which biometric system to deploy, if any, will depend on the particular needs and priorities of the organisation, including the location and purpose of the system, and the number and nature of the people who will be using it. Policy makers are faced with a wide range of considerations when deciding whether or not to implement biometric solutions to identity-related crime. On the one hand, policy makers must evaluate a considerable and ever-increasing body of technical evidence relating to the performance of biometric technologies, while on the other hand a range of social, legal, and practical considerations need to be addressed including privacy, data security, user acceptance, and cost. Compelling evidence of performance should not, however, overwhelm these non-technological considerations. There are nine key issues to consider. Enrolment At the outset, it is important to bear in mind that the traditional ways in which identity is established will still be required during the enrolment phase. A person’s biometric does not, by itself, provide evidence of identity: ‘biometric systems can only confirm or 3 4 determine a claimed identity – one established upon a system at enrolment – as opposed to revealing a “true” identity’ (International Biometric Group 2003, p. 17). To this end, it is important to ensure that appropriate identification documents are still provided, and background checks made prior to enrolment as the integrity of a biometric system is only as good as the quality of the enrolment data (Dunstone 2003). If care is not taken in the enrolment phase, it will continue to be possible for a person to defraud the system, defeating the purpose for which biometrics are used. Moreover, a real danger arises if a person can successfully bind their biometric data to a stolen identity, because this will allow them to continue using the stolen or fabricated identity for a variety of fraudulent purposes, with little risk of detection. It will generally be accepted that because a person can provide a biometric that matches the assumed identity, they must be that person. This may have a significant detrimental effect on the person whose identity has been stolen, and can take a long period of time to correct. Some of the most difficult issues relate to the creation of databases of personal information, and the enrolment of individuals within them. The techniques of deception often reach back to the very creation of the records entailed in a system of identification. If a false identity can acquire the trappings of legitimacy from the very start, it makes a much more effective tool for the identity fraudster, being practically impossible to detect. With established databases, such as those held by the Australian Taxation Office or state offices of Births Deaths and Marriages, there is an on-going need to cleanse the data to ensure that the information recorded about individuals is correct. Some changes that occur may be legitimate, such as changes of name on marriage or through formal change of name procedures. Others, however, are dishonest. In Australia, the Report of the House of Representatives Standing Committee on Economics, Finance and Public Administration (2000), Numbers on the Run, reported the finding of the Australian National Audit Office that there were 3.2 million more Tax File Numbers than people in Australia and 185,000 potential duplicate tax records for individuals. The ATO currently believes that there are over 25,000 duplicate tax file numbers which form part of 6.5 million inactive TFN records (Australian Taxation Office 2004, p. 6). Performance There are a number of ways in which the performance of biometric systems can be measured. These include collecting and analysing the: • failure to enrol rate (FTER) – this measures the proportion of users who, for some reason, cannot enrol in a particular biometric system; • failure to acquire rate (FTAR) – this measures the proportion of cases where a user seeks to provide a biometric to match against their previously enrolled template, but the system cannot acquire an image of sufficient quality; • false match rate (FMR) – this measures the probability that a sample will be falsely declared to match the template of another person; • false non-match rate (FNMR) – this measures the probability that a sample will be falsely declared not to match the template of the user who provided the sample; 4 5 • • • false accept rate (FAR) – this measures the proportion of cases in which an impostor is falsely accepted by a biometric system; false reject rate (FRR) – this measures the proportion of cases in which a genuine user is falsely rejected by a biometric system; and the equal error rate (EER) – this is usually the point at which the false reject and false accept rates are equivalent. In some cases, it can also refer to the point at which the false match and false non-match rates are equivalent. It should be noted that while these measures are widely used, their use is not always consistent. Evaluations are also often carried out within the industry promoting the technology in question, casting doubts on the objectivity of some reports. This makes it vital for policy makers to inspect any biometric evaluation report closely before accepting its results. In addition, while each of the abovementioned measures can be used to evaluate biometric technologies, they are not the only possible ways of evaluating biometric systems. Some of the other issues that need to be considered are as follows. Data Security The implementation of biometrics in any case is not a simple matter, and the two models of storage of biometric information each offer their own challenges and risks. On the one hand, biometrics applications for verification may compare the individual’s body – whether it be a fingerprint, face, hand, or iris – directly to the template recorded on a card or other portable medium. If this is the case then defeat of that medium’s security features may allow replacement or alteration of the template, unbeknown to the system administrators. For facial recognition, depending on the model used, this might be as straightforward as photo substitution or as complex as cracking strongly encrypted data. If, on the other hand, the comparison is to a template held on a central database, that database would represent a high-profile target for hackers, organised criminals and other parties. Securing that information, for example using public key encryption, and ensuring inside parties are not able to access and alter information inappropriately, represents a major challenge. In recent times, various large-scale information databases, including client details held by major credit card companies and respected commercial providers, have been defeated or compromised by outsiders and insiders alike. To anticipate the points of susceptibility to interference is the essential challenge of implementing biometric applications on a wide scale. Spoofing Spoofing refers to techniques developed or adapted to challenge the biometric in question. There are three main ways in which a system can be attacked (Thalheim et al. 2002): The first involves the creation of an artificial biometric. This involves putting artificially created data into the regular sensor technology of the system. For example, a fake finger could be used to deceive a fingerprint scanner, or a photograph used to deceive a facial recognition system. For this approach to work, it is necessary for the 5 6 impostor to obtain a copy of the biometric that they wish to use. This could be done, for example, by taking a photograph of the person to be imitated. The second, known as relay attacks, involves the use of artificially created data. However, instead of obtaining the relevant data by copying the biometric to be used, this method involves capturing the relevant data as they are input into the sensor, through use of a device such as a sniffer program. This is a device which can be attached to the back of a computer (for example, in the USB port), which can obtain information as it is input into the computer. The data captured can then be replayed, to deceive the system. A researcher at the Australian National University in 2002, for example, demonstrated how fingerprint verifiers could be circumvented by presenting a template resembling that electronically stored in the device (Baker 2002). Finally, there are database attacks which seek to compromise the databases in which the data are stored. This will usually need to be done by someone who has administrator rights over the database, although it could be done through an external attack on the database (hacking). One way such an attack could take place is where an individual who works on the development of the system forges user data that are reactivated at a later date to their advantage. For example, a person could match his or her own fingerprint to a false identity, which could then be used for fraudulent purposes. The first of these methods is the main focus of literature in this area, as it is the most likely avenue for spoofing. People are more likely to attempt to spoof a fingerprint recognition system by using a rubber finger, or a voice recognition system by using an audio recording, than they are to hack into the database itself. Many technology developers have attempted to create countermeasures to prevent such attacks. The most common of these countermeasures is what is known as ‘liveness’ testing. Many systems have some form of technology capable of checking that the biometric characteristic being measured belongs to a live person. For example, fingerprint scanners may test the temperature of the finger, while iris scanners may search for a pulse or rapid eye movements. Such systems have a dual advantage: they can help to prevent spoofing, as well as potentially preventing some forms of crime displacement (see below). Unfortunately, there is little research on the steps that have been taken to prevent the less common forms of attack. There are also few evaluations that measure the ability of systems to repel concerted attacks. One of the few studies in this area was conducted by three German researchers who attempted to spoof a number of different devices (Thalheim et al. 2002). While some of the devices caused them slight difficulties, with a little persistence they managed to compromise each device investigated. Facial recognition systems, for example, have yet to be tested against people seriously motivated to evade detection through prosthetic and cosmetic adjustments to their facial shape and size. Indeed, there have been reports that transplant surgery and immune system drugs may very soon enable faces to be transplanted. This might not currently be considered a realistic threat, and in any case is unlikely to occur on a sufficiently wide 6 7 scale to affect most systems. Yet, in view of the global awareness of how far terrorists may go in furtherance of their aims, such concerns need to be taken into consideration. Privacy While some have claimed that biometrics can be a privacy-enhancing technology (Biometrics Institute 2002), there is a general perception that the use of such technologies is likely to invade privacy. Some of the main privacy concerns include fears that biometric information will be gathered without permission or knowledge, or without explicitly defining the purpose for which it is required; that information may be used for a variety of purposes other than those for which it was originally acquired (‘function creep’); shared without explicit permission; or used to track people across multiple databases to amalgamate information for the purpose of surveillance or social control (General Accounting Office, United States 2002). In recent times with ever-present concerns over terrorism, a number of countries have decided to issue compulsory identity cards, some of which include a biometric identifier. Hong Kong, for example, has developed multi-use ID ‘smartcards’ which contain basic biometric information such as thumb prints and a photograph, and are capable of multiple functions including use as drivers’ licences and as library cards (Benitez 2002; South China Morning Post 2002). A pilot program for a biometric ID card has also been implemented in Britain, in relation to asylum seekers (McAuliffe 2002). Such proposals face vocal opposition from advocates of privacy who raise the grave consequences of essential information being misused such as occurred during the Nazi regime in the Second World War. One writer refers to ‘the singular ease with which population registration systems have been mobilized for genocidal purposes’ (Seltzer 1998, p. 544). Any use of biometric systems needs to comply with privacy principles and privacy legislation (Crompton 2002). In Ontario, for example, policies were developed to govern the use of biometric systems designed to prevent welfare fraud. The following privacy protective provisions were included in the Canadian Social Assistance Reform Act (Cavoukian 1999, p. 5): • • • • • any biometric information collected under the Act must be encrypted; the encrypted biometric cannot be used as a unique identifier, capable of facilitating linkages to other biometric information or other databases; the original biometric must be destroyed after the encryption process; the encrypted biometric information only can be stored or transmitted in encrypted form, then destroyed in a prescribed manner; and no program information is to be retained with the encrypted biometric Information. The Act also required that systems should be unable to be implemented that could reconstruct or retain the original biometric sample from encrypted biometric information, 7 8 or that could compare it to a copy or reproduction of biometric information not obtained directly from the individual. These largely reflect the Information Privacy Principles that currently exist in Australia under the Privacy Act 1988 (Cth). User Acceptance Past experience has shown that the efficiency and accuracy of biometric systems can be reduced if those required to use the system are not willing to accept the technology: ‘user attitude can make or break the implementation of a biometric system’ (United Kingdom Biometrics Working Group 2002, p. 7). Some people may find the process of providing personal information in public distasteful. This was one reason given for the reluctance of retailers to make use of a cheque fraud prevention initiative which required customers to leave their fingerprint on cheques before they would be accepted by retailers (see Pidco 1996). Similarly, users may associate fingerprints with policing and criminality and feel reluctant to use fingerprinting systems. Still others may believe that systems which scan irises or retinas may harm their eyes (despite clear evidence to the contrary). Both end users and administrators of systems may be reluctant to make use of them. Accordingly, the need arises to educate users about the reasons why the system has been introduced and how it might benefit them. User concerns relating to privacy and security of data storage, as well as the safety of using some devices especially eye-based biometrics), would also need to be addressed. Rectification Another problem associated with biometrics arises from the fact that once a system has been compromised, it may be difficult to rectify the problem. While a new PIN can always be issued, new fingerprints cannot. Any error, corruption or systematic failure of the biometric identifier will be as permanent and irrevocable as the ‘correct’ identification is supposed to be. Even if the enrolment process remains error-free, a biometric is effectively a ‘PIN you can never change’ – and compromised once, is compromised for all time (Biometrics Institute 2002). Meanwhile, the ostensibly greater security afforded by the use of the biometric may lead to overconfidence in its accuracy, which could make any cases of successful identity fraud that much more damaging. Cost There are a wide range of costs involved in the implementation and use of biometric systems. It is especially important to consider recurrent costs, which can often outweigh the costs of infrastructure and initial implementation. Often however, these differing cost considerations are grouped together and presented in simplistic charts, such as that contained in a Report from the International Biometric Group (2003) which seeks to compare a range of considerations relating to different biometrics in one chart. Although simple to read, it could be grossly misleading. Displacement 8 9 Finally, the use of biometrics as a crime reduction strategy, like many other crime reduction techniques, carries with it the risk that displacement may occur. Displacement has been defined as: ‘a change in offender behaviour, along illegitimate means, which is designed to circumvent either a specific preventive measure or more general conditions unfavourable to the offender’s usual mode of operating’ (Gabor 1990, p. 66). One academic described the problem of displacement of crime as follows: Fear of displacement is often based on the assumption that offenders are like predatory animals - they will do what ever it takes to commit crimes - just as a rat will do whatever it takes to steal food from the cupboard (Eck 1998). If it is assumed that potential offenders act on the basis of some rational calculation in which they balance up the likely risks and benefits to be derived from a potential course of conduct, then as some types of crime are seen to become too difficult to commit, other easier targets may be considered. The use of biometrics for logical access control could result in offenders obtaining access to computers through bribery or coercion of IT personnel or other gatekeepers within organisations, or forcing users under threat of violence to permit the offender to have access by presenting their biometric under duress. We have already seen the occurrence of this with duress being used by offenders against users at ATMs to compel them to withdraw cash. Failure to comply has even resulted in users being killed in some countries. Conclusions This paper has identified some of the considerations that need to be taken into account when deciding whether or not to implement biometric systems. The concern is that the introduction of a new technology may make matters worse – either with respect to the specific crime problem sought to be addressed, or by creating new risks through the infringement of privacy or displacement to other forms of more serious crime. Careful thought also needs to be given to what biometric technologies cannot do. Of greatest importance is the fact that they cannot validate identity upon initial enrolment. If checks are not in place to validate the evidence of identity produced upon enrolment, then the subsequent use of a biometric authentication system may make identity-related crime easier to perpetrate, and more difficult to detect. Offenders may simply put their energies into creating a false identity and using that when enrolling in a biometric system. The decision about which biometric technology, if any, to implement is not a simple one. There is a wide range of factors which need to be considered, including the performance of the system, its ability to be compromised, the ease of using such systems, and user concerns about issues such as privacy and security. In addition, the cost-effectiveness of rolling out a biometric system must be determined, with consideration given to both implementation and recurrent costs. These should be carefully weighed against the cost 9 10 of using other viable alternatives methods of identification, such as knowledge or tokenbased systems, in light of the particular security needs of the organisation. Of great importance from the point of view of policy makers, however, is the need to balance the evidence that exists in support of, and against any given system in relation to each of the various considerations outlined above. Policy makers should avoid the temptation solely to focus on the seemingly convincing evidence of technical performance provided by the industry concerned. Technical performance is only one criterion, and even this can be measured in a wide range of ways. Instead, evidence needs to be sought out and scrutinized concerning the range of other legal, social, and ethical considerations governing the use of any given system. Unfortunately, it is these aspects which have yet to be fully researched. References Australian Payments Clearing Association 2003, Payment Statistics http://www.apca.com.au Baker, L. 2002, ‘Rule of Thumb: Don’t Rely on New Security Systems’, ANU Reporter, vol. 33, no. 9, 7 June, http://www.anu.edu.au/pad/reporter/volume/33/09/acrobat.pdf. Australian Taxation Office 2004, Compliance Program 2004-05, Australian Taxation Office, Canberra. Benitez, M. A. 2002, ‘ID Card Contract Awarded’, South China Morning Post (Hong Kong), 27 February, p. 2. Biometrics Institute 2002, ‘The Impact of Biometrics on Privacy’, an Interview with Dr Roger Clarke, http://www.biometricsinstitute.org/bi/interviews.htm. Cavoukian, A. 1999, Privacy and Biometrics, http://www.ipc.on.ca/docs/pri-biom.pdf Crompton, M. 2002, ‘Biometrics and Privacy: The End of the World as we Know it or the White Knight of Privacy?’, presented at Biometrics-Security and Authentication Biometrics Institute Conference, Sydney, 20 March. Cuganesan S. and Lacey D. 2003. Identity Fraud in Australia: An Evaluation of its Nature, Cost and Extent, SIRCA, Sydney. Dunstone, T. 2003, The use of Biometric Technology in Airports http://www.biometricsinstitute.org/bi/Articles/0303_AirportReview1.pdf Eck, J. 1998, ‘Preventing Crime at Places’, in Sherman, L. W., Gottfredson, D., Mackenzie, D., Eck, J., Reuter, P. and Bushway, S. What Works, What Doesn’t, What’s Promising, National Institute of Justice, Washington. 10 11 Gabor, T. 1990, ‘Crime Displacement and Situational Prevention’, Canadian Journal of Criminology. General Accounting Office, United States 2002, Technology Assessment: Using Biometrics for Border Security, GAO-03-174, November, Washington DC, http://www.gao.gov/new.items/d03174.pdf. House of Representatives Standing Committee on Economics, Finance and Public Administration 2000, Number on the Run: Review of the ANAO Audit Report No 37, 1998-99 on the Management of Tax File Numbers, Parliament of the Commonwealth of Australia, Canberra. International Biometric Group 2003, Biometric Market Report 2003–2007. Krim, J. and Barbaro, M. 2005, '40 Million Credit Card Numbers Hacked', Washington Post, 18 June, p. A01. http://www.washingtonpost.com/wpdyn/content/article/2005/06/17/AR2005061701031_2.html McAuliffe, W. 2002, ‘Asylum Seekers Get First UK Biometric ID Cards’, ZDNet Australia, 5 February. http://www.zdnet.com.au/newstech/security/story/0,2000024985,20263301,00.htm Nash, B. 2005, 'Utilising the Latest in Biometrics Technology to Enhance Your Forensic Capability', paper presented at the IIR Conference Combating Identity Fraud, 1 November, Sydney. Pidco, G. W. 1996, ‘Check Print: A Discussion of a Crime Prevention Initiative that Failed’, Security Journal, vol. 7, pp. 37-40. Seltzer, W. 1998, ‘Population Statistics, the Holocaust, and the Nuremberg Trials’, Population and Development Review, vol. 24 no. 3, pp. 511-552. South China Morning Post (Hong Kong) 2002, ‘ID Card Plans Raise Issue of Carrier Privacy’, 17 January, p. 11. Smith, R. G. 1999, ‘Identity-Related Economic Crime: Risks and Countermeasures’, in Trends and Issues in Crime and Criminal Justice, No. 129, Australian Institute of Criminology, Canberra. Thalheim, L., Krissler, J., and Ziegler, P-M. ‘Body Check: Biometric Access Protection Devices and Their Programs Put to the Test’, c't Magazine (Germany), No. 11, May, http://www.heise.de/ct/english/02/11/114/. United Kingdom, Biometrics Working Group 2002, Use of Biometrics for Identification: Advice on Product Selection, http://www.cesg.gov.uk/site/ast/biometrics/media/Biometrics%20Advice.pdf. 11
© Copyright 2026 Paperzz