A Policy Analysis of Phishing Countermeasures Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Engineering and Public Policy Xinguang (Steve) Sheng B.S., Computer Engineering, University of Pennsylvania M.S., Computer and Information Science, University of Pennsylvania Carnegie Mellon University Pittsburgh, PA December 2009 c Copyright by Xinguang (Steve) Sheng December 2009 All Rights Reserved i To my wife Phoebe Chao, for her loving support every step along the way. ii ACKNOWLEDGMENTS I would like to thank all the people who have helped and inspired me during my doctoral study. I especially want to thank my advisor, Prof. Lorrie Cranor, for her guidance during my research and study at Carnegie Mellon. Her warm and encouraging spirit, enthusiasm in research, and invaluable research insights had motivated all her advisees, including me. I am grateful for the rest of my thesis committee. They are Profs. Alessandro Acquisti, Jason Hong and Adrian Perrig. All of them have offered innumerable comments, suggestions, and feedback about the research presented in my thesis. I had the privilege of working with many bright and fun collaborators. I would like to thank Bryant Magnien, Patrick Kelly, Elizebeth Nunge and Ponnurangam Kumaraguru for Anti-phishing Phil; Yue Zhang, Brad Wardman, Gary Warner, Chengshan Zhang for the phishing blacklist study; Mandy Holbook and Julie Downs for the mechanical turk study; and Ponnurangam Kumaraguru for the expert interview study. All my lab mates at the Cylab Usable Security Privacy and Security Laboratory (CUPS) made it a convivial place to study. In particular, I would like to thank Ponnurangam Kumaraguru, Serge Egelman, Justin Crenshaw, Kami Vaniea, Patrick Kelley, Janice Tsai, Rob Reeder and Robert McGuire for their friendship and help in the past five years. Thanks. I owe my deepest gratitude to my family for their unflagging love and support throughout this dissertation is simply impossible without them. I am indebted to my father, Xiangchen Sheng, for his care and love and I cannot ask for more from my mother, Fengzhi Wang, as she is the most loving mother that I have met. My wife Phoebe has been an encourager, motivator, and counselor along every step of this thesis, in fact this thesis cannot be completed without her, therefore I will dedicate this thesis to her. I would regret my doctoral years at Carnegie Mellon if I did not join Antioch cell group at Pittsburgh Chinese Church Oakland (PCCO). I cherished the fellowship and support between me and them, and the friendships with my Christian brothers and sisters at PCCO. I treasured all precious moments we shared and would really like to thank them. I would also like to acknowledge the support of National Science Foundation under grant CCF0524189, the Army Research Office under grant number DAAD19-02-1-0389. Last but not least, thanks be to God for my life through all tests in the past five years. You have enlarged my tent and made my life more bountiful. May your name be exalted, honored, and glorified. iii ABSTRACT Phishing is a kind of attack in which criminals use spoofed emails and fraudulent web sites to trick people into giving up personal information. This thesis looks at the phishing problem holistically by examining various stakeholders and their countermeasures, and by surveying experts’ opinions about the current and future threats and the kinds of countermeasures that should be put in place. It composed of four studies. In the first study, we conducted semi-structured interviews with 31 anti-phishing experts from academia, law enforcement, and industry. We surveyed experts’ opinions about the current and future of phishing threats and the kind of countermeasures that should be put in place. Our analysis led to eight key findings and 18 recommendations to improve phishing countermeasures. In the second study, we study the effectiveness of popular phishing tools that are used by major web browsers. We used fresh phish that were less than 30 minutes old to conduct two tests on eight anti-phishing toolbars. We found blacklists were ineffective when protecting users initially. The tools that uses heuristics to complement blacklists caught significantly more phish than blacklistonly tools with very low false positives. In the third study, we describe the design and evaluation of Anti-Phishing Phil, an online game that teaches users good habits to help them avoid phishing attacks. We used learning science principles to design and iteratively refine the game. We evaluated Anti-Phishing Phil through laboratory and real-world experiments. These experiments showed that people trained with Anti-Phishing Phil were much better at detecting phishing websites, and they retain knowledge after one week. In the fourth and final study we present our results of a roleplay survey instrument administered to 1001 online survey respondents to study both the relationship between demographics and phishing susceptibility, and the effectiveness of several anti-phishing educational materials. Our results suggest that women are more susceptible than men to phishing iv and participants between the ages of 18 and 25 are more susceptible to phishing than other age groups. We explain these demographic factors through a mediation analysis. Educational materials reduced users tendency to enter information into phishing webpages by 40% percent; however, some of the educational materials we tested also slightly decreased participants tendency to click on legitimate links. v TABLE OF CONTENTS Page ABSTRACT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 2 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.1 1.2 1.3 Thesis statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Thesis contribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Outline of the thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 3 4 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 2.1 2.2 2.3 2.4 2.5 2.6 3 iii Anatomy of Phishing . . . . . . . . . . . . . . . 2.1.1 Planning . . . . . . . . . . . . . . . . . 2.1.2 Setup . . . . . . . . . . . . . . . . . . . 2.1.3 Attack . . . . . . . . . . . . . . . . . . . 2.1.4 Collection . . . . . . . . . . . . . . . . . 2.1.5 Fraud . . . . . . . . . . . . . . . . . . . Why people fall for phishing . . . . . . . . . . . Cost of phishing . . . . . . . . . . . . . . . . . . Recent developments in phishing . . . . . . . . . Phishing countermeasures . . . . . . . . . . . . 2.5.1 Legal solutions . . . . . . . . . . . . . . 2.5.2 Technology countermeasures . . . . . . . 2.5.3 Social response: awareness and education Economics of Information Security . . . . . . . . 2.6.1 Security investment . . . . . . . . . . . . 2.6.2 Security as externality . . . . . . . . . . 2.6.3 Misaligned incentives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 5 8 9 14 16 16 18 19 22 22 23 26 27 28 28 29 Improving Phishing Countermeasures: An Analysis of Expert Interviews . . . . . . 31 3.1 3.2 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 vi Page 3.3 3.4 3.5 3.6 4 Stakeholders . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3.1 Primary victims: . . . . . . . . . . . . . . . . . . . . . . . . 3.3.2 Infrastructure providers: . . . . . . . . . . . . . . . . . . . . 3.3.3 For-profit protectors: . . . . . . . . . . . . . . . . . . . . . . 3.3.4 Public protectors: . . . . . . . . . . . . . . . . . . . . . . . Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4.1 Recruitment and Participants . . . . . . . . . . . . . . . . . . 3.4.2 Interview Protocol . . . . . . . . . . . . . . . . . . . . . . . 3.4.3 Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4.4 Limitations . . . . . . . . . . . . . . . . . . . . . . . . . . . Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5.1 Evolving threat . . . . . . . . . . . . . . . . . . . . . . . . . 3.5.2 Stakeholder incentives . . . . . . . . . . . . . . . . . . . . . 3.5.3 What stakeholders should do . . . . . . . . . . . . . . . . . . 3.5.4 Law enforcement and education . . . . . . . . . . . . . . . . Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.6.1 Applicability of the Recommendations against Spear-phishing 3.6.2 Summary of findings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 34 34 35 35 35 36 37 37 38 39 39 42 47 56 62 62 63 Case Study of Browser-based Anti-phishing Solutions . . . . . . . . . . . . . . . . . 73 4.1 4.2 4.3 4.4 Background and Related Work . . . . 4.1.1 Anti-Phishing Heuristics . . . 4.1.2 Phishing blacklists . . . . . . 4.1.3 Related Work . . . . . . . . . Methodology . . . . . . . . . . . . . 4.2.1 Anti-phishing Testbed . . . . 4.2.2 Phishing Feed . . . . . . . . . 4.2.3 Evaluation Procedure . . . . . Results . . . . . . . . . . . . . . . . . 4.3.1 Length of Phishing Campaign 4.3.2 Blacklist Coverage . . . . . . 4.3.3 False Positives . . . . . . . . 4.3.4 Accuracy of Heuristics . . . . 4.3.5 Total Protection . . . . . . . . Discussion . . . . . . . . . . . . . . . 4.4.1 Limitations . . . . . . . . . . 4.4.2 Opportunities for Defenders . 4.4.3 Improving blacklists . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74 75 76 77 78 78 79 80 82 82 83 86 88 89 91 91 91 93 vii Appendix Page 4.4.4 5 Anti-Phishing Phil: A Case study in User education . . . . . . . . . . . . . . . . . . 95 5.1 5.2 5.3 5.4 5.5 6 Use of heuristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94 Introduction . . . . . . . . . . . . . . . . . . . . . Background and Related Work . . . . . . . . . . . 5.2.1 Why people fall for phishing . . . . . . . . 5.2.2 Tools to protect people from phishing . . . 5.2.3 Anti-phishing education . . . . . . . . . . Design of Anti-phishing Phil . . . . . . . . . . . . 5.3.1 Game Design Principles . . . . . . . . . . 5.3.2 Game Description . . . . . . . . . . . . . 5.3.3 Training Messages . . . . . . . . . . . . . 5.3.4 Pilot Test . . . . . . . . . . . . . . . . . . 5.3.5 Modified Game . . . . . . . . . . . . . . . Evaluation 1: Lab Study . . . . . . . . . . . . . . 5.4.1 Study design . . . . . . . . . . . . . . . . 5.4.2 Participant Recruitment and Demographics 5.4.3 Results . . . . . . . . . . . . . . . . . . . Evaluation 2: Anti-Phishing Phil Field Study . . . 5.5.1 Study design . . . . . . . . . . . . . . . . 5.5.2 Participants . . . . . . . . . . . . . . . . . 5.5.3 Results . . . . . . . . . . . . . . . . . . . Phishing Susceptibility Study 6.1 6.2 6.3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95 97 97 98 99 100 100 103 105 109 110 111 111 113 114 121 121 121 122 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126 Background and related work . . . . . . . . . . . . 6.1.1 Demographics and Phishing Susceptibility 6.1.2 Susceptibility vs. Risk Behavior . . . . . . 6.1.3 Security User Education . . . . . . . . . . Study Design . . . . . . . . . . . . . . . . . . . . 6.2.1 Recruitment . . . . . . . . . . . . . . . . . 6.2.2 Roleplay . . . . . . . . . . . . . . . . . . 6.2.3 Education Materials . . . . . . . . . . . . 6.2.4 Previous Experiences and Demographics . 6.2.5 Knowledge and Technical Background . . 6.2.6 Risk Perceptions . . . . . . . . . . . . . . Results . . . . . . . . . . . . . . . . . . . . . . . . 6.3.1 Measuring User Performance . . . . . . . 6.3.2 Regression Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127 127 128 128 130 130 130 134 135 136 137 137 137 137 viii Appendix Page 6.4 7 6.3.3 Gender and Falling for Phish 6.3.4 Age and Falling for Phish . 6.3.5 Effects of Education . . . . DISCUSSION . . . . . . . . . . . . 6.4.1 Limitations . . . . . . . . . 6.4.2 Summary of findings . . . . 6.4.3 Role of education . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139 141 143 145 145 145 146 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156 LIST OF REFERENCES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160 APPENDIX Appendix I: List of Recommendations . . . . . . . . . . . . . . . . . . 172 1 Chapter 1 Introduction Phishing is a kind of attack in which criminals use spoofed emails and fraudulent web sites to trick people into giving up personal information. Victims perceive these emails as associated with a trusted brand, while in reality they are the work of con artists interested in identity theft [57]. Phishing is a widespread problem that is impacting both business and consumers. In May 2009, MessageLabs estimated that 0.41% of the 3.3 billion emails going through their system each day were phishing emails [87]. Microsoft Research recently estimated that 0.4% of email recipients are victimized by phishing attacks [36]. The annual cost to consumers and businesses due to phishing in the US alone is estimated to be between $350 million and $2 billion [43, 92]. To reduce the damage due to phishing, stakeholders have implemented their own countermeasures: major web browsers have built-in filters (e.g. [10], [46], [90]), Internet service providers filter suspicious phishing emails, law enforcement officers find and prosecute phishers, and US government agencies and corporations now educate consumers on phishing. Phishing scams have also evolved, sometimes at a faster pace than countermeasures. Phishers launch attacks on specific groups (e.g. users of social networking sites) through multiple channels (e.g. phone, instant messaging), and phishing toolkits and compromised credentials are readily available for sale at low prices on Internet black markets [38]. Sophisticated phishing schemes such as man-in-the-middle attacks and malware are becoming more frequent [62]. As the battle against phishing continues, many questions remain about where stakeholders should place their efforts to achieve effective prevention, speedy detection, and fast action. Do 2 stakeholders have sufficient incentives to act? What should be the top priorities for the antiphishing community? 1.1 Thesis statement This dissertation aims to provide insights in answering the objectives raised above through four studies. This thesis presents recommendations about how to better fight phishing; these recommendations are informed by empirical data on the effectiveness of current approaches as well as systematic analyses of stakeholder interests and the phishing life cycle. Semi-structured expert interviews were used to rank and prioritize the recommendations. In addition, we used case studies on the effectiveness of web browser anti-phishing toolbars and anti-phishing education to provide empirical data for our analysis. The centerpiece of the thesis is an expert analysis of phishing countermeasures. We conducted semi-structured interviews with 31 anti-phishing experts from academia, law enforcement, and industry. We surveyed experts’ opinions about the current and future of phishing threats and the kind of countermeasures that should be put in place. Experts discussed technical countermeasures, education, and law enforcement, which led to eight key findings and 18 recommendations to improve phishing countermeasures. One of the findings from the expert analysis is that experts think education and awareness are important. However, not all experts agree on the effectiveness of end-user security education. To investigate this issue further, we conducted two in-depth studies. Firstly, we studied phishing susceptibility with a role-play survey administered to 1000 users of Mechanical Turk. This studies showed different demographic factors’ impact on phishing susceptibility. In the second study, we designed and evaluated Anti-Phishing Phil, an online game that teaches users good habits to help them avoid phishing attacks. 3 Another key area where experts commented on is the strategic position of browsers. Several experts noted that organizations are conservative about filtering and warning about phish because they are worried about false positives. To investigate this further, we studied the effectiveness of popular phishing tools that are used by major web browsers. 1.2 Thesis contribution This thesis is both timely and needed to reduce the negative consequences of semantic attacks on society. The education component of this research can potentially help reduce the increasing number of people who fall for phishing and other semantic attacks, the policy recommendations from this research could help government and various stakeholders to better prioritize their resources and manage their risks to fight for phishing and other semantic attacks. This thesis work builds on existing knowledge in the fields of computer security and privacy, human computer interaction, and economics, and adds to the literature with the following contributions. 1. We designed and evaluated Anti-Phishing Phil, an online game that teaches users good habits to help them avoid semantic attacks. People trained with Anti-phishing Phil were much better at distinguishing phishing website and legitimate websites, and retain their knowledge after one week. The Anti-Phishing Phil game has been played over 110,000 times world-wide and is being commercialized by Wombat Security Technologies. This research showed that computer users can be trained to make better online trust decisions if the training materials are presented in a fun and interactive manner and grounded in learning science principles. 2. We conducted semi-structured interviews with 31 anti-phishing experts from academia, law enforcement, and industry on phishing countermeasures. Our analysis led to eight key findings and 18 recommendations to improve phishing countermeasures. 3. We studied the effectiveness of popular phishing tools that is used by major web browsers. We found blacklists were ineffective when protecting users initially, the tools that uses heuristics to complement blacklists caught significantly more phish than blacklist-only tools 4 with very low false positives. We recommend toolbars vendors use heuristics to complement blacklists to speed up phishing detection. 4. We studied demographics and phishing susceptibility with an role play survey administered to 1001 users of Mechanical Turk. This research is the first study that studied demographics factors contributing to susceptibility to semantic attacks. We also demonstrated the successful use of mechanical turk to conduct online experiments. 1.3 Outline of the thesis The next chapter introduces the fundamentals of phishing attacks and some of the related work that builds the foundation for the thesis; Chapter 3 discusses the expert interviews study in depth, and presents the key findings and recommendations; Chapter 4 discussed our study on the effectiveness of popular phishing tools that is used by major web browsers; Chapter 5 described the design and evaluation of Anti-Phishing Phil, an online game that teaches users good habits to help them avoid phishing attacks; Chapter 6 discussed the role-play survey conducted with mechanical turk users to study demographics and phishing susceptibility. Finally, Chapter 7 presents conclusions from this thesis work and offers recommendations for public policy makers. 5 Chapter 2 Background In this Chapter, I will discuss the literature on phishing, relevant countermeasures, and the literature on economics of information security. 2.1 Anatomy of Phishing Phishing attacks usually take the following steps: planning, setup, attack, collection, fraud and abuse, and post attack [35]. To fight phishing better, we need to understand the nuts and bolts of the attack better. In this section, I described in detail how phishing attacks work. I model in detail both the attackers and the defenders in each of the phishing steps. Previous works have also modeled phishing attacks. The DHS report on phishing [105] separates phishing into seven steps and discusses countermeasures based on the model. The model does not consider IM phishing and voice over IP phishing, nor does it include stakeholders other than banks. My model includes different stakeholders and provides specific recommendations to counter those two attacks. The FSTC report [35] discusses the phishing life cycle, but does not provide details of attacks, or stakeholders. My analysis addresses these problems. 2.1.1 Planning Phishers first need to decide whom to target, and what information to steal. A recent study [38] shows that a large and thriving underground ecrime economy of highly specialized criminals exists. Based on this insight, I model phishers as rational agents. They pick and choose their targets to 6 Networth of the clients Security measures of the institution + + - User base of the institution - Difficulty of launching attacks Phisher Cost of launching attacks Figure 2.1 A model of a phisher’s decision process maximize their gain and minimize their cost and risks. Figure 2.1 lists the likely factors they consider when they plan the attack. • User base of the institution. The larger the institution’s user base, the higher the percentage that targets receiving phishing emails will actually have an account relationship with the institution. Today, each US household carries an average of 6.3 credit cards, and the issuance of these credit cards is concentrated among five banks.1 Based on my estimate, there is a 15%-45% chance that a household receiving a phishing email from one of these top banks is an actual customer of the bank2 . This rule does not apply to targeted attacks in which phishers have specific information about customers’ account relationships. • Net worth of clients. The higher the worth of the client, the higher the returns of an attack. As countermeasures become more widely available, phishers will launch more attacks against and put high net worth clients such as executives and small business merchant accounts at special risk. 1 For more information about credit card usage, see FDIC’s review at http://www.fdic.gov/bank/analytical/banking/2005nov/article2.html 2 The number of households in US are 110 million in 2010, the top creditcard issue bank (Citibank) have 48 million active cards, the lowest of five (Bank of America) have 20 million accounts. 7 • Security measures or processes implemented by the institution. The stronger an institution’s security measures and processes are, the harder it is for phishers to penetrate, commit fraud, and launder money. Phishers can learn about an institution’s preparedness through previous attack experience and publicly available information (news and press releases). • The value of credentials. The more valuable the credentials are, the more frequent the attacks will be. Currently small business accounts and checking and money market accounts are for sale at a high price in the Internet black market, whereas credit card numbers are cheap. This means that attacks to steal these accounts will continue to grow. As time goes by, credentials such as social networking data may become more valuable as phishers team up with malware writers to deliver malware and launch phishing. This analysis of phishers motives and their operating environments yields a few important insights: • As the underground black market continues to develop, phishing will become an operation that involves multiple parties with different specialties. We can further predict that the economics of scope and scale will be in effect, where phishing operations consolidate into a few phishing gangs to increase profit and reduce cost. This is both good and bad news for law enforcement. The good news is that there are fewer phishers to catch. The bad news is that these gangsters will be more technically capable and advanced enough to hide their trails. • If a bank is robbed once, we blame the robbers. However, if the same bank is repeatedly robbed, there must be some problem with its security. Simply put, if a particular institution is the repeat target of phishers, its institutional risk control and methods for handling incidents needs to be scrutinized. • Different categories of targets face different risks and therefore would require different countermeasures. 8 phishing gang roolkit provider domain registrar legitimate websites botnet herders (or host providers) spammers 1) decide on target 2) buy toolkit and websites 3a) register malicious or random domain 3b) hack into legitimate sites and plant phishing 4) buy hosts to set up attack machinary 5) ask spammers to send phishing emails Figure 2.2 Details of Phishing attack planing and setup including stakeholders 2.1.2 Setup After deciding on the target, phishers will set up the attack infrastructure. Figure 2.2 models how phishers do this and how various stakeholders are involved. The key insight here is that phishers rely on hackers, spammers, botnet herders, and pay loaders to launch a large-scale phishing attack. A few other insights: • Phishers will go to great lengths to reduce the probability of being caught. To that end, they will systematically exploit registrars that have weak security or process loopholes (for example, recent rock phish gang exploits of the .hk domain which has a weak verification process), operate from countries that have inadequate law enforcement resources and laws, and deploy proxies to hide their true destinations. • As shown in Figure 2.2, domain registrars are the first line of contact for phishers. If the registrars improve their security process for registration and fraud detection (for example, 9 Phisher Spam relays botnets Recipient mail gateways Mail storage Internet service provider User’s mail client 1) send phishing emails 2) Deliver phishing emails 3) spam filtering 4) MTX move email to storage 5) User check email 6) Establish Connection to Provider 7) email transferred 8) email transferred Figure 2.3 Phishing attack via email simple check against rock phish-type registration), it would sever phishers abilities to get domain names. • Botnets are the crucial machinery for launching and covering-up phishing attacks. Bots configured as proxies hide the trails of phishing attacks and make it very difficult for law enforcement to investigate. • All of the criminals meet and trade at the Internet black market. This means that efforts and research to disrupt the Internet black market will not only reduce phishing, but other types of attacks as well. 2.1.3 Attack Once the machinery is set up, attacks are launched through various vectors. Email and website phishing are the two most common attack vectors. Attacks using instant messenger and voice over IP are also increasing. In this section, I analyze these attacks separately. 10 Vector: email Figure 2.3 details the steps of attack through the email and web vectors. First, spammers send phishing emails through spam relays, botnets, anonymous mailers, and other common spam techniques (for a detailed treatment of these techniques, see [142]). The packets arrive at the ISP’s mail gateway where emails are put back together. The gateway then performs filtering. After that, mail transfer agents move emails to storage. When users check their email, mail clients on user machines connect to the mail storage through user ISPs, and download emails to their personal machines. If the end user’s mail client is a web client, it connects directly to the mail storage through a web interface and retrieves the email. Figure 2.4 and 2.5 shows an example of phishing email and website from eBay. Mail providers are in a unique position to combat phishing. They are the first point of contact for phishing emails. Their filtering effort will reduce the magnitude of the problem later on. Their reporting effort will reduce the time for blocking phishing websites. However, we need to consider that mail providers’ primary worry is spam. It consumes their bandwidth, takes up server space, and annoys customers. The amount of phishing they receive is only about 1% of the overall spam, so it is likely that they will consider phishing as part of the spam problem. To this end, they would be happy to process phishing if the spam filters also catch them, but they may not be willing to add additional phishing filters that consume their resources. Similarly, they also lack incentive to report phishing emails when it would take manual work to do so. There are over one thousand commercial free mail providers3 around the world and many, many more corporate or educational institution email providers. Not all of them are equally resourceful. Major mail providers, such as Yahoo and Microsoft, have resources to invest heavily into antiphishing and anti-spam technology. Smaller providers with limited IT budgets may have to rely on open source filters such as spam assassin, which is not very effective when it comes to phishing.4 3 Statistics compiled from free email providers’ guide [39] Research has shown that the standard configuration of spam assassin currently only catches about 70% of phishing emails [34]. 4 11 Figure 2.4 An example phishing email from eBay asking users to login to update an account. It warns users that failure to comply will lead to account suspension. The email address is spoofed, and the URL link is spoofed as well. Figure 2.5 An example phishing website from eBay that people see once they clicked on the link in the email example above. 12 spammer bots Internet Chatroom IM Servers IM clients User People on the buddy list 1) Look for IM usernames Get IM usernames 2) persuade add to buddy list added to buddies 3) send IM message with software exploits 4) message relayed 5) receive message 6) decision run malware 7) phishing message sent Figure 2.6 Phishing attack via Instant Messenger Instant messaging phishing In the instant messaging phishing case, the attack method is very similar to SPIM (spam over IM) and malware. Figure 2.6 documents one particular way to launch the attack. This attack method assumes that not all instant messaging users have configured their clients to receive messages only from their buddy lists. In this example, phishers first get IM names from various chat networks or trick users to add them to their buddy lists. Phishers then send software exploit to hijack account holders’ instant messenger lists. If the users run the exploit, phishing messages are sent out as if coming from the users. Alternatively, phishers can also send exploits through email, like a regular malware or virus attack. Instant messaging phishing and regular email phishing have some similarities and differences: • In email phishing, users control when they will read messages, whereas in instant messaging, users will generally read message when they are sent. This difference means that IM countermeasures have a shorter window of opportunity to detect and stop messages. 13 • IM is more contextualized than email. This means that an IM phishing attack targeting a bank will probably not be very effective because users would regard it out of context. However, a message purported from the messaging software vendor (AOL, Yahoo, Microsoft) asking users to update account information would be much more appealing. A message pretending to be sent from a friend to ask to view a picture (while installing a malware that alters the local DNS) could also be highly effective in IM. I expect IM attacks to be more sophisticated and deploy malware-based phishing rather than just regular phishing. • IM networks are more connected than email networks. If IM networks are attacked, potential infection can have a greater ripple effect than email. Some research has shown that theoretically it would only take 30 seconds to infect 500,000 machines that have IM clients [80]. • IM is easier to control than email. In IM, a few companies own the infrastructure and the delivery channel of the IM network. This means that it is easy to implement control measures at the IM gateway level if messages are routed in peer to server mode [49]. Clients can also connect in peer to peer mode, bypassing the server altogether. Vector: phishing over VoIP There are two ways to launch phishing attacks with voice over IP. In the first method, phishers create an exact replica of the target company’s voice system, obtain an 800 number from a VoIP service provider, and write and send an email messages that instructs recipients to call the 800 number and verify their credentials. Figure 2.7 illustrates this attack. In the second method, phishers directly dial victims’ phone numbers. Phishers use a random digit dialer or other means to acquire lists of phone numbers to dial (similar to spam over VOIP). Once they have a list, they dial numbers using prerecorded messages, usually alerting consumers of fraud and asking them to verify personal information such as social security numbers and bank pin numbers. Consumers may be tricked into giving credentials on the spot, or by calling back. In my expert interview, I will ask experts about what can be done regarding VoIP phishing. 14 Figure 2.7 Phishing attack via VoIP combined with traditional email scam methods. [132] 2.1.4 Collection The next step is the collection process. Figure 2.8 shows the process of a user falling for an email-based phishing attack. Other attacks that do not involve the email vector but have the web vector, can also be applied. Users, email client vendors, browser vendors, and ISPs all have a stake in developing and deploying countermeasures. Below is a list of possible measures. • Email clients are the first intervention points that can present visible warnings to users. Effective warnings presented here will alert users who were previously unaware of the issue. So far little research exists on the effectiveness of these warnings. • When users visit phishing websites, most of them have already committed time to read the emails, and it is likely that most will follow the steps of the website and submit information if no warnings were given to them. Studies have suggested that this is indeed the case [68]. 15 end user email client browser on user local machine internet service provider phishing web site 1) request email 2) present message 3) read message 4) Click on the link in email 5) open browser 6) connect to phishing site 7) connect to phishing site 8) phishing site loaded 9) phishing site loaded into browser 10) user reads phishing site 11) user think about it 12) enter information into browser 13) send to phishing site 14) send to phishing site Figure 2.8 Attack and Data collection process 16 This means that warnings should be focused on earlier in the process, at the email client level. Browser warnings must make it very difficult by design for users to bypass the warnings. • Browsers are suited to implement solutions, as there are only a few browser products available, compared with tens of thousands of registrars, and ISPs. Currently, major browsers like Internet Explorer and Firefox have phishing filters built into their systems. However, these filters are not so good. For example, in an recent study, Internet Explorer only catches about 50 80% of the phishing urls during the first 6 hours (see figure ??). Research should be conducted to benchmark and improve its performance. 2.1.5 Fraud The final process is fraud. Once the information is stolen, it is usually sold on the Internet black market. It is also possible that phishers themselves use the information to defraud customers. The fraud takes place in three steps: • Phishers use credentials to open new credit cards, or penetrate into user accounts. They would either do this themselves, or employ a cashier. • Phishers recruit money mules with accounts in the same institution and transfer the money to the mule’s account. If they are able to obtain credit cards, they will buy expensive items and have them shipped to the mule. • Mules take out the money and transfer to the phishers in other countries using Western Union and Money Gram. Phishers may also repackage the expensive goods and ship to a third location. 2.2 Why people fall for phishing Phishing is part of a larger class of attacks known as semantic attacks. Rather than taking advantage of system vulnerabilities, semantic attacks take advantage of the way humans interact with computers or interpret messages [123]. In the phishing case, attacks exploit the fact that users 17 tend to trust email messages and web sites based on superficial cues that actually provide little or no meaningful trust information [23, 26]. Research on phishing have shown that people are vulnerable for several reasons. First, people tend to judge a website’s legitimacy by its “look and feel”, which attackers can easily replicate [23]. Second, many users do not understand or trust the security indicators in web browsers [140]. Third, although some consumers are aware of phishing, they do not link that awareness to their own vulnerability or to strategies for identifying phishing attacks [26]. Fourth, the perceived severity of the consequences of phishing does not predict their behaviors [27]. Below I summarize some of the seminal research in understanding why people fall for phishing. Dhamija et al showed twenty-two participants twenty web sites and asked them to determine which were fraudulent. Participants made mistakes on the test set 40% of the time. Many of the participants rely on the content of the webpage (logos, layout, graphic design) to determine its legitimacy. The authors noted that 23% of their participants ignored all cues in the browser address bar and status bar as well as all security indicators [23]. Two related study done by Wu et al and Jakobsson showed similar results [54, 140] . Downs et al have described the results of an interview and role-playing study aimed at understanding why people fall for phishing emails and what cues they look for to avoid such attacks. There were two key findings in their work. First, while some people are aware of phishing, they do not link that awareness to their own vulnerability or to strategies for identifying phishing attacks. Second, while people can protect themselves from familiar risks, people tend to have difficulties generalizing what they know to unfamiliar risks [26]. In a follow up study, Downs et al surveyed 232 computer users to reveal predictors of falling for phishing emails, as well as trusting legitimate emails. Their data suggested that deeper understanding of the web environment, such as being able to correctly interpret URLs and understanding what a lock signifies, is associated with less vulnerability to phishing attacks. However, they also found that the perceived severity of the consequences of phishing does not predict behavior, suggesting that educational efforts should aim to increase users’ intuitive understanding, rather than merely warning them about risks [27]. 18 2.3 Cost of phishing Phishing exerts both direct and indirect cost to the society. Examples of direct loss include consumers losing money, and banking fraud, etc. Examples of indirect cost include erosion of consumer trust of the Internet, negative impact to businesses’ brand, , an increase in service call center complaints volume etc. Estimating either cost is hard, as there are many stages of the attack and it is difficult to collect good data. Three reports attempted estimating direct costs. Gartner Research conducted a survey of 5000 Internet users in August 2006 asking whether consumers have received, clicked or given information in phishing emails. Based on this survey, they estimated that 24.4 million Americans have clicked on a phishing e-mail in 2006, while 3.5 million have given sensitive information. They calculated that the economic loss be 2.8 billion dollars in 2006 [42]. A follow up survey in 2007 with similar methodology estimated that 3.2 billion dollars is lost in 2007 [43]. The above studies rely on people’s survey responses. Psychology literature has shown that there is often a wide discrepancy between people’s stated choices and their actual behavior. Moore and Clayton empirically studied phishing websites using PhishTank data. They found that a phishing site lives for 61 hours on average. Using the web log data of some of these phishing sites, they estimated that on average 18 users would fall for phishing on the first day when the site was up, and subsequently 8 users per day afterwards. The total cost to consumers per year was estimated around 320 million dollars [92]. Florencio and Herley [36] instrumented Microsoft’s anti-phishing toolbar to send notifications back to Microsoft every time a password was re-used on more than one site. They cross-checked sites that reused password against Microsoft’s blacklist, and estimated that 0.4% of people are falling for phishing (the data set had 500k people across a few months). Using this data, and an average of $572 per victim, the estimated yearly cost of phishing to consumers is around 350 million dollars. The above empirical studies use separate methodologies to estimate phishing costs. Their research designs are reasonable, and their estimates agree remarkably well. We can thus treat 350 19 million as a lower bound on the direct economic loss of phishing to consumers. The actual cost of phishing exceeds this amount as we consider cost to businesses, and other indirect costs. It is worth noting that there is disagreement of the magnitude of phishing losses. For example, Florencio and Herley recently used stylized economic models to predict that phishing is a classic example of tragedy of the commons, in which there is open access to a resource that has limited ability to regenerate. Since each phisher independently seeks to maximize his return, the resource is over-grazed and yields far less than it is capable of. The situation stabilizes only when the average phisher is making only as much as he gives up in opportunity cost. They estimate the annual loss to phishing around 60 million dollars [50]. Other security experts have criticized both their research assumptions and results [21]. Such disagreement highlights the lack of empirical evidence for phishing losses. Last but not least, the cost of phishing is disproportionally borne by consumers. The fraud is relatively small in most phishing scams (medium loss per victim around 200 dollars [43]), but the psychological fear, anxiety it causes the victims, and the time it takes to restore identity can be substantial. However many organizations only consider direct monetary loss to them in their calculation. 2.4 Recent developments in phishing Traditional phishing is delivered through email, where phishers send mass email to consumers asking them to visit a website to enter information. However, recently attacks have become more sophisticated. In this section, I will talk about three recent developments: voice over IP phishing, spear phishing , and some new phishing techniques such as rock phish and fast flux. VoIP Phishing In April 2006, phishers started to use Voice over IP to scam consumers (a.k.a vishing). The attacks work as follows. First, phishers set up a voice-mail system using voice over IP and private branch exchange software (such as open-source PBX software Asterisk). They then use an automatic dialer to call a long list of people and play a recorded message, or simply send emails 20 asking them to call a number to update their account [112]. When consumers respond, they hear an automated message asking them to enter their account information [138]. Using VOIP, phishers can achive economics of scale by dialing through a long list of numbers, it also makes them harder to track down than using regular phones. Vishing is growing. MessageLabs has observed the increase frequency of such attacks toward the end of 2007 [86]. Vishing is more damaging than other phishing methods because research has shown that customers generally trust the phone channel more than the email channel [54], and they are well accustomed to enter credit card numbers through automated systems. In my proposed work, I will elicit experts’ opinion on how to better counter vishing attacks. Spear Phishing Spear phishing is a highly targeted phishing scam. Instead of sending mass phishing emails to everyone, phishers exploit the social context to send spoofed emails to consumers that appear to come from someone they know. For example, phishers can send emails impersonating an organization’s system administrator asking people to update their passwords, or impersonate one of your friends in social networking sites. Attacks like these are possible because a myriad of information about consumers exists on the Internet and in many cases are readily available through basic mining of the web. For example attackers can obtain information about consumers’ bidding history or shopping preference from eBay, what banks they use (discoverable through web browser history [56]) or even social security number [53]. A recent incident of spearphishing targeted wall street executives. Phishers sent emails to middle and upper senior management of some wall street firms. These emails appeared to be complaints from the Better Business Bureau, and it contained an a .doc attachment where a spying trojan was embedded [86], many opened the attachment and later found they became victims of fraud. Research studies have shown that spear phishing can be highly effective. In a recent study by Jagatic et al at Indiana University, they send email to students impersonating their friends in the social networking site. The email asked them to enter their secure University credentials. They 21 found that 72% of the time, users would enter their correct credentials, four times more effective over the traditional phishing methods [53]. Studies at West Point Military Academy showed similar results [33]. In my proposed work, I will also elicit experts’ opinion on how to counter spearphishing attacks. Rock phish and fast-flux Rock phish refers to phishing committed by the rock phish gang. They were referred to as the Rock phish gang because early versions of their attacks contained the word rock. The rock phish gang has employed several techniques that make them more difficult to defeat than other phishers. First they use stolen credentials to register multiple short and random domain names at multiple registrars. They then host their own DNS servers, and provide name-to-IP service for each of the fraudulently registered domain. The name-to-IP matches to a farm of compromised computers, which do not host the phishing site, but merely act as a proxy to a handful of servers that host phishing sites [81]. Techniques like these pose challenges. Random domain registrations make it hard for automatic detection. Layers of redundancy makes it hard to shut them down quickly, especially if it is from different jurisdictions. A recent study by Moore and Clayton showed that rock phish domains last almost three times longer than regular phishing domains [92]. Another technique the rock phish gang use is fast flux. It is a DNS technique to evade blacklists. It works the following way: multiple nodes within a network register and deregister their address to a single DNS record. Some of these registration lasts only a few hours. Techniques like these pose challenges to blacklists as phishers cycle through hundreds of addresses in a day. In summary, phishers continue to improve themselves using new techniques such as rock phish and fast-flux, targetting specific groups, and using alternative channels to attack. In my proposed work, I will elicit the experts’ recommendations and stakeholder’s countermeasures addressing specifically these three new threats, and other evolving threats. 22 2.5 Phishing countermeasures Phishing countermeasures can be categorized as legal solutions, technology countermeasures, and social responses. In this section, I will survey each of these types of solutions briefly. 2.5.1 Legal solutions In the wake of increasing publicity about phishing scams, both federal and state legislatures acted. In January 2005, Virgina added phishing to its computer crimes act, categorizing the use of a computer to obtain personal information “through the use of material artifice, trickery or deception” a Class 6 felony punishable by prison sentences of up to five years and fines of up to $2,500 [137]. Similar statues has been enacted in New Mexico [103], and New York [130]. By February 2007, half of the 50 U.S states have enacted laws addressing phishing [98]. At the federal level, anti-phishing laws have been passed in the House of Representatives in 2004 [133], 2005 [134] and 2007 [135], but the Senate failed to act upon it. Some prosecutions has been made. In 2006, a Florida man has been indicted in Pennsylvania for a phishing scam that mimicked a Hurrican Katrina relief website [73]. In 2004, Zachary Keith Hill plead guilty in a Texas federal court to crimes related to phishing activity and was sentenced to 46 months imprisonment [45]. The U.S. Department of Justice has successfully prosecuted several other defendants in U.S. courts [118]. However, criminal law does a poor job of deterring phishing because phishers are so hard to find [14]. Law enforcement authorities have little time to track down the criminal through the fraudulent site because on average they only live a few days. Once the site is shut down, the e-mail is the only remaining evidence, and phishers often cover their tracks using such tools as anonymous remailers [14]. In light of this, some legal experts argue that Internet service providers be made liable for part of the Internet “insecurity”. Their argument is that “Internet service providers control the gateway through which Internet pests enter and reenter the public computer system. They should therefore bear some responsibility for stopping these pests before they spread and for helping to identify 23 Figure 2.9 Taxonomy of phishing technical countermeasures individuals who originate malicious code in the first place” [74]. Service providers, however, have largely been immune to such liability. Because they are distributors of the content, not publishers of the content, and as long as they exercise due diligence to remove these materials, they are be not held liable. How can law enforcements be more effective in the fight against phishing? In my proposed work, I plan to address this issue with further analysis of laws and cases, and by interviewing several law enforcement experts from the Department of Justice and Federal Bureau of Investigation. 2.5.2 Technology countermeasures Anti-phishing services are now provided by Internet service providers (ISPs), built into mail servers and clients, and available as web browser toolbars. In this section, I will review some commercial offerings as well as academic research. Drawing from the life cycle of phishing attacks, we can categorize countermeasures into the following stages: prevention, detection/analysis, shutdown, block emails/websites and warning users (see Figure 2.9). I discuss each of these stages briefly below. • Prevention: As shown in Figure 2.9, the first step to fight phishing is to prevent attacks before they are materialized. Law enforcement officers can catch and prosecute phishers before they launch the attack. Registrars can monitor domain registrations and analyze suspicious registrations. Mail providers can use email verification solutions such as SPF to drop Stages Prevention Table 2.1 Summary of commercial phishing countermeasures Techniques Used Examples of Companies offering the Service 1) monitor domain registrars for suspicious MarkMonitor, Brandimensions, registrations; 2) register domain names defensively to protect a brand; 3) Sender Policy Cyveillance, Framework or similar technologies to validate InternetIdentity.com, email senders; 4) email encryption using GoDaddy.com, Verisign, TimerWeed S/Mime or PGP Communications, RSA Metrics for Effectiveness number of criminals caught and prosecuted, number of phishing attacks stopped Detection / analysis 1) Set up honeynet or spam traps to collect phishing emails; 2) Scan mail provider’s incoming mails; 3) Scan through company weblogs for suspicious activities; 4) User report phishing scams; 5)Scan the web to find malicious websites MarkMonitor, RSA, Cyveillance detection time, true positives vs false positives Block emails / websites 1) blacklist; 2) heuristics internet service providers true positives and false positives Warn User 1) Email client warning; 2) Browser antiphishing toolbars Microsoft, Google, CloudMark, Earthlink reach (market share), time to warn, true positives and false positives Shutdown Contact ISPs, CERTs or necessary authorities to shutdown malicious website RSA time to shutdown, cost of shutdown Verisign, GlobalSign, RSA, Tricipher To be determined 24 Authentication1) Extended Validation Certificates (EV Certs) and fraud 2) Two factor authentication (smart card, detection tokens) 3) fraud detection system 25 unverified traffic. The more effective the prevention is, the smaller the phishing problem will be. • Detection: Once phishing attacks are launched, the best defense is to detect and analyze them as early as possible. Internet service providers can add detection systems in their e-mail processing and storage systems to detect suspicious emails. Anti-phishing tool providers can set up spam traps or honeynets to receive early notice of new waves of attacks. Once suspicious emails and websites are identified, analysis will follow, usually combing automatic analysis with human expertise. Table 2.2 lists some of the current state of the art techniques used in automated detection. • Shutdown: Once attacks are verified, service providers can be contacted to shutdown websites. • Block emails / websites: Shutting down websites may take a few days, especially if they are on foreign domains. However, mail providers can delete phishing emails from storage (or move them to a separate spam or phishing folder). Internet service providers can block their customers access to these websites, and replace them with generic education messages. • Warning users: Browsers and, email clients are in a unique position to warn users because their warnings are most visible and direct. • Authentication and fraud detection: This is the last line of defense. Correctly implemented two factor authentication systems can stop phishers from defraud financial institutions, fraud detection system can also discover the scam and stop it. In summary, the objective is to prevent phishing attacks as much as possible; to detect attacks as early as possible; to shutdown operations as quickly as possible; and to warn users as effectively as possible. Table 2.1 summarizes some of the commercial offerings by the stages, and Table 2.2 summarizes major academic contributions to detection. In my proposed work, I plan to gain more understanding of the effectiveness of these solutions. Specifically, I will test ten anti-phishing toolbars empirically. 26 Authors Fette et al [34] AbuNimeh et al [1] Yue et al [148] Anthony et al [40] Table 2.2 Meta analysis of proposed detection methods by academics Vectors Vectors Addressed Test Results Addressed Emails Machine Learning: 860 phishing emails, True positive: 96.1%, Random Forests and 6950 non-phishing false positive: 0.1% Approach. 6 features emails Emails Compared multiple ML 1171 phishing emails logistic regression techniques, large and 1718 legitimate lowest false positive feature set emails 4.89%, random forest lowest false negative: 11.12% Websites Content based 100 phishing and 100 True positive: 97%, approach on TD-IDF legitimate URLs false positive 10% with five heuristics Websites Use a linear 1000 real webpage, 9 precision: 99.87%, programming model to phishing page recall: 88.88% assess the visual features’ similarity Figure 2.10 Active warning toolbars in Mozilla Firefox blocking a known phishing site 2.5.3 Social response: awareness and education Despite claims by some security and usability experts that user education about security does not work [31], there is evidence that well designed user security education can be effective [68]. 27 Web-based training materials, contextual training, and embedded training have all been shown to improve users’ ability to avoid phishing attacks. A number of organizations have developed online training materials to educate users about phishing [28, 32]. In a previous study, we tested the effectiveness of some of these online materials and found that, while these materials could be improved, they are surprisingly effective when users actually read them [70]. Several studies have adopted a contextual training approach in which users are sent simulated phishing emails by the experimenters to test users’ vulnerability to phishing attacks. At the end of the study, users are given materials that inform them about phishing attacks. This approach has been used in studies involving Indiana University students [53], West Point cadets [33], and New York State employees [104]. In the New York State study, employees who were sent the simulated phishing emails and follow-up notification were better able to avoid subsequent phishing attacks than those who were given a pamphlet containing information on how to combat phishing. A related approach, called embedded training, teaches users about phishing during their regular use of email. In a previous laboratory experiment to evaluate our prototype embedded training system, we asked our participants to role play and respond to the messages in an email inbox that included two training emails designed to look like phishing emails. If a participant clicked on a link in a training email, we immediately presented an intervention designed to train them not to fall for phishing attacks. We created several intervention designs based on learning sciences, and found that our interventions were more effective than standard security notices that companies email to their customers [68]. A follow up study showed that people were able to retain what they learned in the training as well [69]. 2.6 Economics of Information Security To fight phishing, companies have to allocate resources. As companies make decisions as how to best allocate resources, it is important to understand incentives and tradeoffs of these decisions. There is a growing field of literature on economics of information security that broadly addresses some of these questions. 28 2.6.1 Security investment In general, security investment studies seek to answer the question: what is the optimal amount of investment for information security for a given company? What are the incentives and disincentives that affect the security investment? There are two approaches to answer this question. The first type is through quantitative models, and the second approach is qualitative studies. Seminal work in quantitative economic modeling includes works by Gordon and Loeb [47], and by Cavusoglu and Raghunathan [15]. For qualitative studies, Row and Gallaher [117] conducted a series of interviews with large organizations in a variety of sectors. Based on the interview, they derived a conceptual approach for security investment in organizations (see Figure 2.11). The paper made three key observations. First, various internal and external incentives (drivers) affects organizations to adopt countermeasures. Second, some organizations tend to adopt more proactive countermeasures while others more reactive. In my proposed work, I extend their model and study the incentives for different stakeholders in phishing countermeasures. 2.6.2 Security as externality Externalities can be found when we analyze security investment, as protection often depends on the efforts of many principals. Anderson and Moore used the following analogy to explain this externality. Consider a medieval city. If a main threat is a siege, each family is responsible for maintaining and guarding one stretch of the wall, then the city’s security will depend on the efforts of the laziest and most cowardly family. If, however, disputes are settled by single combat between champions, then its security depends on the strength and courage of its most valiant knight. But if wars are a matter of attrition, then it is the sum of all the citizens’ efforts that matters [4]. Does successfully combat phishing depends on the efforts of the laziest and most cowardly family (weakest link)? or is it depends on the most valiant knight? or sum of efforts? 29 Kunreuther and Heal notes that security investments can be strategic complements: An individual taking proactive measures creates positive externalities for others that in turn may discourage their own investment [72]. An example is airline security where airlines may decide not to screen luggage transferred from other carriers that are believed to be careful with security. For phishing countermeasures, externality is clearly an issue. Banks and online merchants suffer loss because the Internet is not secure. On the other hand, the efforts of Internet service and mail providers may reduce the incentive for banks to invest more in phishing, in other words, make them free ride. 2.6.3 Misaligned incentives Anderson and Moore indicates that incentive misalignment significantly undermines information security [4]. For example, in the United Kingdom, banks are liable for financial fraud only when it is proven that they are at fault. The the burden of proof is with consumers. Therefore customers complaints are not taken seriously, and this leads to lots of fraud [4]. Would there be misaligned incentives among stakeholders of phishing countermeasures? Would those in the best position to fight the phishing lack incentives to do so? 30 Figure 2.11 Diagram of Cybersecurity Investment by Row and Gallaher [117] 31 Chapter 3 Improving Phishing Countermeasures: An Analysis of Expert Interviews This chapter is joint work with Alessandro Acquisti, Lorrie Cranor, Jason Hong, and Ponnurangam Kumaraguru. An earlier version of the content in this chapter is published at 2009 eCrime Researchers Summit [126]. 3.1 Introduction As the battle against phishing continues, many questions remain about where stakeholders should place their efforts to achieve effective prevention, speedy detection, and fast action. Do stakeholders have sufficient incentives to act? What should be the top priorities for the antiphishing community? To provide insights into these questions we conducted 31 in-depth interviews with anti-phishing experts between May 2008 and May 2009. We selected experts from academia, Computer Emergency Response Team (CERT) centers, the Anti-Phishing Working Group (APWG) officers, law enforcement, and key industry stakeholders. We sought their expertise on the current and future state of phishing attacks, countermeasures that should be implemented to fight phishing more effectively, and incentives that various stakeholders have in their fight against phishing. The experts we interviewed agreed that phishing is evolving into a more organized effort. It is becoming part of a larger crime eco-system, where it is increasingly blended with malware and used as a gateway for other attacks. Some of the experts suggested that incentives for fighting phishing may be misaligned, in the sense that the stakeholders who are in a position to have the 32 largest impact do not have much incentive to devote resources to anti-phishing efforts. In terms of countermeasures, experts identified improving law enforcement and shutting down money trails as top priorities. They also identified operating systems vendors, web application providers, browsers, and Internet service providers as stakeholders with key technology influence on phishing. Finally, experts agreed that education is an important factor that is not emphasized enough; however, they did not agree on the extent of the impact that education may have. We present these findings and a set of recommendations to improve countermeasures. Although previous reports have studied phishing and issued recommendations, to the best of our knowledge this is the first study that synthesizes the opinions of experts from different fields, and examines the incentives of various stakeholders to contribute to anti-phishing efforts. 3.2 Related Work In response to the growing phishing problem, government agencies, industry groups, and con- sumer groups have conducted studies and issued recommendations [35, 52, 100, 105]. The Financial Services Technology Consortium’s report is the first report that analyzed how phishing works by articulating the life cycle of phishing. It also encouraged financial institutions to assess the costs and risks associated with phishing, develop better intelligence on phishers through improved sharing, and invest and adopt in better mutual authentication. However, the report did not issue recommendations for non-financial institutions who also have high stakes in the phishing problem [35]. The Identity Theft Technology Council report also analyzed different phases of phishing and recommended a set of 21 technical countermeasures citeiitc:phishing-report. We selected a subset of recommendations from this report as a starting point for discussion in our expert interviews. However, we updated the set to address non-technical countermeasures as well as new and evolving threats that were not discussed in the report. In addition to discussing the set of recommendations, we also studied the incentives that stakeholders have to implement them as well as how the incentives can be increased. 33 Table 3.1 Phishing stakeholders. Primary victims suffer direct losses from phishing. Infrastructure providers have technical capabilities to mitigate the problem. For-Profit protectors sell solutions to primary victims and infrastructure providers. Public protectors include law enforcement officials, computer emergency response teams, and academic researchers. Categories Examples of key stakeholders Roles Consumers – Primary victims Organizations Military, Universities, Corporations Financial Institutions Bank of America, Citibank, Paypal Merchants Online merchants (eBay, Amazon), offline merchants Registrars and GoDaddy, Verisign Infrastructure Registries providers Internet Service AT&T, Comcast, AOL, Universities Providers Email Providers Gmail, Yahoo!Mail, Hotmail Browsers Internet Explorer, Firefox, Safari Software Vendors Symantec, RSA, MarkMonitor, Cyveillence For-profit protectors Law Enforcement Federal Bureau of Investigation (FBI), Secret Public Protectors Service state and local enforcement Computer Emergency CERT-CC, CSIRTs Response Teams Academia In addition to these reports, the Anti-Phishing Working Group (APWG) has issued a set of best practices and recommendations for hacked website owners [8], registrars [6], and ISPs and mailbox service providers [85]. Each of these reports focus narrowly on one particular area. In our analysis, we analyzed the phishing issue holistically and asked our experts to prioritize their recommendations based on their importance and effectiveness. 3.3 Stakeholders Phishing involves many stakeholders, including consumers, financial institutions, online merchants, Internet Service Providers (ISPs), mail client and web browser vendors, and law enforcement. In this paper, we have classified stakeholders into the following categories: primary victims, 34 infrastructure providers, for-profit protectors, and public protectors. Table 1 describes these stakeholders and their roles. We used it to select experts and structure our interviews. 3.3.1 Primary victims: In most cases, consumers, organizations, financial institutions, and merchants are direct targets of phishing attacks. Each of them is negatively affected by phishing in a different way. Consumers who fall for phishing can potentially become victims of identity theft: they not only suffer monetary loss, but also psychological costs (e.g. fear, anxiety). Generally speaking, consumers fall for phishing because they have incorrect mental models about what constitutes trustworthy emails and websites [23] and they are susceptible to manipulation and social engineering. Organizations such as the military and corporations worry that phishing may lead to further compromise of credentials that can be used to steal key intellectual property or conduct corporate espionage. Financial institutions lose money from fraud conducted with credentials acquired through phishing. They may also suffer indirect losses such as increased customer service cost, damage to reputation, etc. Some argued that indirect losses are much greater than the direct losses, although this claim has not been independently verified. Merchants lose money because these financial institutions eventually charge them for the fraudulent transactions. In general, these entities are most impacted by phishing, and have the strongest incentive to protect against phishing. However, as shown later in the result section, some of them have limited capabilities to counter phishing attacks. 3.3.2 Infrastructure providers: Internet service providers, email providers, browsers, domain name registrars, and registries are infrastructure providers. In most cases, phishers do not go after these providers for their money; instead, they seek to gain access to the entities’ infrastructures so that phishers may launch their attacks. For example, phishers register fake domain names with registrars. Phishers use compromised machines from Internet Service Providers as part of a botnet to launch phishing campaigns, 35 sending emails to end user mailboxes or compromising mail provider accounts to send phishing emails. These stakeholders are important to study, as they are in a better position than most victims to protect against phishing. However some infrastructure providers do not lose money from phishing, so they may not have sufficient incentives to devote resources to combating phishing. In our interview study, we asked experts what these stakeholders can do and examined whether or not they have incentives to do so. 3.3.3 For-profit protectors: Certain organizations actually benefit from phishing because it is an opportunity to develop and sell products to other stakeholders. These include companies that sell spam filters and antivirus software, as well as companies that take down phishing websites. As they are the frontline defenders against phishing, we selected a few of our experts from these companies. Table X also However, as they make money from combating phishing, it could somewhat bias their recommendations. We discuss these potential biases in detail in the methodology section. 3.3.4 Public protectors: In contrast to anti-virus vendors and spam filter companies who are for-profit protectors, law enforcement, computer emergency response teams (CERT), and academics are public protectors. There are some para-organizations such as the Anti-Phishing Working Group (APWG) and the Message Anti-Abuse Working Group (MAAWG) that aim to bring different stakeholders together to fight more effectively against phishing. Some of the experts we interviewed hold positions in these organizations. However, we did not consider these organizations as separate stakeholders in our analysis. 3.4 Methodology During May 2008 and May 2009, we conducted in-depth interviews with 31 experts involved in phishing countermeasures. In this section, we discuss how we selected the experts, the interview process, and the steps taken to analyze the data. 36 Table 3.2 Anti-phishing experts interviewed. For confidentiality purposes, all partipants are anonymized. Affiliation No. of Experts CERT 4 Academic researchers 5 APWG officers 3 Law enforcement 5 Registrars, Registries 3 Financial institutions 4 Internet service providers 3 Browser vendors 1 Other experts 3 Total 31 3.4.1 Recruitment and Participants We recruited experts in several ways. First, we compiled a list of frequent speakers from 2004 through 2007 APWG member conferences and generated a list of well-known experts in academia and industry. To recruit law enforcement officers, we attended the 2008 Digital PhishNet conference. To recruit experts in Internet service providers, registrars, and technology vendors, we solicited recommendations from APWG’s Internet Policy Committee (IPC), which is composed of 90 members from various stakeholders. Finally, we recruited additional interviewees through our own network of contacts. In order to obtain a variety of views, we tried to select candidates from different organizations who worked at different levels of company hierarchy. We recruited a total of 31 experts responsible for, or knowledgeable of, operational or policy decisions with regard to phishing and malware prevention in their organizations. Most of the interviewees chose to remain anonymous. Table 2 shows the organizational profiles of these experts. 67% of the experts interviewed had a technical background, 20% had a policy or business background, and the remainder had a background in law or law enforcement. In addition to the 31 experts interviewed, we also had a short interview with a legal expert on the question of liability for false positives. 37 The sample size of 31 balances the resource-intensive demands of in-depth interviews and analysis against the marginal return of new insights from additional participants. We had multiple participants who shared similar views on most of the topics we discussed in our interviews, suggesting that theoretical saturation was likely achieved, even with our small sample. 3.4.2 Interview Protocol We used a semi-structured interview protocol. The protocol allowed us to ask structured questions that enabled comparable responses across participants, while providing the interviewer flexibility in drilling down on areas of particular relevance to each participant [115]. Each interview typically lasted 60 minutes (min =25, max = 90) and was recorded for transcription. Some interviews were conducted in-person, while others were conducted over the phone. We began each interview by asking each expert to describe his or her background and responsibilities. We then asked a set of open-ended questions about how phishing impacts their organizations, amount of losses, current and future state of phishing, and the effectiveness of current countermeasures. We then asked them specifically to comment on a set of 31 recommendations broken into six categories that we compiled through our research. Experts prioritized the recommendations in each category and provided feedback on them. Finally, at the end of each interview, we asked experts to provide additional recommendations, and if they did, we summarized and added them to our list of recommendations and asked experts about them in subsequent interviews. 3.4.3 Analysis After completing each interview, we transcribed the audio recordings and recorded the answers to each question in a spreadsheet. We then analyzed the interview results and synthesized a series of findings and accompanying recommendations. In our analysis, we synthesized experts’ opinions by selecting themes that recurred most frequently across all interviews. We also report some of the comments that were discussed by only one or two experts, but that we found particularly useful in thinking about phishing countermeasures. 38 3.4.4 Limitations Before turning to the empirical findings, it is important to note the scope and limitations of this study. Most of the experts interviewed were from the US, but we also had some from Japan, Hong Kong, Italy and Britain. Thus, while there is some international representation, for the most part these interviews represent a US-centric view. It is also reasonable to assume that this set of interviewees was influenced by some degree of self-selection. Registries, for example, are more likely to respond favorably to an interview request about their phishing countermeasures if they have policies in place that are at least on par with other registries, if not better. With that said, some of the organizations we interviewed are not known for having outstanding records with regard to phishing. Our findings reflect how stakeholders themselves describe what they are doing and why. In other words, we report on the perceptions of the interviewees, not the independent assessment of their actions and the factors driving them. Whenever possible, we did crosscheck information provided to us against the information from other interviews and against publicly available data, such as reports, surveys and research publications. In addition, the interviewees are not experts in all areas, and they have biases of their own. For example, take-down vendors are more likely than others to recommend that more efforts should be focused on take-downs. We address this in a few ways. During our interviews, we let interviewees select the two to three areas in which they are most experienced to comment on. Whenever possible, we asked them to provide evidence to support their positions and recommendations, and in some instances, we tried to probe experts further by presenting a counter-argument for experts to respond to. Despite these limitations, our approach is an important complement to purely technical analysis of phishing (e.g. [52] ). First, our interview approach synthesizes the opinions of experts from many different fields. It would be difficult to obtain this information through other methods. Second, our interviews examine the incentives of various stakeholders to contribute to anti-phishing 39 efforts, an important consideration in producing workable solutions. For example, past qualitative research in information security investments has proven to be a valuable complement to the knowledge generated through quantitative modeling or analysis (e.g. [119], [117]). In the next sections we present the findings from our interviews. We classified our findings into four topical categories: the evolving threat, stakeholder incentives, what stakeholders should do, and law enforcement and education. We also provide a set of recommendations based on these findings. Table 3.3 presents the high-level findings from the interviews. Finally, this paper does not discuss some relevant technologies such as email authentication (SPF, DKIM), extended validation certificates. These technologies were rarely mentioned by the experts we interviewed and we found no consensus on the effectiveness of these technologies. 3.5 3.5.1 Results Evolving threat Categories Evolving threat Stakeholder incentives What stakeholders should do Law enforcement and education Table 3.3 High-level findings. Findings A. Phishing is evolving to be more organized and targeted. It is becoming part of a large crime eco-system. B. Phishing and malware are increasingly blended together. A. Stakeholders have varying incentives to fight phishing. B. Sometimes stakeholder incentives are misaligned. A. Operating systems vendors, web application providers, browser vendors, and Internet service providers are stakeholders with key technology influence over phishing. B. Organizations are conservative about filtering and warning about phish because they are worried about false positives. C. Registries and registrars can play an important role in fighting against phishing. A. Law enforcement should be emphasized; but law enforcement lacks the necessary tools, personnel, and resources to catch phishers. B. Shutting down money trails is very important to defeat phishers. C. Education and awareness are important factors that are not emphasized enough. However, not all experts agree on the effects of education. 40 Phishing is evolving to be more organized and targeted. It is increasingly used as a gateway to other attacks. We asked experts to describe the phishing attack trends they have observed and predict how phishing attacks will continue to evolve. Experts observed that phishing attacks are becoming more organized. One technical expert in law enforcement explained: These are criminal organizations that exist that perpetrate these types of fraud. It is not likely your teenage hacker like in the old days. They are criminal organizations with business plans and contingency plans. They are typically involved in other crimes besides phishing. It could be malware, it could be hosting other content, possibly child pornography, and it could be the old 419 scams and mule schemes. What we see is that these types of folks don’t just do one thing. They either do other things or work with groups that do other things. One example of an organized group mentioned frequently by experts is the rock phish group, which is believed by many experts to originate from a group of phishers in Eastern Europe. One academic researcher said 88% of the one million URLs his research group processed in October 2008 had rock phish characteristics. Published studies have also analyzed the frequency of fast flux phishing attacks. For example, Moore et al. found that 68% of the phishing emails in their study sample were sent using fast flux techniques [95]. Another trend that experts observed is that phishing is increasingly used as a gateway to other attacks. One expert from a major browser vendor said: We are seeing a lot of blended attacks, where a piece of the infrastructure is a phishing attack, but that’s not necessarily the end goal. . . . It is malware, it’s affiliate advertising, it’s spam as form of advertising, scams, and ring tones, there is a number of ways to monetize. But the goal is to look for not only the traditional stuff but ways to monetize groups of users. And you know, stealing a password is a pretty good way to tag into real people, real networks, so we see the social network site is being targeted very heavily, and it’s the result of that. 41 One of the experts from a major US bank agreed, and added that his institution had been seeing an increasing amount of cross channel fraud, where credentials harvested through traditional phishing attacks were being used to commit fraud in other channels such as telephone banking. Finally, experts agreed that phishing attacks are evolving into be more targeted attacks, which are very effective and harder for spam filters to detect. Recent phishing attempts to defraud top executives are examples of these targeted attacks. Past research has demonstrated the effectiveness of spear phishing attacks. For example in a study at Indiana University, 16% of participants fell for regular phishing emails, but 72% fell for spear-phishing emails [53]. Phishers kept moving to new targets as traditional targets of phishing attacks have devised response plans. Some experts thought that small and medium brands would become the next victims. Others speculated that credit unions, social network sites, and Xbox live accounts would be increasingly targeted. Phishing and malware are increasingly blended together. Experts mentioned that malware attacks that use phishing emails are on the rise and pose a serious threat. One academic researcher framed phishing and malware as different expressions of the same problem. He said: You will see social engineering aspects of malware and high automation aspects of phishing. At some point, it might be hard to tell them apart . . . To the attackers, it doesn’t matter what they use. They know social engineering has an effect on the end user, they know script and code and have some effect on the user’s machine. It is just a matter of putting what they know and what they have. Some of the experts we interviewed believe that malware now poses a bigger threat than phishing. Their reasoning is that due to vulnerabilities in operating systems and web applications it is easy for computers to get infected with malware, and that even security-conscious users may have difficulty avoiding infection. 42 3.5.2 Stakeholder incentives Stakeholders have varying incentives to fight phishing. We asked experts how phishing impacts their organizations. Their responses provided insights into their organizations’ incentives to fight phishing. In general, we found that the primary victims have incentives to invest resources to protect against phishing as they suffer direct losses from phishing. Nonetheless, there is evidence that not all potential primary victims have made this investment. One expert from academia said that many midsize and smaller banks he talked to did not have a strategy for phishing, as they had never been targets: “There is low chance that those banks are targeted, but if they are targeted, they could lose a lot of money.” The stakeholders who do invest in anti-phishing protection sometimes feel that they are carrying a disproportionate share of the burden. One expert said: After speaking to many service providers such as financial institutions, there is one thing that stands out very clearly, a sense of “injustice,” that they are often carrying the cost for something they have no ability control or even measure. For example, financial service providers, they are not able to determine if their clients, the end users, have appropriate anti-virus software or not. So one way to align the incentives is for service providers be able to audit the security posture of user clients. Our interviews revealed information on the incentives of several types of stakeholders, described below. Financial institutions. Financial institutions are among the primary victims of phishing as they lose money from fraud committed with compromised accounts. Currently, over 79% of phishing attacks target financial institutions [131]. A major US bank told us that over the past 12 months, their loss due to phishing and malware was $4.5 million, accounting for 25% of their fraud loss through online channels. Financial loss and regulatory oversight are both drivers for adopting anti-phishing technologies. One electronic fraud risk manager from a major bank in Asia mentioned that their loss to phishing 43 and electronic crime is less than 1% of their overall fraud loss. However, they still invest a lot of money in anti-phishing efforts because regional regulators demand two-factor authentication and require comprehensive analysis for electronic crime incidents. Thus, stakeholder incentives may vary depending on local regulations. Finally, reputation was also mentioned by some as a factor. This same risk manager mentioned that another major reason his bank was spending a lot of money in this area was that bank management wanted to position their electronic banking service as the safest in the region. It is worth noting the inherent difficulty of obtaining accurate phishing loss figures for financial institutions. It is difficult to separate phishing from other electronic fraud, such as malware. Furthermore, such losses impact a variety of different parts of a company, such as customer service, and thus may not be fully accounted for by the fraud department. Finally, it is difficult to quantify indirect loss such as damage to one’s reputation. Even if financial institutions have accurate phishing loss estimates, they often do not have incentives or regulatory requirements to disclose them. They may prefer not to disclose these losses due to fear of brand erosion due to negative publicity. This leads to a wide range of loss estimates that differ by an order of magnitude (e.g. [92] vs. [43]). Merchants. Merchants lose money because financial institutions eventually charge them back for fraudulent transactions. When a phisher makes a purchase using a stolen credit card, the credit card company usually charges the merchant for the loss. With online and telephone transactions known as “card-not-present” transactions, merchants assume this liability directly if cardholders dispute a charge. The Merchant Risk Council estimates that merchants who manage their risk well still lose about 1% of their revenue to credit card fraud [84]. Internet Service Providers: The ISPs we interviewed all considered phishing as part of the spam problem, which is their number one concern. Since phishing usually represents less than 1% of the spam they receive, their typical response is to filter out phish with spam. For example, one University ISP expert said, “We filter as much as we could and we would like [our users] not be sending their credit card and social security numbers online, but we don’t see that as our 44 responsibilities to protect those numbers, it is their personal data to protect.” Other experts from academia echoed this sentiment as well. ISPs do have an incentive when phishing targets their own mail systems. These phishing attacks typically seek to compromise users’ webmail accounts hosted by these ISPs and use them to send out more spams. ISPs have the incentive to ensure mail flows properly and avoid having their mail servers being blocked by blacklists. When it comes to fixing compromised machines that are often used as part of a botnet to send out phishing emails, ISPs currently do little. These compromised machines sometimes form a fast flux network, in which a domain name that phishers use has multiple IP (Internet Protocol) addresses assigned to it. The phishers switch those domains quickly between the addresses (often compromised machines) so that it is not as easy to find or shut down the phishing sites. One expert from a major US ISP recognized that compromised PCs cause major problems, and told us that close to 10% of their customers’ machines were infected with malware. However, when asked why his company does not remove these computers from the network he said, “Well, they are paying [a monthly fee] . . . for Internet access.” Experts from other ISPs made similar comments and noted that fixing infected computers can be costly. Infected computers may need to have their operating systems reinstalled. One expert from an ISP mentioned that customer service is the largest cost for the ISP. However, most experts who did not work for ISPs identified infected machines on ISP networks as a major problem that needs to be fixed. Domain Registrars: Registrars have been generally regarded as lagging in terms of phishing countermeasures. One expert claimed that registrars actually have a disincentive to fight phishing as criminals still pay them for registering phishing domains. However, another expert familiar with the registrars disagreed, saying, “Registrars would get charge back eventually because phishers are usually using fake credit cards to register these domains.” Some other experts suggested that registrars lacked the capability to detect and shutdown phishing fraud, as they work on small profit margins. 45 Stakeholder Capabilities and Incentives are Often Misaligned. Economists have suggested that liability should be assigned to the party that can do the best job of managing risk [136]. However, throughout the interviews, we found that the party that can do the best job is not always managing the risk. For example, in Asia, if banks can prove that a customer acted with negligence, the bank is not liable for a phishing loss. The difficulty is to prove that customers acted with negligence. One participant from a major bank in Asia said that when his bank was first attacked by phishers, the bank reimbursed victims. However, he said, “We’ve then since done a lot of education and we have joined the association of banks for a series of community bank education programs. After that, if customers do not pay attention to the education, we consider that to be negligent, so we try not to reimburse them. Of course, if the customer starts to yell and complain to the regulators, then it is entered into a fueled debate.” As another example, experts mentioned that merchants are held liable when phishers use fake credit card credentials to buy goods from them. When banks find out about the fraudulent charges, they will charge the merchant for it and sometimes also charge fines. This liability can be shifted if merchants implement the “Verified by Visa” program, but many merchants do not because of usability concerns. Furthermore, one expert argued that it is very difficult for merchants to notice that a credit card is stolen, noting that banks are at a much better position to make that judgment because they possess more information about the given credit card and a history of the transactions that make it easier for them to spot fraudulent charges. As a third example, some experts claimed that ISPs are in the best position to protect their network and clean up compromised machines, but are not willing to take proactive measures because they would incur high costs while generating little benefit. One expert said: The ISP is in a good position to inspect and identify some machines that are sending out spam and launching denial of service attacks. . . . There are quarantine devices that exist. . . . ISPs have it, but even for the ISPs using them, it is not used much. It is expensive for ISPs. If you put the user on quarantine, you end up having high customer 46 cost, the person will call the help desk, and you have to walk them through everything. The benefit to the ISP is very low compared to the cost. This is because the ISP did not bear the cost of compromised machines, putting externalities, hosting spam, it is not infecting the ISPs bottom line, but it is impacting every one else’s bottom line. We asked experts to comment and prioritize on a set of recommendations on the issues of incentives. We discuss the first recommendation with our experts and introduced the second recommendation based on our findings. Recommendation (R1): Financial institutions should produce more accurate estimates of phishing losses and report these statistics. As we mentioned earlier, accurate estimates of the phishing threat are difficult to come by, but very important. For example, it is difficult for law enforcement to open cases if they do not have a good idea of the amount of loss or the type of damages. Similarly, without quantifying damages, it is hard for corporations to manage the risks. For a corporation to obtain these figures, experts suggest two possible steps: first, law enforcement should collect and preserve forensics data when the phishing servers or drop accounts (email accounts used to gather stolen credentials) are seized, provide detailed information about the accounts stolen, and collaborate with banks to double check these fraud cases. Second, fraud managers within the organization should look at the organization as a whole when estimating damages, not just the online fraud itself. For example, they could examine how phishing increases customer service costs. The cost to financial institutions for implementing these policies include researching the damage to the institution holistically, implementing measures to record the losses if no measure is in place. The immediate benefit to the financial institutions is that they will have a clear picture how phishing impacts their organization. The larger benefit, however is given to other stakeholders in that they can make their decisions better with more accurate data. The obstacles for implmenting this recommendation is that currently many financial institutions do not have incentives to report estimates of phishing losses, and fear of negative publicity serves as a disincentive. One way to address this is mandatory anonymous reporting, such as in the case 47 of the UK payment association (APACS), which requires its members to report their losses and aggregate them together. Recommendation (R2): Regulators and academic researchers need to investigate the issue of incentives further. As mentioned in our findings, some stakeholders (such as consumers or merchants) are not really equipped to protect themselves against fraud, so placing the liability or burden of proof on them would do little to help fight against phishing. On the other hand, ISPs who are in a better position to clean the compromised machines do not have incentives to do so. Further research is needed to develop incentive models and determine where incentives are misaligned and ways to realign the incentives. 3.5.3 What stakeholders should do Experts identified operating system vendors, web application providers, browser vendors and Internet service providers as stakeholders with key technology influence over phishing. Experts identified operating system vendors, web application providers, browser vendors, and Internet service providers as being best positioned to fight phishing. Operating systems are crucial because their security or insecurity has far reaching effects. Experts generally praised Microsoft for their efforts in hardening their operating systems, but pointed out more to be done in this area. They gave a few recommendations that we will cover in the later part of this section. Experts pointed out the insecurity of web applications as a significant hurdle. One technical expert charged web application vendors for the current state of the problem: [Phishers] are losing on the email; the majority of the places are running filtering now, spam and antivirus filtering. But if I want to compromise the end-user, I am going to send them a URL and redirect them to some website that hosts malware. The stuff that can become most widespread is SQL injection of some legitimate server, and users will see an iframe that loads a malware onto it. 48 Experts also commented on the strategic position of the browsers in the fight to protect consumers. First, web browsers can warn users directly and effectively. A recent laboratory study showed that when Firefox 2 presented phishing warnings, none of the users entered sensitive information into phishing websites [29]. This study also recommended changes to Internet Explorer’s phishing warnings, and Microsoft has already acted on some of them to improve IE 8’s warning mechanism. Second, the browser market is fairly concentrated, with two browsers (Internet Explorer and Firefox) accounting for 95% of the total market [101]. Solutions implemented by these two browsers would provide the majority of users with a defense against phishing. Finally, experts pointed out that ISPs are in the best position to clean up compromised machines, as described earlier. We asked experts to comment on and prioritize a set of recommendations for securing the computing environment. Experts ranked the following as top priorities. Recommendation (R3): OS vendors should continue to secure operating systems by implementing secure coding practices, investing in secure vulnerability patching, and building anti-malware capability directly into the operating systems to enhance default security. To secure the operating system, experts suggested Microsoft protect the hosts file in Windows XP and earlier versions, as done by some Antivirus software [149], to prevent pharming attacks. Another way to secure the operating system is by constantly patching with the latest updates, as a fully patched computer with firewall enabled provides a strong defense against exploit-based malware. However, one of the problems with patching is that distributing a patch provides information to criminals about the security vulnerability that is being patched. Even if the description is vague, a patch can be disassembled and compared to the code that it replaces. Once a new exploit is known, a malware exploit can be quickly crafted using pre-built components. It currently takes less than three days – sometimes only a matter of hours – between the time a patch is released and the time a malicious exploit appears. After this short period of time, most computers are still vulnerable to infection. Research and application development into securely delivering patches to computers, possibly using public-key cryptography, would help alleviate the problem [52]. 49 Finally, some experts suggested building anti-virus and anti-malware capability directly into the OS. Experts pointed out that XP service pack 2 has a security center with firewalls enabled and suffers fewer attacks than service pack 1 [12]. These experts also praised Microsoft’s effort to distribute malware removal tools and updated malware signatures monthly, and argued that Microsoft should provide some default protection to computer users who do not buy anti-virus software. Recommendation (R4): Stakeholders should focus on improving the security of web applications, providing support and incentives for fixing applications. Currently, over 70% of phishing websites are hosted on hacked websites or free hosting sites. Many vulnerabilities for web applications exist (e.g. SQL injection, cross site scripting, remote code execution), making them a tempting target for criminals. Experts suggested a few ways to improve the security of web applications. One expert felt that technical authorities such CERT or APWG should produce a list of most frequently hacked websites and notify the website operators of their vulnerability. However, not all website operators have the technical capability or incentives to fix the problem. A recent paper by Moore and Clayton showed that 25% of the hosts used for phishing end up being compromised again within a couple of months [94]. If the compromise is due to a lack of technical capability, then there needs to be a way to provide tools and educational resources to help them secure their web application. On the other hand, if repeated compromises are due to a lack of incentives to fix, then there needs to be a way of punishing transgressors, with escalating consequences. Another approach is to involve the hosting provider. For example encourage these providers run intrusion detection on the applications they are hosting, and scanning newly created pages for phishing and malware. Recommendation (R5): Web browser vendors should continue to improve the performance of integrated browser anti-phishing warning systems, with a goal to catch 90% of phishing URLs within an hour after they go online. As mentioned previously in this section, web 50 browsers is at a strategic position as they can warn users effectively, and faster than other methods. Currently, browser-integrated phishing warning systems catch only 40-60% of the URLs 3 hours after the attacks are launched [128]. To provide the majority of Internet users with adequate protection, these warning systems should be improved. To accomplish this, the key is heuristics. Currently major browsers only use human-verified blacklists. To raise detection rates significantly, heuristics need to be used to supplement existings blacklists and block attacks more quickly [128]. Another way to improve the coverage of the blacklists is to gather phishing feeds from multiple sources to maximize their coverage [93]. However, as discussed in the next section, browser vendors are extremely cautious in using heuristics because of false positives, incorrectly labeling a legitimate site as phishing, which could potentially expose them to costly lawsuits. We present recommendations to address this issue in the next section. Recommendation (R6): Academics and for-profit protectors should develop better techniques to quickly identify botnets and proxies, shut down botnet command and control, and clean compromised machines. To shut down botnets, experts recommended that we either go after their command and control centers or clean the bot machine themselves. In November 2008, a hosting company named McColo that hosted a bot command and control center was disconnected by its upstream providers, causing a nearly 70% drop in spam volume [65]. More efforts to identify and shutdown comand and control centers would diminish the usefulness of other bots. However, we have to be mindful that criminals will continue to regroup and attack again. A good illustration is that two months after the McColo case, the spam volume was back to the previous level [20]. Spammers find other bot command and control centers, and they are getting more sophisticated in using P2P tools to control bots instead of traditional IRC commands. Defenders need to learn from successes and failures to ensure faster reaction in the future. The McColo case offers several lessons. There invariably exists some rogue hosting companies (also known as bullet-proof hosting), so persuading them to clean up their network would be difficult and likely have limited effect. Therefore it is important to involve upstream connectivity 51 providers. However, these providers face some challenges for proactive monitoring. For example, the infrastructure for monitoring is expensive, the legal justification is unclear, and because of contractual agreements, they are likely to be very cautious. So other stakeholders such as public protectors or for-profit companies needs to help provide as much evidence as possible. Second, media can play an important role. In the case of McColo, a Washington Post report played a critical role in pursuading the upstream providers. Similarly, the media played an important role in having the Russian authorities shut down the Russian business network, a known hosting provider for Internet miscreants [30]. Finally, the higher the level of coordination between stakeholders, the better they are at identifying and shutting down these rogue providers. Another approach focuses on cleaning up individual machines. This is a much more challenging task as there are millions of compromised machines to fix. ISPs need to be involved. Recognizing the disincentives mentioned in section VI, one expert suggested a notice and take down approach: certain third parties can notify an ISP that a certain computer on its network is in a botnet or doing something malicious. Once the ISP receives the notification, it becomes obligated to clean up the machine. The cost for ISP in this instance is the cost of cleaning up the compromised machine, elevated customer service costs, and potential costs due to customer leaving. The benefit to ISP in this case is little, however the benefit to other stakeholders are more pronounced. Therefore, it is still necessary to implement a notice and take down approach. Another challenge for the notice and take down approach is who is providing the notice? and whether ISP will trust the notice served. With some kind of safe harbor regulation similar to DMCA’s notice and takedown provision, this problem can be solved. Finally, efforts are needed to automate the clean up process. Experts suggested that we won’t see much of an impact on crime rate until we clean up a large fraction of compromised machines. Hence, better automatic solutions are needed to complement the notice and take-down approach. Although no actions have been taken so far, the ISPs we interviewed acknolwedged that compromised machines are a big problem. During the interviews, they asked about academic research 52 on automated tools to quarentine these compromised machines. We suggest conducting more research and development focusing on automated mitigation of malware-infected computers. Organizations are conservative about filtering and warning about phish because they are worried about false positives. However, this often leads to repeated efforts and slow reaction. The issue of false positives came up frequently during our interviews. Generally speaking, phishing detection falls into two categories: blacklist-based methods in which humans verify suspicious phishing URLs, and heuristic approaches that utilize HTML or content signatures to identify phish automatically. In our interviews, we found that providers favor blacklists over heuristics, and even those who do use heuristics are using them conservatively. For example, an expert at an ISP told us that they had a system that warns users if a certain email appears to be phish (based on blacklists and heuristics), but they did not delete these emails because they consider their falsepositive rate to be too high. Browser vendors are also extremely concerned about false positives. The expert from a major browser vendor said that they take false positives very seriously and manually verify each URL on their blacklist to avoid false positives. All of the major browsers appear to favor human-verified blacklists with extremely low false positives over heuristics that may potentially have higher false positives. Registries consider false positives as their biggest concern in implementing anti-abuse policies. One registry told us that they do not take act on phishing URLs submitted by third parties (such as takedown vendors) until the URLs have undergone a review process to determine if they are really phishing URLs. In other words, a phishing site is verified multiple times by different parties before action is taken, wasting precious time. Infrastructure providers are concerned about potential liability from mislabeling or taking down legitimate websites. There have been cases where companies have attempted to hold service providers responsible for false positives, but as of yet no company has been held responsible. 53 For example, in a 2005 court case, Associated Bank-Corp sued Earthlink after the Earthlink antiphishing software ScamBlocker blocked the bank’s legitimate page [11]. Earthlink was able to fend off the suit on the basis that it was using a blacklist of phish provided by a third party, thus, under a provision in the Communication Decency Act (CDA), it could not be held liable as a publisher when that information is erroneous. Although the bank apparently did not sue the provider of the blacklist, the court opened the door for them to do that. False positives based on heuristics have more subtle concerns. If heuristic-based software blocks a phish that turns out to be a false positive, the vendor may be regarded as a publisher under the CDA, and thus not immunized. Because of these fears, heuristics are not favored in integrated browser phishing protection. It is unclear, however, how future cases, if any, will be handled. One legal expert thought there was no case to be made. He said: I think everything will depend on what statements are made about the blocked site by the anti-phishing software. For example, when it says, ‘we think this site might be a phishing site,’ unless they were grossly negligent (in which case the thinking would not be reasonable), there would probably be no liability. If it said ‘This site is absolutely a phishing site’ it would be a whole different story. It is worth noting that vendors have developed blacklist processes and heuristics with extremely low false positive rates. One software vendor told us at their current false positive rate is so low that a user would encounter a false positive only once in a few years. Another takedown provider told us that they only had one or two false positives in the past four or five years, and even those false positives were arguably true positives. Recent academic work has shown that heuristics seem to detect websites with near zero false positives ( [75], [128]). It is therefore, unclear why vendors remain so reluctant to use heuristics more aggressively. To address this issue, we introduce three recommendations based on our findings. Recommendation (R7): Clarify the legal issues surrounding false positives of blacklists and heuristics. Companies are adopting conservative strategies to avoid false positives for fear of 54 liability, even when false positives occur rarely. This is hurting phishing protection, especially when heuristics offer real-time protection against phishing and have considerable benefits over blacklists. We encourage more discussion on liability surrounding the use of phishing blacklists and heuristics. So far, there has been no test case on this matter. The question at hand is at what level of accuracy heuristics can be applied to block phish and not be held liable? Some experts argued that zero false positive is the only acceptable level, but most of the experts interviewed feel that it would be reasonable to block with less-than perfect accuracy if a procedure were in place to correct errors. Safe harbor legislation, which immunizes providers from liability if they meet certain standards, may be necessary to make companies comfortable that they will not be held liable. Clarifying liability is important because lack of clarity on these matters could further reduce vendors’ incentives to use heuristics to detect phishing and get protections in place rapidly. Major browser vendors and ISPs potentially take on liability for false positives, but do not lose money directly from phishing. Therefore, an uncertain legal situation may reduce their willingness to be proactive. Recommendation (R8): Create a central clearinghouse to quickly verify phishing reports coming into APWG and on vendor blacklists. Currently there is a great deal of duplicated effort as phishing reports end up getting verified by multiple sources. For example, many vendors and service providers will not trust phishing reports until they have verified them themselves. A verification organization could serve as a clearinghouse for phishing reports and allow these reports to be verified rapidly using a standard process in which the evidence supporting each report is fully documented. In addition, it is important to report whether each phishing site is a domain setup for phishing or a legitimate domain that has been hacked. This distinction is important for registrars and registries, as these cases require different actions to be taken. Recommendation (R9): Researchers should focus on heuristics that minimize false positives. A sampling of published research has found that current anti-phishing heuristics have a false 55 positive rate of 0.43% - 12% [128]. However, to make sure these heuristics are used, the false positive rate needs to be extremely low. Since billions of websites are visited each day, even if a heuristic has a 1% false positive rate, it means millions of webpages are falsely labeled. For heuristics to be used widely, the false positive of heuristics needs to be at near zero levels. Recent efforts such as [143] and [109] is a good start. Registrars and registries can play an important role in fighting phishing. As mentioned earlier, registrars and registries have been generally regarded as lagging in terms of phishing countermeasures, but many experts interviewed agreed that they could play a more active role. For example in the case of fast flux attacks, registrars need to be prepared to suspend phishing domains. The Anti-Phishing Working Group produced a set of recommendations for registrars and registries [6]. One key player is the Internet Corporation of Assigned Names and Numbers (ICANN). It is responsible for managing the root zone DNS, setting and negotiating contractual standards for registrars and registries. ICANN is not a regulatory body like the Federal Communication Commission (FCC) and it has limited capabilities to regulate. Going forward, many experts think that ICANN can and should play a more active role in combating phishing and other crimes. Experts suggested that ICANN establish a minimum set of standards for registrars and registries, coupled with self-regulation and better enforcement. However, experts acknowledged that ICANN needs to play a delicate role and achieve consensus with the parties involved to avoid backlash. We asked experts to comment on and prioritize a set of recommendations for registrars and registries. Experts ranked the following recommendations as top priorities. Recommendation (R10): ICANN should improve enforcement of domain abuse. Experts agree that one thing ICANN can do better is to enforce compliance. One expert familiar with ICANN said: Some registrars . . . are very good at enforcing compliance. Other registrars are very good at looking as if they can’t do it. KnujOn lists top 10 registrars with domain 56 abuses. Most of my anecdotal research, we see those same names that come up again and again. But they are just confident enough to keep their accreditation. ICANN has been improving their efforts. In October 2008, they de-accredited one of the illbehaving registrars. Experts think more of these efforts would be good, because de-accreditation produces a credible penalty for non-compliance, as it essentially terminated the registrar’s business. Recommendation (R11): ICANN should encourage registries to adopt anti-abuse policies. Several registries have implemented anti-abuse policies, and anecdotal evidence [7] suggests that registries who have implemented anti-abuse policies have much less fraud than those who have not. An expert who works for a registry that recently adopted anti-abuse policies told us his company adopted these policies after they observed how similar policies helped other registries. However, some registries may not have enough incentives to adopt anti-abuse policies because adding policies creates overhead. ICANN can provide some incentives. One way to encourage adoption is for registries who have adopted anti-abuse policies to share their stories and explain how they led to cost savings and how they handle the issue of false positives. To some extent this is already being done, but ICANN can encourage this further. Another inducement to adopt anti-abuse policies is for ICANN or APWG to publish phishing data based on different registries’ performance on phishing takedowns, and to share this information regularly with registrars and registries. Finally, as a stronger incentive, ICANN could use anti-abuse metrics as part of their evaluation criteria for future registry applications, for example approving new gTLDs. 3.5.4 Law enforcement and education Experts agreed that law enforcement should be emphasized, but law enforcement lacks the necessary tools, personnel, and resources to catch phishers. Experts agreed that law enforcement is essential to deter phishers, and the top priority for law enforcement anti-phishing efforts is to catch organized phishing operations such as rock phish, which are responsible for more than 50% of the phishing attacks. One expert commented: 57 If we can take out the major hubs, it is not going to solve the problem, but it can show that law enforcement can catch them . . . On top of that, these criminals have complex network, and it is not easy to set up. If we can get these gangs, then we may still have the coding kiddies, but those are a lot easier to catch. However, experts acknowledged that law enforcement face significant challenges: International nature of the problem. Experts acknowledged that the underground economy is very specialized. One gang is using compromised web servers in many countries that launch attacks with victims in multiple countries. Currently the Mutual Legal Assistance Treaty (MLAT) forms the basis for cooperation between different nations. However, the law enforcement experts that we interviewed complained that this process is very slow. Proxies. Phishers use proxies so that it is difficult to catch them when they check balances on compromised accounts. This problem is hard to overcome, as there are estimated to be over 10,000 active proxies and it is necessary for law enforcement agents to perform network monitoring of the proxy machine to catch phishers. However, a warrant is required for law enforcement to legally monitor proxy machines, and by the time a warrant has been issued, the phisher has moved on to a different proxy. Lack of accuracy in Whois data: Phishes are aware that law enforcement uses Whois data to trace illegal activity, so phishes fabricate contact information when they register domain names using stolen credit cards. Lack of analytical capabilities: Law enforcement often lacks the ability to analyze the data they have. One law enforcement officer that we interviewed said: It takes a lot to identify a criminal. There is a lot of data submitted to us from members of APWG or DPN (Digital PhishNet). We don’t have time to look at it all. We have to pick out a few variables we know historically told us that is a good target. But the question is that what are we missing? Is there something on that phishing kit are we missing? 58 Lack of case development tools to process the subpoena request: Multiple law enforcement agents commented on the lace of case development tools. One local law enforcement agent commented: When we issue subpoenas, some will give searchable PDFs, others give us Microsoft Access database, and some even give us paper. We need tools to conform to the same form of dataset. This is usually done case by case. If law enforcement has a centralized place to do that so that agents all over the country can use it. We asked experts to comment on and prioritize a set of recommendations for more effective law enforcement. Experts ranked the following recommendations as top priorities. Recommendation (R12): Improve and invest more into law enforcement, specifically for international cooperation. Experts commented that it is currently fairly difficult to cooperate with different law enforcement in different jurisdictions because there is often not a lot of money set aside for cooperation. At this time, the cooperation is through the MLAT process, which is very slow. One way to improve on this is to have a joint-task force between two police jurisdictions. Recommendation (R13): The US Government should invest in technologies to provide law enforcement with better analytical capabilities to prioritize and manage cases. There are over 40,000 classic phishing attempts every month, and prioritizing which cases to pursue is critical. One expert said: Just speaking on [our organization’s] behalf, we get a lot of information in, but we are overloaded. People can share data now, that’s occurring, but what’s not happening is the analysis piece. We have limited resources . . . We do it manually. We need resources, software and hardware to enable that, also more bodies looking at it. There is no magic about the data, but the magic is in the analysis. . . taking institutional knowledge and applying some data mining algorithms. 59 Recommendation (R14): Get more corporations to aggregate and submit fraud data to law enforcement to identify proxies. Currently, most phishing attacks are from botnets and proxies and almost all criminal organizations use proxies to check account balances of phished accounts. Aggregating these data from various sources will help law enforcement to determine where to request subpeonas for wire taps. One way to do this is by having corporations work together and give law enforcement fraud data with a single list of IP addresses that have checked balances on compromised accounts. Another way is for Internet service providers who have information to share that with law enforcements. Recommendation (R15): Continue to strengthen collaboration between public protectors, private protectors, and between law enforcement in different countries. Collaboration is key to catch phishers due to the international nature of phishing. It is vitally important for law enforcement to develop good relationships with their peers in other countries. One noteable effort is the Digital PhishNet conferences that NCFTA and Microsoft organize each year. More efforts like these are needed. Experts agree that shutting down money trails is very important to defeat phishers. Experts said that shutting down the money trail can make phishing less attractive. For example, phishers often use “money mules,” persons recruited to receive stolen funds (or goods bought using stolen funds) and then transfer the money out of the country. Mules are recruited by a variety of methods, including spam emails, advertisement on genuine recruitment web sites and newspapers, approaching people who have their CVs available online, and instant messaging. To shut down money trails, one expert recommended we find out where the mules typically are and how mules are recruited. Another expert suggested that banks and take-down organizations put more effort into shutting down mule recruitment websites. He mentioned recent research that mule recruitment sites takes much longer to shutdown than normal phishing websites. 60 Another expert proposed a clearinghouse of accounts where each participating bank submit accounts that have been used as mules. Currently, bank fraud systems can detect some suspicious transactions to mule accounts, but there is no system in place to share this information with other banks. If this list of suspicious accounts were shared, a lot of money laundering could be stopped. Education and awareness are important factors that are not emphasized enough. However, not all experts agree on the effects of education. Most experts agreed that anti-phishing education for end users needs to be implemented better. However, some experts strongly endorses it, while others say eductaion should not be a focus. Both sides have strong words to say. For example, one expert in favor of more education said: There needs to be some accountability on Internet users . . . . People still click on URLs they shouldn’t. So we need to stress user education, and a little bit of common sense. We are a society that becomes desensitized to our responsibility. You really end up paying for this over time. You are going to end up paying high interest rates. So you really do need to pay more attention. Another expert who has worked on anti-phishing campaigns at a large US institution doubted the efficacy of such efforts: My experience of education is that it won’t make that much difference. You have to do it, because if you don’t, consumers will get mad at you. There is trust and there is safety. You have to balance both of them. . . . However, education doesn’t impact phishing losses, or make it less. It doesn’t do any of that, what it does is making people feel safer. If your goal is to improve security, then education shouldn’t be of top priority.” Based on these comments, we introduced a set of recommendations. Recommendations (R16): Academic researchers and industry should continue to make education fun, engaging and up to date. Current academic research shows that popular online 61 user education materials are effective if users actually read them. For example, Kumaraguru et. al asked users to read four popular training materials online and tested their ability to recognize phishing websites. They found that users were able to distinguish phishing websites from legitimate ones much better after reading these training materials [70]. However, the problem is that users normally don’t read security training materials [69]. To make education more effective, we recommend developing more innovative ways to make education fun, engaging, and up to date (e.g. [127], [67]). Recommendation (R17): Launch an education campaign to educate the public about mules, and encourage social networking sites to take the initiative to educate their customers about phishing. Experts mentioned the need to educate money mules, some of whom unknowingly become accomplices to crimes. To educate mules, experts recommend we find out where the mules typically are and how mules are recruited. Finding out where they are recruited can help determine whether national campaigns or if targeted campaigns are needed. Experts also thought social networking sites should take the initiative to educate their customers about phishing, as they are increasingly becoming targets of phishing campaigns. Recommendation (R18): Complement education with other countermeasures such as filtering and better user interfaces. Where possible, efforts should focus on automatic filtering that does not require user knowledge, and designing better user interfaces that make it more obvious to users what the right trust decision is. However, education remains an important part of combating phishing because it is unlikely that any automated system will ever be completely accurate in detecting phishing attacks, especially when detection requires knowledge of contextual information. There will still remain many kinds of trust decisions that users must make on their own, usually with limited or no assistance. 62 3.6 Discussion 3.6.1 Applicability of the Recommendations against Spear-phishing In this chapter, we reported on 18 recommendations from 31 qualitative interviews with antiphishing experts. These recommendations are effective for combating generic phishing. However, as spear-phishing increases, what are the unique challenges that it poses? Can we combat it by applying our anti-phishing recommendations? In the concluding section of this chapter, we address these questions. Compared with traditional phishing, spear-phishing poses two unique challenges. First, unlike traditional phishing scams that send mass phishing emails to everyone, spear-phishers send fewer, more targeted emails. This poses challenges to the current signature-based email filtering systems, which rely on large number of emails for fingerprinting. Second, Spear-phishing is a highly targeted phishing scam. Phishers exploit the social context to send spoofed emails to consumers that appear to come from someone they know. These attacks pose a severe threat for the end users, who normally use social context as cues in determining email legitimacy [50]. As a result, users fall for more spear-phishing attacks compared to regular phishing attacks [53]. Although spear-phishing poses these problems, the majority of our recommendations are likely not affected. Our recommendations attack the root problem of phishing by improving law enforcement (R12 - R15), improving incentives for stakeholders with better statistics and more research (R1, 2), and hardening the underlying infrastructure to make phishing less easy to conduct (R3,4, 6, 10,11). All of these efforts can lead to the reduction of both generic phishing and spear-phishing. A few of our recommendations would be particularly useful in terms of combating spearphishing. Heuristics would be very important in identifying spear-phishing emails, as it does not use signature-based fingerprinting that relies on a large number of emails to be accurate. Therefore the two recommendations on improving heuristics would be particularly helpful in combating spear-phishing (R7,9). 63 The majority of our recommendations on education will be effective against spear-phishing as well, although for recommendation R16, educators need to additionally incorporate elements of spear-phishing into their education curriculum. Finally, spear-phishing poses challenges to two of our recommendations: R5, for web browser phishing protection and R8, for a central clearinghouse to quickly verify phishing reports. The challenge is that spear-phishes are harder to detect, and therefore may take a longer time to verify and warn. However, by deploying heuristics more aggressively, the deficiencies of these two recommendations can be overcome. 3.6.2 Summary of findings In this chapter, we reported on seven findings (summarized in Table 3) and 18 recommendations (summarized in Appendix A) from 31 qualitative interviews with anti-phishing experts. Our findings suggest that phishing is evolving into a more organized effort. It is part of a larger crime eco-system, where it is increasingly blended with malware and used as a gateway for other attacks. Experts identified several places where incentives for fighting phishing may be misaligned, in the sense that the stakeholders who are in a position to have the largest impact do not have much incentive to devote resources to anti-phishing. To resolve this, we recommend further study of these misalignments and development of policy alternatives to correct them. In terms of countermeasures, experts identified improving law enforcement and shutting down money trails as top priorities. We identified key difficulties that law enforcement organizations face, and recommend investment into specific types of technologies made to equip law enforcement to better prioritize cases. Collaboration is the key in these investigations, so we recommend ways to foster it. Experts agreed that education is an important factor that is not emphasized enough, however, they did not agree on the effects of education. We recommend developing more innovative ways to make education fun, engaging and up to date and propose content areas that education needs to be focused on. 64 Finally, we qualitatively analyzed the challenges and obstacles for implementing these recommendations, their associated costs, and benefits, and actionable items that stakeholders can do to (see Table 3.4). Table 3.4: Obstacles and Challenges, benefits, costs, and actionable items for the recommendations Recommendation 1. Financial institutions should produce more accurate estimates of phishing losses and report these statistics. 2. Regulators and academic researchers need to investigate the issue of incentives further (a study comparing different phishing liability regimes around the world) Obstacles and Challenges 1. Financial institutions do not have incentives to report estimates of phishing losses, and fear of negative publicity serves as a disincentive. 2. It is hard to separate phishing from other kinds of losses such as malware. 3. Phishing losses appear in different units of the company and could be difficult to compile. 1. Data hard to get from financial institutions. 2. Regulatory environments are different around the world. Costs Benefits Actionable Items Cost to FI: 1. researching the phishing damage holistically. 2. Implementing measures to record the losses if no measures are in place. Benefit to FIs: they will have a clearer picture how phishing impacts their organization. Federal regulators draft rules to require mandatory anonymous reporting, such as in the case of the UK payment association (APACS). Costs: Time and resources of academicians and regulators for the research Benefits: Solid research can help regulators to assign liability to the party who is most capable of fixing the problem. Benefit to others: They can make more informed decisions about the investment and management of the risk. Regulators in different regions compel financial institutions to provide the data. 65 Table 3.4: Obstacles and Challenges, benefits, costs, and actionable items for the recommendations Recommendation 2a. Regulators develop a notice-takedown approach for botnet C&C removal 3. OS vendors should continue to secure operating systems by implementing secure coding practices, investing in secure vulnerability patching, and building anti-malware capability directly into the operating systems to enhance default security. Obstacles and Challenges Challenges: privacy and contractual considerations for ISPs and hosting providers; potential for abuse Costs Benefits Actionable Items Costs: time to address concerns of opponents and negotiate compromises; cost of enforcement Challenges: 1. Secure coding takes time to mature. 2. OS vendors may lack expertise and experience in antivirus and anti-malware tools. Costs to OS vendors: investment of resources (time, personnel) Benefits: faster notice- Regulators develop a protakedown of botnet com- cess for takedown and apmand and control would peal. reduce the effectiveness of botnets dramatically in the short term Benefits to OS vendor: Improved security and visibility of the operation system. Benefits to others: a cleaner network environment with default security enabled 66 Table 3.4: Obstacles and Challenges, benefits, costs, and actionable items for the recommendations Recommendation 4. Stakeholder should focus on improving on the security of web applications. 5. Web browser vendors should continue to improve the performance of integrated browser anti-phishing warning systems, with a goal to catch 85-95% of phishing URLs within an hour after they go online. Obstacles and Challenges Challenges: 1. The total number of web applications needs to be fixed is large and owners may not know about them. 2. Attacks are continuous, so requires constant vigilance. 3. Website application owners lack expertise or may not care. 4. Hosting providers lacks incentive to proactively scan their network Challenges: browsers are conservative in using heuristics because of false positives Costs Benefits Actionable Items Costs to technical authority: gather knowledge and tools for reporting them. Benefit to web applications: reduce the risk of being blacklisted, improve the security. Cost to web application operators: time, resource and expertise to fix the vulnerabilities Benefit to others: Overall improvement in the general security. 1. technical authorities such CERT or APWG produce a list of most frequently hacked websites and notify the website operators of their vulnerability. 2. Provide educational resources for those who lack technical capability. 3. Punishing continuing transgressors, with escalating consequences such as a reputation-based systems. Cost to browsers: continual investment in improving anti-phishing capacity with better feeds Benefits: significant default protection offered to the end user. 1. Browsers use heuristics as a way to label websites for blacklist review. 2. Legal authorities clarify the liabilities surrounding the use of heuristics. 67 Table 3.4: Obstacles and Challenges, benefits, costs, and actionable items for the recommendations Recommendation Obstacles and Challenges 6. Academics and Challenges: for-profit protectors 1. Botnet C&C is very should develop better adaptable and tend to retechniques to quickly group after being shutidentify botnets and down. proxies, shut down botnet 2. Hosting providers are command and control, cautious because infrasand clean compromised tructure for monitoring is machines. expensive, the legal justification is unclear, contractual agreements could pose problems. 3. We need to fix a significant amount of machines to significantly impact to ecrime infrastructure. 4. There are privacy concerns of sharing fraud data between institutions 7. Clarify the legal issues Challenges: Determining of the false positives of the right level of false blacklists and heuristics. positives; legal risks for companies who are the test case. Costs Benefits Actionable Items Costs to ISP: the cost of cleaning up the compromised machine, elevated customer service costs, potential costs due to customer leaving. Benefits to ISP: little. 1. Other stakeholders such as public protectors or forprofit companies need to help provide as much evidence as possible. 2. The higher the level of coordination between stakeholders, the better they are at identifying and shutting down these rogue providers. Cost: legal research and proceedings Benefits: extremely high for stakeholders such as browsers and ISPs Benefit to others: significant reducing in the key ecrime infrastructure. APWG set the standard for acceptable level of false positives. 68 Table 3.4: Obstacles and Challenges, benefits, costs, and actionable items for the recommendations Recommendation 8. Create a central clearinghouse to quickly verify phishing reports coming into APWG and on vendor blacklists. 9. Academics should focus heuristic research on reducing false positives. 10. ICANN should improve enforcement on domain abuse. Obstacles and Challenges Challenges: Providing phishing feed is a legitimate business, a central clearinghouse would likely drive these out of the business; also likely to be reinventing the wheels Costs Benefits Actionable Items Costs: building the system and the ongoing administration of the system and verifying of phishing feeds Benefits: A single source reduce the duplicated efforts by various organizations and provides uniform protections for its users. Challenges: Transforming research into production is nontrivial. 1. ICANN has limited capability regulating registrars and registries. 2. The ICANN consensus process could be timeconsuming. Costs: Time and resources for the research Benefits: Low false positive heuristics would benefit browsers, email providers greatly. Benefits: deterrence effect for criminals and registrars who opt to play with them. NOTE: These obstacles means that there would be little incentives for APWG or other parties to take initiatives on this; a more likely scenario is for APWG to define certain performance metrics and certify the existing feed providers NSF or industries provide more research funding. Costs to ICANN: developing technical capabilities for spotting domain abuse. Action: ICANN should define metrics for domain abuse, and devise incentives to reward registrars with low abuse rates. 69 Table 3.4: Obstacles and Challenges, benefits, costs, and actionable items for the recommendations Recommendation 11. ICANN should encourage registries to adopt anti-abuse policies. Obstacles and Challenges Registries concern for false positives would slow their action time. 2. Registries may push the responsibilities to registrars. 12. Improve and invest more into law enforcement, specifically for international cooperation. Challenges: Phishers hide their traces in many countries; ecrime cases in other countries may have a low priority. Costs Benefits Actionable Items Cost to registries: building the system, receiving and verifying the phishing feed, and dealing with false positives. Benefits to registries: Improved security, competitive advantage. Benefits to others: fewer entities for takedown companies to interface with and faster takedown time. 1. Registries who have adopted anti-abuse policies to share their stories and explain how they led to cost savings and how they handle the issue of false positives. 2. ICANN or APWG to publish phishing data based on different registries’ performance on phishing takedowns. 3. ICANN provide incentives to registries who have implemented abuse policies, for example giving them priority for new gTLDs applications. Action items: FBI to establish a joint-task force between two police jurisdictions. 70 Table 3.4: Obstacles and Challenges, benefits, costs, and actionable items for the recommendations Recommendation 13. US Government should invest in technologies to provide law enforcement with better analytical capability to prioritize and manage cases. 14. More corporations aggregating fraud data and submit to law enforcement to identify proxies. 15. Continue to strengthen collaboration between law enforcement in different countries, public and private protectors. Obstacles and Challenges Challenges: Which law enforcement agencies to invest? Costs Benefits 1. Corporations may not be willing to share because of privacy, and consumer trust concerns (reminiscent of telecom’s wiretapping scandal after 9/11). 2. Corporations may not share for competitive reasons. Challenges: Law enforcement in different countries may not know each other, hard to find the right people to handle the case; phishing and ecrime cases in other countries maybe of low priority. Costs to law enforcements: costs to set up the system and cost of analysis Benefits: law enforcement would be able to determine which proxies to place wiretaps, significantly improving the opportunity to identify the criminals’ originating machine. Costs: organizing and subsidizing conferences, supporting mutual exchanges, Benefit: Getting the good people organized better is crucial in fighting cybercrime. Actionable Items US government invest in tools for better case management and better digital evidence processing; Expand scholarship programs to recruit graduates in computer science Action items: FBI to produce a list of fraud data variables that it wants financial institutions to share. 71 Table 3.4: Obstacles and Challenges, benefits, costs, and actionable items for the recommendations Recommendation 16. Academic researchers and industry continue to make education fun and engaging and up to date. 17. Launch education campaign to educate the public about mules, and encourage social networking sites to take initiative to educate their customers. 18. Complement education with other countermeasures such as filtering and better user interface design. Obstacles and Challenges Challenges: Lack resources; quickly evolving nature of the phishing threat so need continual education Challenges: some of the mules knowingly participate in the crime; educating people about mules may make some more likely to become mules N/A Costs Benefits Actionable Items Costs: Resources to design and disseminate these materials Benefit: education that are engaging and fun is important as otherwise users will not proactively read them. Benefit: Mule education will help those who are unaware to be more cautious; education on social network phishing can reduce people falling for them N/A Industry and government to fund licensing and deployment of some of the training materials that are proven to be effective. Action: FTC, APWG, US Postal service and other industry groups take lead in designing the materials, fund licensing and deployment of existing materials that are proven to be effective. N/A Cost: developing materials and disseminating them N/A 72 73 Chapter 4 Case Study of Browser-based Anti-phishing Solutions This chapter is largely a reproduction of a paper co-authored with Lorrie Cranor, Brad Wardman (University of Alabama), Gary Warner, and Chengshan Zhang, and published at CEAS 2009 [128]. As discussed in Chapter 2, to reduce phishing damage, stakeholders have enacted their own countermeasures. Internet service providers, mail providers, browser vendors, registrars and law enforcement all play important roles. Due to the strategic position of the browser and the concentration of the browser market, web browser vendors play a key role. Web browsers are at a strategic position at which they can warn users directly and effectively. In addition, the browser market is fairly concentrated with two browsers (Internet Explorer and Firefox) accounting for 95% of the total market [101]. Solutions that these two browsers implement provide the majority of users with a defense against phishing. A recent laboratory study shows that when Firefox 2 presented phishing warnings, none of the users entered sensitive information into phishing websites [29]. This study also recommended changes to Internet Explorer’s phishing warnings, and Microsoft has already acted on some of them to improve IE 8’s warning mechanism. For browsers to truly realize their potential to protect users, their warnings need to be accurate (low false positives) and timely. Currently, most browsers with integrated phishing protection or anti-phishing browser toolbars rely on blacklists of phish and, sometimes, heuristics to detect phishing websites. Perhaps because toolbar vendors are striving to avoid potential lawsuits from mislabeling websites, blacklists are favored over heuristics due to their low false positives. 74 In this chapter, we study the effectiveness of phishing blacklists. We used 191 fresh phish that were less than 30 minutes old to conduct two tests on eight phishing toolbars. We found that 63% of the phishing campaigns in our dataset lasted less than two hours. Blacklists were ineffective when protecting users initially, as most of them caught less than 20% of phish at hour zero. We also found that blacklists were updated at different speeds, and varied in coverage, as 47% - 83% of phish appeared on blacklists 12 hours from the initial test. We found that two tools using heuristics to complement blacklists caught significantly more phish initially than those using only blacklists. However, it took a long time for phish detected by heuristics to appear on blacklists. Finally, we tested the toolbars on a set of 13,458 legitimate URLs for false positives, and did not find any instance of mislabeling for either blacklists or heuristics. To the best of our knowledge, this paper is the first attempt to quantitatively measure the length of phishing campaigns and the update speed and coverage of phishing blacklists. Based on these measurements, we discuss opportunities for defenders, and propose ways that phishing blacklists can be improved. The remainder of the document is organized as follows: section 2 introduces the background and related work, section 3 discusses the test setup, section 4 presents our results, and section 5 discusses ways in which phishing blacklists and toolbars can be improved. 4.1 Background and Related Work Efforts to detect and filter phish can be implemented at the phishing e-mail level and at the phishing website level. To prevent phishing emails from reaching potential victims, traditional spam-filter techniques such as bayesian filters, blacklists, and rule based rankings can be applied. Recently, some phishing-specific filters were developed as well [1, 34]. In addition to these efforts, some protocols have been proposed to verify the identities of email senders [24, 124]. Although these efforts are promising, many users remain unprotected. Filtering techniques, are imperfect and many phishing emails still arrive in users’ inboxes. Thus, we need to make an effort to detect phishing websites as well. 75 Generally speaking, research to detect phish at the website level falls into two categories: heuristic approaches, which use HTML or content signatures to identify phish, and blacklist-based methods, which leverage human-verified phishing URLs to reduce false positives. Our research on blacklist measurement contributes to understanding the effectiveness of blacklists to filter phish at the website level. 4.1.1 Anti-Phishing Heuristics Most of these heuristics for detecting phishing websites use HTML, website content, or URL signatures to identify phish. Machine learning algorithms are usually applied to build classification models over the heuristics to classify new webpages. For example, Garera et al. identified a set of fine-grained heuristics from phishing URLs alone [41]. Ludl et al. discovered a total of 18 properties based on the page structure of phishing webpages [75]. Zhang et al. proposed a content-based method using TF-IDF and six other heuristics to detect phish [148]. Pan et al. proposed a method to compile a list of phishing webpage features by extracting selected DOM properties of the webpage, such as the page title, meta description field, etc [107]. Finally, Xiang and Hong described a hybrid phish detection method with an identity-based detection component and a keyword-retrieval detection component [144]. These methods achieve true positive rates between 85% and 95%, and false positive rates between 0.43% and 12%. The heuristics approach has pros and cons. Heuristics can detect attacks as soon as they are launched, without the need to wait for blacklists to be updated. However, attackers may be able to design their attacks to avoid heuristic detection. In addition, heuristic approaches may produce false positives, incorrectly labeling a legitimate site as phishing. Several tools such as Internet Explorer 7 and Symantec’s Norton 360 include heuristics in their phishing filters. Our research examines the accuracy of these heuristics in terms of their ability to detect phish and avoid false positives. In addition, we examine how anti-phishing tools use heuristics to complement their blacklists. 76 4.1.2 Phishing blacklists Another method web browsers use to identify phish is to check URLs against a blacklist of known phish. Blacklist approaches have long been used in other areas. Blacklists of known spammers have been one of the predominant spam filtering techniques. There are more than 20 widely used spam blacklists in use today. These blacklists may contain IP addresses or domains used by known spammers, IP addresses of open proxies and relays, country and ISP netblocks that send spam, RFC violators, and virus and exploit attackers [61]. Although a spam blacklist of known IP addresses or domain names can be used to block the delivery of phishing emails, it is generally inadequate to block a phishing website. One reason is that some phishing websites are hosted on hacked domains. It is therefore not possible to block the whole domain because of a single phish on that domain. So a blacklist of specific URLs is a better solution in the phishing scenario. Compiling and distributing a blacklist is a multi-step process. First, a blacklist vendor enters into contracts with various data sources for suspicious phishing emails and URLs to be reviewed. These data sources may include emails that are gathered from spam traps or detected by spam filters, user reports (eg. Phishtank or APWG), or verified phish compiled by other parties such as takedown vendors or financial institutions. Depending on the quality of these sources, additional verification steps may be needed. Verification often relies on human reviewers. The reviewers can be a dedicated team of experts or volunteers, as in the case of Phishtank. To further reduce false positives, multiple reviewers may need to agree on a phish before it is added to the blacklist. For example, Phishtank requires votes from four users in order to classify a URL in question as a phish. Once the phish is confirmed, it is added to the central blacklist. In some instances, the blacklist is downloaded to local computers. For example, in Firefox 3, blacklists of phish are downloaded to browsers every 30 minutes [122]. Doing so provides the advantage of reducing network queries, but performance may suffer between blacklist updates. A number of these blacklists are used in integrated browser phishing protection [10, 46, 90], and in web browser toolbars [16, 17, 102]. Although blacklists have low false positive rates, they generally require human intervention and verification, which may be slow and prone to human 77 error. Yet this is the most commonly used method to block phish. Our research investigates the speed of blacklist updates and the accuracy of blacklists? 4.1.3 Related Work Several authors have studied the effectiveness of phishing toolbars. In Nov 2006, Ludl et. al used 10,000 phishing URLs from Phishtank to test the effectiveness of the blacklists maintained by Google and Microsoft [75]. They found that the Google blacklist contained more than 90% of the live phishing URLs, while Internet Explorer contained only 67% of them. The authors concluded that blacklist-based solutions were “quite effective in protecting users against phishing attempts.” One limitation of this study is that the freshness of the data feed was not reported. We overcome this weakness by using a fresh phish feed less than 30 minutes old and by using an automated testbed to visit phishing websites nine times in 48 hours to study the coverage and update speed of blacklists. We arrive at a different conclusion from this chapter. In a related study, Zhang et al. [147] tested the effectiveness of 10 popular anti-phishing tools in November 2006 using data from Phishtank and APWG. Using 100 URLs from each source and 516 legitimate URLs to test for false positives, they found that only one tool was able to consistently identify more than 90% of phishing URLs correctly, but with false positive rates of 42%. Of the remaining tools, only one correctly identified over 60% of phishing URLs from both sources. This study had a similar weakness to the first study, and it also had a small sample of false positives URLs. We based our study on this setup, but made the following improvements. First, we used a source of fresh phish less than 30 minutes old. Second, we extend the methodology by separately analyzing phish caught by heuristics versus blacklists. Third, we tested phish nine times over 48 hours to study the coverage and update speed of blacklists; Finally, we used a much larger sample to test for false positives. Other researchers have studied the effectiveness of spam blacklists [58, 61, 111]. For example, Ramachandran et al. measured the effectiveness of eight spam blacklists in real time by analyzing a 17-month trace of spam messages collected at a “spam trap” domain [111]. In their study, whenever a host spammed their domain, they examined whether that host IP was listed in a set of DNSBLs 78 Figure 4.1 High-level system architecture for our anti-phishing evaluation test bed. The Task Manager (1) gets an updated list of URLs from a phishing feed, and then (2) sends that URL to a set of Workers. Each worker (3) retrieves a web page and checks whether the web page was labeled as a phishing scam or not, and (4) sends the result back to the Task Manager, which aggregates all of the results. The Task Manager and Workers are grouped together because they can be run on the same machine or on separate machines. in real time. They found that about 80% of the received spam was listed in at least one of eight blacklists, but even the most aggressive blacklist had a false negative rate of about 50%. In addition to research work introduced above, a number of industry efforts were used to measure the effectiveness of phishing toolbars as well [59, 82, 88]. 4.2 Methodology In this section we describe our anti-phishing testbed, explain how we collected phishing URLs for testing, and describe our evaluation methodology. 4.2.1 Anti-phishing Testbed We used the anti-phishing testbed developed by Yue et al. [148]. The testbed has a client-andserver architecture. It includes a task manager and set of workers, each of which is responsible for evaluating a single tool. During the test, the task manager first retrieved a list of potential phishing sites to test against. The task manager then sent each URL to a set of workers, each of which 79 was running a separate tool. To reduce the number of machines needed, we ran each worker on a virtual machine. Each worker downloaded the specified web page, examined whether its tool had labeled the web page as phishing or not using a simple image-based comparison algorithm, and returned that value back to the task manager. The image-based comparison algorithm works as follows: each tool has several known states (e.g., a red icon if it has detected a phishing site and a green icon if it has not), and each tool can be set up to be in a known location in the web browser. We capture screenshots of the tools and compare relevant portions of those images to screenshots of the tools in each of their known states. The task manager aggregated all of the results from the workers and tallied overall statistics, including true positives, true negatives, false positives, false negatives, and sites that no longer exist. 4.2.2 Phishing Feed We obtained the phishing URLs for this study from the University of Alabama (UAB) Phishing Team’s data repository. UAB has relationships with several sources who share their spam as part of the UAB Spam Data Mine. One of the largest sources is a spam-filtering company that provides services ranging from small business to the Fortune 500 companies located in more than 80 countries. This company reviews well over one billion emails each day and uses a combination of keyword searching and proprietary heuristics to identify potential phish. They then extract the URLs from these emails and send these URLs to UAB in batches every four minutes. UAB manually tested the URLs they received from the spam-filtering company to determine if they were phishing URLs. If a URL was a phish and had not been reported to UAB before, it was put on a list to be tested by the testbed. UAB sent this list to the testbed every 20 minutes.1 The testbed began testing each batch of URLs within 10 minutes of receipt. Because UAB received phish URLs every four minutes, they were able to label each URL with the four-minute time segments in which it was seen. Thus they could identify the first segment in which a URL was seen and identify subsequent time segments in which the same URL was 1 Sometimes randomization was introduced to URLs to attempt to defeat exact matching. We do not consider two URLs as unique if their difference is only in the attribute portion of the URLs. 80 Table 4.1 The top 10 brands that appear in our data set. Total phish: 191 Institutions # of Percentage Victimized phish Abbey 47 24.9% Paypal 21 11.1% Lloyds TSB 17 9.0% Bank of America 14 7.4% Halifax 13 6.9% Capital One 11 5.8% New Egg Bank 11 5.8% HSBC 7 3.7% eBay 6 3.2% Wachovia 6 3.2% Wellsfargo 6 3.2% reported. This approach to recording phishing URLs allows us to determine the length of each spam campaign — the time period over which phishers send out emails with the same phishing URL. If the spam campaign lasts for only one day, the effectiveness of anti-phishing tools on subsequent days is not as important as effectiveness on day one. While some users will read phishing emails days after the initial email send time, most users will read phishing emails within a few hours. Thus the most critical time to protect is when emails are still being actively sent by the spammer. We collected and tested a total of 191 verified phishing URLs during this study. Table 4.1 lists the top 10 brands that appear in our data set. 4.2.3 Evaluation Procedure Tools tested: We tested eight anti-phishing toolbars that use various blacklists and heuristics. They are Microsoft Internet Explorer version 7 (7.0.5730.11), version 8 (8.0.6001. 18241), Firefox 2 (2.1.0.16), Mozilla Firefox 3 (3.0.1), Google Chrome (0.2.149.30), Netcraft toolbar (1.8.0), McAfee Siteadvisor (2.8.255 free version), and Symantec Norton 360 (13.3.5). Except for Internet Explorer 7 and Symantec, all of these tools use blacklists only. Those two toolbars that use heuristics to complement their blacklists trigger different warnings when a phish is detected by heuristics 81 versus blacklist. We configured all tools with their default settings, except for Firefox 2, in which case we used the “Ask Google” option to query the central blacklist server every time instead of downloading phishing blacklists every 30 minutes.2 Testbed setup: We configured 4 PCs running Intel Core 2 CPU 4300 @ 1.80 GHz. Each PC ran two instances of VMware, each configured with a 720MB RAM and 8GB hard drive. For each toolbar, we ran the task manager and workers on the same machine to avoid network latency. Since some of the toolbars use local blacklists, we left every browser open for six to eight hours before each test to download blacklists, and we left the browser open for 10 minutes between every run during the test. We chose the eight-hour period because the necessary blacklists would download reliably within this time. Thus we are investigating the best case scenario for blacklist effectiveness. Test period: We ran the test for two to three hours on October 2, 8, and 9, 2008 and on December 3, 4, 5, and 15, 2008. During this time, batches of new unique phish were sent to the testbed every 20 minutes. The testbed began testing them 10 minutes after receiving the phish, leaving a total lapse time of approximately 30 minutes. Each worker opened up the desired browser with toolbars for 30 seconds before taking the screenshot. For each URL, we tested the toolbars’ performance at hour 0, 1, 2, 3, 4, 5 12, 24 and 48. We cleared the browser cache every hour. We collected and tested 90 URLs in October and 101 URLs in December. Post verification: After the data was compiled, we manually reviewed every website that toolbars labeled as legitimate. This step was necessary because some host companies did not issue 404 errors when taking down a phish. Instead, they replaced it with their front page. In this case, the toolbar will mark the website as legitimate, but in fact it was the phishing website being taken down. 82 Figure 4.2 Length of phishing campaign, measured as the time between the first and last appearance of the phish in our source report. The graph on the left shows length of phishing campaigns in days. The graph on the right shows length of phishing campaigns in hours for those campaigns that last one day or less. 4.3 Results 4.3.1 Length of Phishing Campaign We define the length of a phishing campaign (LPC) as the time lapse between the first time a phish appeared in our source report and the last time that phish appeared in our source report. As mentioned in Section 3.2, we received reports from our source every 4 minutes. Of the 191 phish we used to test phishing blacklists, 127 of them, 66%, had an LPC less than 24 hours, indicating that their corresponding phishing campaign lasted less than 24 hours. A total of 25 URLs had an LPC between 24 and 48 hours, and the remaining URLs had an LPC between 3 and 23 days. Examining the first day’s data more closely, we found that 109 URLs were spammed only in a two-hour period, accounting for 63% of the URLs in this dataset. To validate our finding, we calculated the LPC for 5491 phish provided by the same source and verified by UAB from February 17 through April 13, 2009. Similar to our testbed dataset result, we found that 66% of these phish had an LPC less than 24 hours, 14.5% had an LPC between 24 and 48 hours, and the remaining 19% of URLs had an LPC between 3 and 47 days. We found that 44% of the URLs had an LPC less than two hours. Figure 5.4 shows the LPC combined LPC results for our two datasets. 2 This feature is no longer available for versions after Firefox 2 update 19. 83 Table 4.2 Website takedown rate vs. length of phishing campaign (LPC). LPC is measured as the time between the first and last appearance of the phish in our source report. Website takedown rate at each hour is measured by the number of phish taken down at that hour divided by total phish. Hours % of website % Phishing taken down Campaign finished 0 2.1% 0% 2 7.9% 63% 4 17.8% 67% 5 19.9% 70% 12 33.0% 72% 24 57.6% 75% 48 72.3% 90% It is important to note that the LPC does not necessarily correspond to the time a phishing site is live. In fact, we found that compared to the length of a phishing campaign, the time to take websites down is generally much slower. By hour 2, 63% of phishing campaigns in our dataset were finished, but only 7.9% of those phish were taken down. As shown in Table 4.2, on average, 33% of the websites were taken down within 12 hours, around half were taken down after 24 hours, and 27.7% were still alive after 48 hours. Our LPC findings demonstrate the freshness of our data and show that current takedown efforts lag behind phishing campaigns. In the test conducted by Ludl et al., 64% of the phish were already down when they conducted their test [75], whereas in our sample, only 2.1% of phish were aleady down in our initial test. 4.3.2 Blacklist Coverage In this section, we present the results of two tests performed in October and December of 2008 (Figures 4.3 and 4.4). We found that blacklists were ineffective when protecting users initially, as most of them caught less than 20% of phish at hour zero. We also found that blacklists were updated at different speeds, and varied in coverage, as 47% to 83% of phish appeared on blacklists 12 hours from the initial test in October. 84 Figure 4.3 Percentage of phish caught by various blacklists in October 2008 data. This percentage is defined as the total number of phish on the blacklist divided by the total phish that were alive. URLs that were taken down at each hour were excluded in the calculation. Total phish at hour 0 was 90. At any given hour, we define the coverage of the blacklist as: No. of phish appearing on blacklist T otal phish − phish that were taken down We found that coverage rates of some of the blacklists were highly correlated. Firefox 2, 3 and Google Chrome appear to use the same blacklists. Internet Explorer 7 and 8 also share a blacklist. In our analysis, we combined the results for those tools that use the same blacklists. In our October test, all of the blacklists contained less than 20% of the phish initially. New phish appeared on the blacklists every hour, suggesting that the blacklists were updated at least once every hour. One notable improvement is the Symantec blacklist. In hour 0, their blacklist caught as much phish as the others, but in hour 1 it caught 73% of the phish, 2 to 3 times more than the rest of the toolbars. This difference is also statistically significant until 12 hours from the initial test.3 One possible explanation is that Symantec uses results from their heuristics to facilitate rapid blacklist updates [5]. 3 ANOVA, p < 0.05 85 We observed that the coverage of the Firefox and Netcraft blacklist is consistently highly correlated. Five hours after our initial test in October, 91% of the URLs that appeared in the Netcraft blacklist also appeared in the Firefox blacklist, and 95% of the URLs that appeared in the Firefox blacklist also appeared in Netcraft. The two blacklists are consistently highly correlated every hour except for our initial test in December. This suggests that the two blacklists have overlap in some of their data sources or have data sources with similar characteristics. Others were less correlated, phish on Internet Explorer only appear 45% of time on Firefox blacklist and 73% vice versa, suggesting they use different feeds with not much overlap. We found that the Firefox blacklist was more comprehensive than the IE blacklist up to the first 5 hours, and the Symantec blacklists performed significantly better than the rest of the toolbars from hour 2 to 12. After 12 hours, the differences were no longer statistically significant. Figure 4.3 shows this result in detail. In our December dataset, we observed similar trends in terms of coverage for some toolbars. However, Firefox and Netcraft performed much better here than in October. The Firefox blacklist contained 40% of phish initially and by hour 2, 97% of phish were already on the blacklist. One reason for this difference could be that during this period, the two tools acquired new sources that were similar to our feed. Finally we did not observe statistically significant improvement in other toolbars. Finally, we examined phish that the IE 8 blacklist and Firefox blacklist missed five hours after our initial test in October. We observed that at hour 5 the IE 8 blacklist missed 74 phish, of which 73% targeted foreign financial institutions. The Firefox blacklist missed 28 phish, of which 64% targeted foreign financial institutions. However, given our limited sample size, we did not observe a statistically significant difference in the speed at which phish targeting US institutions and foreign institutions were added to the blacklist. There were some notable differences between the phish missed by the IE8 blacklist and Firefox. For example, IE8 missed 21 Abbey Bank phish while Firefox missed only 4 Abbey Bank phish. 86 Figure 4.4 Percentage of phish caught by various blacklists in December 2008 data. This percentage is defined as the total number of phish on the blacklist divided by the total phish that were alive. URLs that were taken down at each hour were excluded in the calculation. Total phish at hour 0 was 101. 4.3.3 False Positives We compiled a list of 13,458 legitimate URLs to test for false positives. The URLs were compiled from three sources, detailed below. A total of 2,464 URLs were compiled by selecting the login pages of sites using google’s inurl function. Specifically, we used Google to search for pages where one of the following login-related strings appears in the URL: login, logon, signin, signon, login.asp. A script was used to visit each URL to determine if it was running and also whether it included a submission form. These pages were selected to see whether tools can distinguish phishing sites from the legitimate sites they commonly spoof. Ludl et al. also used this technique to gather their samples [147]. A total of 994 URLs were compiled by extracting 1000 emails reported to APWG on August 20, 2008. Out of the 1000 emails we scanned, we removed URLs that were down at the time of testing or URLs used in spam campaigns through a spam URL blacklist service uribl.com. This left us with 1076 URLs, which comprised a host of phish, malware, some spam and legitimate sites. We manually checked each of these URLs and removed phishing URLs, leaving 994 verified 87 non-phishing URLs. We ran the test for false positives within 24 hours after retrieval. The list was selected because it represented a source of phishing feeds that many blacklist vendors use, and thus we would expect to have more false positives than other sources. While spam messages may be unwanted by users, the URLs in these messages should not be classified as phishing URLs. Similarly, we compiled 10,000 URLs by extracting non-phishing URLS from the list of spam, phish, and malware URLS sent to UAB’s spam data mine on December 1-15, 2008. We tested these URLs within one week of retrieval. Again, this represents a source of phishing feeds that blacklist vendors would likely receive, and thus we would expect this source to have more false positives than other sources. We did not find a single instance of mislabeling legitimate login sites with phish. Among the 1,012 URLs from APWG, there was one instance where a malware website was labeled as a phish by the Firefox blacklist. Finally we did not find any false positives in the 10,000 URLs from the UAB spam data mine. Compared with previous studies [147], our study tested an order of magnitude more legitimate URLs for false positives, yet our findings on false positives are the same: phishing blacklists have close to zero false positives. Our results differ from a 2007 HP research study [88] in which the author obtained the Google blacklist and checked each entry to see if it was a false positive. This study reports that the Google blacklist contains 2.62% false positives. However, the methodology for verifying false positives is not fully explained and the list of false positives is not included in the report. In our test of false positives, we manually verified each URL labelled as phish and double-checked it with one of the known repositories of phish on the Internet. It is also possible that Google changed their techniques or sources for phishing URLs since 2007. For future work, we would like to verify the Google blacklist using the same method used in the HP study [88]. However, Google’s blacklist is no longer publicly available. 88 Table 4.3 Accuracy and false positives of heuristics Detected Detected false posiby by tives blacklist heuristics at hour 0 IE7 - Oct 08 23% 41% 0.00% Symantec - Oct 08 21% 73 % 0.00% IE7 - Dec 08 15% 25% 0.00% Symantec - Dec 08 14% 80% 0.00% 4.3.4 Accuracy of Heuristics Heuristics are used in Symantec’s Norton 360 toolbar and Internet Explorer 7. In this section, we report on their performance. We found that tools that use heuristics were able to detect significantly more phish than those that use only blacklists. At hour 0, Symantec’s heuristics detected 70% of phish, while Internet explorer 7’s heuristics caught 41% of phish. This is two to three times the amount of phish caught by the blacklists in that period. Furthermore, the heuristics triggered no false positives for the 13,458 URLs we tested. Table 4.3 summarizes these results. We also found that IE 7 and Symantec use heuristics somewhat differently. Both tools display a transient and less severe warning for possible phish detected by heuristics. However, Symantec’s toolbar introduced a feedback loop. When a user visits a possible phish which is detected by heuristics and is not on the blacklist then the URL is sent to Symantec for human review [5]. In our test, 95% of the phish detected by Symantec heuristics appeared on the Symantec blacklist at hour 1, while none of the phish detected by IE7 heuristics appeared on the IE blacklist at hour 1. This feedback loop is important at the user interface level. If a phish is detected by heuristics, toolbars display less severe, passive warnings to avoid potential liability. However, once the phish is verified as a phishing site by human, toolbars can block the content of the web page completely (active warnings). A recent laboratory study [29] showed that users only heed active phishing warnings and ignore passive warnings. 89 Figure 4.5 Protection rate for the October run of 91 phishing URLs. Protection rate is defined as total number of phish caught by blacklist or heuristic plus phish taken down divided by the total number of phish. 4.3.5 Total Protection Finally, we consider protection offered to users by phishing toolbars. We define protection rate as: phish on blacklist + detected by heuristics + taken down T otal phish Figures 4.5 and 4.6 present our findings. We found that at hour 0, tools that use heuristics to complement blacklists offered much better protection than tools that use only blacklists. By hour 48 a large fraction of phishing sites are taken down, and the tools we tested detected most of thelive phishing sites. In the December test we found that by hour 48 most tools offered nearperfect protection. 90 Figure 4.6 Protection Rate for the December run of 101 phishing URLs. Protection rate is defined as total number of phish caught by blacklist or heuristic plus phish taken down divided by the total number of phish. 91 4.4 Discussion 4.4.1 Limitations There are a few limitations to our study. First, all of our URLs came from a single anti-spam vendor, therefore the URLs received may not be representative of all phish. Second, all the URLs were detected by a spam vendor and presumably never reached users protected by that vendor. However, as not all users are protected by commercial spam filters, it is important that browsers also detect these phishing URLs. Second, these URLs were extracted only from email and did not include other attack vectors such as Internet messenger phishing. 4.4.2 Opportunities for Defenders The window of opportunity for defenders can be defined as the length of the phishing campaign plus the time lapse between the time a user receives a phishing email and the time the user opens the email. Users are protected if they either do not receive any phish or if, by the time they click on a phish, the website is blocked by browsers or taken down. As shown in Section 4.1, 44% of phishing campaigns lasted less than 2 hours. Recent research shows that, for a non-negligible portion of the Internet population, the time between when a user receives and opens a phishing email is less than two hours. For example, Kumaraguru et al. sent simulated phishing emails to students and staff at a U.S. University and educated them once they clicked on the link in the email. They found that 2 hours after the phishing emails were sent, at least half the people who would eventually click on the phishing link had already done so; after 8 hours, nearly everyone (90%) who would click had already done so [66]. Their study also found that people with technical skills were equally likely to fall for phish than their non-technical counterparts. In a recent national survey, AOL asked 4,000 email users aged 13 and older about their email usage. The survey found that 20% of respondents check their email more than 10 times a day, and 51% check their email four or more times a day (up from 45% in 2007) [9]. Assuming that those who check their emails do so at a uniform rate, 20% of people check their emails once 92 Figure 4.7 High-level view of sources of URLs for phishing blacklists. Potential phishing URLs can be collected from (1) URLs extracted from spam and phishing filters at mail exchange gateways, (2) URLs extracted from user reports of phishing email, (3) phishing websites identified by heuristics, and finally (4) user reports of phishing websites. every hour and half, and 51% check their email once every four hours4. These findings suggest that the critical window of opportunity for defense is between the start of a phishing campaign and 2 to 4 hours later. Our findings have several implications for phishing countermeasures. First, anti-phishing efforts should be more focussed on upstream protections such as blocking phish at the email gateway level. At the browser level, this effort should be focused on updating the blacklist more quickly or making better use of heuristic detection. Secondly, more research and industry development efforts to effectively educate users (eg. [68, 127]) and to design trusted user interfaces (eg. [22, 116, 145, 146]) are needed to overcome the initial limited blacklist coverage problem. 4 Assuming eight hour sleep time. 93 4.4.3 Improving blacklists The first step to improving blacklists is earlier detection of more phishing URLs. As shown in Figure4.7, potential phishing URLs can be gathered from URLs extracted from spam and phishing filters at e-mail gateways, URLs extracted from users’ reports of phishing emails or websites, and phishing websites identified by toolbar heuristics (Figure4.7). Each of these sources have different coverage. We first discuss ways to improve each source. E-mail gateway filters are the first point of contact with phishing emails. Given the limited window of opportunity for defenders, as discussed in section 4.1, vendors should focus their gathering efforts here. However, regular spam filters are not sufficient as they contain a lot of spam that would require much human effort to filter. To improve detection of phish at this level, we recommend using spam filters as the first line of defense, and then applying heuristics developed to detect phishing websites as a second layer. Once a suspicious URL is marked by both sources, it should be submitted for human review. As residential email accounts and business email accounts receive a different distribution of emails, to get the widest coverage vendors should collect URLs from a variety of sources. User reports of phishing emails and websites are likely to contain phish that spam filters missed. Therefore user reports should be use to complement email gateway spam filter data. However, users may lack incentives to report and verify phish. User incentives (e.g. points, prizes) may help overcome this problem. Finally, we recommend browser anti-phishing tools use heuristics to improve their blacklists. This method is analogous to early warning systems for disease outbreaks. When a user visits a possible phish that is detected by heuristics and is not on the blacklist, the tool can send the URL for human review and adds the URL to the blacklist once verified. This system would be likely to succeed based on the fact that some users check their email much more frequent than others [9]. 94 4.4.4 Use of heuristics As shown in Section 4.4 and 4.5, the two tools using heuristics to complement blacklists caught significantly more phish initially than those using only blacklists. Given the short length of phishing campaigns, there is great value in using heuristics. However, vendors may be concerned about the greater possibility of false positives when using heuristics and potential liability for mislabeling websites. In a court case in 2005, Associated Bank-Corp sued Earthlink after the Earthlink anti-phishing software ScamBlocker blocked the bank’s legitimate page [11]. Earthlink was able to fend off the suit on the basis that it was using a blacklist of phish provided by a third party, thus it cannot be held liable as a publisher when that information is erroneous under a provision in the Communication Decency Act. However, if a toolbar uses heuristics to detect and block a phish that turns out to be a false positive, the toolbar vendor may be regarded as “a publisher” under CDA, and thus not immunized. In our testing, we did not detect any false positives triggered by either the blacklists or heuristics. However, it is the potential of false positives that worries vendors. To overcome this liability issue, we recommend vendors first use heuristics to detect phish and then have experts verify them. We also encourage more discussion about the liability associated with providing phishing blacklists and heuristics. So far, there has been no test case on this matter. Lack of clarity on these matters could further reduce vendors’ incentives to apply heuristics. Major vendors such as Microsoft or Firefox, which offer protection to the majority of users, do not lose money directly from phishing. However, if they implement heuristics and get sued, they could potentially lose millions of dollars in restitution and legal fees. 95 Chapter 5 Anti-Phishing Phil: A Case study in User education This chapter is joint work with Alessandro Acquisti, Lorrie Cranor, Jason Hong, and Ponnurangam Kumaraguru. An earlier version of the content in this chapter was published at SOUPS 2007 [127] . 5.1 Introduction Phishing is a kind of attack in which criminals use spoofed emails and fraudulent web sites to trick people into giving up personal information. Victims perceive these emails as associated with a trusted brand, while in reality they are the work of con artists interested in identity theft [57]. These increasingly sophisticated attacks not only spoof email and web sites, but they can also spoof parts of a user’s web browser [55]. Phishing is part of a larger class of attacks known as semantic attacks. Rather than taking advantage of system vulnerabilities, semantic attacks take advantage of the way humans interact with computers or interpret messages [123], exploiting differences between the system model and the user model [139]. In the phishing case, attacks exploit the fact that users tend to trust email messages and web sites based on superficial cues that actually provide little or no meaningful trust information [26, 55]. Automated systems can be used to identify some fraudulent email and web sites. However, these systems are not completely accurate in detecting phishing attacks. In a recent study, only one of the ten anti-phishing tools tested was able to correctly identify over 90% of phishing web sites, and that tool also incorrectly identified 42% of legitimate web sites as fraudulent [147]. It 96 Figure 5.1 Anti-Phishing Phil game screen. Phil, the small fish near the top of the screen, is asked to examine the URL next to the worm he is about to eat and determine whether it is associated with a legitimate web site or a phishing site. Phils father (lower right corner) offers some advice. The game is available at: http://cups.cs.cmu.edu/antiphishing_phil/ is also unlikely that any system will ever be completely accurate in detecting phishing attacks, especially when detection requires knowledge of contextual information. While it makes sense to use automated detection systems as one line of defense against semantic attacks, our philosophy is that there will still remain many kinds of trust decisions that users must make on their own, usually with limited or no assistance. The goal of our research is not to make trust decisions for users, but rather to develop a complementary approach to supportusers so that they can make better trust decisions. More specifically, one goal of our research is to find effective ways to train people to identify and avoid phishing web sites. In this paper we present the design, implementation, and evaluation of Anti-Phishing Phil, a game we developed to teach people how to protect themselves from phishing attacks. AntiPhishing Phil teaches people how to identify phishing URLs, where to look for cues in web browsers, and how to use search engines to find legitimate sites. In Section 2, we present background information and related work on why people fall for phishing, and approaches to protecting 97 them. In Section 3, we describe the design of Anti-Phishing Phil, and present the ways in which we applied learning principles in designing the game. In Section 4, we present the methodology we used to evaluate the game. In Section 5, we present the results of our evaluation, which shows that the game is more effective than a tutorial we created or existing online training materials at teaching people to identify phishing web sites accurately. We discuss the effect of anti-phishing training in Section 6. Finally, we present our conclusions in Section 7. 5.2 Background and Related Work In this section, we present background on anti-phishing research, why people fall for phishing, and approaches to protecting people from falling for phishing attacks. Previous work on phishing falls into three categories: studies to understand why people fall for phishing attacks, tools to protect people from such attacks, and methods for training people not to fall for phishing attacks. 5.2.1 Why people fall for phishing Downs et al have described the results of an interview and role-playing study aimed at understanding why people fall for phishing emails and what cues they look for to avoid such attacks. There were two key findings in their work. First, while some people are aware of phishing, they do not link that awareness to their own vulnerability or to strategies for identifying phishing attacks. Second, while people can protect themselves from familiar risks, people tend to have difficulties generalizing what they know to unfamiliar risks [26]. Dhamija et al showed twenty-two participants twenty web sites and asked them to determine which were fraudulent. Participants made mistakes on the test set 40% of the time. The authors noted that 23% of their participants ignored all cues in the browser address bar and status bar as well as all security indicators [23]. This study did not present users with the email messages that might lead users to visit the web sites presented, so it provides no data on whether users pay attention to, or how they interpret, email cues. Wu et al. studied three simulated anti-phishing toolbars to determine how effective they were at preventing users from visiting web sites the toolbars had determined to be fraudulent. They 98 found that many study participants ignored the passive toolbar security indicators and instead used the site’s content to decide whether or not it was a scam. In some cases participants did not notice warning signals, and in other cases they noticed them but assumed the warnings were invalid. In a follow-up study, the authors tested anti-phishing toolbars that produced pop-up warnings that blocked access to fraudulent web sites until overridden by the user. These pop-up warnings reduced the rate at which users fell for fraudulent sites, but did not completely prevent all users from falling for these sites. The authors concluded that Internet users are not very good at interpreting security warnings and are unfamiliar with common phishing attacks, and recommended educating users about online safety practices [140]. Our work builds on these previous studies. We incorporated many of the lessons learned from this past work into our game. For example, we teach people not to trust the content of the web page but examine the URL instead. Our evaluation methodology is also closely based on Dhamija et al.’s work [23]. 5.2.2 Tools to protect people from phishing Anti-phishing services are now provided by Internet service providers, built into mail servers and clients, and available as web browser toolbars. However, these services and tools do not effectively protect against all phishing attacks, as attackers and tool developers are engaged in a continuous arms race [147]. Furthermore, Internet users who are unaware of the phishing threat will be unlikely to install and use an anti-phishing tool, and may ignore warnings from anti-phishing tools provided by their ISPs or built into their web browsers. Even users who understand anti-phishing warnings may ignore them [140]. Where possible, anti-phishing tools should be applied, but—as noted in the introduction—there will always be cases where people have to make trust decisions on their own. Other research has focused on the development of tools to help users determine when they are interacting with a trusted site. Ye et al. [145] and Dhamija and Tygar [22] have developed prototype “trusted paths” for the Mozilla web browser that are designed to assist users in verifying that their browser has made a secure connection to a trusted site. Herzberg and Gbara have developed 99 TrustBar, a browser add-on that uses logos and warnings to help users distinguish trusted and untrusted web sites [51]. Other tools, such as PassPet and WebWallet, try to engage users by requiring them to interact actively with the tool before giving out sensitive information [139], [141], [140]. However, even these solutions ultimately rely on the user’s ability to make the right decision. In addition, these approaches require either end-users, web servers, or both to install special software. In contrast, our training method only relies on teaching people what cues to look for in existing web browsers. 5.2.3 Anti-phishing education Despite claims by security and usability experts that user education about security does not work [31], there is evidence that well designed user security education can be effective [68]. Webbased training materials, contextual training, and embedded training have all been shown to improve users’ ability to avoid phishing attacks. A number of organizations have developed online training materials to educate users about phishing [28], [32]. In a previous study, we tested the effectiveness of some of these online materials and found that, while these materials could be improved, they are surprisingly effective when users actually read them [70]. Several studies have adopted a contextual training approach in which users are sent simulated phishing emails by the experimenters to test users’ vulnerability to phishing attacks. At the end of the study, users are given materials that inform them about phishing attacks. This approach has been used in studies involving Indiana University students [53], West Point cadets [33], and New York State employees [104]. In the New York State study, employees who were sent the simulated phishing emails and follow-up notification were better able to avoid subsequent phishing attacks than those who were given a pamphlet containing information on how to combat phishing. A related approach, called embedded training, teaches users about phishing during their regular use of email. In a previous laboratory experiment to evaluate our prototype embedded training system, we asked our participants to role play and respond to the messages in an email inbox that included two training emails designed to look like phishing emails. If a participant clicked on a 100 link in a training email, we immediately presented an intervention designed to train them not to fall for phishing attacks. We created several intervention designs based on learning sciences, and found that our interventions were more effective than standard security notices that companies email to their customers [68]. We designed our anti-phishing game to complement the embedded training approach, which trains people while they are performing their primary task (checking email). If users are interested in devoting some additional time to learning more about phishing, they can play the Anti-Phishing Phil game. The embedded training approach trains users to identify phishing emails, while the game teaches users to identify phishing web sites. The game emphasizes that phishing web sites often can be identified by looking at their URLs, and teaches users about the various parts of a URL. This training may also help users analyze URLs in suspicious email messages. 5.3 Design of Anti-phishing Phil In this section we present: the objectives of the game; learning science principles that we applied in designing the game; the story, mechanics, and technology of the game; and results from some of the pilot studies that we conducted, as we iterated on the game design. We used an iterative design process to develop the game. Our early iterations made use of paper and Flash prototypes to explore various design alternatives. After a great deal of play-testing and feedback from our research group, both on the content of the game (what to teach) and the game design itself (presentation), we developed a working prototype that we tested with actual users. We then iterated on the design several more times based on user feedback and behavior, focusing on improving the game mechanics and messages. Finally, we created a more polished look and feel using attractive images and enticing sounds. 5.3.1 Game Design Principles In this section, we present the objectives for the game and the learning science principles that we applied in implementing these objectives. 101 Our objective in developing the anti-phishing game was to teach users three things: (1) how to identify phishing URLs, (2) where to look for cues for trustworthy or untrustworthy sites in web browsers, and (3) how to use search engines to find legitimate sites. We believe that search engines can be an effective tool in identifying phishing web sites. For example, users can search for a brand name in a search engine and see whether the link that appears in the top search results is the same as a potentially suspicious link received in an email. By far, the top search engine results are legitimate web sites [148]. To achieve the above-mentioned objectives, we applied several learning science principles to the game design. Learning sciences theory suggests that training will be effective if the training methodology is goal-oriented, challenging, contextual, and interactive [110]. In goal-oriented training, learners have a specific goal to achieve and in the process of achieving the goal they are challenged and trained. Training is most effective if the materials are presented in a context users can relate to, and if the materials are presented in an interactive form. There also exists a large body of literature on the effectiveness of games for interactively teaching conceptual and procedural knowledge [44]. Conceptual knowledge is knowledge about concepts or relationships that can be expressed as propositions (e.g., URLs have a protocol part and a domain name part). In contrast, procedural knowledge (also referred as declarative knowledge) is the step-by-step knowledge that one uses to solve a given problem (e.g., check the URL in the address bar, and if it contains an IP addresses, you are likely visiting a phishing site) [3]. The Anti-Phishing Phil game conveys both conceptual and procedural knowledge. Research in learning science has established that interactive environments, in particular games, are one of the most effective training methods and are highly motivational for users, especially when they adhere to design principles for educational games [44], [110], [113]. We applied three learning science principles to the design of the Anti-Phishing Phil game: reflection, story-based agent, and conceptual–procedural. Reflection principle. Reflection is the process by which learners are made to stop and think about what they are learning. Studies have shown that learning increases if educational games include opportunities for learners to reflect on the new knowledge they have learned [18]. This principle is employed in our anti-phishing game by displaying, at the end of each round, a list of 102 web sites that appeared in that round and whether the user correctly or incorrectly identified each one (as shown in Figure 2). This helps users reflect on the knowledge gained from the round they just completed. Story-based agent environment principle. Agents are characters that help in guiding learners through the learning process. These characters can be represented visually or verbally and can be cartoon-like or real-life characters. The story-based agent environment principle states that using agents as part of story-based content enhances learning. We applied this principle in the game by having the user control a young fish named Phil, who has to learn anti-phishing skills to survive. People learn from stories because stories organize events in a meaningful framework and tend to stimulate the cognitive process of the reader [64], [83]. Studies have demonstrated that students in story-based agent conditions perform better in learning than in non-story-based agent conditions [79], [96]. Conceptual–Procedural principle.This principle states that conceptual knowledge and procedural knowledge influence one another in mutually supportive ways and build in an iterative process [60]. In the first version of our game, we taught users specific procedural tips such as “URLs with numbers in the front are generally scams,” or “a company name followed by a hyphen is generally a scam.” We did not teach any conceptual knowledge in the game. Users were able to remember the procedural tips, but without a full conceptual understanding of URLs. Hence, some users applied the lessons learned from the game incorrectly. For example, some users misapplied the rule about IP addresses and thought www4.usbank.com was a phishing site because the URL contained the number 4. Other users misapplied the rule “company name followed by hyphen usually means it is a scam” to web-da.us.citibank.com (a legitimate site). In the most recent version of our game, we added conceptual knowledge of URLs, explaining the different parts of an URL and which parts are the most important. We also applied this principle by providing information about how to search for a brand or domain and how to decide which of the search results are legitimate (procedural knowledge) after mentioning that search engines are a good method to identify phishing web sites (conceptual knowledge). In this way, we present conceptual and procedural knowledge iteratively. 103 5.3.2 Game Description Here, we describe our game in three parts: story, mechanics, and technology. Story The main character of the game is Phil, a young fish living in the Interweb Bay. Phil wants to eat worms so he can grow up to be a big fish, but has to be careful of phishers that try to trick him with fake worms (representing phishing attacks). Each worm is associated with a URL, and Phil’s job is to eat all the real worms (which have URLs of legitimate web sites) and reject all the bait (which have phishing URLs) before running out of time. The other character is Phil’s father, who is an experienced fish in the sea. He occasionally helps Phil out by giving Phil some tips on how to identify bad worms (and hence, phishing web sites). Mechanics The game is split into four rounds, each of which is two minutes long. In each round, Phil is presented with eight worms, each of which carries a URL that is shown when Phil moves near it (see Figure 1). The player uses a mouse to move Phil around the screen. The player uses designated keys to “eat” the real worms and “reject” the bait. Phil is rewarded with 100 points if he correctly eats a good worm or correctly rejects a bad one. He is slightly penalized for rejecting a good worm (false positive) by losing 10 seconds off the clock for that round. He is severely penalized if he eats a bad worm and is caught by phishers (false negative), losing one of his three lives. We developed this scoring scheme to match the real-world consequences of falling for phishing attacks, in that correctly identifying real and fake web sites is the best outcome, a false positive the second best, and a false negative the worst. The consequences of Phil’s actions are summarized in Table 2. There are four rounds in the game, each one harder than the previous and focusing on a different type of deceptive URL. Table 1 shows the focus of each round. Our implementation selects eight URLs from a pool of twenty for each round, including 12 URLs consistent with the round’s focus. The eight URLs illustrate concepts from other rounds to maintain continuity between rounds. To make the game more engaging and challenging, Phil has to avoid enemy fish while moving around the screen. If Phil comes in contact with an enemy, it eats him and he loses a life. Early 104 Table 5.1 This table shows the scoring scheme and consequences of the user’s actions (through Phil) Good worm Phil Eats Correct, points Round 1 2 3 4 Phishing worm gains 100 False negative, gets phished and loses life Phil False positive, loses Correct, Rejects 10 seconds gains 100 points Table 5.2 Focus of each round of the game with examples Focus Examples Bumper Sticker Message IP address http://147.46.236.55/PayPal/ “Don’t trust URLs with all URLS numbers in the front” Sub do- http://signin.ebay.com.ttps.us “Don’t be fooled by the word main ebay.com in there, this site beURLs longs to ttps.us.” Similar http://www.msn-verify.com/ “A company name followed and de- http://www.ebay-accept.com/ by a hyphen usually means it ceptive is a scam site” “Companies domains don’t use security related keywords in their domains” All pre- eBay sites combining all of above. vious methods together 105 versions of the game included several fast-moving enemies in each round. However, we found that players became distracted by the enemies and had trouble learning. We reduced the number of enemies to one and made them slower so that they did not interfere with learning in later versions of the game. Players have to correctly recognize at least six out of eight URLs within two minutes to move on to the next round. As long as they still have lives, they can repeat a round until they are able to recognize at least six URLs correctly. If a player loses all three lives the game is over. The game includes brief tutorials before each round and end of round summary, as shown in Figure 3. Technology The game is implemented in Flash 8. The content for the game, including URLs and training messages, are loaded from a separate data file at the start of the game. This provides us with a great deal of flexibility and makes it easy to quickly update the content. In each round of the game, four good worms and four phishing worms are randomly selected from the twenty URLs in the data file for that round. We also use sound and graphics to engage the user better. This includes sound effects to provide feedback on actions, background music, and underwater background scenes. 5.3.3 Training Messages In this section, we discuss details about the training messages that were shown to the users, and the presentation of these training messages. What to teach Our main focus is to teach users how to identify phishing URLs, where to look for cues in web browsers, and how to use search engines to find legitimate sites. To teach users to distinguish phishing URLs from legitimate ones, we first sampled a representative list of phishing URLs from the millersmiles.co.uk phishing archive [91], and organized them into three categories: IP-based phishing URLs, long URLs (with sub-domains), and similar and deceptive domains. Next we designed training messages for each type of URL. We iterated on these messages using the philosophy that they should be messages one could place on a bumper 106 Table 5.3 List of training messages in between rounds; these information helped users to perform better and connect these information with the information presented when they were playing the game. In Between # of Concepts How to do it? Round Tip printed pages Tip 1: Don’t 1 - Highlight and point to the forget about address bar in the browser. the URL. Tip 2: The 5 - Highlight the different parts - Look at the text between the Middle part of the URL (Prefix, address http:// and the first /.The of the URL and file name). text before the first / (this tells you the might be with a .com or .org) name of the is the main domain name. site. Tip 3: When 6 - A search engine is a useful - Type the domain name or in doubt, use tool to check the legitimacy of the organization name into a search ena web site. Google search engine. The gine! top result is usually legitimate website. Tip 4: Know 1 - Scammers register domains - Design and logos can be the enemies similar to real sites. spoofed. Links in the fraudtricks! - They copy logos and con- ulent website might take to letents from real sites to draw gitimate website. you attention. - They request sensitive information. - They point all links to real sites to deceive you. 107 Figure 5.2 An example training message in between rounds. In this message, Phils father (left) teaches Phil (right) how to identify different parts of a URL and which parts are important. sticker on a car. For example, for IP-based phishing URLs, we teach “Don’t trust URLs with all numbers in the front.” Table 1 shows a list of bumper sticker messages in the game. To teach users where to look for cues in the browsers, we created a tip that highlighted the browser’s address bar. To teach users how to use search engines to find legitimate sites, we originally used help messages from Phil’s father during the game play. However, as will be discussed in the next section, we found that this was not very effective, so we used a tutorial in between rounds instead. Where to teach them Training messages are embedded in the following places in the game: (1) feedback during the game, (2) help messages from Phil’s Father during the game, (3) end of the round score sheets, and (4) anti-phishing tips in between rounds. Feedback during the game: When Phil eats a good URL or rejects a phishing one, we provide some visual feedback such as “yummy” and “got ya” to tell Phil that he got it right. When he eats a phishing URL, he gets phished and is drawn upward by a fishing line and hook. At this point, Phil’s father provides a short tip explaining why the URL is a phishing URL. 108 Figure 5.3 “Round over screen”. This screen reviews the URLs shown in the round with an indication as to which ones the player identified correctly. The screen also shows a tip to figure out whether the URL is legitimate. This helps provide an opportunity for self-reflection. Messages from Phil’s Father’s: Phil can also ask his father for help at any time (by pressing T in the game). His father will provide hints as to what to look for to differentiate good worms from bad ones. Phil’s father will also occasionally use a “search engine” and tell Phil the results of the search based on the URL. This is to show Phil how to use a search engine properly to determine the legitimate domain name for a company. This also provides the information players need to determine whether to eat or reject a worm, even if they do not know what the legitimate domain name is for a particular company. In pilot tests of the game, we found that not many users used this option, suggesting that this may not be the most effective way to deliver training. End of round score sheets: We provide players with an opportunity to reflect on what they learned at the end of each round with a score sheet, as shown in Figure 2. This screen reviews the URLs used in that round, indicates whether or not the player identified each URL correctly, and displays a tip that explains how to figure out whether the URL is legitimate. In our pilot and user 109 study, we found that people often spent a great deal of time on this screen looking over the things they missed. This applies the reflection principle described in Section 3.1. In previous iterations of the game, we focused solely on teaching people how to discriminate between legitimate and phishing URLs. However, we observed that people needed more scaffolding to help them understand issues like what to look for in the web browser, and how specifically they could use search engines to find real sites. In our current iteration, we added several short tutorials between each round to teach them these kinds of topics. This applies conceptual-procedural principle described in Section 3.1. 5.3.4 Pilot Test We pilot tested our game with eight users recruited by posting flyers around the Carnegie Mellon University campus. We tested our participants’ ability to identify phishing web sites from a set of real and phishing web sites before and after playing the game. The study is a think aloud study where participants talked about strategies they use. The results were encouraging, but highlighted some areas where the game needed improvements. We found that the game was somewhat effective at teaching users to look at the URL in their browser’s address bar when evaluating a web site. Users looked at the address bar when evaluating 14% of the web sites before playing the game and 41% of the web sites after playing the game. The false negative rate decreased from 31% to 17% after users played the game. However, the false positive rate increased from 37% to 48%, in part due to users misinterpreting the URLs they examined. We observed that users learned some of the URL-related concepts we tried to teach, but not all of them. For example, most users seemed to understand that URLs that have all numbers in the front are usually a sign of scam. However, many users could not properly parse a long URL and did not seem to understand that the most important part of the URL is the right hand side of the domain name. This led them to mis-identify wellsfargo.com.wfcnet.net as a legitimate site and scgi.ebay.com and onlineast1.bankofamerica.com as phishing sites. 110 We also observed that some users applied the lessons learned from the game incorrectly. For example, some users misapplied the rule about IP addresses (in Table 1) and thought www4.usbank.com was a phishing site because the URL contained the number 4. Other users misapplied the rule “company name followed by hyphen usually means it is a scam” to web-da.us.citibank.com. Finally, many participants used wrong strategies to determine the web site legitimacy. For example, one common strategy consisted of checking whether the web site was designed professionally. However, this is not a useful strategy as many phishing sites are exact replicas of professionally designed legitimate sites. Although participants adopted this technique less frequently after the game, some of them still employed a variant of this strategy while using a search engine: they compared the two sites’ design (logos, colors) and the exact match of the URL to determine the legitimacy. We believe this is due to users not knowing exactly what to look for to determine web site legitimacy when they use search engines. To summarize, from the pilot test we observed that it is insufficient to teach users how to look for in the URL. We modified our game according to the lessons learned from the pilot testing. 5.3.5 Modified Game We realized that the initial version of the game focused almost entirely on procedural knowledge. However, some conceptual knowledge about the parts of a URL might have helped users avoid some of the mistakes they made. We added animated messages in between each round of the game to address some of the problems we observed in the pilot study. These messages teach users about the parts of URLs, how to use a search engine to check a suspicious URL, and common tricks used by scam web sites. We designed these messages in a story-like format, in which Phil’s father teaches him about URLs at home before he can explore Interweb Bay on his own. Table 3 presents the summary of the training messages that were provided to the user in between rounds, and Figure 4 gives a screenshot of one of the training messages. 111 5.4 Evaluation 1: Lab Study In this section, we describe the methodology we used to test the game for its effectiveness in training users in a laboratory study. 5.4.1 Study design We based the design of our user study on Dhamija et al.’s study, trying to recreate their experiment as much as possible (however, the original materials for Dhamija’s study have been lost) [23]. Participants were given the following scenario: “You have received an email message that asks you to click on one of its links. Imagine that you have clicked on the link to see if it is a legitimate web site or a spoofed web site.” We then presented participants with ten web sites and asked them to state whether a web site was legitimate or phishing, and to tell us how confident they were in their judgments (on a scale of 1 to 5, where 1 means not confident at all, and 5 means very confident). After evaluating the ten URLs, participants were given fifteen minutes to complete an anti-phishing training task. Finally, participants were shown ten more web sites to evaluate. After finishing this evaluation, participants were asked to complete an exit survey. We selected twenty web sites (shown in Table 5) to test our participants’ ability to identify phishing web sites before and after training. Ten of the sites we selected were phishing sites from popular brands. The other ten were legitimate web sites from popular financial institutions and online merchants, as well as random web sites. We divided the twenty web sites into two groups (A and B), with five phishing sites and five legitimate sites in each group. We randomized the order in which the two groups of URLs were presented so that half the participants saw group A first, and half saw group B first. We hosted the phishing web sites on the local computer by modifying the host DNS file. Thus, our participants were not actually at risk and we were able to show them phishing sites even after they had been taken down. We told participants that they could use any means to determine a web sites’ legitimacy other than calling the company. We also let participants use a separate web browser if they wanted, without prompting them about how or why this might be useful. Some participants used this other 112 web browser to access a search engine to help determine whether a web site was legitimate or not. We used Camtasia Studio to record our participants’ computer screens and spoken comments. We used a between-subjects experimental design to test three training conditions: 1. Existing training material condition:In this condition, participants were asked to spend fifteen minutes reading eBay’s tutorial on spoofed emails [28], Microsoft’s Security tutorial on Phishing [89], the Phishing E-card from the U.S. Federal Trade Commission [32], and a URL tutorial from the MySecureCyberspace portal [97]. We reused the training material condition from our previous study as a control group [70]. 2. Tutorial condition: In this condition, participants were asked to spend up to fifteen minutes reading an anti-phishing tutorial we created based on the Anti-Phishing Phil game. We include this condition to test the effectiveness of the training messages separate from the game. The tutorial included printouts of all of the between-round training messages. It also included lists of the URLs used in the game with explanations about which were legitimate and which were phishing, similar to the game’s end-of-round screens. The 17-page tutorial was printed in color. We designed the tutorial to resemble the game as closely as possible. 3. Game condition: In this condition, participants played the Anti-Phishing Phil game for fifteen minutes. This study was conducted in two phases, separated by five months. For the existing training materials condition, we used data collected during a previous study in September 2006 that measured participants’ improvements after reading existing training materials, as compared with a control group that spent fifteen minutes playing solitaire [70]. For the tutorial and game conditions participants were recruited in February 2007 and randomly assigned to these groups. The same procedures were used in September and February for recruiting, screening, and conducting the experiments, although it is possible that the five month delay between the two phases of the experiment introduced some selection bias.. 113 Table 5.4 Participant demographics in each condition hello Existing Tutorial Game Training Group Group Material Gender Male 29% 36% 50% Female 71% 64% 50% Age 18-34 93% 100% 100% >34 Education High School College Undergrad College graduate Post. Graduate school Years on the Internet 3- 5 years 6-10 years > 11 years 5.4.2 7% 0% 0% 14% 50% 7% 78% 7% 50% 14% 21% 7% 7% 21% 21% 23% 69% 8% 23% 70% 7% 14% 78% 7% Participant Recruitment and Demographics In this section, we present the process that we used in recruiting participants for the study; we also describe the demographics of the participants. We recruited fourteen people for each condition via flyers posted around campus, and with recruitment email on university bulletin boards, and on craigslist.com. We screened participants with respect to their knowledge of computers in general, aiming to recruit only participants who could be considered “non-experts.” We recruited users who answered “no” to two or more of the following screening questions: 1) whether they had ever changed preferences or settings in their web browser, 2) whether they had ever created a web page, and 3) whether they had ever helped someone fix a computer problem. These questions have served as good filters to recruit non-experts in other phishing-related studies [26], [68]. A summary of demographics is shown in Table 4. 114 5.4.3 Results In this section, we present the results from the user study. We found that participants in the game condition performed better than the other two conditions in correctly identifying the web sites. We also found that there was no significant difference in false negatives among the three groups. However, the participants in the game group performed better overall than the other two groups. 5.4.3.1 Correlation between Demographics and Susceptibility to Phishing In this section we present results regarding the correlation between demographics and susceptibility to phishing, user performance, user confidence rating, user feedback, and places where game can be improved. We found no significant correlation between the participants’ performance (measured by total correctnesss) and gender (rho = -0.2, n = 42, p = 0.19), age (spearman rho = 0.008, n = 42, p = 0.96), education (spearman rho = 0.06, n = 42, p = 0.708), race (spearman rho = 0.13, n = 42, p = 0.406), number of hours spent online per week (rho = -0.10, n = 42, p =0.588). Other studies have also found no correlation between these demographics and susceptibility to phishing [23], [140]. The score is positively correlated with years on the Internet (rho = 0.341, n = 42, p = 0.031). 5.4.3.2 User Performance We measured the effectiveness of user training by examining false positives and false negatives as well as the total percentage of correct sites identified before and after the test. A false positive is when a legitimate site is mistakenly judged as a phishing site. A false negative is when a phishing site is incorrectly judged to be a legitimate site. Our game condition performed best overall. It performed roughly as well as the existing training material condition in terms of false negatives, and better on false positives. The tutorial condition also performed better than the existing training material in terms of false positives and total correctness. However, these latter results were not statistically significant. 115 Table 5.5 Percentage of total correct answers for the training group before and after the game Website Real Description Pre Post or Game % Game Fake Correct %correct (average (average conficonfidence) dence) Paypal Real Paypal login page 83 (4.6) 100 (4.7) Bank of Real Bank of America home page; URL: onlin- 66 (3.5) 100 (4.3) America east.bankofamerica.com Wellsfargo Spoof Faked Wellsfargo home page; layered informa- 83 (3.6) 87 (4.5) bank tion request; sub domain deception with URL online.wellsfargo.wfosec.net Citibank Real Citibank login Page; URL: web-da.us.citibank.com 83 (3.6) 75 (4.5) Barclays Spoof Faked Barclays login page; layered information re- 83 (4.2) 100 (4.7) quest; IP address URL AOL Spoof AOL account update, deceptive domain myaol.com 100 (3.3) 75 (3.4) Etrade Real Etrade home page 100 (4.0) 100 (4.3) PNC Bank Spoof Bank account update; pop-up window over the real 66 (4.0) 50 (5.0) PNC Bank web site; security lock; requesting credit card number eBay Real eBay register page; requesting lots of information 66 (4.2) 62 (4.0) Halifax Spoof Halifax bank login page; deceptive domain halifax- 83 (2.8) 100 (4.5) Bank cnline.co.uk. Card Finan- Real Card Financial Online (part of MBNA); domain 50 (3.5) 66 (4.5) cials Online name has nothing to do with MBNA. Citicards Spoof Citicard account update; lock on the page; requesting 50 (4.0) 100 (4.6) a lot of information Chase Real Online banking login page; URL: chaseon- 100 (4.2) 100 (4.1) online line.chase.com Desjardins Real Account login page; unfamiliar foreign bank 50 (3.0) 83 (3.8) Royal Bank Spoof Sign in online banking page; layered information re- 37 (4.0) 100 (4.1) of Canada quest; URL has no resemblance with the bank. Chase Stu- Real Primitive looking page with few graphics and links 37 (3.0) 66 (3.7) dent HSBC Spoof Internet banking login page; layered information re- 50 (4.0) 100 (5.0) quest; IP address URL US Bank Real Online banking login page; URL: 75 (3.5) 100 (4.6) www4.usbank.com eBay Spoof Faked eBay login page; IP address URL 75 (3.8) 100 (5.0) PayPal Spoof Fake URL bar displaying the real Paypal URL; not 50 (3.2) 0 (4.0) requesting much information 116 Figure 5.4 User performance in the experimental conditions: existing training materials, tutorial only, game, and control condition. N=14 in all conditions. The graph on the left shows false negative rates.The existing training material performed best on false negatives. However, the difference is not statistically significant. The graph on the right shows False Positive Rate. The false positives increased in the existing materials condition, and decreased in both the tutorial and game condition, with the game condition showing the highest reduction. Post test false negative rates in all three groups decreased significantly from the pre test values. For the existing training materials condition, the false negative rate fell from 0.38 to 0.12 (paired t-test: 1=0.38, 2=0.12, p = 0.01); for the tutorial condition, it changed from 0.43 to 0.19 (paired t-test: 1=0.43, 2=0.19, p < 0.03); for the game condition, it changed from 0.34 to 0.17 (paired ttest: 1=0.34, 2=0.17, p <0.02). There is no statistical difference between the three groups in either the pre test (oneway ANOVA, F(2,41)=0.52, p=0.60), or post test (oneway ANOVA, F(2,41)=0.81, p=0.45). These results are shown in Figure 5. Post test false positive rates decreased significantly in the game condition (paired t-test: 1=0.30, 2=0.14, p < 0.03). The one-way ANOVA revealed that false positive rates differed significantly in 117 the post test (F(2, 41) = 4.64, p < .02). The Tukey post-hoc test revealed that the game condition has significantly lower false positives than the existing training materials. No other specific posthoc contrasts were significant. The results are shown in Figure 6. Combining false positive and false negatives we derived a measure for the total correctness. We found in the post test that the game condition performed better than the existing training material condition (2 sample t test, p<0.02). We did not find the tutorial condition improvement to be significant over the existing training material condition; however, this is likely due to our small sample size. These results are shown in Figure 7. False negative rates. N = 14 in all conditions. The existing training material performed best on false negatives. However, the difference is not statistically significant. False Positive Rate. N = 14 in all conditions. The false positives increased in the existing materials condition, and decreased in both the tutorial and game condition, with the game condition showing the highest reduction. Total correctness for the test groups. N = 14 in all conditions. The game condition shows the greatest improvements. 5.4.3.3 User Confidence Rating Users became more confident about their judgments after the game or the tutorial conditions. We did not observe the existing training material improving user confidence in a statistically significant way. The average user confidence rating in the game condition increased from 3.72 (variance = 0.09) to 4.42 (variance = 0.10). This change is statistical significant (paired t –test, p < 0.001). In contrast, user confidence in the existing training material condition did not improve in a statistically significant way: the average confidence rating was 4.18 pre test (variance = 0.18) and 4.32 post test (variance = 0.15). 118 5.4.3.4 User Feedback In the post test, we asked participants to measure on a 5- point Likert scale how much they felt they had learnt and how important was the information they learnt. Ninety-three percent of the users either agreed or strongly agreed that they had learned a lot (u = 4.21, std = 0.58), and 100% of them agreed or strongly agreed that they had learned a lot of important information (u = 4.36 std=0.50). On a five point scale, we also asked them to rate the educational and fun levels of the game. Ninety-three percent of the user felt the educational value of the game was very good or excellent (u=4.28, var = 0.61). Fifty percent of the users considered the fun level of the game as very good or excellent (u = 3.7 var = 0.78). We asked similar questions about educational value and fun level in the existing training material condition. Ninety-three percent of the users also felt the educational value of the existing training material was very good or excellent (u=4.28 var = 0.59), where as only twenty-nine percent of the users considered the fun level of the existing training materials to be very good or excellent (u = 2.8 var = 1.36). 5.4.3.5 Where the Game is Failing We found that users in the game group and the tutorial group performed worse when examining two websites. The first website is a fake address bar attack, where we showed users a Paypal website with the address bar spoofed. Six of the users in the game condition were unable to identify this attack in the post test, whereas only three users in the existing training material condition fell for it. We hypothesize that users are more prone to this kind of attacks because, after the training, they look specifically for clues in the URL, and if the clues confirm their belief, they do not look further. (Luckily, current browsers now address this kind of vulnerability.) Two users also fell for the “similar domain attack” after the game condition, in which we showed them myaol.com for account updates. This is an easy attack to identify if users notice the large amount of information requested, because of this reason, none of the users fall for it in the pre test. This problem highlights two lessons: first, some users still have problems with phishing 119 domains that are similar to the real ones; second, they tend to look less for other clues other than the URL, and if the URL does not raise suspicion, they do not look further. 5.4.3.6 Effect of Training Security education plays an important role in increasing users’ alertness towards security threats. Alert users are cautious, and less likely to make mistakes that will leave them vulnerable to attack (false negatives). However, cautious users tend to misjudge non-threats as threats (false positives) unless they have learned how to distinguish between the two. Thus, good user security education should not only increase users’ alertness, but also teach them how to distinguish threats from nonthreats. In this section we use signal detection theory (SDT) [78, 120] to quantify the ability to discern between signal (phishing websites) and non-signal or noise (legitimate websites). We use two measures: sensitivity (d’) and criterion (C). In our user studies, we define sensitivity to be the ability to distinguish phishing websites from legitimate websites, which is measured by the distance between the mean of signal and non-signal distributions. The larger the value of d’, the better the user is at separating signal from noise. Criterion is defined as the tendency of users towards caution when making a decision. More cautious users are more likely to have few false negatives and many false positives, while less cautious users are likely to have many false negatives and few false positives. Figure 5.5 shows example distributions of user decisions about legitimate and phishing websites. The criterion line divides the graph into four sections representing true positives, true negatives, false positives, and false negatives. Training may cause users to become more cautions, increasing C and moving the criterion line to the right. Alternatively, training may cause users to become more sensitive, separating the two means. In some cases training may result in both increased caution and increased sensitivity or in decreased caution but increased sensitivity. We calculated C and d’ for the participants in our user study, Table 5.6 presents the results. We found that in the existing training material condition, the sensitivity increases from 0.81 in pre test to 1.43 in post test. This increase is significant (p < 0.05). We also found that users became cautious after the training, as the d’ changes from 0.03 in pre test to -0.51 in post test (p <0.025). This result (users becoming more cautious) was also shown by Jackson et. al [2]. In contrast 120 Figure 5.5 Applying signal detection theory (SDT) to anti-phishing education We treat legitimate websites as “non signal,” and phishing websites as “signal.” Sensitivity (d’) measures users’ ability to distinguish signal from non-signal. Criterion (C) measures users’ decision tendency (C < 0 indicates cautious users , C = 0 indicates neutral users, C > 0 indicates liberal users). As a result of training users may a) become more cautious, increasing C; b) become more sensitive, increasing d’; or c) a combination of both. to the existing training material condition, the sensitivity increased from 0.93 to d’ post = 2.02 (p<0.025) in the game condition. Also, the decision criterion did not change significantly (C pre = 0.06, C post = 0.06) in the game condition. This shows that the improvement in the performance is due to learned ability to better distinguish phishing websites and real websites. Table 5.6 Results from the Signal Detection Theory analysis. This shows that users had a greater sensitivity with Anti-Phishing Phil, meaning that they were better able to distinguish between phishing and legitimate sites. Consequently, users were able to make better decisions in the game condition compared to the users becoming conservative in the other condition. Sensitivity (d’) Criterion (C) Pre test Post test Delta Pre test Post test Delta Existing train- 0.81 1.43 0.62 0.03 -0.51 ing materials * 0.54 ** Anti-phishing 0.93 2.02 1.09 0.06 0.06 0 Phil ** * p <0.05, ** p < 0.025 121 5.5 Evaluation 2: Anti-Phishing Phil Field Study In this section, we discuss new results from data we collected in a real-world deployment of Anti-Phishing Phil. Our results provide more evidence that Anti-Phishing Phil is effective for knowledge acquistion and knowledge retention. 5.5.1 Study design We recruited participants for an online study through online mailing lists postings offering participants a chance to win a raffle for a $100 Amazon gift certificate. We used a between-subjects design to test two conditions. In the control condition, participants saw 12 websites and were asked to identify whether each website seen was phishing or not. After doing this, the participants were taken to the game. In the game condition, participants were shown six websites before playing the game (pre-test) and another six websites after they finished playing the game (immediate post-test). To measure retention, we emailed participants seven days later and asked them to take a similar test (delayed post-test). In total, we tested each participant in the game condition on 18 websites divided into three groups of three phishing websites and three legitimate websites. We randomized the order of websites within each group, and the order in which the groups were shown to each participant. 5.5.2 Participants Over the course of two weeks (Sep 25, 2007 to Oct 10, 2007), 4,517 people participated in the study. In the game condition, 2,021 users completed both pre-test and immediate post-test, 674 of whom also came back one week later for the delayed post-test. In our analysis we focus on people who completed pre-test, immediate post-test, and delayed post-test. We had 2,496 participants in the control condition. Among the total participants, there were 78% male, 15.6% female, and 6.4% did not give their gender; 4.8% were 13 - 17 years old, 43.7% were 18 - 34 years old, 44.3% were 35 - 64 years old, 0.5% were more than 65 years, and 6.8% did not provide their age. 122 Figure 5.6 False negative and false positive for Anti-Phishing Phil in the real-world. Novice users show greatest improvement in false negative and false positive 5.5.3 Results Our results demonstrate that users are able to more accurately and quickly distinguish phishing websites from legitimate websites after playing the game, and that they retain knowledge learned from the game for at least one week. We classified the game condition participants into three categories based on their pre-test scores: novice (0 - 2 correct), intermediate (3 - 4 correct) and expert (5 - 6 correct). As illustrated in Figure 5.6, novice users showed the greatest improvement, with false positive rate decreasing from 42% to 11.2% (paired t-test, p < 0.0001), and false negative rate decreasing from 28.3% to 11.2% (paired t-test, p < 0.0001). The intermediate group also showed statistically significant improvements, although not as large as the novice group. Finally, we did not observe any statistically significant improvements for the expert group. Delayed post-test scores did not decrease from immediate post-test scores; demonstrating that participants retained their knowledge after one week. Participants were able to determine website legitimacy more quickly after playing the game. The mean time users in the game group took to determine a webwebsite’s legitimacy before the game was 21.2 seconds. After the game, it decreased to 11.2 seconds (paired t-test, p < 0.0001). The mean scores for the control group does not change in a statistically significant way (pre - 18.5 seconds, post - 18.6 seconds). Those who did not come back for the delayed post-test performed slightly worse than those who did come back. Their immediate post-test score is 83.8% for those who did not come back 123 and 89.1% for those who did come back one week later (two sample t-test, p < 0.001). One possible explanation is that those who were more confident in their performance were more likely to come back. To validate this hypothesis, we conducted a Chi-square test of the percentage of novice, intermediate and expert users completed the immediate post-test, or delayed post-test. We found that there were more experts and fewer intermediate and novices in the delayed post-test group (p < 0.001). Before playing the game mean accuracy scores for males were significantly higher than for females (males = 75.5%, females = 64.4%, two sample t-test, t = 8.48, p < 0.0001). However, the two groups improved similarly after playing the game (two proportion test, 14.2% versus 12.4%, p = 0.192). There was also a significant difference in pre-test performance between different age groups (one way ANOVA F = 7.29, p < 0.01). A Turkey simultaneous 95% confidence interval test reveals that participants whose age is less than 18 performed worse than those who are between 18 and 64. There is no statistical difference in performance between the ages groups 18-35 and 36-64. We observed similar trends in immediate post-test performance (one way ANOVA, F = 23.05, p < 0.01). These results suggest that teenagers may be particularly susceptible to phishing attacks. The mean scores for the age group 13-17 years was 3.9 while the mean score was 4.6 for both 18-34 and 35-64 age groups. We used the data from the game to determine which types of URLs are most difficult for people to identify correctly. Especially challenging ones are the URLs longer than the address bar and deceptive URLs that look similar to legitimate URLs with some added text (e.g. http://www.msnverify.com/). The more challenging the URL, the more likely game players are to use the game’s help feature (r = -0.645, p < 0.001). From the game data, we found that users are most confused with long URLs. This confusions makes them susceptible to sub-domain attacks such as (https://citibusinessonline.da-us.citibahnk.com/cbusol/signon.do). Users are also confused with very similar URLs. For example, www.citicards.net (as opposed to www.citicards.com), www.eztrade.com (as opposed to www.etrade.com). This suggests for further investigation on ways to teach to remove these confusions among users. 124 5.5.3.1 Effect of Training Using the Signal detection method that we introduced in section 5.4.3.6. We calculated C and d’ for our evaluation of existing online training materials, PhishGuru retention and transfer study, Anti-Phishing Phil laboratory study, and Anti-Phishing Phil field study, as summarized in Table 5.7. We found that after reading existing training materials, users became significantly more cautious without becoming significantly more sensitive. Thus these materials serve to increase alertness, but do not teach users how to distinguish legitimate websites from fraudulent ones. After playing Anti-Phishing Phil, users became both significantly more sensitive and liberal, indicating that performance improvements from playing the game are due to learning. (Note, in the laboratory study we did not observe the Criterion change that we observed in the field study.) PhishGuru embedded training increased both sensitivity and caution, but these results are not statistically significant due to the small number of user decisions considered in the analysis. The pre-test Criterion for the existing training and Anti-Phishing Phil studies indicate these users started off more cautious than those in the PhishGuru study. This is likely due to the fact that users were primed to think about security in the former studies and not in the latter study. Table 5.7 Signal Detection Theory analysis. PhishGuru and Anti-Phishing Phil increased user’s sensitivity, while existing training materials made users more cautious. * indicates statistically significant differences (p <0.05). Sensitivity (d’) Criterion (C) Prepost- Delay Prepost- Delay test test test test Existing train- 0.81 1.43 – 0.03 – ing materials 0.51* Anti-Phishing 0.93 2.02* – 0.06 0.06 – Phil laboratory study Anti-Phishing 1.49 2.46* 2.47 -0.35 0.02* 0.0 Phil field study 125 Conclusions and Future Work In this paper, we presented the design and evaluation of Anti-Phishing Phil, a game that teaches users not to fall for phishing attacks. Our objective in developing the anti-phishing game was to teach users three things: (1) how to identify phishing URLs, (2) where to look for cues in web browsers, and (3) how to use search engines to find legitimate sites. In particular, the game teaches users about identifying three types of phishing URL’s: IP based URLs, sub domain, and deceptive. We conducted two user studies. In the first study, we compared the effectiveness of the game with existing online training materials and a tutorial we created based on the game. We found that participants who played the game performed better at identifying phishing websites than participants who completed the two other types of training. Using signal detection theory, we also showed that while existing online training materials increase awareness about phishing (which can help people avoid attacks), our game also makes users more knowledgeable about techniques they can use to identify phishing web sites. In the second study, we tested Phil from data we collected in a real-world deployment of AntiPhishing Phil. Our results provide more evidence that Anti-Phishing Phil is effective for knowledge acquistion and knowledge retention. Our results show that interactive games can be a promising way of teaching people about strategies to avoid falling for phishing attacks. Our results suggest that applying learning science principles to training materials can stimulate effective learning. Finally, our results strongly suggest that educating users about security can be a reality rather than just a myth [48]. 126 Chapter 6 Phishing Susceptibility Study This chapter is joint work with Mandy Holbrook, Julie Downs, Lorrie Cranor, and Ponnurangam Kumaraguru. An earlier version of the content in this chapter was submitted to CHI 2010 [125]. Phishing attacks, in which scammers send emails and other messages to con victims into providing their login credentials and personal information, snare millions of victims each year [43]. A variety of efforts aim to combat phishing through law enforcement, automated detection, and end-user education. Researchers have studied why people fall for phishing attacks; however, little research has been done to study demographic factors in susceptibility to phishing. By determining which groups are most susceptible to phishing, we can determine how best to focus anti-phishing education. In this paper, we present the results of our roleplay phishing study, administered to 1001 online survey respondents in order to study demographics and phishing susceptibility. The rest of the paper is organized as follows. In the next section, we present background and related work on why people fall for phishing. Then we describe the design of our experiment and present the results of our study, identifying several important demographic factors that affect phishing susceptibility and describing the effects of education in bridging these gaps. Finally we discuss the implications of our study for designing anti-phishing tools and improving public policy. 127 6.1 Background and related work Research has shown that people are vulnerable to phishing for several reasons. First, people tend to judge a website’s legitimacy by its “look and feel,” which attackers can easily replicate [23]. Second, many users do not understand or trust the security indicators in web browsers [140]. Third, although some consumers are aware of phishing, this awareness does not reduce their vulnerability or provide useful strategies for identifying phishing attacks [26]. Fourth, the perceived severity of the consequences of phishing does not predict users’ behavior [27]. 6.1.1 Demographics and Phishing Susceptibility To the best of our knowledge, there has been no study dedicated to understanding what demographic factors correlate with falling for phishing, and the effectiveness of educational interventions in bridging the demographic divide. We highlight here a few studies that have measured susceptibility to specific types of phishing attacks or have studied the effectiveness of anti-phishing education while reporting at least some data on gender and other demographic factors. Jagatic et al. performed a spear phishing experiment at Indiana University to quantify how reliable social context would increase the success of a phishing attack. They launched an actual (but harmless) phishing attack targeting college students aged 18–24 years old by using information harvested from social networking sites. In their study of 487 participants, female students fell for 77% of the spear phishing attacks, while male students fell for 65% [53]. In a related study, Kumaraguru et al. conducted a real-world phishing study with 515 participants to study the long-term retention of PhishGuru anti-phishing training [69]. They did not find significant differences based on gender, but did find that participants in the 18-25 age group were consistently more vulnerable to phishing attacks. They also did not explain the reason behind this finding. Finally, Kumaraguru et al. [71] conducted a study of 5182 Internet users measuring the effectiveness of Anti-Phishing Phil, an interactive game that teaches people not to fall for phish. They found that men were more likely to correctly distinguish phishing and legitimate websites than 128 women (75.5% correct vs. 64.4% correct). They collected only very coarse-grained information on the age of participants, but found that people under the age of 18 performed worse than those above 18. Although past studies have found differences in phishing susceptibility based on gender and age, they generally did not collect enough information about study participants to isolate these variables from other potentially confounding factors. In addition, previous studies did not address why these demographic factors correlate with falling for phishing. In our paper, we address these research questions. 6.1.2 Susceptibility vs. Risk Behavior The risk literature has shown reliable demographic differences in risk perceptions on various topics, with relatively oppressed groups (e.g., women, racial and ethnic minorities, and less wealthy people) perceiving more risk in the world around them [37, 129]. Such perceptions may be linked to these groups’ experiences of a riskier world, perhaps due to lower degrees of control over risky processes. Age has also been linked to risky behavior, with adolescents tending to engage in riskier behaviors on average, perhaps as a function of their ongoing learning about the world around them [25,114]. Because real-world risk behaviors are complex and subject to such varied predictors as knowledge, goals, and benefits associated with what is perceived to be risky behavior, there have been relatively few studies with the power to assess multiple mediators of demographic effects on risky behavior. The current paper takes a specific, well-defined behavior as a context in which to identify content-specific factors that may explain effects of age, gender, and ethnic background. 6.1.3 Security User Education Despite claims by some security and usability experts that user education about security does not work [48], there is evidence that well-designed user security education can be effective in the real world [67, 127]. Web-based training materials, contextual training, embedded training, and interactive games have all been shown to improve users’ ability to avoid phishing attacks. 129 A number of organizations have developed online training materials to educate users about phishing [28, 32]. In a previous study, Kumaraguru et al. tested the effectiveness of some of these online materials and found that, while these materials could be improved, they are surprisingly effective when users actually read them [70]. Several studies have adopted a contextual training approach in which users are sent simulated phishing emails by the experimenters to test users’ vulnerability to phishing attacks. At the end of the study, users are given materials that inform them about phishing attacks. This approach has been used in studies involving Indiana University students [53], West Point cadets [33], and New York State employees [104]. A related approach, called embedded training, teaches users about phishing during their regular use of email. This trainer sends phishing email to users and, if users click on phishing links, immediately presents an intervention designed to train them not to fall for phishing attacks. Kumaraguru et al. created several intervention designs based on learning sciences, and found that these interventions were more effective than standard security notices that companies email to their customers [68]. The researchers continued to refine the most successful intervention, a comic strip featuring a character named PhishGuru. A follow-up study showed that people were able to retain what they learned from this training [69]. Finally, Sheng et al. designed Anti-Phishing Phil, an online game that teaches users good habits to help them avoid phishing attacks. The researchers used learning science principles to design and iteratively refine the game. Their evaluation showed that participants who played the game were better able to identify fraudulent web sites compared to participants in other conditions [127]. We studied the effectiveness of several of these educational approaches in bridging the demographic divide. The materials we tested included a set of popular web-based training materials, Anti-Phishing Phil, a PhishGuru cartoon, and the combination of Anti-Phishing Phil and a PhishGuru cartoon. 130 6.2 Study Design In this online study, participants provided demographic information, answered survey questions to assess their knowledge about phishing, and completed a roleplay task to assess their behavioral susceptibility to phishing, prior to receiving one of several possible forms of training. Participants then completed a second roleplay task to assess reductions in phishing susceptibility as well as any changes in participants’ tendencies to be suspicious of legitimate emails. Participants were assigned randomly to a control condition or one of four experimental conditions. The conditions varied based on the type of training participants were exposed to (or no training). The ordering of the survey questions relative to the initial roleplay was also counterbalanced. 6.2.1 Recruitment Participants were recruited through Amazon.com’s Mechanical Turk (mTurk), a marketplace for work requiring human intelligence. In this online environment, requesters post tasks known as HITs (Human Intelligence Tasks), and workers are paid for completing these HITs. We offered to pay participants four dollars for those that qualified and twenty cents to those who did not. In total, 1001 participants qualified and completed the entire study as detailed in Table ??. To disqualify people who were hoping to earn money for completing the study without actually paying attention to the study tasks, we asked all participants a series of questions about an email message that discussed an upcoming meeting. We used two of these questions, both of which could be answered correctly by a careful reading of the email, to screen out those participants who were not paying attention to the email content. We also asked basic demographic questions (such as occupation and age) so that participants would not be able to easily identify qualifying questions. 6.2.2 Roleplay Behavior was measured by performance in a roleplay task, with two equivalent exercises administered before and after training (the order of which was counterbalanced). This task is based 131 Table 6.1 Participant demographics by conditions. There is no statistical significant of demographics between different conditions. Characteristics Control Popular AntiPhishGuru Antitraining Phishing Cartoon Phishing materi- Phil Phil with als PhishGuru Sample Size 218 217 166 201 199 Gender Male 50% 48% 54% 45% 45% Female 50% 52% 46% 55% 55% Average age in 30 30 29 30 31 years Education High school or 10% 8% 7% 7% 8% less Some college 33% 32% 37% 39% 36% Competed 4-year 29% 29% 30% 30% 27% college degree Some 11% 12% 10% 6% 10% Post-graduate education Have master or 17% 19% 16% 18% 17% Ph.d degree Percentage from 74% 71% 73% 78% 80% US? Percentage 25% 26% 31% 20% 25% student? Average years on 13 12 12 13 13 the Internet Average emails 44 44 32 57 43 per day on an established roleplay exercise that has been shown to have good internal and external validity. [27]. Participants were told to assume the role of Pat Jones, who works at Baton Rouge University and uses the email address [email protected] for both work and personal emails. Each roleplay showed participants fourteen images of emails along with context about Pat Jones that may help to interpret the emails. Images matched the participant’s operating system and browser (e.g. Firefox on a Mac or Internet Explorer on a PC) so that all images and cues would be familiar 132 Figure 6.1 One of the emails that Pat encounters in her email box to the participant. Participants were asked to indicate how they would handle the emails if they received them in their own email inbox, whether that would be forwarding the email to someone else, replying by email, or any other action from a list of responses generated through earlier qualitative work [26]. Table 6.2 details the list of possible responses. The first email was created to familiarize the participant with the procedure. It was a short message from the same domain as Pat’s email address. This message from the BRU Information Security Office announced a scavenger hunt for National Cyber Security month. The participants continued through the roleplay task by viewing a combination of real, phishing, malware and spam 133 email images. Table 6.3 lists a representative sample of the emails that Pat encounters in one of the roleplays. Each email contained a link to a web page (e.g. Figure 6.1), shown with the mouse pointer positioned on the link and the actual URL destination displayed in the status bar, as it would be if users prepared to actually click on the link on their own computer. For individuals who indicated that they would click on the link or otherwise end up at the web page, an image of that web page was displayed. Each web page requested information to be entered and participants were asked to indicate if they would click on a link on the page, enter the requested information, bookmark the page, visit another related web page, close the website, or take other action. No matter what other actions the user indicated, those who said that they would enter the requested information Figure 6.2 The corresponding website is shown when Pat chooses to the “click on the link” option in the email 134 Table 6.2 List of possible responses for emails in the role play survey reply by email contact the sender by phone or in person forward the email to someone else delete the email keep, save or archive the email click on the selected link in the email (the one that the browser hand is pointing to) copy and paste the selected URL (the www address) from the email into a web browser, if a URL is selected in this email type the selected URL into a web browser, if a URL is selected in this email click on a different link in the email (please specify which link(s) you would click on) Other (please specify) were coded as having fallen for phishing or complied with a legitimate email, corresponding to the legitimacy of the email in question. 6.2.3 Education Materials Participants were randomly assigned to the control condition, or to view one of four types of educational materials on ways to avoid falling for phishing attacks: a PhishGuru cartoon, AntiPhishing Phil, several popular web-based training materials, and a combination of Anti-Phishing Phil plus a PhishGuru cartoon. For popular web-based training, we selected three consumer oriented education materials from the first page of search results from google using keyword “phishing.” They are Microsoft Online safety [89], OnGuardOnline phishing tips [106], and National Consumer League Fraud tips [99]. In total, these materials have 3107 words, and would take roughly 15 minutes to complete reading with a scanning speed of 250 words per minute. In the Anti-Phishing Phil conditions, participants were taken through three levels of the game and allowed to exit at any point. For the educational web page conditions, participants were asked at the end of each of three pages if they would like to read more information or move to the next 135 Table 6.3 A representative sample of emails in Pat’s inbox from one of the roleplays Email Subject Legitimacy Relevant features of email and websites Earn Bonus real win a prize in an online scavenger hunt Points #1 from BRU Information Security Office link: https://www.bru.edu/iso/aware/ncsam/hunt/bonus Picture from last possible impersonal greeting weekend’s party malware link:http://picasaweb.google.com/stevewulitzer/Partypics/ actual url: http://128.3.72.234/Partypics.jpg.exe No obligation spam text of link: “Apply online now” bankruptcy actual url: https://www.bankruptcylawyerfinder.com/... consultation Bandwidth phishing misspelling in url and .org domain Quota Offer link http://wwwbrubandwithamnesty.org/bandwidth/agree.htm actual url: same eBay Accounts phishing threatens account suspension Security link: https://signin.eBay.com/ws/... actual url: http://www.security-validation-your-account.com/ Your real problem with shipping Amazon.com link: www.amazon.com/help/confirmation Order actual url: same (#103-06075556895008) Your eBay item real text of link: “Send Invoice Now” sold! actual url: http://payments.ebay.com/eBayISAPI... part of the study. The PhishGuru conditions provided participants with one page of materials and then participants moved on to the next part of the study. All participants who viewed any of the educational materials were asked how likely they would be to visit that specific educational tool again and how likely they would be to recommend it to someone else, on a scale ranging from 1 (not at all likely) to 7 (extremely likely). 6.2.4 Previous Experiences and Demographics Along with asking participants extensive demographic related questions, all participants were asked to complete a series of questions about their online experiences, including questions about their choice of websites for recent purchases, their use of online banking and their prior exposure to anti-phishing educational materials. Participants also indicated any negative consequences such 136 as having information stolen or compromised in some way by entering it into a web site. Table 3 presents basic demographics of the sample. 6.2.5 Knowledge and Technical Background Knowledge questions asked participants to choose the best definition for four terms related to computer security: ‘cookie,’ ‘phishing,’ ‘spyware,’ and ‘virus.’ Participants were given the same list of eight possible definitions to choose from for each, as well as choices to indicate lack of familiarity with the word. Each term had one correct answer on the list. The options included: 1. Something that protects your computer from unauthorized communication outside the network 2. Something that watches your computer and send that information over the Internet (spyware) 3. Something websites put on your computer so you don’t have to type in the same information the next time you visit (cookie) 4. Something put on your computer without your permission, that changes the way your computer works (virus) 5. Email trying to trick you into giving your sensitive information to thieves (phishing) 6. Email trying to sell you something 7. Other software that can protect your computer 8. Other software that can hurt your computer 9. I have seen this word before but I don’t know what it means for computers 10. I have never seen this word before 11. Decline to answer 12. Other (please specify) To assess the level of their technology background, participants were asked if they had an Information Technology-related degree and any experience with programming languages, and they self-rated how technologically savvy they were on a scale ranging from 1(not at all savvy) to 7 (extremely savvy). 137 6.2.6 Risk Perceptions To evaluate participants’ risk perceptions, we presented them with a series of statements taken from the Domain-Specific Risk-Taking scale of adult populations (DOSPERT) [13], drawing on the categories of financial risk and health & safety risk. These questions asked participants to rate the risk associated with statements such as betting a day’s income at the horses races and riding a motorcycle without a helmet, on a scale ranging from 1 (not at all risky) and 7 (extremely risk). 6.3 Results 6.3.1 Measuring User Performance We measured participants’ susceptibility to phishing by examining two kinds of errors before and after education interventions: falling for phish and false positives. A false positive is when a legitimate email or website is mistakenly judged as a phish and users refuse to follow the desired actions. Falling for phish occurs when a phishing email or website is incorrectly judged to be legitimate and users click on the email and submit information to the website. In our analysis, we consider falling for phishing as giving information to phishing websites, unlike previous studies that have used the close correlate of clicking on links in phishing emails. In previous studies and this one, around 90% of the participants who clicked on the phishing link end up giving information to the phishing website [68, 69]. We used giving information to phishing sites as a stricter measure for falling for phishing. 6.3.2 Regression Analysis To explore factors that predict phishing susceptibility, we performed a multivariate linear regression. This section explains the steps we took to build the model and discusses the results from the linear regression. We used factor analysis to reduce the dimensionality of our variables on participants’ online experience (eight variables), participants’ technical knowledge and experience (5 variables), and 138 Table 6.4 Regression analysis with parameters that are significant at p < 0.01 Model Parameters Standardized Coefficients Ever seeing information on avoiding .19 phish before this test Gender .14 Age -.12 Participants’ technical knowledge -.10 Risk perception of financial invest- -.08 ment participants’ risk perception(12 variables). The factor analysis using principle component and varimax rotation reduced our list of variables from 40 to 22. We then ran the regression predicting falling for phish from the 22 variables. In Table 6.4, we report variables that are statistically significant at p <= 0.01. Participants’ degree of prior experience with phishing education significantly predicts how much phishing they will fall for (B = 0.189, p <0.01). Participants who have seen training material before (56.6% of total participants) fell for 2.4 phishing websites (40%), whereas those who have not seen training before fell for 3.6 phishing websites (60%), t = -9.02, p < 0.001. This factor had the most impact on phishing susceptibility, suggesting that exposure to education may play a larger role than other important factors. Women fall for more phish than men (B = 0.140, t = 3.98, p < 0.01), an average of 53.1% phishing emails, compared to just 41% for men, t(981) = -5.48, p < 0.001. We explore reasons for women’s greater susceptibility in the next section. Participants’ age linearly predicts their susceptibility to phishing (B = -0.116, p < 0.01). An analysis of variance (ANOVA) comparing age groups found a significant overall effect, F(4, 996) = 9.65, p < 0.001, driven by participants aged 18 to 25 falling for phishing more than other age groups (all post-hoc tests comparing this group to other groups significant at p<.01; no other groups significantly different from one another). 139 Participants’ self-rated knowledge about technology also significantly predicts whether they will fall for phishing. For each standard deviation higher the tech knowledge score, participants fell for [how many: raw number] fewer phish (3.6%). Finally, participants’ risk aversion, as measured by reactions to risks of financial investments, also predicts whether they will fall for phishing. The more risk-averse a participant is, the less likely he or she will fall for phish. For each standard deviation increase in their risk perception score, participants fell for [how many: raw number] fewer phish (2.8%). 6.3.3 Gender and Falling for Phish In order to better understand why women appear to be more susceptible to phishing, we examined clicking on phish, giving information to phish, clicking on legitimate URLs, and giving information to legitimate websites with respect to gender. We found that, before training, women were more likely than men to click on phishing links and enter information on phishing websites. On average, women clicked on 54.7% of phishing emails, compared to just 49% for men, t(981) = 2.69, p < 0.01. After clicking on a phishing link, women continued on to give information to the corresponding phishing website 97% of the time, compared to 84% for men, t = 5.42, p< 0.001. This further exacerbates the gender differences in clicking on links. These results are consistent with previous real world phishing studies [67], where 52.3% of participants clicked on the simulated spear phishing emails they sent and subsequently 40.1% gave information to phishing sites. The similarity in our results suggested the validity of the roleplay survey instrument. In an attempt to explain these gender effects, we did a mediation analysis using all the key predictors as potential mediators. Mediation analysis explains “how” an effect occurred by hypothesizing a causal sequence. The basic mediation model is a causal sequence in which the independent variable (X) causes the mediator(s) (M) which in turn causes the dependent variable (Y), therefore explaining how X had its’ effect on Y [76, 77]. Mediational processes are common in basic and applied psychology. 140 tech training 0.49** 0.23** Total effect: 0.72 ** falling for phish sex Direct effect: 0.43** -0.71** tech knowle dge -0.25** Figure 6.3 Mediation of the effect of gender on falling for phishing through participants’ tech knowledge and tech training. Table 6.5 Mediation analysis for gender. Each path is quantified with unstandardized regression coefficients. The direct effect of gender on phishing susceptibility (measured by number of phishing websites participants’ giving information to) is calculated as total effect minus all the effect through each of the mediators, which is calculated as the product of coefficients in the paths. Point Percentile estimates 95% CI Lower Upper Total Effect of gender on falling for 0.72 phishing Total effect of various mediators 0.29 0.18 0.42 tech knowledge 0.17 0.10 0.27 tech training 0.12 0.02 0.21 We used the multiple mediator model developed by Preacher and Hayers [63] for our mediation analysis. For gender, we used tech knowledge and tech training as mediators; our hypothesis is that women have less technical experience than men and therefore fall for phishing more. We report the mediation statistics in Table 6.5 and Figure 6.3 shows the results of the analysis, which are consistent with the hypothesis. 141 As shown in Figure 2, the effect of being female on falling for phishing drops from a total effect of 0.72, p<0.01, down to a direct effect of just 0.43, p<0.01. The difference between these effects represents the total indirect effect through the two mediators, with a point estimate of 0.29, and a 95% CI of 0.18 to 0.42 (see Table 6.5). Thus, women in our study have less technical training and have less technical knowledge than men, which appears to partially account for their greater susceptibility to phishing. The mediation relationship is only partial, as the direct effect is still statistically significant. This partiality suggests that there are other factors that are not captured by our survey instruments; these factors might be explored in future work. We included several other predictors that did not mediate this relationship. For example, women may fall for phishing more because they have fewer opportunities or are less motivated to learn about phishing. However, prior exposure to phishing education did not turn out to be significant mediator. In fact, in our sample, more women than men claimed to have seen phishing education before the study. Neither were income or education significant mediators for the effect of gender on phishing susceptibility. Other factors that we did not measure might potentially explain the remaining tendency for women in our study to be more susceptible to phishing than men. Factors that may be worth further exploration include differences in the way men and women use the Internet, differences in the way men and women make trust decisions, and differences in the tendency of men and women to be cooperative or comply with instructions. 6.3.4 Age and Falling for Phish As described above, people in the 18 – 25 age group were more likely to fall for phish than people of other ages. We used the multiple mediator model to determine why younger people are more frequently falling for phishing. We report the mediation statistics in Table 6.6 and Figure 6.4. Taken as a set, participants’ prior exposure to phishing, numbers of years on the Internet, financial risk perception, and education mediate the effect of age on falling for phishing. As can 142 Exposed to training before 1.00** 0.078** -0.30* -0.30** Education -0.16** 0 16** Total effect: 0.34 4 ** age bracket falling for phish Direct effect: 0.12 -1.31** -1.31 Year on internet -0.06* 0.06 -0.14** -0.13* financial _risk_ investing Figure 6.4 Mediating the effect of age with prior exposure to training, education, years on the Internet and risk perception for financial investment. Each of the paths is quantified with unstandardized regression coefficients. Table 6.6 Total effect of age on falling for phishing and the effect of various mediators that are statistically significant at p<0.01. Point Percentile estimates 95% CI Lower Upper Total Effect of age on falling for 0.34 phishing Total effect of various mediators 0.23 0.16 0.29 Prior exposure 0.08 0.04 0.12 years on Internet 0.08 0.03 0.13 education 0.05 0.02 0.08 risk financial investing 0.02 0.00 0.04 143 be seen in Figure 3, the total effect of age on falling for phishing fell from 0.34, p < 0.01, down to 0.12 (not significant). The difference between the total and direct effects is the total indirect effect through the four mediators, with a point estimate of 0.23, and a 95% CI of 0.16 to 0.29 (see Table 6). Because younger people have a lower level of education, fewer years of experience with the Internet, less exposure to training material, and less of an aversion to financial risks, they tend to be more susceptible to phishing. 6.3.5 Effects of Education All of the training materials we tested reduced participants’ tendency to click on phishing links in emails by 13-17 percentage points. There is no statistical difference between each education material, F(3,779) = 1.28, p = 0.28. The control group, which received no training during the study, showed no statistically significant improvement between the first and second roleplay. We also did not find the ordering of the knowledge survey affected the users’ performance, so in our analysis we collapsed across orders. All training materials reduced participants’ tendency to enter information into phishing webpages by about 16-21 percentage points, and there is no statistically significant improvement for the control group. Anti-Phishing Phil, Phishguru cartoon and Anti-Phishing Phil with Phishguru cartoon did not decrease participants’ tendency to click on legitimate links and go to legitimate websites. However in the popular training condition, participants’ tendency to click on legitimate links was slightly reduced, t(216) = 2.01, p< 0.05, suggesting that improvements in avoiding phish may merely reflect an avoidant strategy and not better detection. Since the various education materials perform similarly in reducing people not falling for phishing, to study the effect of education in bridging the demographic gaps, we combined all the training conditions together. Before the training, participants on average fell for 2.8 phishing websites out of 6, or 47%. After the training, this number is reduced to 1.6 out of 6, or 26%, a 21 percentage point improvement or 42% improvement. In terms of demographics, we found that women learned more than men 144 during the training about avoiding phishing links (t (767) = 5.63 p < 0.01); after training women and men perform equally well in not clicking on phishing links in emails t(767) = -0.05, p = 0.96 ). In entering information into phishing sites, women and men learned similarly, t(767) = -1.51, p = 0.13). Women’s higher rate of entering this information before the training carried over, and they still fell for more phish after the training than men, t(767) = -4.22, p< 0.001). Finally, people of different age groups learned similarly in training, leaving no statistical difference between age groups’ performance increase, F(4,778) = 1.66, p = 0.16. Participants between ages 18 and 25 were the most susceptible group in pretest, and they remained more susceptible to phishing in posttest. People in different education groups also learned similarly, F(5,763) = 1.4, p = 0.20). We also found no significant effect for education or race. We also analyzed the amount of time user spent on education materials. We found that users in the game conditions (Anti-phishing phil alone and Anti-Phihing Phil with Phishguru cartoon) spent the longest time, averaging 8.6 minutes. Although the popular education were designed to last as long as the game condition, users only spent 1.8 minutes on average (Table 6.7). Table 6.7 Time user spent on education materials Education Materials Popular training materials Anti-Phishing Phil PhishGuru Cartoon Anti-Phishing Phil with PhishGuru Cartoon Estimate time user would spent 12 min Average time user spent 10 min 8.68 min (SD = 5.70) .50 min (SD = 1.05) 8.55 min (SD = 5.50) 2 min 12 min 1.80 min (SD = 2.09) 145 6.4 DISCUSSION 6.4.1 Limitations There are several limitations to the current study. First, the sample was drawn from mTurk users and is not expected to be representative of the larger population of email users. Our sample of mTurk users tends to be younger, more educated and more tech savvy than the general public. A second limitation of this study is the lack of direct consequences for user behavior. Participants might be more willing to engage in risky behavior in this roleplay if they feel immune to any negative outcomes that may ensue. Similarly, participants are not risking opportunity costs from being too conservative in their behavior. However, performance on this roleplay has been validated with real-world behavior, showing that, if anything, people are more conservative in their roleplay responses than they are with their actual email inboxes [121]. Furthermore, there is no reason to believe that the predictors described here should differ in their relationship to roleplay behavior compared to real-world behavior. 6.4.2 Summary of findings Prior exposure to phishing education is associated with less susceptibility to phishing, suggesting that phishing education may be an effective tool. Also, more risk-averse participants tended to fall for fewer phish. Gender and age are two key demographics that predict phishing susceptibility. Specifically, women click on links in phishing emails more often than men do, and also are much more likely than men to continue to give information to phishing websites. In part, this appears to be because women have less technical training and less technical knowledge than men. There is also a significant effect for age, in which participants aged between 18 and 25 are much more likely than others to fall for phishing. This group appears to be more susceptible because participants in this age group have a lower level of education, fewer years on the Internet, less of an exposure to training materials, and are less of an aversion to risks. Educators can bridge this gap by providing anti-phishing education to high school and college students. 146 All the education materials in our study reduce users’ tendency to enter information into phishing webpages by about 16-21 percentage points. However, some education material decreased participants’ tendency to click on legitimate links, this suggests that educator need to do a better job of teaching people how to distinguish phish from non-phish so that they avoid false positives. Demographics such as age, gender, race, and education do not affect the amount of learning, suggesting that training can provide some benefit for all groups, if provided with the right materials. Although the 46% reduction in phishing susceptibility after training is substantial, even after training participants fell for 26% of the phishing messages in our roleplay. This finding shows that education is effective and needed but is not a cure all. In our study, 61% of the U.S participants have seen phishing education before; the task for the various stakeholders is to reach out to the 39% of the population who have not been exposed to training. However, even with the best educational materials, participants in our study still fell for around 28% of phish after training. Women and younger populations such as college students are especially vulnerable. These findings show that education should be complemented with other countermeasures such as filtering and law enforcement. 6.4.3 Role of education As phishing continues to evolve, what is the role of education in combating it? Specifically, what problems can education solve, and how does education fit into a layered approach to combat phishing? We discuss these questions in the concluding section of this chapter. Generally speaking, strategies for protecting people from phishing fall into three major categories: silently eliminating the threat, warning users about the threat, and training users not to fall for attacks. These categories of anti-phishing strategy mirror the three high-level approaches to usable security: build systems that just work” without requiring intervention on the part of users, make security intuitive and easy to use, and teach people how to perform security-critical functions [19]. Our view is that these three approaches should complement each other. Today, the majority of phishing emails are filtered at email gateways, and forwarding the future more efforts are needed to 147 filter as many phishing emails as possible, as quickly as possible, and with as few false positives as possible. Without this first layer of defense, even the best-educated users would be inundated with phishing messages that could paralyze their decision-making. It is also important to strengthen the browser, OS, and application security. Since it would be very difficult even for the experts to notice a compromised browser URL bar, user education would do little to alleviate the problem. In the same vein, users’ computers can be infected with malware even without any user action. As a result, where possible, the first layer of defense should always be automated solutions to filter and increase the default security offered to users’ computers and web applications. However, we also need to acknowledge that these systems are not completely accurate in detecting phishing attacks. It is unlikely that any system will ever be completely accurate in detecting phishing attacks, especially when detection requires knowledge of contextual information. While it makes sense to use automated detection systems as one line of defense against semantic attacks, there will still remain many kinds of trust decisions that users must make on their own, usually with limited or no assistance. Thus, the second line of defense is to develop a complementary approach to support users so that they can make better trust decisions. There are two options for this approach: teach people not to fall for phish, or build easy-to-use software and interfaces that prevent users from falling for phishing. User education is a low-hanging fruit. In our study, 61% of the U.S participants have seen phishing education before, and those who have seen education on average fell for 40-50% less phishing. Therefore efforts to reach out to the 39% of the population who have not been exposed to training would be likely to quickly reduce phishing susceptibility. Finally, User education has its limits as well. Even with the best educational materials, participants in our study still fell for around 28% of phish after training. Women and younger populations such as college students are especially vulnerable. Therefore, the last step of defense is to build easy-to-use software and interfaces. Examples such as integrated web browser warnings [29] and foolproof phishing solutions are promising [108]. Appendix 148 Email contest Legitimacy Relevant features of email and sites real Win a price in an online scavenger hunt From BRU Information Security Office link: https://www.bru.edu/iso/aware/ncsam/hunt/bonus National real Pat has an account. City text of link: “view your statement” actual url: http://www.nationalcity.com/statements party possible impersonal greeting malware link: http://picasaweb.google.com/stevewulitzer/Partypics/ actual url: http://128.3.72.234/Partypics.jpg.exe verify email phishing threatens account deactivation account asks for password in text of email no link in email bankruptcy spam text of link: “Apply online now” actual url: https://www.bankruptcylawyerfinder.com/ bandwidth phishing misspelling in url link http://wwwbrubandwithamnesty.org/bandwidth/agree.htm actual url: same eBay phishing threatens account suspension link: https://signin.eBay.com/ws/eBayISAPI.dll... actual url: http://www.security-validation-your-account.com/... Amazon real problem with shipping link: www.amazon.com/help/confirmation actual url: same National phishing system upgrade City link: http://service-nationcity.org actual url: http://210.7.78.331/SITE/natcity/ summary real sender from bru.edu and in Pat’s address book. report summaryreport.doc attached help desk phishing threatens account termination link: http://bruwebmail.org/password/change.htm actual url: same eBay real text of link: “Send Invoice Now” actual url: http://payments.ebay.com/eBayISAPI... networking: phishing .org domain link: http://batonrougenetworking.org/summer09/register.html actual url: same As seen of spam dot com written out in email text Television Table 6.8 Emails in Pat Jones’ Inbox: Roleplay A 149 Rotated Component Matrix Component 1 2 purchased anything .804 .024 on the web online banking: ever .258 .842 used online banking bills online: ever -.177 .885 paid bills online credit card stolen: .339 .115 ever happen ssn stolen: ever hap- .240 .086 pen info stolen: ever hap- .284 .162 pen lose money: did you -.113 -.009 permanently lose money paypal account: ever .754 .020 had a paypal account 3 -.057 4 -.039 -.136 -.027 .070 .113 .168 .670 .810 .081 -.705 .120 -.147 .832 .049 .140 Table 6.9 Factor analysis for various Internet experience variables. Rotation Method: Varimax with Kaiser Normalization. Rotation converged in 5 iterations. They are (1) “web purchase experience” (averaging purchasing at the web or whether had a paypal account); (2) “online banking” by averaging ever used online banking and online bill pay; (3) “ssn stolen” that is whether they had their ssn stolen, and (4) and “creditcard stolen” that averages the credit card stolen and ever lose money. 150 Rotated Component Matrixa Component 1 programming -.254 languages techology .850 spectrum tech savvy .820 security -.569 preference adjusted computers -.153 daily IT degree .047 2 .733 -.170 -.288 .032 .376 .861 Table 6.10 Factor analysis for various Internet experience variables. Rotation Method: Varimax with Kaiser Normalization. Rotation converged into two factors in five iterations. They are (1) ”tech knowledge” by averaging tech spectrum and tech savvy, and(2) called ”tech training” by averaging programming languages and IT degree (for tech training, lower numbers mean more training). 151 Rotated Component Matrixa Component 1 2 risk: Betting a days .083 .901 income at the horse races risk: Investing 10 of -.008 -.030 your annual income in a moderate growth mutual fund risk: Drinking heav- .446 .415 ily at a social function risk: Betting a days .132 .911 income at a high stake poker game risk: Investing 5 of .051 .082 your annual income in a very speculative stock risk: Betting a days .140 .894 income on the outcome of a sporting event risk: Engaging in un- .628 .179 protected gender risk: Driving a car .800 .062 without wearing a seat belt risk: Investing 10 of .154 .142 your annual income in a new business venture risk: Riding a motor- .681 .120 cycle without a helmet risk: Sunbathing .755 .042 without sunscreen risk: Walking home .740 .065 alone at night in an unsafe area of town 3 .059 .701 -.042 .058 .829 .129 -.045 .019 .764 .218 .068 .073 Table 6.11 Principle Component analysis for various Internet experience variables. Rotation Method: Varimax with Kaiser Normalization. Rotation converged in 5 iterations. 152 Model Summary Model R 1 .440a R Square .194 Adjusted R Std. Error Square of the Estimate .174 1.90464 Table 6.12 Regression statistics ANOVA Model Sum of Squares 1 Regression 799.840 Residual 3330.196 Total 4130.036 b. Dependent Variable: pre test phish giveinfo df 22 918 940 Mean Square 36.356 3.628 F Sig. 10.022 .000a 153 Table 6.13: Complete list of variables for regression Variable age numeric sexsurvey education recode OCCU student hispanic racewhite countryindia countryusa income avoidphish Descriptions What is your age? What is your gender? 1 = Male 0 = Female What is your highest education? 1 = High school or less 2 = Some college 3 = Completed 4-year college degree 4 = Some Post-graduate education 5 = Have masters or Ph.D degree 6 = Decline to answer Are you currently a student? 1 = YES 0 = NO Are you Hispanic? 1 = YES 2 = NO What’s your race (white or Caucasian?) 1 = YES 2 = NO Do you currently reside in India? 1 = YES 2 = NO Do you currently reside in US? 1 = YES 2 = NO What is your annual household income? 1 = < $20,000 2 = $20,000 - $39,000 3 = $40,000 - $59,000 4 = $60,000 - $79,000 5 = $80,000 - $99,000 6 = >100,000 7 = Decline to answer Have you ever seen information to avoid phish before this study? 1 = YES Statistics M = 30.1 SD = 10.6 475 508 79 349 284 97 169 5 247 736 66 842 641 342 145 838 739 244 203 196 181 99 74 74 156 556 154 Table 6.13: Complete list of variables for regression Variable Descriptions 1.5 = NOT SURE 2 = NO computersdaily Do you use computers daily? 1 = YES 2 = NO emailperday numericOn average, how many emails do you receive a day? tech knowledge Tech knowledge scale from Factor analysis (1 – 7) tech training Tech training scale from factor analysis (1 – 2) risk health safty How do you perceive the following risks (1– 7)? risk financial bettingHow do you perceive the following risks? risk financial investing How do you perceive the following risks? magcomputer What magazines do you frequently read (computers and electronics?) 1 = YES 2 = NO internet numeric At what year did you first use Internet? online banking Online banking scale from Factor analysis (1 – 2) creditcard stolen Have you ever had your creditcard stolen online? 1 = YES 1.5 = NOT SURE 2 = NO web purchase Web purchase experience scale from factor analysis (1 – 2) Statistics 85 342 867 116 M = 44, SD = 81 M = 5.3 SD = 1.2 M = 1.7 SD = 0.36 M = 5.5, SD = 1.0 M = 5.8 SD = 1.3 M = 4.1 SD = 1.1 335 648 M = 1996, SD = 3.7 M = 1.17 SD = 0.33 34 25 924 M = 1.12 SD = 0.27 155 Coefficients Model 1 (Constant) age numeric sexsurvey education OCCU student hispanic racewhite countryindia countryus income avoidphish computersdaily emailperday numeric tech knowledge tech training risk health safty risk financial betting risk financial investing magcomputer internet numeric online banking creditcard stolen web purchase Unstandardized Coefficients B Std. Error -90.519 40.864 -.023 .007 .586 .148 -.126 .053 -.090 .170 .068 .173 -.324 .162 .074 .292 .018 .222 -.060 .031 .851 .147 .100 .201 -.002 .001 -.173 .061 .496 .208 .103 .067 .110 .054 -.153 .061 .213 .156 .046 .020 -.084 .222 -.102 .324 -.250 .287 Standardized Coefficients Beta -.116 .140 -.080 -.019 .012 -.074 .012 .004 -.060 .189 .015 -.073 -.103 .085 .050 .067 -.080 .048 .083 -.013 -.010 -.032 Table 6.14 Complete output of the regression analysis. t Sig. -2.215 -3.142 3.964 -2.391 -.533 .390 -2.005 .253 .080 -1.942 5.787 .495 -2.324 -2.840 2.388 1.530 2.055 -2.518 1.366 2.275 -.377 -.315 -.871 .027 .002 .000 .017 .594 .697 .045 .801 .936 .052 .000 .621 .020 .005 .017 .126 .040 .012 .172 .023 .706 .753 .384 156 Chapter 7 Conclusions As phishing and related crimeware continue to evolve, what should be the US government’s role? In the concluding chapter of this thesis, we answer this question based on our analysis of the interests of phishing stakeholders, experts’ recommendations, and insights from case studies. We discuss several measures that the US government can take to combat phishing. First, experts in our study agreed that catching criminals would provide a strong deterrent as it shows the determination and capability of law enforcement. However, in our interviews with law enforcement, we found that they face three major challenges: lack of necessary analytical capabilities in determining investigative priorities (including qualified investigators and software tools), the international nature of the crime, and sophistication of criminals to hide their traces. To address the first challenge, we recommend the US government invest in tools for better case management and better digital evidence processing. To attract more talent, the US government could provide incentives to recruit from top computer science programs around the country for digital forensics and analysis, possibly through the expansion of DHS scholarship program. To address the second challenge of the international nature of the crime, we recommend the US government fund international operations and facilitate the communication and connection of law enforcement in various countries. To address the third challenge we recommend that the US government facilitate greater information sharing between law enforcement and industry. Second, botnets are major pieces of crimeware infrastructure that have greatly enabled spamming and phishing operations. Our analysis shows that Internet service providers, who are in the best position to address this issue, do not have enough incentives to address this problem. In the short run, we recommend the US government institute a notice and takedown regime which 157 mandates ISPs or upstream providers to disconnect bot command and controls once they were identified. Such policies will relieve the ISPs of potential liability and will be applied to all ISPs. It is not wise for the government to set standards for ISPs to clean up the compromised machines as currently such methods are costly and will have little benefits unless a majority of the compromised machines are fixed. To secure this key piece of infrastructure, the government can also leverage some of the technologies used in national defense and apply them to fixing ISP networks. In the long term, research is needed on automatically cleaning compromised machines. Finally, fixing botnets in the US alone will not be likely to solve the problem, as there are compromised machines overseas as well. To address this issue, relevant agencies in the US government need to establish close working relationship with other countries to share intelligence of the botnets. Third, experts in our study agreed that better statistics about phishing and related electronic crime are necessary for law enforcement to prioritize and for corporations to manage their security risks better. However, as of today, little accurate statistical information exists because financial institutions are not required to report, and they have little incentives to do so, thus making loss estimates differ by orders of magnitude. To correct this misalignment of incentives, we recommend that the US government institute some mandatory reporting. Fourth, our case study showed that the window of opportunity for defenders is limited as close to half of the phishing campaigns lasted less than 2 hours. Although leveraging heuristics is a key solution, experts in our study pointed out major legal issues with false positives that hinder the use of heuristics. However in our case study we found that the group of heuristics that we tested yielded extremely low false positives. Therefore, we conclude it is the “fear” of false positives, not “actual” false positives that is hindering the adoption of heuristics. We recommend that legislators step in to clarify the legal issues surrounding the false positives, and provide incentives such as safe harbor legislation. Fifth, experts in our study disagreed on the effectiveness of user education. In our mTurk study and other studies, we showed that currently the best education materials could reduce the number of people falling for phishing by 40 - 50%. This finding shows that education is effective and needed but is not a cure all. In our study, 61% of the US participants have seen phishing education 158 before; the task for the various stakeholders is to reach out to those 39% of the population who have not been exposed to training. However, even with the best educational materials, participants in our study still fall for around 28% of phish after training. Women and younger populations such as college students are especially vulnerable. These findings show that education should be complemented with other countermeasures such as filtering and law enforcements. Last but not least, we detailed the recommendations that various stakeholders should take to better fight for phishing as summarized in Table 7.1. By implementing these measures, we can drastically reduce phishing and other related electronic crimes that use the same infrastructure. 159 Stakeholders Financial Institutions OS vendors Web browser vendors ISPs US Government US Government US Government Corporations Various stakeholders Academic researchers and industry Academic researchers and industry Legal Community Table 7.1 Summary of Recommendations Recommendations Produce more accurate estimates of phishing losses and report these statistics Continue to secure operating systems by implementing secure coding practices, investing in secure vulnerability patching, and building anti-malware capability directly into the operating systems to enhance default security Continue to improve the performance of integrated browser anti-phishing warning systems in order to catch 85-95% of phishing URLs within an hour after they go online Develop or deploy better techniques to quickly identify botnets and proxies, shut down botnet command and control, and clean compromised machines Develop notice and takedown regimes for botnet C&C removal Invest in international cooperations through funding joint operations, facilitating the communication and connection of law enforcement in various countries Invest in tools for better case management and better digital evidence processing; Expand DHS Scholarship program to recruit master students from top computer science schools Aggregate fraud data (proxies) and submit to law enforcement to identify proxies Focus on improving the security of web applications Continue to make education engaging and up to date Focus heuristic research on reducing false positives Clarify the legal issues of the false positives of blacklists and heuristics 160 LIST OF REFERENCES [1] A BU -N IMEH , S., NAPPA , D., WANG , X., AND NAIR , S. A comparison of machine learning techniques for phishing detection. In eCrime ’07: Proceedings of the anti-phishing working groups 2nd annual eCrime researchers summit (New York, NY, USA, 2007), ACM, pp. 60–69. [2] A NANDPARA , V., D INGMAN , A., JAKOBSSON , M., L IU , D., AND ROINESTAD , H. Phishing IQ tests measure fear, not ability. Usable Security (USEC’07) (2007). http://usablesecurity.org/papers/anandpara.pdf. [3] A NDERSON , J. R. Rules of the Mind. Lawrence Erlbaum Associates, Inc., 1993. [4] A NDERSON , R., AND M OORE , T. The economics of information security. Science 314, 5799 (2006), 610–613. [5] A NDY PATRIZIO. Symantec readies phishing tion software, august 7, 2006. visited jan 1, http://www.smallbusinesscomputing.com/news/article.php/3624991. protec2009. [6] A NTI -P HISHING WORKING G ROUP. Anti-phishing Practices Recommendations for Registrars. Report, http://www.apwg.org/reports/APWG_RegistrarBestPractices.pdf. Best 2008. [7] A NTI -P HISHING WORKING G ROUP. Global Phishing Survey: Trends and Domain name use in 2H 2008. Report, 2008. http://www.antiphishing.org/reports/APWG_GlobalPhishingSurvey2H2008.pdf. [8] A NTI -P HISHING WORKING G ROUP. What to Website Has Been Hacked by Phishers. http://www.apwg.org/reports/APWG_WTD_HackedWebsite.pdf. Do If Report, Your 2008. [9] AOL P RESS R ELEASE. It’s 3 a.m. – are you checking your email again? july 30, 2008. visited jan 1, 2009. http://corp.aol.com/press-releases/2008/07/it-s-3-am-are-you-checking-your-email-a [10] A PPLE I NC . . New features in safari. http://www.apple.com/safari/features.html#security. 161 [11] ASSOCIATED BANK-CORP V. EARTHLINK, INC. Memorandum and order, 05-c0233-s. http://www.iplawobserver.com/cases/2005-09-14_Associated_Banc_Corp_CDA_Sectio [12] AVANTGARDE. Time to Live on the Network. http://www.avantgarde.com/xxxxttln.pdf. Tech. rep., Avantgarde, 2004. [13] B LAIS , A.-R., AND W EBER , E. U. A domain-specific risk-taking (dospert) scale for adult populations. Judgment and Decision Making 1, 1 (2006), 33–47 KW –. [14] C ALMAN , C. Bigger phish to fry: California’s antiphishing statute and its potential imposition of secondary liability on internet service providers. Richmond Journal of Law and Technology XIII, 1 (2006). [15] C AVUSOGLU , H., AND R AGHUNATHAN , S. Configuration of Detection Software: A Comparison of Decision and Game Theory Approaches. Decision Analysis 1, 3 (2004), 131–148. [16] C HOU , N., L EDESMA , R., T ERAGUCHI , Y., AND M ITCHELL , J. C. Client-side defense against web-based identity theft. In Proceedings of The 11th Annual Network and Distributed System Security Symposium (NDSS ’04). (2004). [17] C LOUDMARK I NC . Visited jan 1, 2009. http://www.cloudmark.com/desktop/download/. [18] C OMMITTEE ON D EVELOPMENTS IN THE S CIENCE OF L EARNING AND NATIONAL R E SEARCH C OUNCIL. How People Learn: Bridging Research and Practice. National Academies Press, 2000. [19] C RANOR , L. F. A framework for reasoning about the human in the loop. In UPSEC’08: Proceedings of the 1st Conference on Usability, Psychology, and Security (Berkeley, CA, USA, 2008), USENIX Association, pp. 1–15. [20] DANCHEV, D. Google: Spam volume for q1 back to pre-mccolo levels. CBS Interactive, April 2 2009. [21] DANCHEV, D. Microsoft study debunks phishing profitability. ZDNet, January 8 2009. [22] D HAMIJA , R., AND T YGAR , J. D. The battle against phishing: Dynamic Security Skins. In SOUPS ’05: Proceedings of the 2005 symposium on Usable privacy and security (New York, NY, USA, 2005), ACM Press, pp. 77–88. [23] D HAMIJA , R., T YGAR , J. D., AND H EARST, M. Why phishing works. In CHI ’06: Proceedings of the SIGCHI conference on Human Factors in computing systems (New York, NY, USA, 2006), ACM Press, pp. 581–590. [24] DKIM S IGNATURES , RFC 4871 . http://dkim.org/specs/rfc4871-dkimbase.html. Visited jan 1, 2009. 162 [25] D OWNS , J., AND F ISCHHOFF , B. Adolescent Health: Understanding and Preventing Risk Behaviors. John Wiley and Sons, 2009, ch. 5. [26] D OWNS , J. S., H OLBROOK , M. B., AND C RANOR , L. F. Decision strategies and susceptibility to phishing. In SOUPS ’06: Proceedings of the second symposium on Usable privacy and security (New York, NY, USA, 2006), ACM Press, pp. 79–90. [27] D OWNS , J. S., H OLBROOK , M. B., AND C RANOR , L. F. Behavioral Response to Phishing. In eCrime ’07: Proceedings of the 2007 e-Crime Researchers summit (New York, NY, USA, 2007), ACM Press, pp. 79–90. [28] E B AY I NC . Tutorial: Spoof(fake) http://pages.ebay.com/education/spooftutorial/. Emails, 2006. [29] E GELMAN , S., C RANOR , L. F., AND H ONG , J. You’ve been warned: an empirical study of the effectiveness of web browser phishing warnings. In CHI ’08: Proceeding of the twentysixth annual SIGCHI conference on Human factors in computing systems (New York, NY, USA, 2008), ACM, pp. 1065–1074. [30] E VERETT-C HURCH , R. Mccolo and the difficulty of fighting spam. Internet.com, November 20 2008. http://itmanagement.earthweb.com/features/print.php/3786296.. [31] E VERS , J. Security expert: User education is pointless. http://news.com.com/2100-7350_3-6125213.html. CNet News.com, 2006. [32] F EDERAL T RADE C OMMISSION. How Not to Get Hooked by a Phishing Scam, 2006. http://www.ftc.gov/bcp/edu/pubs/consumer/alerts/alt127.htm. [33] F ERGUSON , A. J. Fostering E-Mail Security Awareness: The West Point Carronade. EDUCASE Quarterly, 2005. http://www.educause.edu/ir/library/pdf/eqm0517.pdf. [34] F ETTE , I., S ADEH , N., AND TOMASIC , A. Learning to detect phishing emails. In WWW ’07: Proceedings of the 16th international conference on World Wide Web (New York, NY, USA, 2007), ACM Press, pp. 649–656. [35] F INANCIAL S ERVICES T ECHNOLOGY C ONSORTIUM. Understanding and Countering the Phishing Threat. White Paper, 2005. http://www.fstc.org/projects/docs/FSTC_Counter_Phishing_Project_Whitepaper.pdf. [36] F LORENCIO , D., AND H ERLEY, C. EVALUATING A TRIAL DEPLOYMENT OF PASSWORD RE-USE FOR PHISHING PREVENTION. In eCrime ’07: Proceedings of the 2007 e-Crime Researchers summit (New York, NY, USA, 2007), ACM Press, pp. 26–37. [37] F LYNN , J., S LOVIC , P., AND M ERTZ , C. K. Gender, race, and perception of environmental health risks. Risk Analysis 14, 6 (1994), 1101–1108. 163 [38] F RANKLIN , J., P ERRIG , A., PAXSON , V., AND S AVAGE , S. An inquiry into the nature and causes of the wealth of internet miscreants. In CCS ’07: Proceedings of the 14th ACM conference on Computer and communications security (New York, NY, USA, 2007), ACM, pp. 375–388. [39] F REE E MAIL P ROVIDERS G UIDE. Free email providers list. http://www.fepg.net/providers.html. fepg.net, 2004. [40] FU, A. Y. WEB IDENTITY SECURITY: ADVANCED PHISHING ATTACKS AND COUNTER MEASURES. PhD thesis, CITY UNIVERSITY OF HONG KONG, 2007. [41] G ARERA , S., P ROVOS , N., C HEW, M., AND RUBIN , A. D. A framework for detection and measurement of phishing attacks. In WORM ’07: Proceedings of the 2007 ACM workshop on Recurring malcode (New York, NY, USA, 2007), ACM, pp. 1–8. [42] G ARTNER R ESEARCH. Number of Phishing Adults Nearly Doubles in Just Two Years. http://www.gartner.com/it/page.jsp?id=498245. E-Mails Sent to U.S. Press Release, 2006. [43] G ARTNER R ESEARCH. Gartner survey shows phishing attacks escalated in 2007. Press Release, 2007. http://www.gartner.com/it/page.jsp?id=565125. [44] G EE , J. P. What Video Games Have to Teach Us About Learning and Literacy. Palgrave Macmillan, Hampshire, England, 2003. [45] G OLDMAN , L. Cybercon. http://www.forbes.com/forbes/2004/1004/088.html. Forbes.com, 2004. [46] G OOGLE I NC . Google safe browsing for firefox. visited jan 1, 2009, 2007. http://www.google.com/tools/firefox/safebrowsing/. [47] G ORDON , L. A., AND L OEB , M. P. The economics of information security investment. ACM Trans. Inf. Syst. Secur. 5, 4 (2002), 438–457. [48] G ORLING , S. The myth of user education. In Proceedings of the 16th Virus Bulletin International Conference (2006). [49] G RIMES , R. A. Malicious Mobile Code: Virus Protection for Windows, first ed. O’Reilly & Associates, Inc., Sebastopol CA, USA, 2001. [50] H ERLEY, C., AND F LOR ÊNCIO , D. A profitless endeavor: phishing as tragedy of the commons. In NSPW ’08: Proceedings of the 2008 workshop on New security paradigms (New York, NY, USA, 2008), ACM, pp. 59–70. [51] H ERZBERG , A., AND G BARA , A. Protecting (even) naive web users, or: preventing spoofing and establishing credentials of web sites. Cryptology ePrint Archive, Report 2004/155, 2004. http://eprint.iacr.org/2004/155. 164 [52] I DENTITY T HEFT T ECHNOLOGY C OUNCIL. Online Phishing technology, chokepoints and countermeasures. http://www.antiphishing.org/Phishing-dhs-report.pdf. [53] JAGATIC , T. N., J OHNSON , N. A., JAKOBSSON , M., Commun. ACM 50, 10 (2007), 94–100. AND identity Report, theft: 2005. M ENCZER , F. Social Phishing. [54] JAKOBSSON , M. The Human Factor in http://www.informatics.indiana.edu/markus/papers/aci.pdf, 2006. Phishing. [55] JAKOBSSON , M., AND M YERS , S. Phishing and Countermeasures: Understanding the Increasing Problem of Electronic Identity Theft. Wiley-Interscience, 2006. [56] JAKOBSSON , M., AND S TAMM , S. Invasive browser sniffing and countermeasures. In WWW ’06: Proceedings of the 15th international conference on World Wide Web (New York, NY, USA, 2006), ACM Press, pp. 523–532. [57] JAMES , L. Phishing Exposed, 1 edition ed. Syngress, 2005. [58] J EFF M AKEY. Blacklists compared, april 11, 2009. retrieved april 14, 2009. http://www.sdsc.edu/~ jeff/spam/cbc.html. [59] J OHN E. D UNN. Ie 7.0 tops study of anti-phishing tools , 29 september 2006, techworld. retrieved april 1, 2009. http://www.techworld.com/security/news/index.cfm?newsID=6995&pagtype=sam. [60] J OHNSON , B. R., AND KOEDINGER , K. R. Comparing instructional strategies for integrating conceptual and procedural knowledge. In Proceedings of the Annual Meeting [of the] North American Chapter of the International Group for the Psychology of Mathematics Education (October 2002), vol. 1–4, pp. 969–978. [61] J UNG , J., AND S IT, E. An empirical study of spam traffic and the use of dns black lists. In IMC ’04: Proceedings of the 4th ACM SIGCOMM conference on Internet measurement (New York, NY, USA, 2004), ACM, pp. 370–375. [62] K EIZER , G. Phishers Beat Bank’s Two-factor Authentication. Information Week, 2006. http://www.informationweek.com/news/showArticle.jhtml?articleID=190400362. [63] KJ, P., AND AF., H. Asymptotic and resampling strategies for assessing and comparing indirect effects in multiple mediator models. Behavior research methods 40, 3 (Aug 2008), 879–91. [64] K LEIN , G. Sources of power : How people make decisions? The MIT Press Cambridge, Massachusetts The MIT Press, Cambridge, Massachusetts, London, England, February 1999. 165 [65] K REBS , B. Host of Internet Spam Groups Is Cut Off. Washington Post, November 12 2008. [66] K UMARAGURU , P., C RANSHAW, J., ACQUISTI , A., C RANOR , L., H ONG , J., B LAIR , M. A., AND P HAM , T. School of phish: a real-word evaluation of anti-phishing training. In SOUPS ’09: Proceedings of the 5th Symposium on Usable Privacy and Security (New York, NY, USA, 2009), ACM, pp. 1–12. [67] K UMARAGURU , P., C RANSHAW, J., ACQUISTI , A., C RANOR , L. F., H ONG , J., B LAIR , M. A., AND P HAM , T. School of phish: A real-word evaluation of anti-phishing training. In SOUPS ’09: Proceedings of the 5rd symposium on Usable privacy and security (New York, NY, USA, 2009), ACM. [68] K UMARAGURU , P., R HEE , Y., ACQUISTI , A., C RANOR , L. F., H ONG , J., AND N UNGE , E. Protecting people from phishing: the design and evaluation of an embedded training email system. In CHI ’07: Proceedings of the SIGCHI conference on Human factors in computing systems (New York, NY, USA, 2007), ACM Press, pp. 905–914. [69] K UMARAGURU , P., R HEE , Y., S HENG , S., H ASAN , S., ACQUISTI , A., C RANOR , L. F., AND H ONG , J. Getting users to pay attention to anti-phishing education: evaluation of retention and transfer. In eCrime ’07: Proceedings of the anti-phishing working groups 2nd annual eCrime researchers summit (New York, NY, USA, 2007), ACM, pp. 70–81. [70] K UMARAGURU , P., S HENG , S., ACQUISTI , A., C RANOR , L. F., AND H ONG , J. Teaching Johnny not to Fall for Phish. Tech. rep., Carnegie Mellon University, 2006. [71] K UMARAGURU , P., S HENG , S., ACQUISTI , A., C RANOR , L. F., AND H ONG , J. Teaching Johnny not to Fall for Phish. Transactions on Internet Technology (2009). [72] K UNREUTHER , H., AND H EAL , G. Interdependent http://citeseer.ist.psu.edu/kunreuther02interdependent.html. security. [73] L EYDEN , J. Florida man indicted over Katrina phishing scam. The Register, 2006. http://www.theregister.co.uk/2006/08/18/hurricane_k_phishing_scam/. [74] L ICHTMAN , D., AND P OSNER , E. Holding Internet Service Providers Accountable. CHICAGO JOHN M. OLIN LAW and ECONOMICS WORKING PAPER, 2004. http://www.law.uchicago.edu/Lawecon/index.html. [75] L UDL , C., M CALLISTER , S., K IRDA , E., AND K RUEGEL , C. On the effectiveness of techniques to detect phishing sites. In DIMVA ’07: Proceedings of the 4th international conference on Detection of Intrusions and Malware, and Vulnerability Assessment (Berlin, Heidelberg, 2007), Springer-Verlag, pp. 20–39. [76] M ACKINNON , D. P., AND DWYER , J. H. Estimating Mediated Effects in Prevention Studies. Eval Rev 17, 2 (1993), 144–158. 166 [77] M AC K INNON , D. P., FAIRCHILD , A. J., AND F RITZ , M. S. Mediation analysis. Annual Review of Psychology 58, 1 (12 2006), 593–614. [78] M ACMILLAN , N. A., AND C REELMAN , C. D. Lawrence Erlbaum, 2004. Detection Theory: A User’s Guide. [79] M ALDONADO , H., L EE , J.-E. R., B RAVE , S., NASS , C., NAKAJIMA , H., YAMADA , R., I WAMURA , K., AND M ORISHIMA , Y. We learn better together: enhancing elearning with emotional characters. In CSCL ’05: Proceedings of th 2005 conference on Computer support for collaborative learning (2005), International Society of the Learning Sciences, pp. 408–417. [80] M ANNAN , M., AND VAN O ORSCHOT, P. C. On instant messaging worms, analysis and countermeasures. In WORM ’05: Proceedings of the 2005 ACM workshop on Rapid malcode (New York, NY, USA, 2005), ACM, pp. 2–11. [81] M ARK M ONITOR. Rock Phishing: Characterization of the Threat and Recommended Countermeasures. whitepaper, 2007. http://www.markmonitor.com/resources/docs/wp-rockphish-070824.pdf. [82] M ATTHEW B ROERSMA. Firefox 2 tops ie 7 in anti-phishing study, 15 november 2006, techworld. retrieved april 1, 2009. http://www.techworld.com/security/news/index.cfm?newsid=7353. [83] M AYER , R. E. Multimedia Learning. New York Cambridge University Press, 2001. [84] M ERCHANT R ISK C OUNCIL. Annual e-commerce fraud survey results. Press Release, March 2009. https://www.merchantriskcouncil.org/index.cfm?fuseaction=Feature.showFeature&Featur [85] M ESSAGE A NTI -A BUSE WORKING G ROUP , AND A NTI -P HISHING WORKING G ROUP. Anti-Phishing Best Practices for ISPs and Mailbox Providers. Report, 2006. http://www.apwg.org/reports/bestpracticesforisps.pdf. [86] M ESSAGE L ABS. Messagelabs Intelligence: 2007 Annual Security Report. MessageLabs Intelligence, 2007. http://www.messagelabs.com/mlireport/MLI_2007_Annual_Security_Report.pdf. [87] M ESSAGE L ABS. Messagelabs Intelligence May 2009. http://www.messagelabs.com/intelligence.aspx. Report, May 2009. [88] M ICHAEL S UTTON. A tour of the google blacklist, august 7, 2006. visited jan 1, 2009. http://www.communities.hp.com/securitysoftware/blogs/msutton/archive/2007/01/04/A-T [89] M ICROSOFT C ORPORATION. Consumer awareness page on phishing, 2006. Retrieved Sep 10, 2006. http://www.microsoft.com/athome/security/email/phishing.mspx. 167 [90] M ICROSOFT C ORPORATION. Phishing filter: Help protect yourself from online scams, 2008. http://www.microsoft.com/protect/products/yourself/phishingfilter.mspx. [91] M ILLERSMILES . CO . UK. The web’s dedicated anti-phishing service. Retrieved April 15, 2006, http://millersmiles.co.uk/. [92] M OORE , T., AND C LAYTON , R. Examining the Impact of Website Take-down on Phishing. In eCrime ’07: Proceedings of the 2007 e-Crime Researchers summit (New York, NY, USA, 2007), ACM Press, pp. 1–13. [93] M OORE , T., AND C LAYTON , R. The Consequence of Non-Cooperation in the Fight Against Phishing. In eCrime ’08: Proceedings of the 2008 e-Crime Researchers summit (New York, NY, USA, 2008), ACM Press. [94] M OORE , T., AND C LAYTON , R. Evil Searching: Compromise and Recompromise of Internet Hosts for Phishing. In 13th International Conference on Financial Cryptography and Data Security (February 23-26, 2009 2009). [95] M OORE , T., C LAYTON , R., AND S TERN , H. Temporal Correlations between Spam and Phishing Websites. In 2nd USENIX Workshop on Large-Scale Exploits and Emergent Threats (LEET ‘09) (2009). [96] M ORENO , R., M AYER , R. E., S PIRES , H. A., AND L ESTER , J. C. The case for social agency in computer-based teaching: Do students learn more deeply when they interact with animated pedagogical agents? Cognition and Instruction 19, 2 (2001), 177–213. [97] M Y S ECURE C YBERSPACE . Uniform resource locator (URL), 2007. Retrieved Feb 4, 2007, http://www.mysecurecyberspace.com/encyclopedia/index/uniform-resource-locator-url-. [98] N ATIONAL C ONFERENCE OF S TATE L EGISLATURES. 2007 state legislation relating to phishing. Report, 2007. http://www.ncsl.org/programs/lis/phishing07.htm. [99] N ATIONAL C ONSUMERS L EAGUE. Avoid getting ’hooked’ by phishers, 2006. [100] N ATIONAL C ONSUMERS L EAGUE. A Call for Action: Report from National Consumers League Anti-Phishing Retreat. Report, 2006. http://www.nclnet.org/news/2006/Final%20NC%20Phishing%20Report.pdf. [101] N ET A PPLICATIONS . I NC . . Browser market share q4, 2008. visited jan 1, 2009. http://marketshare.hitslink.com/report.aspx?qprid=0&qpmr=15&qpdt=1&qpct=3&qpcal=1&q [102] N ETCRAFT I NC . Netcraft anti-phishing http://toolbar.netcraft.com/. toolbar. visited jan 1, 2009. [103] N EW M EXICO L EGISLATURE – 2005 SESSION. Sb 720. Law, 2005. http://legis.state.nm.us/Sessions/05%20Regular/final/SB0720.pdf. 168 [104] N EW YORK S TATE O FFICE OF C YBER S ECURITY & C RITICAL I NFRASTRUCTURE C O ORDINATION . Gone Phishing: A Briefing on the Anti-Phishing Exercise Initiative for New York State Government. Aggregate Exercise Results for public release, 2005. [105] H OMELAND S ECURITY, U. D., C OUNCIL , S. I. I. T. T., AND A NTI - PHISHING WORKING G ROUP. The crimeware landscape:malware, phishing, identity theft and beyond. Report, 2006. http://www.antiphishing.org/reports/APWG_CrimewareReport.pdf. OF THE [106] O N G UARD O NLINE. Phishing quickfacts, 2008. [107] PAN , Y., AND D ING , X. Anomaly based web phishing page detection. Computer Security Applications Conference, Annual 0 (2006), 381–392. [108] PARNO , B., K UO , C., AND P ERRIG , A. Phoolproof phishing prevention. In Proceedings of the 10th International Conference on Financial Cryptography and Data Security (FC’06) (Feb. 2006). [109] P ENDLETON , B., X IANG , G., AND H ONG , J. Augmenting the Crowds: Fighting Phishing on a Budget. Under Submission, 2009. [110] Q UINN , C. N. Engaging Learning: Designing e-Learning Simulation Games. Pfeiffer, 2005. [111] R AMACHANDRAN , A., AND F EAMSTER , N. Understanding the network-level behavior of spammers. In SIGCOMM ’06: Proceedings of the 2006 conference on Applications, technologies, architectures, and protocols for computer communications (New York, NY, USA, 2006), ACM, pp. 291–302. [112] R EBBAPRAGADA , N. New VoIP Phishing Scams. PCWorld, http://blogs.pcworld.com/staffblog/archives/001921.html. 2006. [113] R EPENNING , A., AND L EWIS , C. Playing a game: The ecology of designing, building and testing games as educational activities. In ED-Media, World Conference on Educational Multimedia, Hypermedia & Telecommunications (2005), Association for the Advancement of Computing in Education. [114] R EYNA , V. F., AND FARLEY, F. Risk and rationality in adolescent decision making: Implications for theory, practice, and public policy. Psychological Science in the Public Interest 7, 1 (2006), 1–44. [115] ROSENTHAL , R., AND ROSNOW, R. L. Essentials of Behavioral Research, third ed. McGraw Hill, New York, NY, USA, 2008. [116] ROSS , B., JACKSON , C., M IYAKE , N., B ONEH , D., AND M ITCHELL , J. C. Stronger password authentication using browser extensions. In Usenix security (2005). 169 [117] ROWE , B. R., AND G ALLAHER , M. P. Private sector cyber security investment: An empirical analysis. In WEIS 2006 - Fifth Workshop on Economics of Information Security (2006), pp. 18–41. http://weis2006.econinfosec.org/docs/18.pdf. [118] RUSCH , J. J. Phishing and Federal Law Enforcement. Presentation at ABA, 2004. http://www.abanet.org/adminlaw/annual2004/Phishing/PhishingABAAug2004Rusch.ppt. [119] S. DYNES , H. B RECHBUHL AND M. E. J OHNSON. Information Security in the Extended Enterprise: Some Initial Results From a Field Study of an Industrial Firm. In Fourth Workshop on the Economics of Information Security (2006), Harvard University. [120] S ALKIND , N. J. Encyclopedia of Measurement and Statistics. Sage Publications, 2006. [121] S CHECHTER , ., D HAMIJA , ., O ZMENT, ., indicators. sp 00 (2007), 51–65. AND F ISCHER , . The emperor’s new security [122] S CHNEIDER , F., P ROVOS , N., M OLL , R., C HEW, M., AND R AKOWSKI , B. Phishing protection: Design documentation. visited jan 1, 2009. https://wiki.mozilla.org/Phishing_Protection:_Design_Documentation. [123] S CHNEIER , B. Inside risks: semantic network attacks. Commun. ACM 43, 12 (2000), 168. [124] S ENDER P OLICY F RAMEWORK S PECIFICATIONS (RFC 4408). http://www.openspf.org/Specifications. Visited jan 1, 2009. [125] S HENG , S., H OLBROOK , M., K UMARAGURU , P., C RANOR , L. F., AND D OWNS , J. Who falls for phish? a demographic analysis of phishing susceptibility and effectiveness of interventions. In Submission: CHI ’10: Proceedings of the SIGCHI conference on Human factors in computing systems (2010). [126] S HENG , S., K UMARAGURU , P., ACQUISTI , A., C RANOR , L., AND H ONG , J. Improving phishing countermeasures: An analysis of expert interviews. In eCrime Researchers Summit 2009 (Tacoma, WA, USA, 10 2009). [127] S HENG , S., M AGNIEN , B., K UMARAGURU , P., ACQUISTI , A., C RANOR , L. F., H ONG , J., AND N UNGE , E. Anti-phishing phil: the design and evaluation of a game that teaches people not to fall for phish. In SOUPS ’07: Proceedings of the 3rd symposium on Usable privacy and security (New York, NY, USA, 2007), ACM, pp. 88–99. [128] S HENG , S., WARDMAN , B., WARNER , G., C RANOR , L., H ONG , J., AND Z HANG , C. An empirical analysis of phishing blacklists. In 6th Conference in Email and Anti-Spam (Mountain view, CA, July 16 - 17 2009). [129] S LOVIC , P. The Perception of Risk. The Earthscan Risk in Society Series. Earthscan Publications Ltd, 2000. 170 [130] S TATE OF N EW YORK L EGISLATURE – 2007 SESSION. http://assembly.state.ny.us/leg/?bn=A08025&sh=t. A.b 8025. Law, 2007. [131] S YMANTEC . I NC . . Symantec Global Internet Security Threat Report. Tech. rep., Symantec., 2009. http://eval.symantec.com/mktginfo/enterprise/white_papers/b-whitepaper_internet_sec [132] T HERMOS , P., AND TAKANEN , A. Securing VoIP Networks: Threats, Vulnerabilities, and Countermeasures, first ed. Addison Wesley Professional, 2007. [133] US H OUSE OF R EPRESENTATIVES. Internet spyware (ispy) prevention act of 2004. H. R. 4661, 2004. http://thomas.loc.gov/cgi-bin/query/D?c108:5:./temp/~ mdbsui94q6::. [134] US H OUSE OF R EPRESENTATIVES. Internet spyware (ispy) prevention act of 2005. H. R. 744, 2005. http://thomas.loc.gov/cgi-bin/query/D?c109:11:./temp/~ mdbsGYDwP7::. [135] US H OUSE OF R EPRESENTATIVES. Internet spyware (ispy) prevention act of 2007. H. R. 1525, 2007. http://thomas.loc.gov/cgi-bin/query/D?c110:7:./temp/~ mdbs2yQuGo::. [136] VARIAN , H. Managing Online Security Risks, June 2000. New York Times. http://people.ischool.berkeley.edu/~ hal/people/hal/NYTimes/2000-06-01.html. [137] VIRGINIA ACTS OF ASSEMBLY – 2005 SESSION. Chapter 827. Law, 2005. http://leg1.state.va.us/cgi-bin/legp504.exe?051+ful+CHAP0827. [138] WARD , M. Criminals exploit net phone calls. http://news.bbc.co.uk/2/hi/technology/5187518.stm. BBC News, 2006. [139] W U , M. Fighting Phishing at the Interface Level. PhD thesis, Massechusets Institute of Technology, 2006. http://groups.csail.mit.edu/uid/projects/phishing/minwu-thesis.pdf. [140] W U , M., M ILLER , R. C., AND G ARFINKEL , S. L. Do security toolbars actually prevent phishing attacks? In CHI ’06: Proceedings of the SIGCHI conference on Human Factors in computing systems (New York, NY, USA, 2006), ACM Press, pp. 601–610. [141] W U , M., M ILLER , R. C., AND L ITTLE , G. Web wallet: preventing phishing attacks by revealing user intentions. In SOUPS ’06: Proceedings of the second symposium on Usable privacy and security (New York, NY, USA, 2006), ACM Press, pp. 102–113. [142] X, S. Inside the Spam Cartel, first ed. Syngress Publishing, Inc., Rockland, MA , USA, 2004. 171 [143] X IANG , G., AND H ONG , J. An Adaptive Shingling-based Approach using Search Engines for Zero False Positive Phish Detection. Under Submission, 2009. [144] X IANG , G., AND H ONG , J. I. A hybrid phish detection approach by identity discovery and keywords retrieval. In WWW ’09: Proceedings of the 18th international conference on World wide web (New York, NY, USA, 2009), ACM, pp. 571–580. [145] Y E , Z. E., S MITH , S., AND A NTHONY, D. Trusted paths for browsers. ACM Trans. Inf. Syst. Secur. 8, 2 (2005), 153–186. [146] Y EE , K.-P., AND S ITAKER , K. Passpet: convenient password management and phishing protection. In SOUPS ’06: Proceedings of the second symposium on Usable privacy and security (New York, NY, USA, 2006), ACM Press, pp. 32–43. [147] Z HANG , Y., E GELMAN , S., C RANOR , L., AND H ONG , J. Phinding Phish: An Evaluation of Anti-Phishing Toolbars. In Proceedings of the ISOC Symposium on Network and Distributed System Security (2007), Internet Society. [148] Z HANG , Y., H ONG , J. I., AND C RANOR , L. F. Cantina: a content-based approach to detecting phishing web sites. In WWW ’07: Proceedings of the 16th international conference on World Wide Web (New York, NY, USA, 2007), ACM Press, pp. 639–648. [149] Z ONE A LARM. Smart Defense System, 2004. http://smartdefense.zonealarm.com/tmpl/Advisory 172 APPENDIX Appendix I: List of Recommendations This section lists the full list of recommendations that I discussed with experts during my interviews. A.1 Recommendations This section makes a set of recommendations based on the insights from the phishing analysis (chapter 2) and preliminary stakeholder analysis (chapter 3). The recommendations are categorized into the following framework: prevention, detection, block emails/websites, shutdown, and warn user (see graph 2.9). The overall objectives are: 1. Catch phishers before they launch attacks 2. Detect attacks as early and accurately as possible 3. Block phishing emails at mail gateways Figure A.1 Taxonomy of phishing technical countermeasures 173 4. Takedown phishing websites as soon as possible 5. Improve mutual authentication between financial institutions and consumers 6. Minimize money laundering due to phishing 7. Warn and educate users effectively In the section that follows, I outline recommendations to achieve these objectives. A.1.1 Prevention As shown in Figure 2.9, the first step to fight phishing is to prevent attacks before they materialize. Effective law enforcements will reduce the phishers incentive to commit crimes, and will lower the probability of phishers launch attacks after securing personal and corporate resources. Corporate that handle incidents better will be less attractive targets for phishers, and finally proactive measures of anti-phishing from registrars will make setting up phishing attacks much harder. We list the recommendations below. A.1.1.1 Recommendations for effective law enforcement 1. Law enforcement: Continue operations to identify and catch phishing gangs such as the Rock Phish gang. As the underground phishing market improves its efficiency, phishing operations will consolidate and a few organizations will be responsible for most of the phishing. It is estimated that phishing gangs such as Rock Phish are responsible for up to to 50% of phishing. Therefore efforts spent on catching them is necessary. In my interview with law enforcement and other experts, I will consolidate their advice on catching Rock Phish. 2. Law enforcement, industrial organization, and academia: Provide a more accurate measure of the loss due to phishing in general and particular incident. There is a lack of data on the monetary losses caused by phishing attacks. It is hard to get for a variety of reasons: banks do not know whether a fraud charge is due to phishing or other activities such as 174 dumpster driving or malware and the number of people entering information does not mean that the information is correct and can be used by phishers–attacks may not convert to actual losses because banks have sophisticated fraud systems. Confident estimates are important because it is difficult for law enforcement to open cases if they do not have a good idea of the loss. I suggest three possible directions to gather the data: first collect and preserve forensics data when the phishing server is seized, provide a detailed information about the accounts stolen and collaborate with banks to double check these fraud cases; second, study the internet phishing black market for prices of the stolen goods.1 and lastly, conduct empirical measures, not surveys. Recent efforts by Moore and Clayton [92], Florencio and Herley [36] provide innovative ways to investigate this issue–their methods can be easily shared with law enforcements on a case by case basis to measure the monetary loss both in general and in specific phishing cases. 3. Regulators: Push the adoption of the cybercrime conventions around the world. Criminals work their way through the countries that do not comply with the cybercrime convention. To close the loophole, efforts need to be made for countries to ratify the cybercrime convention–a model regulation framework proposed by the European Union. 4. Law Enforcement: Disrupt the underground black market economy. As mentioned in Chapter 2.2, phishers, spammers, botnet herders, and pay loaders collaborate to commit crimes and make trades in the Internet black market. Efforts to disrupt the Internet black market will sever the criminals ability to connect with each other. The paper by Perrig and Franklin [38] has outlined a few possible ways to disrupt the market. I recommend further research and action in this area. In my expert interviews, I will ask law enforcement experts to comment on these proposals and make suggestions that could enhance phishing law enforcement. 1 Economics predicts that markets at equilibrium supply equals demand. It is therefore possible to infer the loss due to phishing from the prices of these commodities sold on blackmarkets 175 A.1.1.2 Recommendations for securing personal and corporate computing resources for anti-phishing Today, phishing attacks are launched through compromised personal and corporate computers around the world. Spam emails are sent through vulnerable open mail relays and susceptible web forms. Hacked machines host half of the phishing websites. Securing personal and corporate computing environments will make it harder for phishers to launch attacks. Below is a list of security recommendations. 1. Technology Vendors: Protect host files on user computers. Some phishing attacks poison DNS records by altering local DNS look up files (except for Windows Vista). Currently, local hosts files are not protected by Windows or by anti-virus software. Protecting these files will help eliminate the DNS poisoning problem and reduce phishing attacks. 2. Website operators: Check and fix the web form vulnerability for mail injection attacks. Mail injection attacks can compromise web mail forms, a means for spammers to relay mail. CERT or APWG can also help by producing a toolkit to discover this vulnerability. 3. Academic institution, CERT, vendors and law enforcement: Continue research and operations to shutdown botnets. Botnet is the crucial machinery for criminals to launch and evade phishing attacks. Shutting down botnets will not only help eliminate phishing, but a variety of other attacks such as DDOS and spam. However, many have argued that shutting down botnets is not worthwhile for three reasons: vulnerabilities in computers are numerous and it only takes one exploit to control computers; users are careless and are easily fooled into installing malware on their computer; there are hundreds of millions of potentially vulnerable computers connected to the Internet. I think all of these are valid points that acknowledge the difficulty of the task, however, it has been shown in the past that it can be done. 4. Researchers, Vendors: Research into secure patching for vulnerabilities. Many computers become infected because of zero day exploits which hackers reverse engineer a patch 176 and produce exploits and infect computers that are not patched. It can be as little as six hours for an exploit to be created from a patch. Although we would like to see software secure by design, it is unlikely that patches will not be needed. Research into secure patching (possibly using public cryptography) would help alleviate the problem of zero day exploits. 5. CERT or APWG: Produce a list of most frequently hacked websites and notify the website operators of their vulnerability. Provide toolkits and educational resources to help website operators secure themselves. Because about 50% of phishing today is on hacked websites2 , this will give incentive for operators to investigate why the websites are hacked and provide them with tools to fix it. A.1.1.3 Recommendations for improving risk management and incident handling for phishing 1. Institutions: If frequently targeted, review security procedures and security processes and establish phishing countermeasures. If a bank is continually being robbed, it means that the security measures in place are inadequate. In the same vein, if phishers continually target an institution, it means that the security measures at the institution need to be improved. 2. Institutions: Identify a list of high-risk clients and provide education and additional measures to protect them. Clients such as account executives and business account holders will be at special risk due of phishing because of their networth and their inexperience. 3. Banking Regulators: Obtain and monitor statistics of the targeted institutions for fraud losses and press the corporations about their security practices if necessary. As mentioned earlier, there is little data available about fraud losses in banks. Banks do not want to disclose these numbers because they do not have any incentives to do so. Without accurate reporting of these fraud losses, regulators would not know the banks’ performance and would find it hard to provide guidance. Requiring banks to report quarterly fraud losses for 2 According the data we compiled from Phishtank during the two week period in July. 177 regulator review will help the banks examine their internal processes of control and also help them better manage the process. The data may not need to be public. A.1.1.4 Recommendations for proactive measures from registrars 1. Academic Institutions or industry groups: Conduct a study on registrars’ preparedness for phishing and other frauds. Produce best practices for registrars and compile case studies for registrars that prevented phishing. 2. Regulators (ICANN): Provide guidance and help registrars to detect phishing registrations. If necessary, issue security standards about phishing for registrars. A.1.2 Detection The earlier the detection of attacks, the shorter the response time for shutdown and blocking. 1. Email Services: Automatically forward suspected phishing emails to antiphishing services at mail gateway level. Since the email gateway is the first point of contact to phishing emails, phishing emails are freshest here. The difficulty is that mail providers lack incentives to report phishing, because their primary concern is spam. Since most filters do not treat spam and phishing differently, reporting phishing emails at the gateway level means manual work to separate phishing from spam first. 2. Academic institutions or open source community: Provide a good set of open source phishing filters to integrate with spamassassin. There are many email providers on the Internet. While large mail providers can deploy sophisticated email filters, smaller and medium size providers usually rely on open source spam filters such as spamassassin. The standard configuration of spam assassin only catches about 70% of phishing emails [34]. To raise the bar for phishing protection, phishers filters should be released to the public domain for free. 178 A.1.3 Filter email / websites 1. Encourage mail providers to scan for phishing at mail storage. In some instances, doing filtering at the mail storage level is preferable–gathering and updating the phishing email signatures take time and some phishing detection techniques require network query (DNS lookup), which would slow down the filter performance dramatically if implemented at gateway (it takes roughly four seconds to process a 10kb email if running network lookup). There is usually a 12-hour lapse between the time mail is in mail storage to the time mail is downloaded to clients computers [100]. Between these stages, some filtering can be applied and mails can be tagged or removed before the client ever downloads them. However, there may be legal and privacy concerns regarding provider examinations of users personal inboxes. 2. Mail clients could be the next step to combating the problem. Regular software clients such as Outlook and Thunderbird can run some phishing tests and warn users when the emails are opened. The benefits of doing it here are that there would be no privacy and legal concerns, and mail clients have more information about senders and others for sophisticated filtering. 3. Web browser vendors: Continue to improve browser anti-phishing toolbar performances, with a goal to catch 85-95% of phishing URLs within an hour. As shown in Figure 9, Internet Explorer 7 was only able to detect less than 50% of phishing websites within 12 hours, and Safari does not have any phishing protection yet. More efforts here are needed in this area. 4. Email Providers: Support email authentication SPF and DKIM. Although email authentication will not solve the problem of email fraud, it does provide accountability in email when used properly. For companies to adopt these methods, email clients must first support them natively. 179 A.1.4 Shutdown / block phishing websites 1. CERT or APWG: Produce a list of most frequently hacked websites and constantly monitor websites’ security for improvements. Roughly 50% of phishing is hosted on hacked websites. By producing these statistics, website holders will be aware of their vulnerabilities. Whenever websites are hosted on hacked sites, site owners should be directly notified so that they can take it down and fix its vulnerabilities. 2. Registrars: Examine solutions to shutdown and suspend Rock Phish domains quickly. A.1.5 Warn and educate user 1. Email clients: Provide effective and integrated warnings for users about phishing messages, and research ways to better present warnings. 2. Government, education, and industry groups: Educate consumers about the risks of instant messaging networks. A.1.6 Minimize money laundering The final step is to minimize money loss due to phishing. To do this, we need to make it harder for third parties to use stolen credentials to commit fraud, and make it more difficult for phishers to launder money even with stolen credentials. My recommendations are: 1. Financial institutions: Work closely with anti-money laundering communities to ensure that anti-money laundering systems are used to detect phishing related fraud. Antimoney laundering systems have been used worldwide for many years. To the best of my knowledge, they have not been used to detect phishing fraud. I recommend that phishing rules be added to the AML systems and focus on phishing gangs behaviors. In my expert interviews, I will ask their opinions on these issues. 180 2. Regulators (FTC): Launch education campaign to educate the public about mules. Mules are a crucial element in the underground market, as they transfer money or redirect goods to criminals. Many of the money mules are unaware that the activities they engage in are illegal. As companied with phishing, there are few educational materials in the media about money mules. I recommend regulators such as the FTC organize a campaign against money mules The campaign could either be a standalone campaign or a combined campaign. The format could be testimonials, actual police cases, and recommendations on how not to become a money mule3 . 3. Industry association: Study money wiring practice of Western Union and Money Gram, especially their security practices about wiring money outside the country. Western Union and Money Gram are one of the key tools that mules use to transfer money. The system is designed to make money transfer easy, which also makes it easy for criminals. I propose a simple study: to investigate security practice validations and authentications, investigators should visit a dozen local Western Union and Money Gram branches and try to transfer money to Eastern Europe. A.1.7 Other recommendations 1. Financial institutions: Implement better mutual authentication systems. Better mutual authentication means banks can be certain that customers they are dealing with are actually customers, and vise versa. Better-implemented systems will make it difficult for phishers to gain access to accounts even though they may have credential such as usernames and passwords. However, better authentication will not make it impossible to eliminate fraud because we can assume that attackers can gain access to all the credentials that regular customers have (in extreme cases). Although this comes at a higher cost to the attacker, it is not impossible. 2. Academia Continue research on mutual authentication. 3 Recently, Phil H. at Verisign also had the idea of a mule-fool campaign. 181 3. Internet service providers: Implement egress and ingress filtering. 4. Internet service providers: monitor outbound network traffic from unpatched computers and request users to update.
© Copyright 2026 Paperzz