Audio Access to Calendars Andy Brown, Caroline Jay and Simon Harper School of Computer Science University of Manchester Kilburn Building, Oxford Road Manchester, UK [andrew.brown-3 | caroline.jay | simon.harper]@manchester.ac.uk ABSTRACT Keywords The rise of ‘Web 2.0’ has brought a much more interactive aspect to the Web: users are no longer just reading pages, but creating them, modifying them, and interacting with them. The Web is increasingly becoming the preferred means of communication, and particularly booking events and appointments; online personal and corporate diaries allow friends and colleagues to arrange meetings and coordinate activities. Many of these types of online activities require users to perform the apparently simple task of entering a date. For sighted people who have access to pop-up calendars, selecting a date is quick and easy. Unfortunately, this facility is not currently available to people with visual impairments, for whom entering a correctly formatted date can be a difficult and time-consuming task, with mistakes having potentially serious consequences. Here we describe the process by which we designed and evaluated an audio interface for entering dates. An eye-tracking study gave insight into how tabular calendars help sighted people enter dates, This understanding was used to design an audio interface that used the cognitive advantages of the visual design, rather than mimicking the visual representation. Iterative testing was followed by an evaluation using participants with visual impairments that highlighted the problems with manual date entry, and which showed the audio system to be effective and popular. Web 2.0, AJAX, Visual Disability, Eye-Tracking. Categories and Subject Descriptors H.1.2 [Models and Principles]: User/ Machine Systems— human factors, human information processing; H.5.4 [ Information Interfaces and Presentation]: Hypertext/ Hypermedia—user issues, navigation; K.4.2 [Computers and Society]: Social Issues—assistive technologies for persons with disabilities General Terms Human Factors, Experimentation Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. W4A2010 - Technical, April 26-27, 2010, Raleigh, USA. Co-Located with the 19th International World Wide Web Conference. Copyright 2010 ACM 978-1-4503-0045-2 ...$5.00. 1. INTRODUCTION The modern Web is more interactive than ever: users are no longer passive consumers, reading pages as they would a book, newspaper or encyclopedia. Instead, information now flows in all directions, allowing it to be used in a much more conversive way. Not only is it used to communicate with friends, it is widely used for many more formal transactions such as ordering and buying goods and services. Examples for individuals (as opposed to businesses) include buying the weekly shopping, paying taxes, arranging insurance, and booking appointments. Many of these are important and must be done correctly, with the consequences of error ranging from minor to major inconvenience, and including breaking the law; it is therefore vital that users are supported in giving information as accurately as possible, and with the minimum room for error. This is particularly the case for users with visual impairments, as it is more difficult for them to review their input. Here we describe the design, implementation, and evaluation of a non-visual interface for entering dates on a Web form. Pop-up calendars, or ‘date pickers’ are common on forms where a date needs to be input. These typically display one month in a tabular format; figure 1 gives some examples. The month is given at the top, with controls on either side to navigate to the next or previous month; the table below has a header row giving the days of the week, followed by the dates represented by numbers with one week per row. The table cell representing the date currently in the input box is often given visual highlighting. A typical mode of operation (used by all the examples in figure 1) is to have a text input box that, when focused, initiates the appearance of a calendar. The calendar usually appears immediately below the text box, and over other content (hence ‘pop-up’). The user then has the choice of typing the date manually, or selecting the date using the calendar. The latter is usually done with the mouse, although some are keyboard accessible. Once a date has been selected, the calendar disappears. Due to their dynamic nature (i.e., they appear on demand, rather than being ever-present), these pop-up calendars are not generally announced or accessible to users with visual impairments. Supporting non-visual date entry does not appear to have had much consideration, probably due to the availability of an apparently simple alternative: typing the date manually. However, as will be demonstrated in section 5.2, this is not as simple as it might seem. Furthermore, 2. (a) Yahoo! User Interface Library (b) Dojo Toolkit (c) Meteora (d) Uize Figure 1: Examples of JavaScript pop-up Calendars from online ‘widget’ libraries. pop-up calendars are presumably of some benefit to sighted users, so it is reasonable to offer those benefits to all users, particularly considering the importance of inputting data like this correctly. Since the appearance of a calendar is effectively a change to the page, an example of a pop-up calendar was included in research looking at dynamic pages (pages that change without requiring a reload). An eye-tracking study was undertaken that sought to understand how sighted users allocated attention when changes took place; in addition to pop-up calendars, this study [5] also looked at tabs, slide shows, auto-suggest lists [2], expansion buttons and tables. The results with respect to calendars were particularly interesting, however, and prompted the work described below. The initial eye-tracking study (§ 3.1) showed that sighted users nearly always used the calendars, and that the way they did so gave them some advantages over manually typing the date (§ 3.2). In particular, users did not need to be aware of the required format, the risk of typing errors was reduced, and the relationships between dates became more explicit. A direct translation of the calendar to audio, however, would not offer all of those advantages to screen reader users. Indeed, the difficulties of interacting with tables in audio are well documented (see § 2). Instead, the understanding of the cognitive benefits tables offer sighted users allowed an interface to be designed that offers those same benefits to users browsing non-visually (§ 4). Iterative testing with screen reader users resulted in a design that evaluation with users with visual impairments showed to be effective and popular (§ 5). BACKGROUND The typical use of calendars to assist with date entry has been described above, where it was noted that they may be considered ‘dynamic’ content, as they appear (and disappear) without reloading the page. This type of behaviour is one of the key characteristics of ‘Web 2.0’, where such chunks of interactive content are known as ‘widgets’. Screen reader users, however, are suffering from a disadvantage with this type of content as their assistive technologies lag behind these developments. A review of how users of different assistive technologies deal with dynamically updating pages [1] showed that older browser / screen reader combinations apparently did not even respond to some updates, in particular those which occurred automatically. Even newer technologies often did not notify the user that the page had changed (it should be noted that technology is changing rapidly, even if users don’t always have the most up-to-date version). The situation is improving, however, notably through the development and deployment of the Accessible Rich Internet Applications Suite (ARIA) from the W3C’s Web Accessibility Initiative1 . The essence of WAI-ARIA [3] is that adding semantic metadata will allow the user-agent/assistive technology partnership to render the rich information appropriately. Although there are situations where it is difficult to find an appropriate solution through ARIA (see, for example, [15, 2]), there are also many situations where it can be applied effectively. The difficulty, however, is designing an interface that allows the ‘widget’ to be used effectively and efficiently. A common approach in the field of accessibility is to make the information available through audio by direct translation. We contend that this is not always appropriate, indeed that it can often be a poor choice. The case of pop-up calendars is a good example: direct translation would give the updated content to the user to browse, and present this as a table. The user could move between the cells of the table until the desired date is found. Tables are, however, difficult structures to navigate non-visually. As Wright [17] has pointed out, tables have no aural equivalent and they can be considered as one of the natural functions of written language. The fundamental problem with presenting them in an audio interface is that tables are multi-dimensional, whereas the audio information stream is essentially one-dimensional [18]. A particular problem is gaining an overview (see, for example, [8]), although this should not apply to calendars as most people are aware of the general layout. However, it is possible that presenting a calendar as a table introduces as many difficulties to the user as manual date entry. If direct translation is not appropriate, what approach should be taken? Clearly users with visual, or other, impairments must be able to access the same raw information as sighted users, but the precise details about how that information is presented need not be maintained. Thus, in the case of calendars, it is important to give users equivalent information — which dates fall on which days of the week, how many days in each month, etc. — but this does not necessarily mean giving them monthly tables of dates organised by day and week. A better approach than direct translation is to develop an understanding of how the information benefits the user (its cognitive benefits), then build 1 Introduced at http://www.w3.org/WAI/intro/aria.php a mechanism for presenting it around that understanding, i.e., attempt to replicate those benefits through the audio interface. As an aside, Wright also notes that tables are cleverly convoluted lists [17], and this is particularly clear in the case of calendar tables. This suggests that, while a tabular arrangement might suit sighted users, something closer to the list arrangement might suit users with visual impairments. However, given that most information on the Web has been designed for an audience of sighted users, a natural first stage in developing an understanding of the cognitive benefits is to examine, in detail, how those users interact with the information. Eye tracking has been used to investigate cognitive processes for over 100 years [12], and is a powerful tool when used in conjunction with Web pages, enabling us to determine those areas that are most salient (attract the most fixations), and those that receive little attention. Eye tracking’s most obvious application is in improving the standard design and layout of Web pages, and evaluating their usability [13]. Studies have also examined the saliency of items on a page under varying conditions (e.g. [4, 10, 14]), how eye movements vary according to information scent [11] and how looking for a menu is influenced by page complexity and prior expectations [9]. Eye-tracking is thus an excellent tool to help develop an understanding of the benefits of certain forms of presentation. 3. EYE-TRACKING EXPERIMENTS In order to gain insight into how sighted users deal with various forms of dynamic information, an eye-tracking experiment was performed. The study investigated how people allocate visual attention to dynamic content in two situations: completing short, directed tasks; and browsing. Each of the directed tasks used in the experiment was designed to prompt interaction with an element containing dynamic content, with the goal of gathering data about how people perceive and use the content in that particular situation. To ensure the data have external validity, the tasks were chosen to reflect what a person might reasonably choose to do with such content. The procedure and results for the complete experiment are reported in full in a technical report [5]; this paper describes only those sections that gave data on how participants interacted with calendars. 3.1 Method 17 male and 13 female volunteers, aged between 18 and 34, took part in the study, which took approximately 30 minutes, and for which they were paid £5. All participants used the Internet on a daily basis. Participants sat in front of a 17” monitor with a built in Tobii 1750 eye tracker, and Tobii Studio Professional edition eye gaze analysis software was used to record and analyse the data. During the experiment participants were required to enter two dates on the Kayak site. Kayak is a travel site incorporating a flight search engine. In order to accurately document user interaction with the site, each participant completed three tasks in order, in a single visit. The first was: ‘Search for a flight for one person from Manchester to New York, leaving on <date> and returning a week later’. Subsequent tasks were to find the cheapest flight and the last flight of the day; these tasks involved interaction with a variety of dynamic content, but not calendars. Departure and return dates are set using text input fields; when these receive focus, a calendar appears below the box (and obscuring other content) showing two months arranged in tables (Figure 2). Although only one example of a Calendar was used in the experiments, the consistency of form (see Figure 1, or wall calendars) is such that we believe that these results are generalisable (the appearance of one month or two is not significant — the results demonstrate that it is the basic layout of a month that is important; see §3.2). In all cases, interaction was documented with live Web sites, as this was necessary for many of the dynamic features to function, meaning that all participants viewed slightly different information2 . 3.2 Results When selecting dates for the flight, in four cases, the dates were already correct, so participants simply checked them visually and continued. Moving the cursor to the date field, or clicking the symbol next to it, prompted a calendar to appear (see Figure 2). One participant viewed the calendar when it appeared, but chose to edit the dates directly. Only two participants seemed to ignore it completely. The remaining 23 participants used the calendar. Of these, 14 requested it directly by clicking the symbol to the right of the date field. 9 participants clicked in the date field initially to edit the departure date, but used the calendar when it appeared, and of these, 6 then went on to click the symbol, to request the calendar, for selecting the return date. The eye-tracking not only gave information about whether participants looked at the calendar, but also about the way in which they used it. When selecting the departure date, they quickly scanned and located the required day. When selecting the return date, specified to them as ‘a week later’, people used the layout of the calendar to work it out. Participants treated the calendar as a table, running their mouse and/or gaze along the row underneath the departure date (which was highlighted in red), before selecting the date directly underneath, or focusing on the departure date, and then moving down the same column by one row (see Figure 3). 3.3 Analysis It can be seen that the tabular layout is organising the dates in a way that makes relationships between dates more explicit, and allows people to jump between them at different levels of granularity. The day is the smallest unit in the calendar, and is represented by a simple number. Context can be obtained by glancing at the top of the column, to give the day of the week, or the top of the table, to give the month and year. Days are grouped into rows, each representing a single week, and weeks are grouped into tables, each representing a month. Thus, when viewing one date, viewing a date one week earlier (or later) requires moving up (or down) a cell in the column. Moving a month requires shifting one’s gaze to the next table, and possibly activating the control to display other months. Finally, and perhaps most importantly, to select a date the user simply clicks it — there is no need to consider the format in which it should 2 Note that this also means that the pages are no longer the same structure, and that the nature of some of the dynamic updates has changed or may change. be entered. We may summarise the main advantages as follows: • Entering a date does not need the user to find, or guess, the date formatting used by the designer. • The possibility for typing errors is reduced. A typical date requires 8 numerals and 2 separators; selecting a date by mouse requires a single click. This could be a particularly significant benefit for users with motorimpairments [16], as multiple actions with potential for typing errors are replaced by a single action with the potential for pointing or clicking errors [7]. • Relationships between dates become explicit: that the 13th day of the month is exactly one week after the 6th can be identified rapidly by their relative position (one above the other) in the table. This reduces the need for simple, but error-prone, mental arithmetic. 4. Figure 2: A participant selecting the departure date from the calendar. The calendar appears when the participant edits the date field, or clicks the symbol to the right of the date field. PRESENTING CALENDARS IN AUDIO The benefits that calendars offer sighted users could also be expressed in terms of the problems caused by needing to type a date in a box, namely: uncertainty about the format; the possibility of typing errors going unnoticed; and the need for mental calculations (or referral to another calendar) to determine the day of the week, or relationships. These problems are as valid for users with visual impairments as they are for those who are sighted; indeed, it could be argued that they are more important. Glancing is more difficult when interacting non-visually, so reading formatting cues (e.g., dd/mm/yyyy under the input box) is more difficult, particularly as the user may not be aware of the existence of the cue. For the same reason, it can also take more effort for users with visual impairments to examine another calendar for help. Finally, it should be noted that checking one’s input can also be more difficult when interacting through audio, again due to the difficulties of glancing. Considering that we wish to replicate the benefits that a calendar for selecting dates offers sighted users, it is possible to enumerate the goals of any non-visual implementation: 1. Automate formatting. 2. Minimise potential for typing errors. 3. Make relationships more explicit. Implementing an interface that fulfils these goals should help screen reader users to enter dates more quickly and with fewer errors than typing manually. 4.1 Figure 3: A participant selecting the return date from the calendar. The participant locates the departure date (21, highlighted), and then moves down one row to select the date a week later. Implementation In order to test a set of mappings for dealing with dynamic content generally, a proof-of-concept system has been developed for evaluation. This is a self-voicing extension to the Firefox Web browser based on the Fire Vox extension3 . This system includes an interface for dealing with calendars. 4.1.1 Introduction Since the date entry interface is designed as part of a system, it is necessary to introduce this system as a whole before describing the details of how users entered dates. The 3 http://www.firevox.clcworld.net/ extension was a prototype screen reader designed to test a set of rules for presentation and interaction with dynamic information in Web pages. This was achieved by a threestage process: 1. Detecting page changes then clustering them into meaningful updates. 2. Classification, according to the attributes of the update and the user’s activity. 3. Presentation, applying the rules appropriate for the class. One particular class was the pop-up calendar. These updates involve dynamic insertion of a calendar onto the page, and occur when users move focus to a date field (i.e., the update is initiated by the user) or when an icon is selected (i.e., the update is requested by the user). Before describing the detection, classification, and presentation of pop-up calendars, it is necessary to introduce the user interface. In particular we describe the methods for navigating around a page. The user interface for the experimental prototype was changed from that used in the standard Fire Vox, on the basis that this was not particularly intuitive, and might take more learning time than would be appropriate in a short (less than two hours) evaluation. The replacement interface was loosely based on that used in the Orca screen reader for the Gnome desktop on Linux. Navigating around a page can be done in two ways: sequential navigation, or elementbased navigation. Sequential navigation allows users to move from element to element as they appear in the DOM (a depth-first traversal of the tree). This can be done at three levels: element, sentence or word.Moving at these three levels of granularity is achieved using the number pad keys found on a typical desktop computer keyboard. Each row relates to a different level of granularity: the top row moves by element, the middle by sentence, and the bottom by word. The columns enable the user to move backwards (left column) or forwards (right column), or to query their current location (middle column). Element-based (or structural) navigation is a method for jumping between elements of a certain type. For example, jumping between, and reading, the headings on a page can be used to give an overview of its content. This type of navigation is achieved by using character keys (e.g., ‘H’ moves between heading elements) when in structural navigation mode. This mode may be toggled between on, off, and automatic. In automatic mode, structural navigation is enabled unless the focus is in a form element requiring typing, such as an input box. 4.1.2 Calendar Initiation The non-visual implementation for pop-up calendars uses some simple heuristics to detect the update type, then employs the grid layout of the navigation keys to recreate some of the benefits of the table layout typical in these updates. An update is assumed to be a pop-up calendar if it appears when entering a text field that has some characteristic of a date. Characteristics include containing text formatted as a date, or having a label, name, or id containing ‘date’. Presentation involves a change of mode, and starts with the announcement ‘Date Entry’ followed by the date in the field (if no date is present, today’s date is spoken). There is also a command to allow this mode to be entered manually. Having identified that the user has entered a field requiring date input, the next stage of the process is to determine the format required. This is done using a set of heuristics. The first stage is to identify the separator, if there is one. Once this has been done it is necessary to determine which segments represent the day, month and year. The date is compared with two patterns, representing the ‘British’ format (day, month, year) or ‘American’ format (month, day, year) using the following regular expressions (assuming the delimiter has been identified as a colon): UK: ^(0?[1-9]|[12][0-9]|3[01]): (0?[1-9]|1[012]): (19|20)?[0-9][0-9] US: ^(0?[1-9]|1[012]): (0?[1-9]|[12][0-9]|3[01]): (19|20)?[0-9][0-9] If only one of these matches, that format is assumed, otherwise an additional test is performed, based on the possibility that the date is today. The segments of the day and month in the field are compared to those of the current date; if both are identical, no conclusions can be drawn; similarly if only one matches. If both match, however, the assumption is that the date is today, and whichever format matches is used. Finally, if none of these are conclusive, the format last used is assumed; if none exists, a default can be used based on user preferences. These rules undoubtedly have scope for further generalisation and refinement. 4.1.3 User Interaction Once the calendar has been initiated, and announced, the user may then change the date a day, week, or month at a time, either forwards or backwards; this is done using the number pad keys, with the top row moving months, middle row moving weeks, and bottom moving days. The current day, date, or month is spoken using the middle column, while the left and right columns move backwards and forwards in time respectively. The ‘enter’ button puts the selected date into the input field and the ‘escape’ button exits the calendar mode without modifying the input field. Similarly, typing the date manually (using the numbers above the character keys on the keyboard) exits this mode and allows ‘normal’ operation using text input. Speech feedback is kept relatively terse. On entering the mode the date is spoken in full, e.g., ‘Friday 21 June 2009’ ; after this the default when navigating is to speak only the day of the month. If, however, a month boundary is passed, the new month will be spoken, e.g., moving a day at a time ‘30’, ‘31’, ‘1 February’, ‘2’. The same applies when passing year boundaries. When the user queries the current date, key 2 gives the day of week and day of month (‘Friday 21’ ), key 5 gives the full date (‘Friday 21 June 2009’ ), and key 8 gives the month and year (‘June 2009’ ). Figure 4 shows these commands, as mapped in the evaluation, with examples of the output. Finally, setting the date is done with the ‘Enter’ button, with confirmation including the full date (‘Date input set to Friday 21 June 2009’ ); at this point, date mode is left and the user can continue normal navigation. The ‘Escape’ button exits date entry mode with no change to the field (even if the user has changed the date). focus on the tasks relevant to date entry. Only the results of the date entry task are discussed. Previous Month 28 January Current Month March 2010 Next Month 28 March Previous Week 21 Current Date Wednesday 28 February 2010 Next Week 7 March Previous Day 27 Current Day Wednesday 28 Next Day 1 March Figure 4: The user interface. A 3 × 3 grid of keys maps to 9 commands. Italicised text gives the output for each command if the date is currently set at February 28 2010. Referring back to the requirements (section 4), it can be seen that they have been fulfilled in the following ways: 1. Formatting is automatically detected and applied to the selected date. 2. Navigation keys are used to move between dates, with a brief confirmation of the new date for each change. This feedback minimises the consequences of any typing error, demanding only another key press to correct. 3. Navigation by week and month preserves some of the relationships that the visual tabular presentation makes explicit. For example, moving one week at a time allows the user to browse all dates that lie on the same day of the week. Furthermore, it is possible to identify some other aspects of this interface that potentially improve upon the visual form. For example, moving by larger units, such as a month, preserves the values of the day, so moving from the fifteenth of May to the fifteenth of September requires four key presses. Performing this visually requires attention to move away from the date (15) to the next month button, pressing this until September appears, then locating the fifteenth of this month. This is arguably a slower and more complex operation. 5. EVALUATION In the same way that the implementation was integrated with a more general system for Web browsing and interaction, evaluation of this interface for date entry was performed as part of an evaluation of the system as a whole [6]. The procedure for the whole experiment is described, with 5.1 Method The aim of the evaluation was to test whether the interaction metaphors developed in the SASWAT project provided better access to dynamic micro-content than screen readers do at present. The study used a within-subjects design, in which all participants completed a series of tasks as part of a holiday booking process on the HCW Travel Company Website (specially designed for the evaluation), once using the Firefox browser with the Fire Vox screen reading extension4 (the base case browser) and once using the Firefox browser with the SASWAT screen reading extension (the SASWAT browser). The Fire Vox browser was chosen as a baseline because it provides access to dynamic micro-content at a level similar to many proprietary screen readers, including JAWS. In brief, most screen readers currently allow access to new content if the user is able to navigate to it, but do not inform the user that the content has changed. The navigation commands (see Section 4.1.1) were identical in both browsers. The differences between the behaviour of the two browsers with respect to date entry are detailed in Section 5.1.3. The goal of the evaluation was to give participants an opportunity to compare the behaviour of both browsers whilst interacting with a variety of dynamic micro-content, so they could later specify which they preferred in a structured interview. As such, they completed the same tasks, on the same website, twice. To control for practice effects, half the participants completed the holiday booking tasks using the base case browser first, the other half using the SASWAT browser first. Two participants (p1m and p2m — details of participants are given in section 5.1.1) completed the tasks on the ‘full’ version of the HCW Travel Website5 that had been used during the iterative evaluation. It was noted that this obliged the participants to spend some time navigating between different elements on the page, so the remaining participants completed the tasks on a reduced version6 , which contained identical dynamic micro-content, but fewer (superfluous) links and images, and a reduction in the amount of text. The evaluation was run on a MacBook Pro running OS X 10.5.8 Leopard. The evaluation was conducted with the Mac system voice ‘Alex’ at the speaking rate ‘normal’. Audio output was provided by the on-board speakers. Participants used a USB Dell keyboard to navigate and provide typed input. The session was recorded with a digital voice recorder. The evaluation sessions were run in meeting rooms at four locations: the Human-Centred Web Lab at the University of Manchester; Macclesfield Eye Society, Macclesfield; Walthew House, Stockport; the Disability Information Bureau, Macclesfield. 5.1.1 Participants Participants were recruited in a variety of ways. P1m was a member of staff and p2m and p3m were students 4 http://firevox.clcworld.net/ http://hcw.cs.manchester.ac.uk/research/saswat/ experiments/hotels/book.php 6 http://hcw.cs.manchester.ac.uk/research/saswat/ experiments/hotels/simple/intro.php 5 Participant P1m P2m P3f P4f P5f P6m P7m P8f P9m P10f P11m P12f Impairment profoundly blind profoundly blind partially sighted registered blind profoundly blind registered blind profoundly blind registered blind profoundly blind registered blind partially sighted registered blind Onset adulthood birth birth childhood adulthood birth adulthood adulthood birth adulthood adulthood childhood Table 1: Nature and onset of visual impairments of participants in the final evaluation of the SASWAT Web browser. Note that all participants who are marked as registered blind had some residual sight. Participant P1m P2m P3f P4f P5f P6m P7m P8f P9m P10f P11m P12f Assistive technology JAWS JAWS ZoomText JAWS JAWS JAWS JAWS ZoomText (no audio) JAWS ZoomText (no audio) Enlarged fonts JAWS Browser IE IE IE IE IE IE IE IE IE IE IE Firefox Table 2: The assistive technology and Web browser normally used by the participants in the final evaluation of the SASWAT Web browser. browsed the Web using the screen reader JAWS (v. 10 or below). One participant used the screen magnifier ZoomText with audio and two used ZoomText without audio. P11m used a standard desktop computer set-up with a large font size, but had had experience of using JAWS at college. Table 2 describes the computer set-up used by the participants. Most participants browsed the Web every day. P4f browsed the Web weekly and p10f monthly. P7m had only browsed the Web twice before, for a short period in college. Common reasons for Web browsing included work, study, shopping and email (see Table 3). 5.1.2 Procedure The investigator read out the Information Sheet and Consent Form, and when participants had indicated that they were happy with the information they had received, they signed the Consent Form. All participants signed the form themselves, with the exception of p2m, for whom a second investigator signed on his behalf. Participants were shown how to use the browser and given the chance to practice using the navigation commands on a shortened version of the Fire Vox User Manual home page. They then completed the tasks with each browser, and answered questions about their experience in the structured interview. Each session was audio recorded. The information sheet and script used for the experiment (including the interview questions) and transcripts of all the session recordings are available to download with the evaluation Technical Report [6]. 5.1.3 Task The task involving date entry was found on the booking page of the Web site. This page allowed participants to enter their destination and departure and return dates. A link was also present that, when selected, expanded a section of the page to display instructions. After viewing these instructions, and entering their destination, participants were given the following task: Set the departure date to the 1st January 2010, and the return date to the 2nd February 2010. at the University of Manchester who were contacted via word of mouth (they had taken part in previous studies and expressed an interest in participating in future ones). These participants had provided feedback in informal iterative evaluation sessions during development, so had thus had experience of using an earlier version of the SASWAT Web browser. The remaining participants were recruited through advertisements placed in Macclesfield and Stockport talking newspapers, and circulated by staff at Macclesfield Eye Society and Walthew House Deaf Blind Services in Stockport. P4f, p5f, p6m, p7m and p11m were service users and p8f, p9m and p10f were volunteers/staff at either Macclesfield Eye Society or Walthew House. P12f was an IT teacher who worked with blind students. P1m, p2m, p3f, p6m and p12f were in the 20-45 age group; the remaining participants were over 45. Participants received an honorarium of £20 for taking part. Three of the participants were partially sighted, four were registered blind with some residual vision and five were profoundly blind. Table 1 summarises the nature and onset of the visual impairments of participants. All participants normally used Windows, and the majority The ways in which the site appeared to the user when performing this task are described below. The visual presentation describes what a sighted user would see; the base case presentation describes what occurred when the simple version of the system was used, and represents a typical method for entering a date; the SASWAT presentation describes again how the interface being tested works. Visual presentation: When the focus entered one of the date fields a pop-up calendar appeared. This was from the Dojo toolkit, styled in the form shown in Figure 1(b). Initially the current date was displayed in the date field (in the format dd/mm/yyyy), and this was highlighted in the calendar by a box. Two controls displaying arrow icons could be used to move to the preceding or next month. Clicking on one of the days entered it into the date field. When a departure date had been entered, the return date was automatically set to this too, providing it had been correctly formatted, otherwise it was set to the current date. Base case presentation: When the focus entered the date field, the browser read out the date it contained (the Participant P1m P2m P3f P4f P5f P6m P7m P8f P9m P10f P11m P12f Browsing Frequency daily daily daily weekly daily daily only twice in college daily daily monthly daily daily Browsing Nature work study, email, Facebook shopping, study, email finding things out study, shopping, ‘everything really’ looking up historical information N/A family tree, shopping (with help) work, study, email, occasional shopping family tree (with help) shopping, price comparison, looking things up work, email, looking things up Table 3: The frequency and nature of Web browsing of the participants in the final evaluation of the SASWAT Web browser. text sent to the speech synthesiser was in the form ‘departure date 26/01/2010’, which was spoken as ‘departure date twenty six, one, two thousand and ten’). The calendar appeared, but the user was not notified of this, and it was not possible to access it in audio. To set the date, the user typed it in manually, using the numbers along the top of the keyboard. SASWAT presentation: When the focus entered the date field, the browser read out the date it contained (as above). It then said, ‘date entry’, followed by the date again (in full, e.g, ‘Tuesday January twenty-six two thousand and ten’) to indicate that it was in a mode that allowed users to access the calendar. The user could move to a different date using the commands described in section 4.1.3. Pressing the enter key placed the current date in the date field and the browser said, ‘date input set to <date>’. 5.1.4 Interview After they had completed the tasks with each browser, participants answered a series of questions about their experience browsing the Web, and their opinions of how the two different browsers dealt with the various types of dynamic micro-content. They were asked for details of the nature and length of their visual impairment, the assistive technology they usually used, the frequency with which they browsed the Web and the kind of things they used the Web for. For each item of dynamic micro-content, participants were asked whether they had encountered it before on the Web, to give their opinion on how easy it was to use, whether they would like to be able to access it, and whether they could think of any ways of improving access to it in audio. Finally, participants were asked which browser they preferred overall and why, whether it was useful to be notified when information on a Web page changed, and whether they agreed with the statement, ‘if information on a Web page is available to someone sighted, it should also be available to someone who is visually impaired’. Throughout the interview, participants were encouraged to expand on an answer if they wished to. 5.2 Results P1m, p3f, p4f, p9m, p11m and p12f said they had come across calendars before on the Web, whilst the remaining participants had not. P5f said she had heard of them, but not accessed them. On the whole, participants found the calendar date entry easier to use than typing the date manually. When typing manually, all participants made either formatting or typing errors, or both, or asked the experimenter what format was needed. Only one third of participants (p1m, p2m, p4f and p11m) did not make any typing errors. It should be noted that all participants were using an unfamiliar keyboard, and that errors were often related to the special character separating the days, months and years. No participants conclusively determined the format to use: 4 of the 12 immediately asked the experimenter what the format should be: p1m: ‘So, format?’ p8f: ‘Do you have slashes in between? Dots?’ p9m: ‘And what punctuation after each?’ p11m: ‘What do you put between them? Anything?’ Another three asked indirectly: p2m: ‘Ah right. It’s. . . it must be 01 and then another 01 and then 10, isn’t it?’ p3f: ‘Do you want me to put like a 01?’ p10f: ‘Shall I just do 1 1 then?’ Three further participants (p5f, p6m, p7m) asked or were told after having difficulties or making errors. Even with such help from the experimenter, some participants submitted the form without the correct date in the correct format. When using the audio date selection tool, all participants correctly entered both dates, without error. Positive comments about the calendar ranged from ‘alright, yeah’ (p4f) and ‘ace’ (p1m) to ‘great, that is brilliant, that is really, really good’ (p7m). P10f thought the audio presentation of the calendar was very good, and would like to have access to it, but thought she would be more likely to type the date in. P2m was also keen to access the calendar, but non-committal about whether he would use this or type the date. P1m thought he would probably use the calendar, but would like the option of entering the date manually. The interface already provides this flexibility, by allowing people to select a date using the number pad, or type it in directly using the numbers along the top of the keyboard. P3f, p5f, p6m, p7m, p8f and p12f all said that using the calendar was much easier than entering the date manually. Part of the reason for this may have been because using the keyboard to type the date was difficult. P6m said, ‘going down the keypad to put in the dates and the week that you wanted to go away I found very useful, whereas if I’m using the numbers along the top of the keyboard, I always get dead confused for some reason. I find out where the numbers are, and then having to find the slash... ’Cause you’re moving from one side of the keyboard to the other side and then moving back again.’ It was not only novice users who had trouble entering the date: p3f, p1m and p12f who use computers at work on a daily basis all found typing the date difficult and time-consuming. Most of the participants wanted to have access to the calendar, but p9m felt it was ‘very fiddly... I wouldn’t like to have to mess around for its own sake’. When asked if he would like the option of either using the calendar or typing in the date, he remained negative about the calendar: ‘I think I’d prefer to type it in, ’cause that’s quicker.’ P11m really liked using calendars when looking at the screen (he has partial sight), but thought he would prefer to type the date when using a screen reader, and was not convinced of the benefit of providing calendars in audio. It is worth noting, however, that all of the participants, including those who were sceptical about the need to use the calendar, either asked for help with formatting the date, or entered the date incorrectly (p9m, in fact, did both of these things). P1m and p9m both thought they would prefer to use the cursor keys for accessing the calendar, although p1m could see the advantage of consistently using the number pad for interaction. P5f also mentioned ‘making it easier with the keystrokes’, although she did not specify any particular changes. 6. DISCUSSION Before discussing these results, and their implications, in more detail, it is worth summarising what the tool offered the user: those who declared a preference for manual input, demonstrate that the task of entering a date in a Web form is far from simple. When this tool was provided, although some participants took a little time to become familiar with the commands for changing the date, all completed the departure and return fields successfully. While it could be argued that the problems addressed by this tool could be determined by a detailed analysis of the process, or by performing studies with users who have visual impairments, we believe the process that led us to design such a tool is of particular interest to the assistive technology research community. Our process started by developing an understanding of how sighted users, i.e., those users for whom the original interface was specifically designed, interacted with information, in this case dynamic Web content. This understanding was then used to map the visual interface of the facility to an audio one. Importantly, however, this mapping is not a direct, or naı̈ve, translation but is based on determining how the user is helped by the information. In this case a direct translation would be a table, but while this would help deal with formatting, it misses the other main point, which is that the user really needs to be able to find the required date efficiently, ideally with a quick way of determining its relationship with another date. This approach can be applied to different types of information, and differences between users with congenital or late-onset blindness, or with less severe impairments do not appear to matter, and it can (and should) be combined with other design approaches to provide optimum accessibility. If interpreted correctly, techniques such as eye-tracking can give real insight into how people use information and how to present it non-visually. 7. Technical reports describe the eye-tracking study [5] and final evaluation [6] in more detail. Available with these reports in the Human Centred Web Lab repository are associated files containing the eye tracking data, participant information sheets, consent forms, questionnaires, and transcriptions. 8. • Automatic identification of date input fields. • Automatic date formatting. • Notification of the date currently set in the box. • Keyboard commands to change the date by day, week and month (the precise keys chosen are not that important, although the 3 × 3 arrangement matches the command requirements well). • Terse speech output. • Confirmation of the date set. • The option to type the date in manually. The evaluation gave clear evidence that providing these features to the user is beneficial. The users liked the tool, and most would choose to use it, while those who prefer manual date entry have that option at no extra cost. Furthermore, the problems encountered by most users, including EXPERIMENTAL RESOURCES ACKNOWLEDGEMENTS This work is part of the Single Structured Accessibility Stream for Web 2.0 Access Technologies (SASWAT) project and is funded by the UK EPSRC (EP/E062954/1). As such the authors would like to thank them for their continued support. 9. REFERENCES [1] A. Brown and C. Jay. A review of assistive technologies: Can users access dynamically updating information? Technical Report, University of Mancester, http://hcw-eprints.cs.man.ac.uk/70/, 2008. [2] A. Brown, C. Jay, and S. Harper. Audio representation of auto suggest lists. In W4A’09: Proceedings of the 2009 Cross-Disciplinary Conference on Web Accessibility (W4A), pages 58–61, 2009. [3] B. Gibson. Enabling an accessible web 2.0. In W4A ’07: Proceedings of the 2007 international cross-disciplinary conference on Web accessibility (W4A), pages 1–6, New York, NY, USA, 2007. ACM. [4] L. A. Granka, T. Joachims, and G. Gay. Eye-tracking analysis of user behavior in www search. In SIGIR ’04: Proceedings of the 27th annual international ACM SIGIR conference on Research and development in information retrieval, pages 478–479, New York, NY, USA, 2004. ACM. [5] C. Jay and A. Brown. User review document: Results of initial sighted and visually disabled user investigations. Technical Report, University of Mancester, http://hcw-eprints.cs.man.ac.uk/49/, 2008. [6] C. Jay, A. J. Brown, and S. Harper. Internal evaluation of the saswat audio browser: method, results and experimental materials., January 2010. [7] S. Keates and S. Trewin. Effect of age and parkinson’s disease on cursor positioning using a mouse. In Assets ’05: Proceedings of the 7th international ACM SIGACCESS conference on Computers and accessibility, pages 68–75, New York, NY, USA, 2005. ACM. [8] J. Kildal and S. Brewster. Non-visual overviews of complex data sets. In CHI 2006, Montreal, Quebec, Canada, 2006. [9] J. McCarthy, A. Sasse, and J. Riegelsberger. Could i have a menu please? an eye tracking study of design conventions. In Proceedings of HCI2003, pages 401–414, 2003. [10] B. Pan, H. A. Hembrooke, G. K. Gay, L. A. Granka, M. K. Feusner, and J. K. Newman. The determinants of web page viewing behavior: an eye-tracking study. In ETRA ’04: Proceedings of the 2004 symposium on Eye tracking research & applications, pages 147–154, New York, NY, USA, 2004. ACM. [11] P. Pirolli, S. K. Card, and M. M. V. D. Wege. Visual information foraging in a focus + context visualization. In CHI ’01: Proceedings of the SIGCHI conference on Human factors in computing systems, pages 506–513, New York, NY, USA, 2001. ACM. [12] K. Rayner. Eye movements in reading and information processing: 20 years of research. Psychological Bulletin, 124(3):372–422, 1998. [13] M. Russell. Using eye-tracking data to understand first impressions of a website. Usability News, 7.2, 2005. [14] A. Schiessl, S. Duda, and R. Fischer. Eye tracking and its application in usability and media research. In MMi-Interativ ’03, pages 41–50, 2003. [15] P. Thiessen and C. Chen. Ajax live regions: chat as a case example. In W4A ’07: Proceedings of the 2007 international cross-disciplinary conference on Web accessibility (W4A), pages 7–14, New York, NY, USA, 2007. ACM. [16] S. Trewin and H. Pain. Keyboard and mouse errors due to motor disabilities. International Journal of Human Computer Studies, Volume 50:109–144, 1999. [17] P. Wright. Tables in text: the subskills needed for reading formatted information. In L. J. Chapman, editor, The Reader and The Text, pages 60–69. Heineman, London, 1981. [18] Y. Yesilada, R. Stevens, C. Goble, and S. Hussein. Rendering tables in audio: the interaction of structure and reading styles. SIGACCESS Access. Comput., (77-78):16–23, 2004.
© Copyright 2026 Paperzz