Evaluation of e-Readers - Research School of Computer Science

Evaluation of e-Readers
A Preliminary Analysis
Ujala Rampaul
u5051329
May 2014
Supervisor: Prof. Tom Gedeon
Human Computer Interaction Project: COMP6474
Acknowledgements
I would like to thank Professor Tom Gedeon for his invaluable support, guidance knowledge sharing and
assistance.
Dr. Weifa Liang (Project Coordinator) for his guidance and organising the community of practice meetings.
Dr. Duncan Stevenson for sharing his knowledge and advice on grounded theory.
Leana Copeland and Dr. Sabrina Caldwell for their assistance with the experiments.
Dr. Dingyun Zhu for his helpful comments on analysis of variance
To all the participants who took part in the evaluations, thank you.
E-Reader evaluation
Ujala Rampaul
Page 2 of 66
Abstract
Evaluation of new consumer goods, particularly electronic, are often done from the perspective of analysing the
latest models, comparing their advantages and disadvantages with respect to price. This style of evaluation is
often performed by one of a few product experts on a wide range of features that may not be applicable to each
user.
This project uses a scenario-based approach to evaluate a number of e-readers. The scenario mimics a user who
is interested in a new product or technology and has a limited budget. The objective is to evaluate from a
purchasing perspective the quality and usability of e-readers available within that budget range.
The e-readers evaluated were operated in multiple ways, which suggests that the interface design should cater
for users with different levels of experience with technology. The results of a large user study with over 70
participants, shows that the popular brands do not necessarily produce the best products in terms of e-readers.
E-Reader evaluation
Ujala Rampaul
Page 3 of 66
1 Contents
Acknowledgements................................................................................................................................................................................2 Abstract.......................................................................................................................................................................................................3 1 Introduction....................................................................................................................................................................................7 1.1 Motivation ............................................................................................................................................ 8 2 Literaturesurvey.......................................................................................................................................................................10 3 Methodology................................................................................................................................................................................16 3.1 3.1.1 E-reader .......................................................................................................................................... 16 3.1.2 Scenario .......................................................................................................................................... 17 3.1.3 Measurement of tasks ..................................................................................................................... 17 3.2 4 Experiment 1 ..................................................................................................................................... 16 Experiment 2 ..................................................................................................................................... 18 3.2.1 E-readers ......................................................................................................................................... 18 3.2.2 Questionnaire .................................................................................................................................. 19 Results............................................................................................................................................................................................20 4.1 Experiment 1 ..................................................................................................................................... 20 4.1.1 Turning on the e-reader................................................................................................................... 20 4.1.2 Finding the first document .............................................................................................................. 20 4.1.3 Opening the document .................................................................................................................... 21 4.1.4 Increasing the font size ................................................................................................................... 21 4.1.5 Navigate to specific section ............................................................................................................ 22 4.1.6 Navigate to a second document ...................................................................................................... 22 4.1.7 Readability on the screen ................................................................................................................ 23 4.1.8 E-Reader preference ....................................................................................................................... 23 4.1.9 Pairwise comparison on e-readers with significant differences ...................................................... 24 4.2 Experiment 2 ..................................................................................................................................... 25 4.2.1 Turn on the e-reader........................................................................................................................ 25 4.2.2 Find the first document ................................................................................................................... 25 4.2.3 Open the document ......................................................................................................................... 25 4.2.4 Increase the font size....................................................................................................................... 26 4.2.5 Navigate to specific section ............................................................................................................ 26 4.2.6 Navigate to second document ......................................................................................................... 26 4.2.7 Readability on the screen ................................................................................................................ 27 4.2.8 e-reader preference ......................................................................................................................... 27 4.2.9 Pairwise comparison for Q6 ........................................................................................................... 27 4.2.10 4.3 Pairwise comparison on Q3 – Q6 ............................................................................................... 28 Discussion........................................................................................................................................... 29 E-Reader evaluation
Ujala Rampaul
Page 4 of 66
5 Conclusions..................................................................................................................................................................................33 5.1 Future work ....................................................................................................................................... 33 6 Bibliography................................................................................................................................................................................34 7 Appendices...................................................................................................................................................................................35 7.1 Appendix 1: Technical specification on e-readers .......................................................................... 35 7.2 Appendix 2: Scenario ........................................................................................................................ 38 7.3 Appendix 3: Questionnaire............................................................................................................... 39 7.4 Appendix 4: Questionnaire for Experiment 2................................................................................. 41 7.5 Appendix 5: ANOVA calculations ................................................................................................... 43 7.5.1 Experiment 1 ................................................................................................................................... 43 7.5.2 Experiment 2 ................................................................................................................................... 49 7.6 Appendix 6: Participants comments for experiment 1................................................................... 54 7.7 Contents of second document (Dates 2013) ..................................................................................... 66 E-Reader evaluation
Ujala Rampaul
Page 5 of 66
List of Figures
Figure 1. Growth of ownership of e-readers and tablets in the United States......................................................... 8 Figure 2. Map of common reading devices between this study and previous study ............................................. 14 Figure 3. Map of tasks between this study and previous study ............................................................................. 15 List of Tables
Table 1. Cost of second-hand e-readers .................................................................................................................. 9 Table 2. Matrix of the e-reader randomisation pair for device A. ........................................................................ 17 Table 3. Matrix of touch screen e-reader pairs for device B................................................................................. 18 Table 4. Mean Ranking and Standard Deviation (SD) for turning the e-readers on ............................................. 20 Table 5. Mean Ranking and Standard Deviation (SD) for finding the document ................................................. 21 Table 6. Mean Ranking and Standard Deviation (SD) on opening the document: significant ............................. 21 Table 7. Mean Ranking and Standard Deviation (SD) for increasing font size: significant ................................. 22 Table 8. Mean Ranking and Standard Deviation (SD) for navigation to section 3: significant ............................ 22 Table 9. Mean Ranking and Standard Deviation (SD) for navigation to second document ................................. 23 Table 10. Mean Ranking and Standard Deviation (SD) for readability ................................................................ 23 Table 11. Mean Ranking and Standard Deviation (SD) for e-reader preference: significant ............................... 23 Table 12. Qualitative, quantitative and overall ranking ........................................................................................ 24 Table 13. Mean Ranking and Standard Deviation (SD) for turning the e-reader on. ........................................... 25 Table 14. Mean Ranking and Standard Deviation (SD) for finding a document .................................................. 25 Table 15. Mean Ranking and Standard Deviation (SD) for opening the document ............................................. 26 Table 16. Mean Ranking and Standard Deviation (SD) for increasing the font size ............................................ 26 Table 17. Mean Ranking and Standard Deviation (SD) for navigating to section 3 ............................................. 26 Table 18. Mean Ranking and Standard Deviation (SD) for navigating to second document: significant ............ 27 Table 19. Mean Ranking and Standard Deviation for readability ........................................................................ 27 Table 20. Mean Ranking and Standard Deviation on user preference .................................................................. 27 Table 21. Qualitative, quantitative and overall ranking for Q6 ............................................................................ 28 Table 22. Qualitative, quantitative and overall ranking for Q3-Q6. ..................................................................... 28 E-Reader evaluation
Ujala Rampaul
Page 6 of 66
1 Introduction
In 2007, Amazon launched the Kindle in the United States and it sold out in five and a half hours (Patel, 2007).
Prior to the Kindle launch there was not a huge demand for e-readers as earlier releases were the Rocket eBook
in 1998, the Sony Libriè in 2004 and Sony Reader in 2006. These earlier devices were unsuccessful because
they were very expensive, had too many technological limitations and lacked available content.
The Kindle was different because it was more affordable, consumers could purchase a wide variety of e-books
via the wireless data connection and it was comparatively easy to use.
Apple again ignited the market in 2010, with the introduction of the iPad, raising the profile of e-readers and
tablets.
Dedicated reading devices (e-readers) differ from the iPad and other tablets. E-readers with e-ink screens
simulate the experience of reading a paper book whereas tablets are multifunctional devices. Tablets allow
users to browse the web, read and write email messages, take and view photographs, view movies, play games
and perform many of the same functions as with a computer.
The advantage of an e-reader is that it is easier to read even in direct sunlight, does not consume a lot of
battery, is lighter in weight, permits undisturbed reading of an e-book and the eyes do not fatigue compared to
devices with backlit LCD displays.
In the United States, the cost of e-readers has continued to fall, making them more accessible to purchase. Data
from Pew Research Centre over a three-year period from May 2010 to September 2013 indicate e-reader
ownership has grown over 20%, see Figure 1 on ownership growth of e-readers and tablets in the United
States.
Overall, 50% of Americans now have a dedicated handheld device – either a tablet computer like an iPad, or
an e-reader such as a Kindle or Nook for reading e-content (Zickuhr & Rainie, 2014).
E-Reader evaluation
Ujala Rampaul
Page 7 of 66
Figure 1. Growth of ownership of e-readers and tablets in the United States
Source: Pew Research Centre (Rainie & Smith, 2013)
The success of e-readers in the United States has influenced the Australian consumer market for e-readers,
which continues to grow despite the popularity of tablets. Although the price of e-readers has fallen over the
years, they are still expensive. For example, the price range of e-readers in 2013 as indicated by Choice, a
consumer advisory group, varied from AUD$100 to $469 for the latest e-reader then on the market.
1.1 Motivation
The objective of this project was to evaluate from a purchasing perspective the quality of e-readers that could be
purchased within a limited budget.
Shoppers of e-readers and e-reading devices are often influenced by advertising, reviews conducted by computer
and consumer sites such as PC Authority and Choice, as well as opinions of friends and families. The
commercial reviews compare the performance, features, file formats and specifications of the latest e-reader
models against the market price to identify the best value for money.
To obtain e-readers for this study a budget based on the cost of a new paper book ($25) was established and
multiplied to set the budget range of approximately $50-75 (excluding shipping costs) for the purchase of an ereader. Establishing a budget reduced the range of e-readers that could be obtained i.e. either, earlier, e-reader
models with a higher specification or later models with a relatively low specification.
E-Reader evaluation
Ujala Rampaul
Page 8 of 66
We argue that if we focused on the quality of the devices available within this budget range, then the differences
in the e-reader model and specifications were not limiting factors. The experiment requires that a realistic range
of e-readers of a high quality could be obtained from eBay and other auction sites for an effective comparison.
The following nine e-readers were selected and mostly purchased from within Australia except for the Sony and
jetBook, at a lower price range than originally estimated being $35 -$70. This includes a selection of e-readers
having buttons only, touch screen only or a combination of both.
e-readers
cost in AUD$
Barnes & Noble Nook
69.90
Kobo
60.00
Amazon Kindle 4G
57.85
Sony PRS-600
44.88
iRiver Story HD
52.00
Elonex 621EB
41.00
Ectaco jetBook
61.28
Pandigital Novel
36.85
EZReader EZ601
55.00
Table 1. Cost of second-hand e-readers
Each device was labeled with an alphabetical letter from A to I, based on their arrival sequence.
E-Reader evaluation
Ujala Rampaul
Page 9 of 66
2 Literature survey
The rise of e-books and e-readers are part of a larger story about the shift from printed to digital material. Early
research focused on academic contributions towards education where students used e-readers in the classrooms
to read textbooks.
There have been numerous approaches taken when evaluating both ebooks and ebook readers in projects such as
Superbook and Electronic Books On-screen interface (EBONI). (Gibson & Gibb, 2011).
Other studies included university and public libraries investigating strategies to support the borrowing of ebooks and e-readers. “In particular, many academic libraries since begun pilot projects using a variety of
different reader devices to investigate the possibilities for simplifying and innovating the content and related
services they offer to users in light of the technologies that are available. Countless libraries have experimented
with offering lending programs for their devices. Many libraries have undertaken these initiatives on their own
whereas others have partnered with either Sony electronics or Amazon to design and carry out their projects”
(Behler & Lush, 2010).
To establish design guidelines for e-readers (Pearson, et al., 2010) conducted a study to determine the usability
concerns with user interface of the e-readers using human computer interaction (HCI) principles. The
experiment was carried out using the Kindle 2, Sony PRS 600 and Sony PRS 300. These three e-readers had
similar screen type (e-Ink technology), screen size (six inches) and resolution (600 x 800). The Sony PRS 600
was a combination of touch and button device whereas the Kindle 2 and Sony PRS 300 were button devices.
Based on the guidelines and principles for ergonomics, consistency, completeness, page numbering, book
marks, annotation and magnification, the Kindle was found to be better than Sony PRS-300 and PRS-600. The
positioning of the buttons in the Kindle made it easier to turn pages for both left and right-handed users. The
Kindle had a full QWERTY keyboard which supported web browsing and location identifiers in relation to the
file instead of page numbers. The availability of full annotation and zooming was only available on the Sony
PRS-600. User feedback on the three e-readers was leveraged from Amazon’s online customer review and
weighted on a Likert scale rating. This could have attributed for the higher ratings for the Kindle, as it is
produced by Amazon.
The study by Gibson and Gibb (2011) evaluated four second-generation e-reading devices, namely Sony
PSRS505; Bookeen Cybook 3g; iRex iLiad and Asus Eee PC 105 HA, with 33 participants. Their evaluation
measured weight, quality of display, size and robustness of the devices based on a five point Likert scale rating.
On overall impressions and functionality, the Asus Eee netbook was the best, followed by Sony PRS 505, iLiad
and lastly Cybook. The Asus Eee netbook was the preferred choice because participants were familiar with the
design and layout compared to the dedicated e-readers (Gibson & Gibb, 2011).
E-Reader evaluation
Ujala Rampaul
Page 10 of 66
Both the Sony PRS505 and Booken Cybook were similar in screen size (six inches) and resolution (600 x 800
pixels) whereas the iRex iLiad had a larger screen size (eight inches) and a larger resolution (768 x 1024 pixels).
All three devices had screens with e-Ink technology. The Asus Eee PC 105 HA, a netbook had a LCD screen, a
much larger screen size (ten inches) as well as a large resolution (600x1024 pixels) and weighed seven times
more than the Booken Cybook, the lightest (175g) device in the trial. This meant the refresh rate for turning the
page were faster in the netbook compared to the e-readers using e-Ink technology. This was an unequal
comparison between three dedicated e-readers and a netbook.
Usability was favourable towards the e-readers for lightness and portability, readability and ease of use.
Participants commented, “the screen was not wearing on the eyes” (iLiad) and that it was a “straightforward
operation” (Sony), and that “the non-glare screen made the text as easy to read as ink” (iLiad) (Gibson & Gibb,
2011).
Based on reported sales and market share in the United States, Richardson & Mahmood (2012) studied five
leading e-readers namely Kindle 3G; iPad 1G; Nook; Kobo N647; and Sony PRS950. Their objective was to
identify the advantages and disadvantages of e-readers and “to compare and contrast the most popular (i.e.
bestselling) devices against a comprehensive, if not exhaustive, set of technical specifications as well as
qualitative judgment of users” (Richardson & Mahmood, 2012).
The results from Richardson and Mahmood’s study showed that the Kindle was the most popular device, though
some participants commented on poor navigation. Some 47% of the participants that undertook this study
owned a Kindle, accounting for a probably biased popularity of the Kindle. There was no evidence for the
technical comparison nor did they elaborate what it consisted of. The comparison between the iPad and the ereaders was probably unjust as the tablet is a multi-function device, whereas e-readers are designed to read and
purchase e-books.
The NextMedia eReading Project conducted a qualitative and quantitative study on reading behaviours and
patterns on e-readers including the effects of age groups. The objective of the main project was to encourage
Finnish consumers to adopt and grow an e-reading community. This eReading Project consisted of multiple
smaller projects that focused on different aspect of the study such as usability of e-reading devices, types of
content, emotional and cognitive reactions.
Heikkilä (2011) usability study of e-reading devices was based on a scenario, which involved the participants
recording their experiences of opening and using a newly purchased e-reading device. A number of pages from
a paper book was the benchmark for the evaluations. The same content being available on the e-reading devices.
Over a week, seventeen participants recorded their reading times, places they had read and their experiences on
using the e-readers. The time-based assessment tasks included opening the device; finding a book; finding a
specific spot in the book and changing the font size.
Heikkilä used the following e-readers in his study Kindle (included both Kindle 2 and Kindle 3); Sony PRS600;
iPad 1G; BeBook Neo; Booken Cybook Opus; Elonex eBook and Samsung galaxy.
E-Reader evaluation
Ujala Rampaul
Page 11 of 66
The outcome of the usability study was the creation of a conceptual model of an e-reader labeled Acola. The
Acola was a combination of touch sensitive and gesture savvy pad on an e-ink device. The pad would handle
page turning and skimming. Swiping to the left would turn the page forward, swiping to the right would move
one page backwards. Swiping fast two times successively, would turn two pages, three times: three pages etc.
Swiping with two fingers simultaneously would move you between chapters. A menu and an OK button could be
situated in the upper and lower border of the pad (Heikkilä, 2011).
There is some similarity between Heikkilä’s (2011) study and my report in the use of a scenario for the
evaluation of the e-reading devices. It is interesting that his study evaluated two tablets and four e-readers to
create their ideal e-reader.
Siegenthaler, et al., (2012) conducted a study to with the following reading devices Sony PRS505, Sony
PRS600 and an Apple iPad, to determine effects of touch screen technology.
Twelve participants tested the three devices sequentially within a session based on a set of time related tasks
such as open a book, open a specific page within the book, highlight a sentence, find a highlighted sentence,
delete a highlighted sentence, change the orientation of the layout and increase the font size. The participants
rated the above tasks on the navigation, design, handiness and handling based on a Likert scale rating.
The results showed “that e-reading devices with touch screens correlate with better navigation rating” and touch
screen technology also has its advantages in terms of a higher intuitiveness and flexibility for adaptations of the
navigation (e.g., due to firmware updates) compared to devices with static buttons. (Siegenthaler, et al., 2012).
The evaluation of handiness and handling was ambiguous, for example the related questions were “How handy
do you rate the reading device?” and “how easy was it for you to handle the reading device?” There was no
clarification to what this meant and the evaluation questions were vague.
We earlier identified the differences between an e-reader, which is solely for reading e-books whereas a tablet is
a multifunction device making this evaluation an unequal comparison.
Siegenthaler, et al, (2012) argued, “E-ink technology has low power consumption, thereby increasing battery
life and allowing for a more lightweight device. Another advantage is that e-ink devices can be used outside
without glare being a big issue. However, e-ink screens have some disadvantages, most of them are black and
white and the pages do not refresh as quickly as devices with an LCD screen”.
The disadvantage of tablets is that they cost considerably more than e-readers, the LCD screens are susceptible
to glare, they do not provide a comfortable reading experience, are heavier and the battery does not last as long.
E-ink technology is designed to emulate printed books. Independent studies on the reading behaviours measured
by eye tracking found no difference between a paper book and a dedicated e-reader (Nielsen, 2009) and
E-Reader evaluation
Ujala Rampaul
Page 12 of 66
(Siegenthaler, et al., 2010).
Similarly, analysis of the reading behaviours between a dedicated e-reader and an e-reading device (tablet)
found no significant difference as measured by eye speed (Nielsen, 2010) and (Siegenthaler, et al., 2012).
However, the iPad “actually scores slightly higher in user satisfaction” (Nielsen, 2010).
Please see Figure 2 which maps the reading devices that are common to this study and to other studies
identified. Most studies shared one common e-reader, for example Richard & Mahmood (2012) share the Nook,
Siegenthaler, et al.,(2012) shared the Sony PRS-600, Siegenthaler, et al., (2010) share the jetBook and Pearson
et al., (2010) share the Sony PRS-600. The exception is Heikkilä (2011) where there two common e-readers
namely the Sony PRS-600 and the Elonex.
Please see Figure 3 which maps the tasks that are common to this study and to other studies identified. All
studies share at least one common task, with Siegenthaler et al, (2010) and (2012) evaluating two similar tasks
namely, open a document (open a book) and increase font size. The other exception is Heikkilä (2011), where
four similar tasks were evaluated that is turn the power on (open device), navigate to first document (find a
particular book), increase font size (change font size) and navigate to specific section within the book (find a
specific place within the book). All studies have commented on readability.
No previous studies have been conducted to evaluate from a purchasing perspective the quality and usability of
e-readers purchased within a budget of up-to $70. This study proposes to conduct the evaluation of the e-readers
based on a scenario with two experiments.
1.
Evaluation of nine e-readers that are button interface devices only, touch screen only and a combination
of both to determine which device is the best
2.
Evaluation of e-readers with touch screens
E-Reader evaluation
Ujala Rampaul
Page 13 of 66
Figure 2. Map of common reading devices between this study and previous study
E-Reader evaluation
Ujala Rampaul
Page 14 of 66
Figure 3. Map of tasks between this study and previous study
E-Reader evaluation
Ujala Rampaul
Page 15 of 66
3 Methodology
The Human Research Ethics Committee at the Australian National University (ANU) approved this study.
The objective of this project was to evaluate from a purchasing perspective the quality and usability of e-readers
purchased within a limited budget.
Participants from the ANU College of Engineering and Computer Science were recruited for the experiments.
The participants evaluated the devices in pairs so one could act as the scribe and observe, while the other
operated the device. On completing the evaluation of the first device, the participants swapped roles to evaluate
the next device. This grouping encouraged discussion amongst themselves on their observations and interactions
with the e-readers.
3.1 Experiment 1
In experiment 1, 72 senior students (3rd year HCI course participants, Honours, Masters and PhD) took part in
the evaluations.
3.1.1
E-reader
Nine e-readers with e-readers having buttons only, touch screen only or a combination of both were used in the
evaluations. Each device was allocated an alphabetical letter based on the arrival sequence for easy
identification in the randomisation matrix, namely:
A. Barnes & Noble Nook
B. Kobo Touch
C. Amazon Kindle 4G
D. Sony PRS-600
E. iRiver Story HD
F. Elonex 621EB
G. Ectaco jetBook
H. Pandigital Novel
I.
EZReader EZ601
Please see Appendix 1 for the full technical specifications of the above e-readers.
The randomisation was similar to a Latin square where two sets of the e-readers were labeled from A to I and
arranged in such a way that no row contained the same letter twice. The pairs were balanced, so that the number
of times a letter came first or second was the same. There were 36 pairs of e-readers for 72 participants to
evaluate, which meant each device was evaluated four times as a first device and then four times as a second
device. Table 2 shows an example of the device ‘A’ evaluated eight times with the other devices.
E-Reader evaluation
Ujala Rampaul
Page 16 of 66
BA
BH
HE
IF
GE
CA
BI
GC
DC
IE
DA
BG
DI
EC
CH
EA
AF
HD
CF
HF
CB
AG
GD
CI
GF
DB
AH
ED
FD
IH
EB
AI
IG
FE
HG
FB
Table 2. Matrix of the e-reader randomisation pair for device A.
The evaluation commenced when each participant pair were given the first e-reader from a randomised pair, and
a questionnaire. The participants received an overview of the study and its objectives. After performing the tasks
on the first e-reader, the tasks were repeated on the second e-reader.
The questionnaire included a scenario, the evaluation tasks based on Likert-like numerical scale, open-ended
questions to gain an insight on improving the device and comments after each device evaluated. It also gathered
optional personal details including the name, gender, age, year of study, familiarity with e-readers and follow-up
contact details for purposes of clarification or removal from the experiment in adherence with The Human
Research Ethics Committee guidelines of the Australian National University. Lastly, the questionnaire requested
the subjective opinions and reasons on whether one of the two devices was preferred.
3.1.2
Scenario
The evaluations were designed around the same scenario to provide context to the experiment. The scenario was
based on a user finding a gift received a couple months ago. Included in the box was a note to register the device
prior to use. The participant had to register the device without the assistance of the user manual, which was
based on a series of tasks.
For both the experiments, the participants had to assume the battery on the e-readers were fully charged, as it
was not a requirement of the scenario.
Please see Appendix 2 for full details of the scenario.
3.1.3
Measurement of tasks
The tasks was measured on a Likert-like summative scale rating ranging from the most positive to the least
positive:

5 = Very good

4 = Good

3 = Ok

2 = Bad

1 = Very bad
E-Reader evaluation
Ujala Rampaul
Page 17 of 66
The tasks were
1.
To turn the device on
2.
Navigate to a document (called “Somedevice XY User Guide”)
3.
Open the document
4.
Increase the font size
5.
Navigate to specific section to find the ‘model number’
6.
Navigate to a second document (“Dates-2013” and use the model number).
7.
Assess readability on the screen
Please see Appendix 3 for the full questionnaire.
3.2 Experiment 2
The results of the first experiment suggested that a comparison be conducted between the touch screen operated
devices. Please see the results section 4.1 for details. The objective of the second experiment was to identify
preferences between e-readers with touch screens excluding button only e-readers.
In experiment 2, 64 mostly junior (first year) students took part in the evaluations. In a few instances, senior
students filled in as partners when a second junior student was not available.
3.2.1
E-readers
Three touch screen e-readers from experiment 1, and two additional e-readers were used in experiment 2
evaluations, namely:
B. Kobo Touch
D. Sony PRS-600
H. Pandigital Novel
J. Nook Touch
K. Kindle Touch
Experiment 2 applied a similar randomisation as used in experiment 1. Table 3 shows an example for device ‘B’
on the randomisation of the touch screen e-reader pairs. The pairs in italics indicate the outstanding data
remaining to be collected for experiment 2.
BD
KH
KD
BH
HD
DJ
JB
KJ2
JH
BD2
JK2
KH2
HD2
DK
JD
BK2
DB
HJ
JB2
KJ
HB2
JD2
BK
KD2
KB2
DB2
HJ2
BH2
HK2
BJ
KB
DH2
JH2
DH
HB
HK
BJ2
DK2
DJ2
JK
Table 3. Matrix of touch screen e-reader pairs for device B
E-Reader evaluation
Ujala Rampaul
Page 18 of 66
3.2.2
Questionnaire
The same scenario and the tasks based on a Likert-like numerical scale rating from experiment 1 was used in the
evaluation of experiment 2.
Explicit instructions were given to not use the recent history, continue reading or the ‘date’ search in the ereaders.
Additional subjective open-ended questions were requested, on likes and dislikes about the e-readers; previous
use or ownership of an e-reader and the model and if they were regular users of a smartphone; tablet or laptop.
These questions were to gain a better understanding into the user’s experience with mobile devices.
Please see Appendix 4 for the full updated questionnaire as used in experiment 2.
E-Reader evaluation
Ujala Rampaul
Page 19 of 66
4 Results
Statistical analysis was performed using F-statistics based on a repeated measures ANOVA for the within factor
of e-readers. A critical p < 0.05 was used for statistical significance in all analyses. Each of the following
sections report the ANOVA results for that question based on the usability tasks measured on a Likert-like
numerical scale rating from 1 (very bad) to 5 (very good) on the e-readers.
4.1 Experiment 1
4.1.1
Turning on the e-reader
There were no significant differences found in the responses to the question on powering on the e-readers. Table
3 shows the results and ranks for the e-readers. The jetBook, Pandigital and the Kobo are ranked at the top.
There are four e-readers (i.e. Nook, iRiver, Elonex and EZReader) ranked at the middle in equal fourth. The
Sony and the Kindle were ranked at the bottom, however there were no significant differences found in the ereaders for turning them on.
e-Reader
Mean
SD
Rank
jetBook
4.1
0.83
1
Pandigital
4.0
0.76
2
Kobo
3.9
1.17
3
Nook
3.8
1.28
=4
iRiver
3.8
1.28
=4
Elonex
3.8
0.71
=4
EZReader
3.8
1.28
=4
Sony
3.4
1.06
8
Kindle
3.1
0.83
9
Table 4. Mean Ranking and Standard Deviation (SD) for turning the e-readers on
4.1.2
Finding the first document
There were no significant differences found in in response to the question on finding the first document. Table 4
shows the results and ranks for the e-readers. Sony and Kobo ranked in the top for finding a document whereas
the iRiver, Pandigital, jetBook, and Nook were ranked in the middle. EZReader and Kindle were ranked at the
bottom but there was no significant difference found in all e-readers.
Mean
SD
Rank
Sony
e-Reader
4.0
0.53
1
Kobo
3.5
1.41
2
iRiver
3.3
1.16
=3
Pandigital
3.3
1.49
=3
jetBook
3.0
1.31
5
E-Reader evaluation
Ujala Rampaul
Page 20 of 66
e-Reader
Mean
SD
Rank
Nook
2.9
1.46
6
Elonex
2.5
0.76
7
EZReader
2.4
1.60
8
Kindle
2.3
0.89
9
Table 5. Mean Ranking and Standard Deviation (SD) for finding the document
4.1.3
Opening the document
An analysis of variance showed that the responses to the question on opening the document was significantly
different between the e-readers, F(8,63) = 3.21, p =0.004. Table 5 shows the results and rankings for opening
the document. The Kobo, Kindle and the iRiver were ranked at the top. The Sony and Elonex were ranked equal
fourth. The jetBook and Pandigital ranked equal sixth. The EZReader and Nook rank at the bottom. This shows
there is a significant difference in the e-readers to perform the task of opening a document.
e-Reader
Mean
SD
Rank
Kobo
4.6
0.52
1
Kindle
4.5
0.93
=2
iRiver
4.5
0.76
=2
Sony
4.4
0.74
=4
Elonex
4.4
0.74
=4
jetBook
4.3
1.39
=6
Pandigital
4.3
0.89
=6
EZReader
3.5
1.20
8
Nook
2.8
1.16
9
Table 6. Mean Ranking and Standard Deviation (SD) on opening the document: significant
4.1.4
Increasing the font size
An analysis of variance showed that increasing the font size was significantly different between the e-readers,
F(8,63) = 2.25, p =0.03. Table 6 shows the results and rankings on the e-readers for increasing the font size. The
Pandigital, jetBook and iRiver were ranked at the top three positions. The Kindle and Elonex were ranked equal
fourth. The Nook was ranked at sixth position. The Kobo and EZReader were ranked at the bottom two
positions as equal eighth. This shows a significant difference was found in the e-readers to perform the task of
increasing the font size.
e-Reader
Mean
SD
Rank
Pandigital
4.6
0.52
1
jetBook
4.1
1.13
2
iRiver
4.0
1.07
3
Kindle
3.9
1.46
=4
Elonex
3.9
1.13
=4
Nook
3.4
1.06
6
Sony
3.3
0.71
7
E-Reader evaluation
Ujala Rampaul
Page 21 of 66
e-Reader
Mean
SD
Rank
Kobo
2.8
1.75
8
EZReader
2.8
1.49
8
Table 7. Mean Ranking and Standard Deviation (SD) for increasing font size: significant
4.1.5
Navigate to specific section
An analysis of variance showed that navigating to a specific section within the document was significantly
different between the devices, F(8,63) = 3.43, p =0.002. Table 7 shows the results and rankings on the e-readers
for navigating to a specific section within the document. The jetBook, Pandigital and the Kindle ranked at the
top. The Sony, iRiver and the Kobo ranked in the middle and the Elonex, Nook and EZReader ranked in the
bottom positions. This shows a significant difference in the e-readers for navigating to specific section within a
document.
e-Reader
Mean
SD
Rank
jetBook
4.0
0.93
1
Pandigital
3.6
1.19
2
Kindle
3.5
1.07
3
Sony
3.4
1.41
4
iRiver
3.3
1.28
5
Kobo
2.8
1.39
6
Elonex
2.6
1.30
7
Nook
1.9
1.46
8
EZReader
1.6
0.92
9
Table 8. Mean Ranking and Standard Deviation (SD) for navigation to section 3: significant
4.1.6
Navigate to a second document
There was no significant differences found in navigating to a second document with the e-readers. Table 8
shows the results and ranks for navigating to a second document. The iRiver, jetBook and the Sony ranked in
the top three positions. The Pandigital, Elonex and the Kobo ranked at the middle and the Nook, Kindle and
EZReader ranked at the bottom three positions. Navigating to a second document was not significantly different
in the e-readers.
e-Reader
Mean
SD
Rank
iRiver
4.0
0.93
1
jetBook
3.9
1.81
2
Sony
3.8
1.04
3
Pandigital
3.6
0.92
4
Elonex
3.4
1.69
5
Kobo
3.1
1.64
6
Nook
2.9
1.46
7
Kindle
2.6
1.41
8
EZReader
2.5
1.31
9
E-Reader evaluation
Ujala Rampaul
Page 22 of 66
Table 9. Mean Ranking and Standard Deviation (SD) for navigation to second document
4.1.7
Readability on the screen
There were no significant differences found in response on assessing the readability. Table 9 shows the results
and ranks for readability. The iRiver was ranked the highest, followed by Kobo, Kindle and jetBook in second
position and the Pandigital in the fifth. The Nook and EZReader were both ranked equal sixth. The Elonex and
Sony ranked lower but no significant difference were found between the e-readers for readability.
e-Reader
Mean
SD
Rank
iRiver
4.5
0.76
1
Kobo
4.0
0.53
=2
Kindle
4.0
0.93
=2
Jetbook
4.0
0.93
=2
Pandigital
3.9
0.35
5
Nook
3.8
0.71
=6
EZReader
3.8
1.28
=6
Elonex
3.5
1.20
8
Sony
3.3
1.28
9
Table 10. Mean Ranking and Standard Deviation (SD) for readability
4.1.8
E-Reader preference
For the question, “Did you like one of the two devices better?” an analysis of variance showed that the
responses were significantly different between the e-readers, F(8,54) = 2.33, p =0.03. Table 10 shows the results
and rankings of the participants’ preference when comparing e-readers. The Kindle and iRiver ranked equal
first, followed by Sony, jetBook and Pandigital in equal third position. The Elonex was ranked sixth, Kobo
seventh and both Nook and EZReader ranked at the bottom. This indicates a significant difference in the ereaders that participants’ preferred the most.
e-Reader
Mean
SD
Rank
Kindle
0.6
0.52
=1
iRiver
0.6
0.52
=1
Sony
0.5
0.53
=3
jetBook
0.5
0.53
=3
Pandigital
0.5
0.53
=3
Elonex
0.4
0.52
6
Kobo
0.3
0.46
7
Nook
0.0
0.00
=8
EZReader
0.0
0.00
=8
Table 11. Mean Ranking and Standard Deviation (SD) for e-reader preference: significant
E-Reader evaluation
Ujala Rampaul
Page 23 of 66
4.1.9
Pairwise comparison on e-readers with significant differences
Pairwise comparison is commonly used to estimate preference values of finite alternatives with respect to a
given criterion. This is part of the model structure of the analytical hierarchy process, a widely used multicriteria decision-making methodology. The main difficulty is to reconcile the inevitable inconsistency of the
pairwise comparison matrix elicited from the decision makers in real-world applications (Choo & William,
2004)
The elimination of designs or candidates can change the tabulated rankings of those designs or candidates that
remain under consideration. The determination of which design is “best” or which candidate is “preferred most”
may well be sensitive to the set of designs considered (Dym, et al., 2002).
The overall ranking we used based on the qualitative and quantitative mean rankings of the questions about ereaders that indicated a significant difference as shown in 4.1.3 opening a document, 4.14 increasing the font
size, 4.1.5 navigating to specific section within the document and 4.1.8 e-reader preference. Table 11 shows the
overall ranking of e-readers where significant difference was noted.
e-Reader
Qualitative ranking
Quantitative ranking
Overall ranking
Kindle
=1
3
=1
Pandigital
=3
1
=1
iRiver
=1
4
=3
jetBook
=3
2
=3
Sony
=3
5
5
Elonex
6
6
6
Kobo
7
7
7
Nook
=8
8
8
EZReader
=8
9
9
Table 12. Qualitative, quantitative and overall ranking
The qualitative rankings were based on the number of yes responses to the question on whether one of the two
devices was better. The quantitative rankings were based on Likert-like numerical scale rating on the individual
tasks and participants’ responses in questions where a statistically significant difference was found in the
responses.
The qualitative and quantitative rankings are similar, within one rank for the majority of e-readers. This suggests
that our rankings can be relied on, as the two forms of data are largely consistent. The exceptions are the Kindle
(1th in qualitative ranking and 3rd in quantitative ranking); Pandigital (3rd in qualitative ranking and 1st in
quantitative ranking); iRiver (1st in qualitative ranking and 4th in quantitative ranking) and Sony (3rd in
qualitative ranking and 5th in quantitative ranking). These differences are discussed in section 4.3.
E-Reader evaluation
Ujala Rampaul
Page 24 of 66
4.2 Experiment 2
The data collection for experiment 2 is not complete, so this analysis is preliminary only.
4.2.1
Turn on the e-reader
Similar to experiment one, there were no significant differences found in the response to the question on
powering on the e-readers. Refer to Table 13 for the Mean Ranking and Standard Deviation (SD) for switching
the e-reader on. The Kobo was in the top first, followed by the Kindle Touch (2nd), Sony (3rd), Pandigital (4th)
and the Nook Touch at the bottom (5th).
e-Reader
Mean
SD
Rank
Kobo
4.3
0.9
1
Kindle Touch
4.2
0.7
2
Sony
4.1
0.9
3
Pandigital
3.5
1.0
=4
Nook Touch
3.5
1.0
=4
Table 13. Mean Ranking and Standard Deviation (SD) for turning the e-reader on.
4.2.2
Find the first document
There were no significant differences found in response to the question on finding the first document. Refer to
Table 14 on the Mean Rankings and the Standard Deviation for finding a document. Sony ranked the highest
(1st) followed by Kobo (2nd), Kindle Touch (3rd), Nook Touch (4th) and the Pandigital (5th).
e-reader
Mean
SD
Rank
Sony
4.0
0.6
1
Kobo
3.8
0.8
2
Kindle Touch
3.6
1.1
3
Nook Touch
3.5
1.4
4
Pandigital
3.0
1.4
5
Table 14. Mean Ranking and Standard Deviation (SD) for finding a document
4.2.3
Open the document
There were no significant differences found in response to the question on opening the document. Refer to Table
15 on the Mean Rating and Standard Deviation on opening the document. The Nook Touch ranked the highest
(1st) followed by Kobo (2nd), Sony (3rd), Pandigital (4th) and Kindle Touch (5th).
e-Reader
Mean
SD
Rank
Nook Touch
4.9
0.4
1
Kobo
4.6
0.5
2
Sony
4.5
0.9
3
Pandigital
4.5
0.5
4
Kindle Touch
4.4
0.7
5
E-Reader evaluation
Ujala Rampaul
Page 25 of 66
Table 15. Mean Ranking and Standard Deviation (SD) for opening the document
4.2.4
Increase the font size
There were no significant differences found in response to the question on increasing the font size. Refer to
Table 16 on the results of the mean ranking and standard deviation for increasing the font size. The Sony was
ranked in the top (1st), followed by Kindle Touch (2nd), Pandigital (3rd), Nook Touch (4th) and Kobo (5th).
e-Reader
Mean
SD
Rank
Sony
4.0
0.7
1
Kindle Touch
3.7
1.3
2
Pandigital
3.5
1.0
3
Nook Touch
3.5
1.3
4
Kobo
3.4
1.3
5
Table 16. Mean Ranking and Standard Deviation (SD) for increasing the font size
4.2.5
Navigate to specific section
There were no significant differences found in response to the question on navigating to a specific section within
the document. Refer to Table 17 on the mean ranking and standard deviation on navigating to a specific section
within the document. The Nook Touch ranked the highest (1st) followed by Sony (2nd), Kindle Touch (3rd) and
Pandigital (5th).
e-Reader
Mean
SD
RANK
Nook Touch
4.0
1.2
1
Sony
3.4
1.3
2
Kindle Touch
3.1
1.3
3
Pandigital
3.1
1.4
4
Kobo
2.9
1.1
5
Table 17. Mean Ranking and Standard Deviation (SD) for navigating to section 3
4.2.6
Navigate to second document
An analysis of variance showed that navigating to a second document was significantly different between the ereaders, F(4,59) = 2.77, p =0.04. Refer to Table 18 on mean ranking and standard deviation for navigating to the
second document. The Nook Touch and the Sony ranked in the two positions. The Kobo and the Kindle Touch
ranked in the middle with third and fourth position. The Pandigital ranked at the bottom in the fifth position.
The results show a significant difference in the e-readers for navigating to a second document.
e-Reader
Mean
SD
RANK
Nook
4.2
1.1
1
Sony
3.8
1.2
2
Kobo
3.6
1.3
3
E-Reader evaluation
Ujala Rampaul
Page 26 of 66
e-Reader
Mean
SD
RANK
Kindle
3.4
1.3
4
Pandigital
2.6
1.5
5
Table 18. Mean Ranking and Standard Deviation (SD) for navigating to second document: significant
4.2.7
Readability on the screen
There were no significant differences found in response on assessing the readability. Refer to Table 19 on mean
ranking and standard deviation for readability. The Nook Touch ranked the highest (1st) followed by Kobo (2nd),
Sony (3rd), Pandigital (4th) and Kindle Touch (5th).
e-Reader
Mean
SD
Rank
Nook Touch
4.6
0.5
1
Kobo
4.4
0.8
2
Sony
4.3
0.7
3
Pandigital
4.3
0.5
4
Kindle Touch
4.1
0.9
5
Table 19. Mean Ranking and Standard Deviation for readability
4.2.8
e-reader preference
No significant differences were found in response to the question on “which device was better”. Refer to Table
20 on the mean rankings and standard deviation for user preference. The Nook Touch ranked the highest (1st)
followed by Kindle Touch (2nd), Sony (3rd), Kobo (4th) and Pandigital (5th).
e-Reader
Mean
SD
Rank
Nook Touch
0.6
0.5
1
Kindle Touch
0.5
0.5
2
Sony
0.4
0.5
3
Kobo
0.3
0.5
4
Pandigital
0.2
0.4
5
Table 20. Mean Ranking and Standard Deviation on user preference
4.2.9
Pairwise comparison for Q6
The overall ranking is based on the qualitative and quantitative mean rankings of the e-readers with a significant
difference as shown in 4.26. Navigate to second document. Refer to Table 21 on the pairwise comparison for
Q6.
There is no change in the qualitative, quantitative and overall rankings for the Nook Touch (1st) and Pandigital
(5th). The Sony remains in the middle (3rd) for both the qualitative and quantitative rankings. The exceptions are
the Kobo (4th in qualitative ranking and 2nd in the quantitative ranking) and the Kindle Touch (2nd in the
qualitative ranking and 4th in the quantitative ranking).
E-Reader evaluation
Ujala Rampaul
Page 27 of 66
e-reader
Qualitative rank
Quantitative rank
Overall rank
Nook Touch
1
1
1
Sony
3
2
2
Kindle Touch
2
4
3
Kobo
4
3
4
Pandigital
5
5
5
Table 21. Qualitative, quantitative and overall ranking for Q6
4.2.10
Pairwise comparison on Q3 – Q6
We have used comparisons on the questions which were significant in the previous experiment, as we assume
that when the full results are collected, questions 3 to 5 will again be significant, and we have added question 6
which is already shown to be significant.
The overall ranking is based on the qualitative and quantitative mean rankings of the e-readers with a significant
difference from experiment 1 (Q3-Q5) and experiment 2 (Q6). Refer to Table 22 on the pairwise comparison for
Q3 -Q6.
There is no difference in the qualitative, quantitative and overall rankings for the Nook Touch (1st) and
Pandigital (5th). Refer to Table 22 on the pairwise comparison for Q3 – Q6. The Kindle Touch and the Kobo are
within one rank of each. The exception is Sony where it has ranked 4th in the qualitative rankings and 2nd in the
quantitative ranking.
e-reader
Qualitative rank
Quantitative rank
Overall rank
Nook Touch
1
1
1
Kindle Touch
2
3
2
Sony
4
2
3
Kobo
3
4
4
Pandigital
5
5
5
Table 22. Qualitative, quantitative and overall ranking for Q3-Q6.
E-Reader evaluation
Ujala Rampaul
Page 28 of 66
4.3 Discussion
Interaction with e-readers should be simple, effective and a pleasant experience as it is emulating a paper book.
The power ON/OFF buttons on the Sony and the Kindle was not easily noticeable. On the Sony, three similar
buttons of the same shape and size were located at the top edge of the device and the label was not distinct due
to the reflective surface and the angle at which the device was commonly held. A comment on the Sony was
“doesn’t stand out from the other buttons on the same area”. The power switch on the Kindle was positioned on
the lower edge of the device and the button was not labeled. Some of the comments on the Kindle power button
were “button hard to find at the bottom - unusual position”; “hard to find switch button”; and “hard to locate the
button, button un-labeled”. The ANOVA results showed no significant differences in the e-readers in response
to the question for turning the e-readers on.
Some devices such as Sony, Elonex, jetBook, iRiver, Pandigital and EZReader stored the recently read e-books
in the current reading or recent history folder. This extremely useful feature allows the reader to go straight to
their e-book without being distracted with cumbersome navigation and searching. In the experiments there was
only one of each e-reader, when an e-reader was available for the next pair to evaluate there was not sufficient
time to refresh the reading history folder, as it generally requires a full factory reset. It was a compromise to
leave the documents within the folder especially if alternate methods for finding the document were supported.
The recent history was not an obvious feature to all participants. While users were instructed that if the e-reader
‘accidentally’ started in the target document, then they should back out of it and navigate to it, it is clear from
the comments that sometimes they ‘cheated’.
Typing a search query in the Kindle was cumbersome and required too much effort to complete the full query as
it involved navigating the keyboard via the 5-way controller. Prevention of keyboard slips and incorrect spelling
was not supported and frustrating to the user. An autocomplete feature would be a suitable design solution to
combat the complexity in typing and error management.
Buttons are designed to afford pressing to trigger a visible change on screen. The search button on the EZReader
was unresponsive and disconcerting as the users were unsure how to continue. If a button is displayed it must be
associated to an action otherwise it leads to user frustration and a waste of valuable real estate on the interface.
Although the participants encountered some difficulty with some of the e-readers on finding the document, the
ANOVA results show there was no significant difference in the e-readers.
Opening a document was found to be either intuitive on some e-readers or challenging on other e-readers. This
is also reflected in the ANOVA results, which showed a significant difference in the e-readers to support
opening a document. For example, tapping on the document title opened the document on the Kobo, Sony and
Pandigital. The iRiver had two “enter” buttons, one situated above the keyboard and the other within the
keyboard for easy selection. The center buttons on the multi-controllers of the Kindle, Elonex and the jetBook
behaved as the “enter” key.
E-Reader evaluation
Ujala Rampaul
Page 29 of 66
Comments on the EZReader for opening a document were “hard to find”, “tried the arrows keys first“, and “a bit
slow”. The “enter” button on the Nook touch panel was not identifiable because it did not meet a user’s concept
of a button and it was not labeled. A comment on the Nook for opening the document was “open button is not
obvious, task a bit to find. Fact screen is not touch is not obvious”. The “enter” key could not be distinguished if
it was a radio button or a decorative element.
The ANOVA results for increasing the font size showed a significant difference in the e-readers to accomplish
this task. The Pandigital supported alternate methods to change the font size for example, within an opened
document from previous task, zoom icons (both plus and minus) were available on the bottom right corner of the
screen and selecting the menu button displayed a toolbar with additional access buttons such as “My Library;
Dictionary; TOC; Bookmarks; Go To; Font size and Next” on the top of the screen.
The font size adjust button was visible on the multi-controller of the jetBook, the keyboard on the iRiver and a
stepper on the Elonex. The option to change the font size in the Kindle was accessible via the “menu” button.
Comments on the Nook on changing the font size were “menu a bit tricky to scroll through” and “not intuitive to
find". The Kobo has a multi-touch screen which meant each area of the screen was reserved for a function, for
example tapping in the middle or top right of the screen turned a page forward and tapping on the top left of the
screen turned the page backwards. A tap to the bottom of the screen was essential to unfold the font size icon
button and a comment was “hidden toolbar, obvious icon”.
In the EZReader a “font” option was displayed via the “option” button but this was ambiguous as only the font
type could be changed and not the font size. Comments on changing the font size on the EZReader were “no
idea where to navigate”; “had to select font first, that wasn't any good so we kept looking through other menus”
and “can't find the setting”. Each e-reader had its own methodology to accomplish the task of changing the font
size which accounts for the difference identified in the analysis of variance calculation.
In the Kobo the ‘table of contents’ options was available via the concealed toolbar, similar to accessing the font
size option. Comments on the Elonex were “first tried to use ‘explore’ button and press keyword ‘model
number’, no response came out. Easy to find when using ‘content’ button”. The ‘content’ button was accessible
via the “menu” button.
There were no visual cues for a search or table of contents feature in the Nook and user’s found it to confusing.
Comments on navigation in the Nook were “couldn't figure out how to scroll the screen”, “not found” and kept
pressing the wrong button (left button sent pages forward)”.
Similar to the Nook there were no visual cues in the EZReader to gain access to the specific section within the
document. Comments on the EZReader to navigate to the section within the document were “cannot find it”,
“we cannot find it”, “needed 15% hint for go to page”, “it's impossible to find that paragraph” and “can't find the
document”. These variations for navigating in the e-readers to a specific section within a document were also
identified in the analysis variance results.
E-Reader evaluation
Ujala Rampaul
Page 30 of 66
Navigating and opening a second document reinforced the learning experience from the previous task for
finding and opening a document and no significant differences were found in the ANOVA results.
There was no significant difference for readability. Almost all the e-readers had e-Ink screen types. Pandigital
had SiPiX and jetBook had a LCD screen type, but these did not affect the readability. SiPiX technology uses
white particles suspended in black ink while e-Ink technology has both black and white particles suspended in a
clear fluid. A comment on Sony was “screen is dull and not enough backlight for contrast” but it was not
significantly different compared to the other e-readers.
A significant difference in the analysis of variance results were found when asked which e-reader was preferred
from the pair evaluated. This was then used to conduct a comparison to determine which e-readers was overall
significantly better, based on counting the number of times each device was considered better than the other
device.
There were four e-readers for which their qualitative rankings were more than one-step different in the
quantitative rankings. The Kindle, Sony and the iRiver were all higher by two to three steps and the Pandigital
was two steps lower. This could be due to the Pandigital being an unknown brand compared to the Kindle and
Sony as popular brands, though this does not explain the iRiver which showed the highest difference between
the qualitative (1st) and quantitative (4th). It is possible that the brand recognition is the primary answer for the
Kindle and Sony. For the iRiver, it appears that the overall interface is visually appealing to the users which
would account for the high qualitative rankings but the functional aspects and the aesthetics of the interface are
not in complete harmony to meet the user’s expectation.
Comments on the iRiver and EZReader were that the keyboards were redundant; the buttons were too small and
overall not very useful. A full list of participants’ comments can be viewed in Appendix 6. Here is a summary of
the comments
1) The e-readers were either non-intuitive or intuitive
2) Some e-readers were found to be user friendly and easy to use
3) Expectations were for faster and more responsive feedback
4) E-readers with un-labeled buttons and use of icons were confusing
5) The lack of contrast colour of buttons and interface was deceptive
6) Preference for a device with touch screen interface
Grounded theory is a systematic generation of theory from systematic research (Corbin & Strauss, 2008).
A “grounded theory” approach was considered to interpret the qualitative data but the comments were found to
be too short and insufficient to code and to form concepts of the code, for a meaningful analysis.
E-Reader evaluation
Ujala Rampaul
Page 31 of 66
In this study, three e-readers, the Pandigital, the Sony and the Kobo used Touch technology. The Pandigital and
the Sony are also button-operated interfaces, while the Kobo was touch only. The qualitative rankings for these
devices were Pandigital (3rd) and Sony (3rd) and quantitative rankings were Pandigital (1st) and Sony (5th). The
Kobo is consistent with both its qualitative and quantitative ranking at 7th. It is clear from these tasks highlighted
in 4.1.9 pairwise comparison, that button operated interfaces are consistently better, that is the three devices
which had touch interfaces are distributed among the top, middle and bottom. Further, the only full touch ereader (Kobo) is among the bottom three.
It is worth noting that most new devices available on the market in 2014 are mostly touch devices, and over 60%
of the participants commented, that they would either preferred a device with touch screen interface or would
improve the device by adding a touch screen interface.
In experiment 2, the preliminary analysis of variance calculation showed a significant difference in the e-readers
only for navigating and opening a second document. This could be attributed to the instruction on to not use the
‘date’ search. That is, mentioning ‘search’ could have predisposed the participants to use the search feature to
find the documents. The second document could be found only by the title search. Content search was not
possible because the second document was a very short document, with very little content to search on. See
Appendix 7.7 on the list of ‘Dates 2013’ in the second document.
The search and table of contents features has not been investigated in previous studies, most likely, because it is
not a feature associated with paper books. However, on an electronic device a search feature is critical to finding
information quickly. The search feature should be investigated in the future.
The overall ranking is based on the qualitative and quantitative mean rankings of the e-readers with a significant
difference from the experiment 1 (Q3-Q5) and experiment 2 (Q6) to understand the difference between the sets.
Preliminary results from experiment 2 show that touch devices are preferred over the Pandigital, which had
ranked equal 1st in experiment 1. There are a number of possible explanations. For example, in the first
experiment there was only one full touch e-reader and the other e-readers were either, button only device or a
combination of button and touch. In the first experiment, perhaps the Pandigital ranked in the top because it was
compared to mostly button devices whereas in the second experiment the Pandigital was compared to touch
devices. That is, the Pandigital’s button interface may be better than the button interface of most e-readers in our
study and perhaps even gain extra ‘points’ for having touch. However, in the second experiment, perhaps the
Pandigital’s touch interface was used more and did not compare well to other touch e-readers.
A second explanation could be the selection difference in the participants. Perhaps, senior students evaluated the
devices more objectively, and were not persuaded by preconceptions. In the second experiment, a few students
commented on brand recognition, for example “I don’t recognise the manufacturer so it can’t be any good”
(about the Pandigital). While comments after the experiment suggested that the Kobo interface is similar to a
smartphone, with the home button at the bottom centre of the interface and that this was “a very clever design”.
E-Reader evaluation
Ujala Rampaul
Page 32 of 66
5 Conclusions
In experiment 1, the ANOVA results showed significant differences between the e-readers for the tasks of
opening the document, increasing the font size, navigating to a specific section within the document, and user
preference on the e-reader from the device pair evaluated.
In experiment 2, the only significant difference in the e-readers identified from the ANOVA results was for the
tasks of navigating to and opening a second document.
In this study the tasks was measured on a Likert-like numerical scale rating the e-readers which was used to
calculated the Mean, Standard Deviation and Analysis of Variance (ANOVA). From ANOVA results, a further
comparison on the qualitative and quantitative rankings were used to determine, which e-readers were overall
better.
In the overall ranking for experiment 1, the Kindle, Pandigital, iRiver and jetBook were in the top four
positions. The Kindle, iRiver and jetBook are button-operated e-readers whereas the Pandigital is combination
of buttons and touch. e-Readers with buttons are intuitive to use. Users know when they have selected a physical
button whereas tapping a button on a touchscreen is not always instantaneous due to the speed of response and
the user may continue to tap the screen.
In the overall ranking for experiment 2, the touch devices ranked higher than the Pandigital. The touch e-readers
are a later release than the Pandigital, which may account for a better performance and speed of response, or it
may be that its touch interface is not as good as its button interface. Almost all the junior participants were
familiar with using a touch screen device and 95% indicated they were regular users of a smartphone.
From the evaluations, it is clear that a good quality device within a limited budget can be obtained and the latest
device is not necessarily the best device on the market that meets a user’s requirements.
5.1 Future work
This study concluded an analysis on nine e-readers and conducted a preliminary analysis on five touch ereaders. The next stage of this project is to complete the evaluation for experiment 2, requiring more data
collection and further analysis. Exploring the direction for further studies could involve:

Updating the scenario to evaluate tasks for continuous reading

Evaluating the e-readers by concealing the brand names

Evaluating the usability of an e-reader to read one-handed whilst drinking coffee, holding an umbrella,
standing on a bus or train and so on

Evaluating the search and table of contents features in e-readers
E-Reader evaluation
Ujala Rampaul
Page 33 of 66
6 Bibliography
Behler, A. & Lush, B., 2010. Are You Ready for E-readers?. The Reference Librarian, 52(1-2), pp. 75-87.
Choo, E. U. & William, W. C., 2004. A common framework for deriving preference values from pairwise
comparison matrices. Computers & Operators, 31(6), pp. 893-908.
Corbin, J. & Strauss, A., 2008. Basics of Qualitative Research. 3rd ed. s.l.:Sage.
Dym, C. L., Wood, W. H. & Scott, M. J., 2002. Rank Ordering Engineering Designs: Pairwise Comparison
Charts and Borda Counts. Research in Engineering Design, Volume 13, pp. 236-242.
Gibson, C. & Gibb, F., 2011. An evaulation of second generation ebook readers. 29(3), pp. 303-319.
Heikkilä, H., 2011. eReading User Experience: eBook devices, reading software & contents. eReading Media
Use, Experience & Adoption, 10 January, Volume 1.
Nielsen, J., 2009. Kindle 2 Usability Review. [Online]
Available at: http://www.nngroup.com/articles/kindle-2-usability-review/
[Accessed 10 March 2014].
Nielsen, J., 2010. iPad and Kindle Reading Speeds. [Online]
Available at: http://www.nngroup.com/articles/ipad-and-kindle-reading-speeds/
[Accessed 10 March 2014].
Patel, N., 2007. Kindle sells out in 5.5 hours. [Online]
Available at: http://www.engadget.com/2007/11/21/kindle-sells-out-in-two-days/
[Accessed 5 April 2014].
Pearson, J., Buchanan, G. & Thimbleby, H., 2010. HCL Design Principles for eReaders. Swansea and London,
ACM, New York, USA, pp. 15-24.
Rainie, L. & Smith, A., 2013. Tablet and E-reader Ownership Update. [Online]
Available at: http://www.pewinternet.org/2013/10/18/tablet-and-e-reader-ownership-update/
[Accessed 8 April 2014].
Richardson, J. & Mahmood, K., 2012. eBook readers: user satisfaction and usability issues. Emerald insight,
30(1), pp. 170-185.
Saaty, T. L., 1990. How to make a decision: The Analytisc Hierarchy Process. European Journal of Operational
Research, Volume 48, pp. 9-26.
Siegenthaler, E. et al., 2012. The effects of touch screen technology on the usability of e-reading devices.
Journal of Usability Studies, May, 7(3), pp. 94-104.
Siegenthaler, E., Wurtz, P. & Groner, R., 2010. Improving the usability of e-book readers. Journal of Usability
Studies, November, 6(1), pp. 25-38.
Siegenthaler, E., Wyss, M., Schmid, L. & Wurtz, P., 2012. LCD vs E-ink: An analysis of Reading Behaviour.
Journal of Eye Movement Research, pp. 1-7.
Zickuhr, K. & Rainie, L., 2014. E-Reading Rises as Device Ownership Jumps. [Online]
Available at: http://www.pewinternet.org/2014/01/16/e-reading-rises-as-device-ownership-jumps/
[Accessed 5 April 2014].
E-Reader evaluation
Ujala Rampaul
Page 34 of 66
7 Appendices
7.1 Appendix 1: Technical specification on e-readers
Name
Nook
ID
A
e-Reader
Cost
69.90
Screen
Screen
Resolution
Button
Touch
Weight
size
type
(pixel)
s
screen
(g)
6 inches
e-Ink
600 x 800
Yes for
Yes
324
page
turning
Kobo
B
60
6 inches
e-Ink
600 x 800
No
Yes
185
600 x 800
Yes
No
170
600 x 800
Yes for
Yes
284
Pearl
Kindle
C
57.85
6 inches
e-Ink
Pearl
Sony
D
41.79
6 inches
e-Ink
navigati
on
E-Reader evaluation
Ujala Rampaul
Page 35 of 66
Name
ID
e-Reader
Cost
Screen
Screen
Resolution
Button
Touch
Weight
size
type
(pixel)
s
screen
(g)
e-Ink
600 x 800
Yes
No
202
e-Ink
600 x 800
Yes
No
186
600 x 800
Yes
No
215
600 x 800
Yes
Yes
264
iRiver
E
52
6 inches
Elonex
F
41
6 inches
Pearl
jetBook
G
61.28
5 inches
e-Ink
Pearl
Pandigital
H
E-Reader evaluation
36.85
6 inches
siPiX
Ujala Rampaul
Page 36 of 66
Name
EZReader
ID
I
E-Reader evaluation
e-Reader
Cost
55
Screen
Screen
Resolution
Button
Touch
Weight
size
type
(pixel)
s
screen
(g)
6 inches
e-Ink
480 x800
Yes
No
236
Ujala Rampaul
Page 37 of 66
7.2 Appendix 2: Scenario
Tutorial and experiment on e-readers
Professor Tom Gedeon, RSCS, ANU
Experiment:
In this exercise you will get some experience in a simple evaluation of a user interface of a device. We have a
number of different e-readers, each with its own user interface. We will form groups of 2 students and give
each group one of these e-readers together with a task to be performed, and then swap over. After the task the
group will discuss their observations about the user interface of their particular e-reader.
A second purpose of this exercise is to be a pilot study in how we might evaluate e-readers. We will collect
your notes and these will help me design a more formal study.
Experiment Setting:
You have been given an e-reader as an unexpected gift, and you haven’t opened the box until now, you have had
it since last year sometime. You find a slip of paper which says “Important check registration deadline – device
not usable if not registered by this date” and tells you to open the “Somedevice XY User Guide” document and
find section 3, paragraph 2, and find the model number, and then look for that number in the “Dates-2013”
document to check how long you have to register the e-reader. You have no manual for the device, so you just
have to try and figure it out as you go along.
Task:
Appoint one of your group as a scribe.
Please find out your registration deadline, and make observations about the device, which the scribe recording
interesting or important observations whether positive, neutral or negative on the form provided.
Select from a Likert scale (V. good / Good / OK / Bad / V. bad) for all of the 7 points below.
Write in some comments into the form on the following pages.
Form sections on following page:
1.
Turning it on
2.
Finding the right document (called “Somedevice XY User Guide”)
[Some hints which you may need on some devices: the author is Unknown; and some devices have multiple
places to store documents/books – it’s definitely on there somewhere, I’ve checked!]
3.
Opening the document
4.
Increasing the font size (please assume it’s too small to read as is)
5.
Navigating to section 3 (about 15% into the book), and finding the end of the 2nd (short) paragraph to find
the ‘model number’
6.
Going to the other document (“Dates-2013” and use the model number from step 5).
7.
Readability of the screen
E-Reader evaluation
Ujala Rampaul
Page 38 of 66
i.
How would you improve this device?
ii.
Any other comments:
7.3 Appendix 3: Questionnaire
0.
Device letter:
1.
Turning it on
V. good
Good
OK
Bad
V. bad
2.
Finding the right document ( “Somedevice XY User Guide”)
V. good
Good
OK
Bad
V. bad
3.
Opening the document
V. good
Good
OK
Bad
V. bad
4.
Increasing the font size
V. good
Good
OK
Bad
V. bad
5.
Navigating to section 3, and finding the 2nd paragraph, model number is:
V. good
Good
OK
Bad
V. bad
6.
Going to “Dates-2013”
write in the date (year) here:
V. good
Good
E-Reader evaluation
Ujala Rampaul
Page 39 of 66
OK
Bad
V. bad
7.
Readability of the screen
V. good
Good
OK
Bad
V. bad
i.
How would you improve this device?
ii.
Any other comments:
Scribe (1st device)
Scribe (2nd device)
Name
Gender
Age
Familiarity with
eReaders
Year of study
Can we contact you to
follow-up this
experiment?
Last question:
Did you like one of the two devices better?
If so, why?
Thank you!
E-Reader evaluation
Ujala Rampaul
Page 40 of 66
7.4 Appendix 4: Questionnaire for Experiment 2
The scenario was the same as the first experiment.
Device letter:
1. Turning it on
V. good
Good
OK
Bad
V. bad
2. Finding the right document ( “Somedevice XY User Guide”)
V. good
Good
OK
Bad
V. bad
3. Opening the document
V. good
Good
OK
Bad
V. bad
4. Increasing the font size
V. good
Good
OK
Bad
V. bad
5. Navigate to section 3, and find the 2nd paragraph, model number is:
V. good
Good
OK
Bad
V. bad
6. Navigate to “Dates-2013”
write in the date (year) here:
V. good
Good
OK
Bad
V. bad
7. Readability of the screen
V. good
Good
OK
Bad
V. bad
i.
How would you improve this device?
ii.
What did you like about this device?
iii.
What did you dislike about this device?
E-Reader evaluation
Ujala Rampaul
Page 41 of 66
iv.
Any other comments:
I need both your names on this form.
Person 1 (usually scribe 1st
Person 2 (usually for 2nd device)
device)
Name
Gender
Age
Do you own or have used an ereader before?
If so, what device was it?
Which of these do you own or use
regularly?
Smartphone
Tablet
Laptop
Year of study
Can we contact you to follow-up
this experiment?
Last question: Did you like one of the two devices better?
If so, why?
Thank you!
E-Reader evaluation
Ujala Rampaul
Page 42 of 66
7.5 Appendix 5: ANOVA calculations
7.5.1
Experiment 1
1) Turning the e-reader on
Anova: Single Factor
Q1
SUMMARY
Groups
Count
Sum
Average
Variance
A
8
30
3.75
1.64
B
8
31
3.88
0.98
C
8
25
3.13
0.70
D
8
27
3.38
1.13
E
8
30
3.75
1.64
F
8
30
3.75
0.50
G
8
33
4.13
0.70
H
8
32
4.00
0.57
I
8
30
3.75
1.64
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
Within Groups
Total
5.944444
8
0.743056
66.5
63
1.055556
72.44444
71
P-value
0.704
0.687
F crit
2.089
2) Finding a document
Anova: Single Factor
Q2
SUMMARY
Groups
Count
Sum
Average
Variance
A
8
23
2.9
2.1
B
8
28
3.5
2.0
C
8
18
2.3
0.8
D
8
32
4.0
0.3
E
8
26
3.3
1.4
F
8
20
2.5
0.6
G
8
24
3.0
1.7
H
8
26
3.3
2.2
I
8
19
2.4
2.6
ANOVA
E-Reader evaluation
Ujala Rampaul
Page 43 of 66
Source of Variation
SS
df
MS
F
Between Groups
20.75
8
2.59375
Within Groups
95.25
63
1.511905
116
71
Total
P-value
1.72
0.11
P-value
F crit
F crit
2.09
3) Opening a document
Anova: Single Factor
Q3
SUMMARY
Groups
Count
Sum
Average
Variance
A
8
22
2.75
1.36
B
8
37
4.63
0.27
C
8
36
4.50
0.86
D
8
35
4.38
0.55
E
8
36
4.50
0.57
F
8
35
4.38
0.55
G
8
34
4.25
1.93
H
8
34
4.25
0.79
I
8
28
3.50
1.43
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
23.75
8
2.96875
Within Groups
58.125
63
0.922619
Total
81.875
71
3.218
0.004
2.089
4) Increasing the font size
Anova: Single Factor
Q4
SUMMARY
Groups
Count
Sum
Average
Variance
A
8
27
3.4
1.13
B
8
22
2.8
3.07
C
8
31
3.9
2.13
D
8
26
3.3
0.50
E
8
32
4.0
1.14
F
8
31
3.9
1.27
G
8
33
4.1
1.27
H
8
37
4.6
0.27
I
8
22
2.8
2.21
E-Reader evaluation
Ujala Rampaul
Page 44 of 66
ANOVA
Source of Variation
SS
Between Groups
Within Groups
Total
df
MS
26
8
3.25
90.875
63
1.44246
116.875
71
F
2.25
P-value
0.03
F crit
2.09
5) Navigating to a specific section within the document
Anova: Single Factor
Q5
SUMMARY
Groups
Count
Sum
Average
Variance
A
8
15
1.88
2.13
B
8
22
2.75
1.93
C
8
28
3.50
1.14
D
8
27
3.38
1.98
E
8
26
3.25
1.64
F
8
21
2.63
1.70
G
8
32
4.00
0.86
H
8
29
3.63
1.41
I
8
13
1.63
0.84
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
Within Groups
Total
41.5
8
5.1875
95.375
63
1.513889
136.875
71
3.43
P-value
0.002
F crit
2.089
6) Navigating to a second document
Anova: Single Factor
Q6
SUMMARY
Groups
Count
Sum
Average
Variance
A
8
23
2.88
2.13
B
8
25
3.13
2.70
C
8
21
2.63
1.98
D
8
30
3.75
1.07
E-Reader evaluation
Ujala Rampaul
Page 45 of 66
E
8
32
4.00
0.86
F
8
27
3.38
2.84
G
8
31
3.88
3.27
H
8
29
3.63
0.84
I
8
20
2.50
1.71
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
19.52778
8
2.440972
121.75
63
1.93254
141.2778
71
Within Groups
Total
P-value
1.26
F crit
0.28
2.09
7) Readability
Anova: Single Factor
Q7
SUMMARY
Groups
Count
Sum
Average
Variance
A
8
30
3.8
0.50
B
8
32
4.0
0.29
C
8
32
4.0
0.86
D
8
26
3.3
1.64
E
8
36
4.5
0.57
F
8
28
3.5
1.43
G
8
32
4.0
0.86
H
8
31
3.9
0.13
I
8
30
3.8
1.64
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
Within Groups
Total
7.944444
8
0.993056
55.375
63
0.878968
63.31944
71
1.13
P-value
0.36
F crit
2.09
8) Number of yes responses
Anova: Single Factor
E-Reader evaluation
No. of Yes responses
Ujala Rampaul
Page 46 of 66
SUMMARY
Su
Groups
Varianc
Count
m
Average
e
A
8
0
0.0
0.00
B
8
2
0.3
0.21
C
8
5
0.6
0.27
D
8
4
0.5
0.29
E
8
5
0.6
0.27
F
8
3
0.4
0.27
G
8
4
0.5
0.29
H
8
4
0.5
0.29
I
8
0
0.0
0.00
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
3.75
8
0.46875
P-value
F crit
0.03501
2.08918
4
5
2.25
0.20833
Within Groups
13.125
63
Total
16.875
71
3
9) Overall Ranking Calculations
For Q3-Q5 with significant differences
No. Yes
Rank
Total AVE
Rank
Overall
Rank
A
Nook
0.00
8
2.7
8
8.0
8
B
Kobo
0.25
7
3.3
7
7.0
7
C
Kindle
0.63
1
4.0
3
2.0
1
D
Sony
0.50
3
3.7
5
4.0
5
E
iRiver
0.63
1
3.9
4
2.5
3
F
Elonex
0.38
6
3.6
6
6.0
6
G
Jetbook
0.50
3
4.0
2
2.5
3
H
Pandigital
0.50
3
4.0
1
2.0
1
I
EZReader
0.00
8
2.6
9
8.5
9
1
2
3
4
5
7
8
E-Reader evaluation
Ujala Rampaul
6
9
Page 47 of 66
Nook
Kobo
Kindle
Sony
iRiver
Elonex
Jetbook
Pandigital
EZReader
Finally, after judgments have been made on the impact of all the elements and priorities have been computed for
the hierarchy as a whole, sometimes, and with care, the less important elements can be dropped from further
consideration because of their relatively small impact on the overall objective. The priorities can then be
recomputed throughout, either with or without changing the remaining judgments (Saaty, 1990).
Q2-Q4,7 ONLY
No. Yes
Rank
Total AVE
Rank
Overall
Rank
A
Nook
0.00
8
3.0
8
8.0
8
B
Kobo
0.25
7
3.5
6
6.5
7
C
Kindle
0.63
1
3.3
7
4.0
5
D
Sony
0.50
3
3.8
2
2.5
2
E
iRiver
0.63
1
3.9
1
1.0
1
F
Elonex
0.38
6
3.5
5
5.5
6
G
Jetbook
0.50
3
3.8
3
3.0
3
H
Pandigital
0.50
3
3.8
4
3.5
4
I
EZReader
0.00
8
2.8
9
8.5
9
1
2
3
4
5
7
8
6
9
Nook
Kobo
Kindle
Sony
iRiver
Elonex
Jetbook
Pandigital
EZReader
It is unclear if a comparison for Q2-Q4 and Q7 is meaningful, because Q2 and Q7 were not significant on the
ANOVA. According to Saaty’s study if we decide which questions seem important then we get a different result
The most important would be finding a document, opening a document, increasing the font size and readability
compared to navigating to a specific section with the document and navigating to a second document. More
analysis would be required.
E-Reader evaluation
Ujala Rampaul
Page 48 of 66
7.5.2
Experiment 2
1) Turning the e-reader on
Anova: Single Factor
Q1
SUMMARY
Groups
Count
Sum
Average
Variance
B
13
56
4.31
0.90
D
12
49
4.08
0.81
H
11
42
3.82
0.96
J
14
49
3.50
1.04
K
14
59
4.21
0.49
ANOVA
Source of Variation
Between Groups
Within Groups
Total
SS
df
MS
F
5.804971
4
1.451243
49.1794
59
0.833549
54.98438
63
P-value
1.741
F crit
0.153
2.528
2) Finding a document
Anova: Single Factor
Q2
SUMMARY
Groups
Count
Sum
Average
Variance
B
13
50
3.8
0.64
D
12
48
4.0
0.36
H
11
33
3.0
2.00
J
14
52
3.7
0.99
K
14
51
3.6
1.17
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
6.673764
4
1.668441
Within Groups
59.76374
59
1.012945
66.4375
63
Total
E-Reader evaluation
Ujala Rampaul
P-value
1.65
F crit
0.17
Page 49 of 66
2.53
3) Opening a document
Anova: Single Factor
Q3
SUMMARY
Groups
Count
Sum
Average
Variance
B
13
60
4.6
0.26
D
12
54
4.5
0.82
H
11
49
4.5
0.27
J
14
68
4.9
0.13
K
14
61
4.4
0.55
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
2.017233
4
0.504308
Within Groups
23.73277
59
0.40225
25.75
63
Total
P-value
1.253717
F crit
0.298407
2.527907
4) Increasing the font size
Anova: Single Factor
Q4
SUMMARY
Groups
Count
Sum
Average
Variance
B
13
44
3.4
1.76
D
12
48
4.0
0.55
H
11
39
3.5
1.07
J
14
49
3.5
1.81
K
14
52
3.7
1.60
ANOVA
Source of Variation
SS
df
MS
Between Groups
2.838661
4
0.709665
Within Groups
82.16134
59
1.392565
85
63
Total
E-Reader evaluation
Ujala Rampaul
F
0.510
P-value
F crit
0.729
Page 50 of 66
2.528
5) Navigating to a specific section within the document
Anova: Single Factor
Q5
SUMMARY
Groups
Count
Sum
Average
Variance
B
13
38
2.9
1.24
D
12
44
3.7
0.61
H
11
34
3.1
1.89
J
14
56
4.0
1.38
K
14
44
3.1
1.67
SS
df
MS
F
ANOVA
Source of Variation
P-value
F crit
10.7868
Between Groups
8
4
2.69672
2
59
1.359544
91
63
1.98
0.11
2.53
80.2131
Within Groups
Total
6) Navigating to a second document
Anova: Single Factor
Q6
SUMMARY
Groups
Count
Sum
Average
Variance
B - Kobo
13
47
3.62
1.59
D- Sony
12
45
3.75
1.48
H- Pandigital
11
29
2.64
1.65
J- Nook Touch
14
59
4.21
1.10
K- Kindle Touch
14
47
3.36
1.63
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
16.41557
4
4.103892
Within Groups
87.44381
59
1.482098
E-Reader evaluation
Ujala Rampaul
P-value
2.77
0.04
F crit
2.53
Page 51 of 66
Total
103.8594
63
7) Readability
Anova: Single Factor
Q7
SUMMARY
Groups
Count
Sum
Average
Variance
B
13
57
4.4
0.59
D
12
52
4.3
0.42
H
11
47
4.3
0.22
J
14
65
4.6
0.25
K
14
58
4.1
0.75
SS
df
MS
F
ANOVA
Source of Variation
Between Groups
1.880396
4
0.470099
Within Groups
26.85398
59
0.455152
Total
28.73438
63
P-value
1.033
0.398
F crit
2.528
8) Number of yes responses
Anova: Single Factor
No. Of Yes responses
SUMMARY
Groups
Count
Sum
Average
Variance
B
16
5
0.31
0.23
D
16
6
0.38
0.25
H
16
3
0.19
0.16
J
16
9
0.56
0.26
K
16
8
0.50
0.27
ANOVA
Source of Variation
Between Groups
Within Groups
E-Reader evaluation
SS
df
MS
1.425
4
0.35625
17.5625
75
0.234167
Ujala Rampaul
F
1.521
P-value
0.205
Page 52 of 66
F crit
2.494
Total
18.9875
79
9) Overall calculations for
Q6 only
No.Yes
Rank
Total Av
Rank
Overall
Rank
Kobo
0.31
4
3.6
3
3.5
4
Sony
0.39
3
3.8
2
2.5
2
Pandigital
0.19
5
2.6
5
5
5
Nook Touch
0.63
1
4.2
1
1
1
Kindle Touch
0.50
2
3.4
4
3
3
No.Yes
Rank
Total Av
Rank
Overall
Rank
Kobo
0.31
4
3.64
5
5
4
Sony
0.38
3
4.06
2
3
2
Pandigital
0.19
5
3.70
4
5
4
Nook Touch
0.63
1
4.12
1
1
1
Kindle Touch
0.50
2
3.74
3
3
2
No.Yes
Rank
Total Av
Rank
Overall
Rank
Kobo
0.31
4
3.63
4
4
4
Sony
0.38
3
3.98
2
3
2
Pandigital
0.19
5
3.43
5
5
5
Nook
0.63
1
4.14
1
1
1
Kindle
0.50
2
3.64
3
3
2
No.Yes
Rank
Total Av
Rank
Overall
Rank
Kobo
0.31
4
3.87
3
4
4
Sony
0.38
3
4.05
2
3
2
Pandigital
0.19
5
3.55
5
5
5
Nook
0.63
1
4.06
1
1
1
Kindle
0.50
2
3.80
4
3
3
Q3-Q5
Q3-Q6
FOR ALL QUESTIONS
E-Reader evaluation
Ujala Rampaul
Page 53 of 66
7.6 Appendix 6: Participants comments for experiment 1
Q1. Turn on the device
e-reader pair
ID
IF
ID
RAN
K
Comment
I
F
ED
E
HF
AH
BI
HE
EB
IE
I
4
looked like On button screen
CH
C
3
Button hard to find at the bottom - unusual position
EA
E
3
slow initialisation
AF
a
5
on screen instructions - easy
F
4
button in standard location but need to be pressed for a while
C
3
hard to find switch button
I
2
I thought it was touch screen because it displayed an "On" icon on screen
G
5
click a button and its ON
G
4
I had to look around for the power button
F
5
The button is onto of the device
E
4
Button should be on the front
F
4
Easy to find 'power on' button but during the 1st press and second press,
not sure if the device is on or not. Need to press button twice.
E
5
Turn on button is easy to find because it is the only button that looks like
a "turn-on" button.
D
3
doesn't stand out from other buttons on the same area
A
2
too long to start
C
2
Hard to locate the button. Button un-labeled
B
5
intuitive and prevents accidental switching
DI
HD
FB
IH
CI
IA
IG
GF
FD
GD
CF
EC
AG
FE
DC
DA
DB
CB
BA
CA
E-Reader evaluation
Ujala Rampaul
Page 54 of 66
e-reader pair
ID
GC
ID
RAN
K
5
G
C
5
Comment
obvious ON/ OFF symbol
obvious power button
GE
HG
BG
Q2. Find the first document
e-reader pair ID
ID
Rating
Comment
IF
I
1
hard to find
F
ED
E
HF
AH
BI
HE
EB
IE
I
4
we cheated, recent doc
CH
C
1
very hard
A
4
"my library" pretty intuitive
F
2
Took a bit to work out how to control screen. Document was in
"unknown" author folder- had to guess this. Would be easy to not find it.
(overlook)
C
2
Hard to find search (just randomly enter the interface). Hard to delete
character, no hotkey
I
1
There is a search button but I cannot use it when I click on it. It's easy to
navigate through folder but I can't find the file
G
2
G
3
user had to search each folder to find the document , there's no search
action
later I understood it was in A-Z order, and the hint helped a lot
F
3
It took me time to go to the end of the document list
e
2
Found in recent documents - took some time
C
3
AG
G
1
search should work with partial characters - required full string "some
device"
feel confused cause there is an icon, which is used to change the font and
seems like usually search icon
FE
F
2
DI
HD
FB
AF
IH
CI
AI
IG
GF
FD
GD
CF
EC
E-Reader evaluation
1. Hard to find the correct button to start finding. The button is black and
the surface of the e-reader is black too. In addition, the button does not
have a label.
Ujala Rampaul
Page 55 of 66
e-reader pair ID
ID
Rating
Comment
2. no search function
3. the screen is slow and always flashing, misleading the user that she
hasn't pressed the button
1. We have to know the rule to move the cursor. 2. Need time to get used
to the move button.
E
4
D
4
A
1
keyboard is horrible, not intuitive
C
3
Controls are obvious but it is hard to navigate through 8 pages of
documents. Search looks through documents instead of titles.
B
2
Touch buttons have low responsiveness but keyboard superior. Browse
interface placed ahead of the search
G
4
hint helped
C
3
took a second to figure that the page turn buttons were necessary
DC
DA
DB
CB
BA
CA
GC
GE
HG
BG
Q3. Open the document
e-reader pair ID
ID
RANK
Comment
IF
I
1
Hard to find
F
ED
E
HF
AH
BI
HE
EB
IE
I
4
tried > key first
CH
C
5
easy
A
3
F
5
"open" button is not obvious. task a bit to find. Fact screen is not
touch is not obvious
move to it and click it
C
5
fast
I
3
a little bit slow
DI
HD
FB
AF
IH
CI
AI
IG
E-Reader evaluation
I
5
straight forward
G
0
Opening the document was easy. Just click 'ok' button to open it
Ujala Rampaul
Page 56 of 66
GF
G&F
5
easy
F
5
easy to open
E
4
1. There is clearly a label on the button to enter the file.
FD
GD
CF
EC
AG
FE
2. When keeping press the 'down' button, it would not go to the next
page but it will jump to the top. We have to use the tiny 'flip page'
button to go next page.
DC
D
4
easy
A
1
DB
B
1
couldn't figure out how to open document needed assistance from
tutor
could be easy found by research
CB
C
5
easy & quick
B
5
easy & quick
G
5
straight forward
DA
BA
CA
GC
GE
HG
BG
Q4. Increase font size
e-reader pair ID
ID
RANK
Comment
IF
I
1
no idea to navigate
E
5
ugly the screen is thinking
D
3
called zoom instead of font size
IE
I
2
had to select font first, that wasn't any good so we kept
looking through other menus
CH
c
5
easy
FB
F
1
couldn’t find the model no.
AF
A
4
Menu a bit tricky to scroll through
F
5
Handy button on side
F
ED
HF
AH
BI
HE
EB
DI
HD
IH
CI
E-Reader evaluation
Ujala Rampaul
Page 57 of 66
e-reader pair ID
ID
RANK
Comment
I
1
can't find the setting
G
4
user just had to click "F" button, however user don't
know how to make it return to normal size
G
5
take few minutes but it is easy to increase the font size
F
5
using + and - is very user friendly
F
2
1. long response time after pressing correct button
AI
IG
GF
FD
GD
CF
EC
AG
FE
2. easy to find the accurate button
3. After pressing the 'increase size' button 6 times the font
returns back to minimal size
E
5
1. It is easy to find the button to increase the font size.
Have clearly label 'AA' - easy to understand the meaning
of
2. After changing the size, the highlight is not correct.
"AA' is not in the right position.
DC
DA
D
3
A
2
Touch screen small for fingers. Not intuitive to find
C
5
first level menu is as intuitive as it gets
B
4
hidden toolbar, obvious icon
G
5
magnifying glass icon was helpful
C
5
easy, clearly described in menu
H
2
We did not try to find an easy way to go to section 3. So
maybe it’s not that bad
DB
CB
BA
CA
GC
GE
HG
BG
Q5. Navigate to section 3 within document
e-reader pair ID
ID
IF
I
RANK
Comment
1
cannot find it
3
It looks like touch but not. It was sluggish and took time to figure it out up & down buttons
F
ED
E-Reader evaluation
E
Ujala Rampaul
Page 58 of 66
e-reader pair ID
ID
RANK
Comment
HF
AH
A
1
not found
BI
I
1
we cannot find it
IE
I
3
Needed 15% hint for go to page
CH
C
2
quite hard
FB
F
1
couldn’t find the model no.
B
2
long res time
AF
A
3
F
4
Scrolling was easy when I tested but file opened to correct page! This
should be fixed for future tests. Side buttons could be a little more ???
and have a little less lag time. Scrolling slow but content links fix this.
Home scrolling through links was also slow.
scrolling bit slow
IH
H
0
could not find it
CI
I
0
it's impossible to find that paragraph
HE
EB
DI
HD
AI
I
1
can't find the document
G
4
user just have to click on the up/down button
G
3
It takes time to find the option and it is slow
F
1
I couldn't find it
E
1
selecting content from menu counter intuitive
F
4
First tried to use ' explore' button and press keyword 'model number', no
response came out. Easy to find when using 'content' button
E
3
Sometimes the highlighted square is disappeared. Need to re-open the
doc. The button to move highlighted square and the button to flip pages
are different. It is confusing.
D
1
cannot find 2nd para, not clearly indicated
A
1
couldn't figure out how to scroll the screen
CB
C
4
intuitive if you are familiar with those devices
BA
A
2
kept pressing the wrong button (left button sent pages forward)
G
4
first icon guessed was correct
C
5
TOC was easy to find
IG
GF
FD
GD
CF
EC
AG
FE
DC
DA
DB
CA
GC
GE
HG
BG
E-Reader evaluation
Ujala Rampaul
Page 59 of 66
Q6. Navigate to second document
e-reader pair ID
ID
IF
I
ED
E
RANK
1
Comment
cannot find it
took much time to find the search function as we misunderstood, we
thought it was in the same doc
HF
AH
BI
BI
HE
EB
IE
CH
DI
HD
FB
AF
IH
F
1
can't find date -2013
B
1
speed is very low& cannot find the menu now to open it
A
5
once I know the navigation - easy
F
4
scrolling bit slow
H
0
could not find it
I
1
can't find file
G
5
user just had to click on the go back button
G&F
5
easy to find the document
G
1
Device froze
E
3
similar to Q3 - however instructions on front page not very clear
C
3
Keyboard slow. Hit done, yet had to hit 'enter' again to search for
DOC - repetitive
F
1
cannot find it
E
5
It is easy to find. However it is because someone had used it recently,
so it is in the "recent''. Even not use the 'recent' hotkey , its easy to
find the doc because it is sorted by title alphabetically
D
1
Scrolling not suitable for big fingers. Did not realise 'date 2013' was a
book. Scroll bar too close to the edge of the hardware lip (raised)
A
2
buttons hard to press
C
1
Complicated search function. Didn't match -
B
5
used library search
G
4
Didn't take too much time
CI
AI
IG
GF
FD
GD
CF
EC
AG
FE
DA
DB
CB
BA
CA
GC
E-Reader evaluation
Ujala Rampaul
Page 60 of 66
C
5
easy to find second book (since we had done it before)
GE
HG
BG
Q7. Access readability and comments on improving the device
e-reader pair ID
ID
BH
B
Comment on Improving
the device
touch sensitivity
H
weight, more responsive
IF
F
ED
E
screen is too small, cannot
control lighting of screen,
no touch screen function
Make it less buttons,
improve navigation
say font size instead of
zoom
D
HF
H
F
AH
A
H
BI
B
I
HE
H
E
EB
E
B
IE
I
E
CH
c
H
E-Reader evaluation
RANK
Comment
needs to be more
responsive to touch
Search function - sort
names by title rather than
author
page design of the buttons
in opposite direction - see
illustration
Add scrollbar to allow
fingers slide the screen
make it more colourful
Improve search
functionality as had to use
2 fingers to type cap
letters.
better usability, touch
screen, response /sensitive
Improve the system
operating speed. Add
number buttons
Improve main navigation
(appears like a roll).
Duplication of arrow?
navigation is always
visible as buttons. Power
button on front - very
laggy unresponsive
faster
Faster loading. Scroll
button can be enter as well
touch screen fewer clicks
to do stuff, improve
sorting
better touch screen
sensitivity, colour, bigger
Ujala Rampaul
Other comments
zoom button opens a
menu rather than actually
zooming
some confusion with
buttons
navigation buttons
weren't immediately
obvious
Really hard to use. Poor
reaction with operation.
Poor arrangement of
screen
accelerate the processing
speed
slightly slow / show time
the touch screen could be
used for this device
on/off response
why must the screen
black whenever the
screen changes
improve Q7 clarity separate document
book opens on last
opened page i.e. sections
already
Page 61 of 66
e-reader pair ID
ID
DI
D
RANK
Comment
Comment on Improving
the device
either have buttons or
touchscreen
Touch screen, menu
options, get specific
location
less buttons, simpler menu
I
HD
H
D
FB
F
Power button should be
placed in a different place,
speed, how to open the
menu?
Add suggestions/ hints for
keys to use. Make screen
touch.
E
A
AF
Touch screen does not
work well.
Counterintuitive buttons
& navigation
Make the screen less like
a touch screen. Reduce
command lag. Add
hardware 'up / down'
buttons in unobstructed
location
Place manuals in folder at
the top of the library
marked manuals
touch screen
A
F
IH
I
H
CI
C
I
AI
A
I
IG
I
G
GF
FD
E-Reader evaluation
G
5
F
5
F
scrolling / swiping
difficult
very slow page changing
Better zooming (font
size). Touch screen
improvements
touchable & adjustable
B
EA
Other comments
easy to read off
the screen
easy to read
Navigation bar is hard to
use. Make the content
dynamic as have to
sequentially access
content.
Not user friendly
Jump n page. Texture is
nice.
Not user friendly
increase the reaction
speed
Add 'delete' button, touch
screen
Change a better CPU, its
running too slow
increase the responding
speed
improve flexibility and
efficiency of use
The button should respond
faster when click. When
click on search button it
should work
fasten feedback, provide
search function
make it touch screen
should have help tips, just
like turn ON the device
back' button is a little
tricky
make it touch screen and
colour display
it is bit faster than
previous device but it
needs to be more faster
it's good the text is clear
each letter & #'s should
have their own button
it takes time to load
make it touchable
Ujala Rampaul
Page 62 of 66
e-reader pair ID
ID
RANK
Comment
Comment on Improving
the device
The arrow icon beside the
power is confusing. Make
it point to the left.
improve the response time
D
GD
CF
G
D
searching time takes way
too long, may improve in
that aspect
C
Colour will improve
usability. Add Directory
like "Home", "back", XX
for users to know where
they are.
functionality to search
document
Content selection in TOC.
Have touch screen included instruction?
Button to move manual
from "recent document"
On screen QWERTY slow
to use with cursor
movement. Also, touch
screen would be better.
Change proportions of
buttons. Make them map
the context
F
EC
E
C
AG
A
G
FE
DC
F
2
E
5
Always
flashing, slow.
The hotkeys at
the sides are
well designed
it's acceptable
D
C
DA
E-Reader evaluation
1
A
3
screen is dull
not enough
backlight for
contrast
the device is too slow
When searching is failed
the device jumps to the
home screen. We think
this design is confusing
How do I zoom in?
Very bad UI design.
Difficult to figure out
how to use.
there are a lot of place
need to improve
stop flashing clear label
on the button
Integrate the "flip' button
and the "more' button to
one button to avoid
confusing.
not use e-paper(better,
faster, display)
better way of entering text
and search should be
optimized
D
Other comments
Brighter whiter
background for contrast.
Device should be bigger.
Make screen flush with
hardware for smoothness
make device touch screen
Ujala Rampaul
Cannot turn the screen
over? At main page
Only use USB from
power. No light for night
time reading
Buttons are a bit
redundant as both sides
and it gets a while to
realise how to use them.
Touch screen could have
been better as display
response is slow.
Controls should be more
visible, standout. It's too
hard to press the buttons.
Device is heavy. Use of
colour to differentiate
buttons & actions
controls are too small and
hard to press. Response is
slow
Page 63 of 66
e-reader pair ID
ID
DB
D
RANK
Comment
Comment on Improving
the device
should remove less use
buttons
function list should be
remap, based on the
frequency of use or some
other criterion
good search
Other comments
keyboard first / improve
screen refresh time
B
faster response time
home button too small
A
make touchscreen, take
away second set of scroll
buttons on side, make
opening documents more
intuitive
to support touch screen
B
CB
BA
CA
C
3
B
3
No backlight
grey display
No backlight
grey display
A
C
GC
to increase the processor
speed
backlight, lighter,
touchscreen
no particular ideas
G
C
GE
G
E
HG
H
3
V Good with
list. V.Bad with
thumbnails
G
BG
B
G
add more 'screen touch
based features'
basic e-reader
to increase the frame per
second (fps)
the buttons on the sides
are confusing
Buttons need to have
more meaningful icons.
Loading time. Can't find
the home back button
The keyboard was not
used. It was difficult to
input numbers, the main
buttons could be bigger on
the interface
My library thumbnails is
far too small. Either make
it better or offer an
intuitive zoom, or have
large thumbnails
it looks quite good to use
screen size needs to be
large
Buttons on screen are
confusing. Hard to search
within documents
Hard to see what device is
selection. Font size button
looks like search button confusing
Good that it is Touch
Q8. User preference from the comparison
Comparison pair ID
BH
Q
Comments for which device is better?
Pandigital better than the Kobo. It responses faster and higher reliability.
Pandigital has better readability than Kobo
IF
Elonex better than iRiver on navigation and ease of use
ED
Sony better than iRiver on navigation and is a touch screen. iRiver is not
intuitive. Sony uses the word 'Zoom' instead of font size, which opens a menu
rather than actually zooming.
E-Reader evaluation
Ujala Rampaul
Page 64 of 66
7
iRiver looks like a touch screen but is not.
AH
Pandigital better than Elonex because it has touch screen, faster response time &
search facility
Neither was any good, hard to do operation but Pandigital better than Nook.
BI
Neither, even though the Pandigital is better than Nook. It still needs to improve
its usability
Kobo better than EZ reader but needs a faster response rate
HF
HE
iRiver is better than Pandigital. The screen is more reliable and response
quicker. The navigation button is flexible.
EB
iRiver is better than Kobo. iRiver is more responsive, intuitive. Even though not
as 'minimal' interface easier to work out what each button does.
IE
iRiver much easier and faster than EZ Reader
CH
Pandigital is more intuitive & has touch, better than kindle
DI
Sony has better navigation & user-friendly than EZReader
HD
FB
None both device Pandigital & Sony are not user friendly. Affordance of Zoom
instead of font size
None, Elonex and Kobo not intuitive. Kobo is slow to respond.
EA
iRiver is better, more user friendly than Nook
AF
Elonex is easier to use, no touch screen, intuitive button controls, compared to
Nook
Pandigital has touch screen and user guide, compared to than EZReader
IH
Kindle faster than EZReader on navigation. Add touch interface to kindle
CI
6
AI
IG
Couldn't locate paragraph in EZReader
Mixed. One preferred Nook because it had touch screen and is better understood
and easier. The other preferred EZReader as found the keyboard was easier to
use & understand
Jetbook rated better than EZReader because user was able to find the required
documents
functions are minimum but can still find documents
buttons actually provide functions when user click on them
GF
FD
GD
Elonex more user friendly than Jetbook. Easy to find documents and it present
the documents nicely.
Sony better than Elonex because it is touchable and convenient for user to
search the documents.
Sony better than Jetbook. Touch screen, user-friendly interface, more fitting into
modern conceptual model of e-readers.
CF
None. Disliked both UI was terrible
EC
AG
Kindle was better than iRiver. Navigating to Document easier on Kindle. iRiver
required 2 separate 4 way nav. buttons to select a section from the TOC. Search
function on Kindle easier to find Docs. Keyboard too complex for an iRiver
suitable for web browsing.
None both devices Nook & jetBook are not user friendly, non intuitive
FE
iRiver better than Elonex because easy to use and beautiful.
DC
None, both devices Sony & kindle is not user friendly /non-intuitive. Preference
for touch, faster, light for night reading
DA
None, both Nook and Sony is non-intuitive. Control on Nook are small and hard
to press.
None, both Sony & Kobo is non-intuitive. Preference for touch screen as iPad
DB
CB
E-Reader evaluation
Kindle better than Kobo. Better feedback, No touchscreen less confusing
navigation
Ujala Rampaul
Page 65 of 66
BA
Kobo better than Nook. I like the function that changing the font size with 2
fingers and other functions like iPad does. Kobo is much more intuitive ,
simpler interface, touch screen, we made a lot less mistakes
CA
Kindle because the Nook is even worse, the operation is confusing.
GC
Kindle better than Jetbook because bit lighter, clearer design, bit faster
GE
jetBook better than iRiver as the keyboard is not user friendly.
HG
Jetbook better than Pandigital because of intuitive navigation. Like the black
sliding button more
Jetbook better than Kobo because better controls, navigation. Kobo is better to
read and the touch screen is good.
BG
7.7 Contents of second document (Dates 2013)
The entire contents of the Dates file:
A1234: 1984
B3942: 2001
C7654: 2013
D3423: 1901
E7777: 1632
F9875: 1492
G1254: 1939
H3110: 1918
I4444: 1995
J7631: 2014
K0078: 2030
E-Reader evaluation
Ujala Rampaul
Page 66 of 66