Research Tools papers

Asian Social Science; Vol. 9, No. 5; 2013
ISSN 1911-2017
E-ISSN 1911-2025
Published by Canadian Center of Science and Education
A Comparison between Two Main Academic Literature Collections:
Web of Science and Scopus Databases
Arezoo Aghaei Chadegani1, Hadi Salehi2, Melor Md Yunus3, Hadi Farhadi4, Masood Fooladi1, Maryam Farhadi1
& Nader Ale Ebrahim5
1
Department of Accounting, Mobarakeh Branch, Islamic Azad University, Mobarakeh, Isfahan, Iran
2
Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University, Najafabad, Isfahan, Iran
3
Faculty of Education, Universiti Kebangsaan Malaysia (UKM), Malaysia
4
School of Psychology and Human Development, Faculty of Social Sciences and Humanities, Universiti
Kebangsaan Malaysia (UKM), Malaysia
5
Research Support Unit, Centre of Research Services, Institute of Research Management and Monitoring (IPPP),
University of Malaya, Malaysia
Correspondence: Arezoo Aghaei Chadegani, Department of Accounting, Mobarakeh Branch, Islamic Azad
University, Mobarakeh, Isfahan, Iran. Tel: 60-17-319-1093. E-mail: [email protected]
Received: February 11, 2013
doi:10.5539/ass.v9n5p18
Accepted: March 18, 2013
Online Published: April 27, 2013
URL: http://dx.doi.org/10.5539/ass.v9n5p18
Abstract
Nowadays, the world’s scientific community has been publishing an enormous number of papers in different
scientific fields. In such environment, it is essential to know which databases are equally efficient and objective
for literature searches. It seems that two most extensive databases are Web of Science and Scopus. Besides
searching the literature, these two databases used to rank journals in terms of their productivity and the total
citations received to indicate the journals impact, prestige or influence. This article attempts to provide a
comprehensive comparison of these databases to answer frequent questions which researchers ask, such as: How
Web of Science and Scopus are different? In which aspects these two databases are similar? Or, if the researchers
are forced to choose one of them, which one should they prefer? For answering these questions, these two
databases will be compared based on their qualitative and quantitative characteristics.
Keywords: web of science, scopus, database, citations, provenance, coverage, searching, citation tracking,
impact factor, indexing, h-index, researcher profile, researcher ID
1. Introduction
Web of Science (WOS) and Scopus are the most widespread databases on different scientific fields which are
frequently used for searching the literature (Guz & Rushchitsky, 2009). WOS from Thomson Reuters (ISI) was
the only citation database and publication which covers all domains of science for many years. However,
Elsevier Science introduced the database Scopus in 2004 and it is rapidly become a good alternative (Vieira &
Gomes, 2009). Scopus database is the largest searchable citation and abstract source of searching literature
which is continually expanded and updated (Rew, 2009). WOS is challenged by the release of Scopus, an
academic literature database which is built on a similar breadth and scale. WOS and Scopus are expensive
products and it may not be feasible to obtain both of them. Therefore, by surfacing the new citation database,
Scopus, scientific libraries have to decide about which citation database can best meet the requests of the
consumers?
The competition between WOS and Scopus databases is intense. This competition has led to improvements in
the services offered by them. Recently, various papers have compared the coverage, features and citation analysis
capabilities of WOS and Scopus (Bakkalbasi, 2006; Burnham, 2006; LaGuardia, 2005; Deis & Goodman, 2005;
Dess, 2006; Li et al., 2010). These comparative studies of WOS and Scopus conclude that these two databases
are permanently improving. They also conclude that the significant advantage of choosing one of these two
sources depends on the particular subject’s area. Some researchers propose undertaking subject’s specific
analysis to find out which database work best for specific fields or time period (Bar-Ilan et al., 2007; Bakkalbasi
et al., 2006; Neuhaus & Daniel, 2008). Based on Lopez-Illescas et al. (2008), prior comparisons of these two
18
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
databases have not exposed a clear winner. They believe that the advantage of one database over another one
depends on what explicitly will be analyzed, the scientific field and also period of analysis. Due to this argument,
the existing issue is: which database is superior to use? Which one should you subscribe to? The answer to this
question is full of equivocation.
In this article, we will briefly analyze the journal evaluation capabilities of WOS and Scopus databases. In fact,
this article concentrates on features and capabilities of each product and these databases will be compared from
different aspects such as: 1) Provenance and coverage 2) Searching and analysis of results 3) Citation tracking
and citation analysis 4) Forming and costs 5) Impact factors 6) Indexing (h-index) 7) Researcher profile and ID
tools. Prior studies which compare these databases will be reviewed in this article.
2. Provenance and Coverage
Web of Science (WOS), products from “Thomson Reuters Institute of Scientific Information” (ISI), arises from
the Science Citation Index created by Eugene Garfield in 1960s. WOS includes above 10,000 journals and
comprises of seven different citation databases including different information collected from journals,
conferences, reports, books and book series. WOS citation databases are Social Sciences Citation Index (SSCI),
Science Citation Index Expanded (SCI Expanded), Conference Proceedings Citation Index Science (CPCI-S),
Arts and Humanities Citation Index (A&HCI) and Conference Proceedings Citation Index-Social Sciences and
Humanities (CPCI-SSH). It has also two chemistry databases named Index Chemicus (IC) and Current Chemical
Reactions (CCR-Expanded). Since WOS is the oldest citation database, it has strong coverage with citation data
and bibliographic data which goes back to 1900 (Boyle & Sherman, 2006). WOS claims it has the most depth and
the most quality however, Scopus burst in research domain in 2004 which claims to have the biggest database with
the wide range of records.
Scopus, officially named SciVerse Scopus, has introduced by Elsevier in November 2004 to the information
market. Scopus is the largest database existing on the market for multidisciplinary scientific literatures (Kuchi,
2004). Scopus covers more than 49 million records including trade publications, open-access journals, and book
series. Almost 80% of these records include abstract. It contains 20,500 peer-reviewed journals from 5,000
publishers, together with 1200 Open Access journals, over 600 Trade Publications, 500 Conference Proceedings
and 360 book series from all areas of science (Rew, 2009). Scopus offers the new sorting and refining feature for
researchers to access above 27 million citations and abstracts going back to 1960s (Boyle & Sherman, 2006).
Most of institutions in all over the world such as Latin America, Europe, North America, Australia, Asia and the
Middle East believe Scopus have positive influence on their researches. Based on Scopus report more than 50%
of its information are from the Middle East, Europe and Africa. Boyle and Sherman (2006) believe that choosing
Scopus is due to its quality of outcomes, time savings, ease of use and possible effect on research findings.
Vieira and Gomes (2009) using the set of Portuguese universities compare tow databases. They conclude that
Scopus provides 20% more coverage than WOS. Scopus covers broader journal range which is limited to recent
articles comparing to WOS. Their results reveal that 2/3 of these records can be found in both databases but 1/3
of the records are only in one database. Vieira and Gomes (2009) suggest the coverage of one journal by Scopus
database can have break. It means Scopus has a partial coverage for some journals. Checking the title list with
journals indexed in Scopus shows that for different journals, the coverage by Scopus is not complete. The
coverage comparison of WOS and Scopus databases provided by free online database comparison tool
‘JISC-ADAT’ is shown in figure 1.
Figure 1. JISC-ADAT coverage comparison of Web of Science and Scopus
19
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
3. Searching and Analysis of Results
WOS and Scopus are powerful databases which provide different searching and browsing options
(Lopez-Illescas et al., 2008). The search options in both databases are the Standard Basic and Advanced. There
are different searchable fields and several document types that permit the user to easily narrow their searching.
Both databases sort the results by parameters such as; first author, cites, relevance and etc. The Refine Results
section in both databases allows the user to quickly limit or exclude results by author, source, year, subject area,
document type, institutions, countries, funding agencies and languages. The resulting documents provide a
citation, abstract, and references at a minimum. Results may be printed, e-mailed, or exported to a citation
manager. The results may also be reorganized according to the needs of the researcher by simply clicking on the
headings of each column. Both databases provide users with the ability to set up a profile for personal settings
and managing their saved searches and alerts all under one tab. A further user option is the ability to browse
individual journals by simply clicking the Journals tab, locate the journal name and select the individual issue.
Figure 2 shows the search screen of WOS and Scopus databases.
Figure 2. Web of Science and Scopus search screen
The analysis of search results is competitive characteristics of these two databases. Therefore, these databases
attempt to improve their features to make searching easier for users. For example, WOS is able to create
histograms and rankings according to different parameters. This feature is a stronger tool than that available in
Scopus. Moreover, The Scientific Web Plus feature available through WOS enables user to see the search results
by clicking on the icon in the search result page. Scientific Web Plus helps users to find scientifically
concentrated open web content, author’s latest information and research scientists’ potential collaborators or
20
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
competitors, and complements searches. Cited Reference Search (CRS) is another features in WOS for finding
previously published and cited articles. WOS also has the “Related Records” feature to detect which authors are
citing the same records. The similar feature in Scopus provides the option to discover documents related to the
authors or keywords in a specific record.
Another capability of Scopus is journal analyzer which allows users to evaluate journals based on the number of
citations, articles published and percentage not cited. One of the weaknesses of Scopus is that it does not have
references before 1996. However, one of the important features that introduced by Scopus to online databases is
the presentation of faceted indexing. This faceted list of metadata has become a main delight to show in users’
training. But, WOS is one step upper than Scopus in representing metadata lists. In fact, search results in WOS are
not presented and analyzed by author, year, subject and document types like Scopus, but also by language,
institutions and countries in WOS.
4. Citation Tracking and Citation Analysis
Citation tracking and analysis have been documented as an important factor for evaluating the influence or
importance of specific record for a period of time. Researchers are able to evaluate records using citation data by
WOS and Scopus Citation Tracker. In fact, Citation tracking indicates the number of times that a particular work,
author or journal have been cited in other works. Citation tracking also allows you to track your own effect and
the effect of your institution. Citation tracking offers complete information for other authors and organizations that
have similar subject and identifies publications which have similar topics (Quint, 2006).
Citation tracking to journal article is another frequent factor of comparing Scopus and WOS (Bergman, 2012).
WOS analyzes the citations by document type, author, funding agency, country, organization’s name, language,
grant number, publication year and research title. WOS citation reports provide two charts, representing citations
in each year and published items in each year. This report also supplies the average citations per item, sum of
times cited and the h-index number (Jacso, 2005). Scopus also provides citation information by date of records’
citation found about specific article or other records. Similar to WOS, it has a feature to find how many times other
authors cited the same topic. Citation overview of Scopus is displayed by option of excluding self-citation with
h-index in a graph format (Jacso, 2005). Therefore, both databases allow authors and researchers to discover how
many citations an author or an article has received, to find the journals and authors who publish in your area of
interest, to explore citations for a specific journal issue, volume or year; to review the citations and work of other
authors (Rew, 2009). There are some studies that compare WOS and Scopus databases from the citation aspects
and they find different results (Levine & Gil, 2009; Haddow & Genoni, 2010; Bergman, 2012). Figure 3 shows
the citation reports in WOS and Scopus.
21
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
Figure 3. Citation reports in Web of Science and Scopus
Levine Clark and Gil (2009) compare two citation databases for citing references from fifteen business and
economic journals. They conclude that Scopus produced slightly higher citation numbers than WOS. Haddow
and Genoni (2010) also find that Scopus have more citations counts than WOS and cover more journals.
However, Bergman (2012) suggest that for the social welfare journals, Scopus commonly provides higher
citation counts than WOS, but the ranking from the highest to the lowest citation count for each database results
in similar outcome. Therefore, WOS and Scopus may likely provide correlated results in terms of ranking based
on citation counts for other journals. Kulkarni et al. (2009) argue that both WOS and Scopus produce
qualitatively and quantitatively diverse citations of articles published in medicine journals. However, Escalona et
al. (2010) find that there is a high similarity between WOS and Scopus in other fields such as Chemical
Engineering.
Jacso (2005) and Falagas et al. (2008) compare the weaknesses and strengths of WOS and Scopus. They confirm
that Scopus covers more journals and its analysis of citations is faster than WOS. However, WOS’s citation
analysis provides more understandable graphics and is more detailed than Scopus’s citation analysis. Meho and
Rogers (2008) investigate the differences of Scopus and WOS databases in the citation ranking, citation counting
and h-index. They find that no significant differences exist between WOS and Scopus if they are compared only
regarding their citation in journals. Levine Clark and Gil (2009) present the result of a comparison of two
databases; WOS and Scopus for fifty journals in business and economics. Citations from both databases are
analyzed to find out if one database is superior to another one, or whether Scopus could replace WOS. They
conclude that researchers may intent to use alternative sources to get a more comprehensive picture of scholarly
impact of one article. Bauer and Bakkalbasi (2005) compare the number of citations in WOS and Scopus for
American Society Information Science and Technology journal for articles published from 1985 to 2000. They
find no difference in citation numbers between WOS and Scopus for articles published in 2000.
5. Forming and Costs
Regarding databases forming and shape, Burnham (2006) argues that Scopus is easy to follow even for beginner
users comparing WOS. LaGuardia (2005) believes that Scopus draws one and attract researcher in at first look.
It's easy to read because fonts and colours provide good contrast and it's less cluttered. Therefore, it is oriented
more toward topics searching than WOS.
Pricing is a critical issue as potential buyers pay attention to the cost and benefit outcomes of their decisions. A
cost comparison between WOS and Scopus is extremely difficult because information about their cost is closely
held by the database producers (Dess, 2006). The results in general show that databases pricing is a complex
issue which depends on the size of the organization, discounts which they negotiated and other elements as well.
Deis and Goodman (2005) claim that the estimation of WOS costs is about $100,000 per year for large
organizations. However, the cost of Scopus database is about 85-95% of the cost of WOS for the same
organizations. LaGuardia (2005) argues that WOS refused to provide pricing information but Scopus pricing is
set according to annual subscription fee with unlimited usage. Pricing varies based on the size of the
22
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
organizations from $20,000 to $120,000 per year. Regarding library budgets, it is highly difficult for any
organization to pay for both of these databases. Therefore, the decision about which one to choose is taken by the
sort of trade-offs of cost versus performance for each database.
6. Impact Factors
The most powerful index which is used by WOS and Scopus databases for ranking journals is the journal’s
impact factor (IF) (Garfield 2006). IF is calculated according to the citations of all articles published in the
journal during two years (Guz & Rushchitsky, 2009). Levine Clark and Gil (2009) define IF as a ratio between
citations and recent citable items published. Therefore, the journal’s IF is calculated through dividing the
numbers of this year citation to the source item which is published in that journal during the prior two years.
Researchers can obtain journals’ IF by using two online databases. WOS, Thomson Reuters’s product, annually
publishes and analyzes indexed journals’ IF. Scopus, Elsevier publisher’s product, also records the information
about journal citations, but does not report indexed journals’ IF, which could be obtained through manual
calculation of journal citations or other tools (Chou, 2012). Nowadays, it is an honour for each journal to indicate
the IF determined by the WOS. However, some scholars have expressed concern that WOS's IF does not
accurately reflect the true influence of social work journals within the discipline partially due to the fact that
many critical journals are left out of the ranking and calculations (Bergman, 2012).
Lopez Illescas et al. (2008) find that for journals indexed in WOS and Scopus databases, those in WOS indicate
higher IFs. On the contrary, those that are only covered by Scopus indicate lower IFs than if they are in both
databases. Furthermore, they find that the differences between two databases regarding citation are much lower
than the differences regarding coverage. Abrizah et al. (2012) compare the ranking, coverage, IF and subject
categorization of Information Science journals based on data from WOS and Scopus. These comparisons are
made according to factor scores reported in 2010 Journal Citation Reports. They use the Library of Congress
Classification System to compare IF and subject categorization. They find that there is high degree of similarity
in rank normalized IF of titles in both WOS and Scopus databases. Pislyakov (2009) reports that the IF ranking
of indexed journals between WOS and Scopus is partially different. However, Gary and Hodkinson (2008)
conclude that there is no significant difference between WOS and Scopus on journals’ IF ranking but these
studies only focus on business and science field. Meho and Sugimoto (2009) investigate differences between
WOS and Scopus to evaluate the scholarly impact of fields which focus on frequently research domains and
institutions, citing journals, conference proceedings, and all citing countries. Their results indicate that when
they assess the smaller citing entities such as journals, institutions and conference proceedings, both databases
produce significantly different results. However, when they assess larger citing entities such as research domains
and countries, they produce similar scholarly impact.
7. Indexing (H-Index)
WOS database is a more scholarly source than Scopus database because of more indexing (Fingerman, 2006).
Some indices are proposed by WOS and Scopus databases such as h-index, g-index, impact factor, the Eigen
factor metric for journal ranking, source normalized impact per paper and relative measure index (Hirsch, 2005;
Egghe, 2006; Bergstrom, 2007; Moed, 2010; Raj & Zainab, 2012). The h-index is a well known metric for
assessing the researcher’s scientific effects (Raj & Zainab, 2012). For measuring h-index, the publication records
of an author, the number of papers published during the selected number of years and the number of citations for
each paper are considered (Moed, 2010). Glanzel (2006) believes that the advantage of h-index is that it
combines the assessment of both the number of papers (quantity) and the impact or citations to these papers
(quality). The h-index is automatically computed in both databases for every author and collections of articles
which are selected by the user.
Jacso (2011) verifies h-index, which is used by researchers to measure research productivity, for the Scopus
database. The results reveal that 18.7 million records in Scopus have one or more cited references, which
represents %42 of the whole database content. The cited references enhanced records ratio is rising from 1996 to
2009. Moreover, for other records of 23,455,354 published after 1995, the h-index is 1,339. Therefore, the total
number of citations should be at least 1,792,921. For the whole Scopus database of 44.5 million records the
h-index is 1,757 (Jacso, 2011). Bar-Ilan (2008) compares the h-index of a list of highly cited authors based on
citations counts recovered from the WOS and Scopus. The results show that there is no difference when citation
tool is used to calculate the h-index of scientists because the results of both databases are very close to each other.
Moreover, it seems to be disciplinary differences in the coverage of the databases. The differences in citation
numbers put science policy makers and promotion committees into the trouble.
23
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
8. Researcher Profile and ID Tools
Both WOS and Scopus databases make available an author search feature by distinguishing the author’s
affiliation, subject and journals to classify variations for the same author. They also have some different features
for researchers to upload their profile information. The Distinct Author Set is a discovery tool in WOS database
which shows several papers which written by one author. In fact, WOS analyzes the citation data, like journals
and subject areas for developing the distinct author set. However, in Scopus, there is a feature called The Author
Identifier. The user can enter the author’s name and will have an exact match of whole name of author, as well as
how his name appears in journal articles and other records. The Scopus Author Identifier (AI) matches author
names according to their affiliation, source title, subject area, address and co-authors. Consequently, the Scopus
AI helps researchers to get accurate and comprehensive information as quickly as possible without wasting time
on searching through long list of author’s names.
Moreover, the research ID feature in WOS allows users to upload their profile information (Vieira & Gomes,
2009). Researcher ID provides a solution to author ambiguity problem in the scholarly research community.
Each member is assigned with unique identifier to enable researcher to manage their publication lists, track their
times cited counts and h-index, identify potential collaborators and avoid author misidentification. Additionally,
your Researcher ID information integrates with the Web of Knowledge and is ORCID compliant, allowing you
to claim and showcase your publications from a single one account. Figure 4 shows Researcher ID feature in
WOS.
Figure 4. Researcher ID in Web of Science
9. Conclusion
Until 2004, the Web of Science (WOS) was the only international and multidisciplinary database available to
obtain the literature of technology, science, medicine and other fields. However, Elsevier introduced Scopus
which is become a good replacement (Vieira & Gomes, 2009). The information provide by these databases
specify the active journals in covering current and relevant research as well as prominent in shaping potential
research fields. The intense competition between these databases motivated researchers to compare them to
identify their similarities and differences. A numerous researches compare these databases from different aspects.
In this article, WOS and Scopus databases are compared based on qualitative and quantitative characteristics
such as provenance, citations, searching and special features by reviewing prior studies.
The comparison of WOS and Scopus discovers that WOS has strong coverage which goes back to 1990 and most
of its journals written in English. However, Scopus covers a superior number of journals but with lower impact
and limited to recent articles. Both databases allow searching and sorting the results by expected parameters such
as first author, citation, institution and etc. regarding impact factor and h-index, different results obtained from
prior studies. Although there is a high association between both databases, researchers interested to know why
authors prefer one database over the other one. For further studies, it is suggested to investigate the perceptions of
authors and researchers on both databases to find the reasons which make them to use one database more than the
other one. It could be helped databases to improve their features to provide better facilities.
24
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
References
Abrizah, A., Zainab, A., Kiran, K., & Raj, R. (2012). LIS journals scientific impact and subject categorization: a
comparison between Web of Science and Scopus. Scientometrics.
Bakkalbassi, N., Bauer, K., Glover, J., & Wang, L. (2006). Three options for citation tracking: Google Scholar,
Scopus and Web of Science. Biomedical Digital Libraries, 3(7). Retrieved from
http://www.bio-diglib.com/content/3/1/7.
Bar-Ilan, J. (2008). Which h-index? – A comparison of Web of Science, Scopus and Google Scholar.
Scientometrics, 74(2), 257-271. http://dx.doi.org/10.1007/s11192-008-0216-y
Bar-Ilan, J., Levene, M., & Lin, A. (2007). Some measures for comparing citation databases. Journal of
Informetrics, 1(1), 26-34. http://dx.doi.org/10.1016/j.joi.2006.08.001
Bauer, K., & Bakkalbasi, N. (2005). An examination of citation counts in a new scholarly communication
environment.
D-Lib
Magazine,
11(9).
Retrieved
November
25,
2007,
from
http://www.dlib.org/dlib/september05/bauer/09bauer.html http://dx.doi.org/10.1045/september2005-bauer
Bergstrom, C. (2007). Eigen factor: Measuring the value and prestige of scholarly journals. College & Research
Libraries News, 68(5), 314-316.
Boyle, F., & Sherman, D. (2006). Scopus: The product and its development. The Serials Librarian, 49(3),
147-153. http://dx.doi.org/10.1300/J123v49n03_12
Burnham, J. F. (2006). Scopus database:
http://dx.doi.org/10.1186/1742-5581-3-1
a
review.
Biomedical
Digital
Libraries,
3(1).
Chou, P. N. (2012). A Comparison Study of Impact Factor in Web of Science and Scopus Databases for
Engineering Education and Educational Technology Journals. Issues in Informing Science and Information
Technology, 9, 187-194.
Deis, L., & Goodman, D. (2005). Web of Science (2004 version) and Scopus. The Charleston Advisor, 6(3).
Retrieved from http://www.charlestonco.com/comp.cfm?id=43
Dess, H. M. (2006). Database reviews and reports: Scopus. Issues in Science and Technology Librarianship,
45(Winter). Retrieved from http://www.istl.org
Egghe, L. (2006). An improvement of the h-index: the g-index. ISSI Newsletter, 2(1), 8-9.
Escalona Fernández, M. I., Lagar Barbosa, P., & Pulgarín Guerrero, A. (2010). Web of Science Vs. Scopus: un
estudio cuantitativo en ingeniería química. Anales de Documentación, 13, 159-175.
Falagas, M. E., Kouranos, V. D., Arencibia-Jorge, R., & Karageorgopoulos, D. E. (2008). Comparison of
SCImago journal rank indicator with journal impact factor. The FASEB Journal, 22(8), 2623-2628.
http://dx.doi.org/10.1096/fj.08-107938
Fingerman, S. (2006). Web of Science and Scopus: Current features and capabilities. Issues in Science and
Technology Librarianship, 48(Fall). Retrieved from http://www.istl.org/06-fall/electronic2.html
Garfield, E. (2006). The history and meaning of the journal impact factor. JAMA-Journal of the American
Medical Association, 295(1), 90-93. http://dx.doi.org/10.1001/jama.295.1.90
Gary, E., & Hodkinson, S. Z. (2008). Comparison of journal citation reports and Scopus impact factors for
ecology and environmental sciences journals. Issues in Science and Technology Librarianship, 54.
Glanzel, W., Schlemmer, B., Schubert, A., & Thijs, B. (2006). Proceedings literature as additional data source for
bibliometric analysis. Scientometrics, 68(3), 457-473. http://dx.doi.org/10.1007/s11192-006-0124-y
Guz, A. N., & Rushchitsky, J. J. (2009). Scopus: A system for the evaluation of scientific journals. International
Applied Mechanics, 45(4), 351-362. http://dx.doi.org/10.1007/s10778-009-0189-4
Haddow, G., & Genoni, P. (2010). Citation analysis and peer ranking of Australian social science journals.
Scientometrics, 85(2), 471-487. http://dx.doi.org/10.1007/s11192-010-0198-4
Hirsch, J. E. (2005). An index to quantify individual’s scientific research output. Proceedings of the National
Academy of Sciences, 102(46), 16569-16572. http://dx.doi.org/10.1073/pnas.0507655102
Jacso, P. (2005). As we may search – Comparison of major features of the Web of Science, Scopus, and Google
Scholar citation-based and citation-enhanced databases. Current Science, 89(9), 1537-1547.
Kulkarni, A. V., Aziz, B., Shams, I., & Busse, J. W. (2009). Comparisons of citation in Web of Science, Scopus
25
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
and Google Scholar for articles published in general medical journals. JAMA, 302, 1092-1096.
http://dx.doi.org/10.1001/jama.2009.1307
Laguardia, C. (2005). E-views and reviews: Scopus vs. Web of Science. Library Journal, 15.
Levine-Clark, M., & Gil, E. L. (2009). A comparative citation analysis of Web of Science, Scopus, and Google
Scholar.
Journal
of
Business
and
Finance
Librarianship,
14(1),
32-46.
http://dx.doi.org/10.1080/08963560802176348
Li, J., Burnham, J. F., Lemley, T., & Britton, R. M. (2010). Citation Analysis: Comparison of Web of Science,
Scopus, SciFinder, and Google Scholar. Journal of Electronic Resources in Medical Libraries, 7(3),
196-217. http://dx.doi.org/10.1080/15424065.2010.505518
Lopez-Illescas, C., Moya-Anegon, F., & Moed, H. F. (2008). Coverage and citation impact of oncological
journals in the Web of Science and Scopus. Journal of Informetrics, 2(4), 304-316.
http://dx.doi.org/10.1016/j.joi.2008.08.001
Meho, L. I., & Rogers, Y. (2008). Citation counting, citation ranking, and h-index of human-computer interaction
researchers: A comparison of Scopus and Web of Science. Journal of the American Society for Information
Science and Technology, 59(11), 1711-1726. http://dx.doi.org/10.1002/asi.20874
Meho, L. I., & Sugimoto, C. R. (2009). Assessing the scholarly impact of information studies: A tale of two
citation databases — Scopus and Web of Science. Journal of the American Society for Information Science
and Technology, 60(12), 2499-2508. http://dx.doi.org/10.1002/asi.21165
Moed, H. F. (2010). Measuring contextual citation impact of scientific journals. Journal of Informetrics, 4(3),
265-277. http://dx.doi.org/10.1016/j.joi.2010.01.002
Neuhaus, C., & Daniel, H. D. (2008). Data sources for performing citation analysis: an overview. Journal of
Documentation, 64(2), 193-210. http://dx.doi.org/10.1108/00220410810858010
Pislyakov, V. (2009). Comparing two thermometers: Impact factors of 20 leading economic journals ac-cording
to
Journal
Citation
Reports
and
Scopus.
Scientometrics,
79(3),
541-550.
http://dx.doi.org/10.1007/s11192-007-2016-1
Quint, B. (2006). Elsevier’s Scopus introduces citation tracker: challenge to Thomson ISI’s Web of Science?
Retrieved from http://newsbreaks.infotoday.com/nbreader.asp?ArticleID=16004
Raj, R. G., & Zainab, Z. N. (2012). Relative measure index: a metric to measure quality. Scientometrics.
http://dx.doi.org/10.1007/s11192-012-0675-z
Vieira, E. S., & Gomes, J. A. N. F. (2009). A comparison of Scopus and Web of Science for a typical university.
Scientometrics, 81(2), 587-600. http://dx.doi.org/10.1007/s11192-009-2178-0
26
Electronic physician; Volume 6, Issue 2, April-June 2014
http://www.ephysician.ir
Letter To Editor
Citation Frequency and Ethical Issue
Nader Ale Ebrahim
Ph.D. of Technology Management, Visiting Research Fellow, Research Support Unit, Centre of Research Services,
Institute of Research Management and Monitoring (IPPP), University of Malaya, Kuala Lumpur, Malaysia
Corresponding Author:
Dr. Nader Ale Ebrahim, Research Support Unit, Centre of Research Services, Institute of Research Management and
Monitoring (IPPP), University of Malaya, Kuala Lumpur, Malaysia. Tel: +60.37967 7812, Fax: +60.37967 7813,
Email: [email protected]
Keywords: citation, impact, ethics
Bibliographic Information of this article:
[Nader Ale Ebrahim. Citation Frequency and Ethical Issue. Electronic Physician, 2014; 6 (2): 814-815]. (ISSN:
2008-5842). http://www.ephysician.ir Available online at: http://www.ephysician.ir/2014/814-815.pdf
This work is licensed under a Creative Commons Attribution 3.0 Unported License.
Received: May.02.2014
Revised: 1st Revision: May.04.2014; 2nd Revision: May.10.2014
Accepted: May.10.2014
Published: May.11. 2014
Dear Editor:
I read your publication ethics issue on “bogus impact factors” with great interest (1). I would like to initiate
a new trend in manipulating the citation counts. There are several ethical approaches to increase the number of
citations for a published paper (2). However, it is apparent that some manipulation of the number of citations is
occurring (3, 4). Self-citations, “those in which the authors cite their own works” account for a significant portion of
all citations (5). With the advent of information technology, it is easy to identify unusual trends for citations in a
paper or a journal. A web application to calculate the single publication h-index based on (6) is available online (7,
8). A tool developed by Francisco Couto (9) can measure authors’ citation impact by excluding the self-citations.
Self-citation is ethical when it is a necessity. Nevertheless, there is a threshold for self-citations. Thomson Reuters’
resource, known as the Web of Science (WoS) and currently lists journal impact factors, considers self-citation to be
acceptable up to a rate of 20%; anything over that is considered suspect (10). In some journals, even 5% is
considered to be a high rate of self-citations. The ‘Journal Citation Report’ is a reliable source for checking the
acceptable level of self-citation in any field of study. The Public Policy Group of the London School of Economics
(LSE) published a handbook for “Maximizing the Impacts of Your Research” and described self-citation rates across
different groups of disciplines, indicating that they vary up to 40% (11).
Unfortunately, there is no significant penalty for the most frequent self-citers, and the effect of self-citation
remains positive even for very high rates of self-citation (5). However, WoS has dropped some journals from its
database because of untrue trends in the citations (4). The same policy also should be applied for the most frequent
self-citers. The ethics of publications should be adhered to by those who wish to conduct research and publish their
findings.
[[email protected]]
Page 814
Electronic physician; Volume 6, Issue 2, April-June 2014
http://www.ephysician.ir
References
1. Jalalian M, Mahboobi H. New corruption detected: Bogus impact factors compiled by fake organizations.
Electron Physician. 2013;5(3):685-6. http://www.ephysician.ir/2013/685-686.pdf
2. Ale Ebrahim N, Salehi H, Embi MA, Habibi Tanha F, Gholizadeh H, Motahar SM, et al. Effective
Strategies for Increasing Citation Frequency. International Education Studies. 2013;6(11):93-9. doi:
10.5539/ies.v6n11p93, http://opendepot.org/1869/1/30366-105857-1-PB.pdf
3. Mahian O, Wongwises S. Is it Ethical for Journals to Request Self-citation? Sci Eng Ethics. 2014:1-3. doi:
10.1007/s11948-014-9540-1
4. Van Noorden R. Brazilian citation scheme outed. Nature. 2013;500:510–1. doi: 10.1038/500510a,
http://boletim.sbq.org.br/anexos/Braziliancitationscheme.pdf
5. Fowler JH, Aksnes DW. Does self-citation pay? Scientometrics. 2007;72(3):427-37. doi: 10.1007/s11192007-1777-2
6. Schubert A. Using the h-index for assessing single publications. Scientometrics. 2009;78(3):559-65. doi:
10.1007/s11192-008-2208-3
7. Thor A, Bornmann L. The calculation of the single publication h index and related performance measures:
a web application based on Google Scholar data. Online Inform Rev. 2011;35(2):291-300.
8. Thor A, Bornmann L. Web application to calculate the single publication h index (and further metrics)
based on Google Scholar 2011 [cited 2014 3 May]. Available from: http://labs.dbs.uni-leipzig.de/gsh/
9. Couto F. Citation Impact Discerning Self-citations 2013 [cited 2014 3 Ma]. Available from:
http://cids.fc.ul.pt/cids_2_3/index.php
10. Epstein D. Impact factor manipulation. The Write Stuff. 2007;16(3):133-4.
11. Public Policy Group L. Maximizing the impacts of your research: a handbook for social scientists. London
School of Economics and Political Science, London, UK.: 2011.
http://www.lse.ac.uk/government/research/resgroups/lsepublicpolicy/docs/lse_impact_handbook_april_201
1.pdf
[[email protected]]
Page 815
"Electronic Physician" is indexed or listed in:
Islamic World Science Citation Center (ISC), Malaysia Citation Center (MCC), Index Copernicus, EBSCO
Host, Directory of Open Access Journals (DOAJ), Genamics Journal Seek , Open-J Gate, Iran Medex
Directory of Biomedical Journals, J Gate, Ind Medica, Geneva Foundation for Medical Education and
Research (GFMER), University of Chicago, Arkansas University, Johns Hopkins Libraries, University of
Miami, University of Arkansas, University of Zurich, Oregon Health & Science University, Directory of
Research Journals Indexing (DRJI), Utrecht University Library, John Brown University online library,
University Library of Regensburg (Electronic Journals Library), Universidad Veracruzana, Library catalog
of the University of Groningen, University library of Saskatchewan, and more.
Asian Social Science; Vol. 9, No. 5; 2013
ISSN 1911-2017
E-ISSN 1911-2025
Published by Canadian Center of Science and Education
Does Criticisms Overcome the Praises of Journal Impact Factor?
Masood Fooladi1, Hadi Salehi2, Melor Md Yunus3, Maryam Farhadi1, Arezoo Aghaei Chadegani1, Hadi Farhadi4
& Nader Ale Ebrahim5
1
Department of Accounting, Mobarakeh Branch, Islamic Azad University, Mobarakeh, Isfahan, Iran
2
Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University, Najafabad, Isfahan, Iran
3
Faculty of Education, Universiti Kebangsaan Malaysia (UKM), Malaysia
4
School of Psychology and Human Development, Faculty of Social Sciences and Humanities, Universiti
Kebangsaan Malaysia (UKM), Malaysia
5
Research Support Unit, Centre of Research Services, Institute of Research Management and Monitoring (IPPP),
University of Malaya, Malaysia
Correspondence: Masood Fooladi, Department of Accounting, Mobarakeh Branch, Islamic Azad University,
Mobarakeh, Isfahan, Iran. E-mail: [email protected]
Received: February 16, 2013
doi:10.5539/ass.v9n5p176
Accepted: March 18, 2013
Online Published: April 27, 2013
URL: http://dx.doi.org/10.5539/ass.v9n5p176
Abstract
Journal impact factor (IF) as a gauge of influence and impact of a particular journal comparing with other
journals in the same area of research, reports the mean number of citations to the published articles in particular
journal. Although, IF attracts more attention and being used more frequently than other measures, it has been
subjected to criticisms, which overcome the advantages of IF. Critically, extensive use of IF may result in
destroying editorial and researchers’ behaviour, which could compromise the quality of scientific articles.
Therefore, it is the time of the timeliness and importance of a new invention of journal ranking techniques
beyond the journal impact factor.
Keywords: impact factor (IF), journal ranking, criticism, praise, scopus, web of science, self-citation
1. Introduction
Citation is the reference text and recognition of article information. Generally, each article has a reference
section at the end including all references cited in that article. Each reference is called a citation. The frequency
of citations to an especial article by other articles means citation count. In addition, citation index as a type of
bibliographic database traces the references in a published article. Citation index shows that how many times an
article has been cited with other articles. In other words, one can find that which later article cites which earlier
articles. Citations are applied to measure the importance of information contained in an article and the effect of a
journal in related area of research activity and publications.
The first citation index for published articles in scientific journals is Science Citation Index (SCI), which is
founded by Eugene Garfield’s Institute for Scientific Information (ISI, previously known as Eugene Garfield
Associates Inc.) in 1960 (Nigam & Nigam, 2012.). Then, it was developed to produce the Arts and Humanities
Citation Index (AHCI) and the Social Sciences Citation Index (SSCI). The SSCI is one of the first databases
developed on the Dialog system in 1972. Small (1973) develops his efforts on Co-Citation analysis as a
self-organizing classification mechanism namely “Research Reviews”. Garner, Lunin, and Baker (1967)
describes that the worldwide citation network has graphical nature. Giles, Bollacker, and Lawrence (1998)
introduce the autonomous citation indexing, which can provide the automatic algorithmic extraction and classify
the citations for any digital scientific and academic articles. Subsequently, several citation indexing services are
created to automate citation indexing including Google Scholar, EBSCOhost, Institute for Scientific Information
(ISI) and Elsevier (Scopus).
Among all citation indexing services, ISI, has dominated the citation indexing career. ISI is a part of Thomson
Reuters, which publishes the citation indexes in form of compact discs and print format. One can access to the
products of ISI through the website with the name ‘Web of Science’ (WOS). It is possible to access seven
databases in WOS including: Social Sciences Citation Index (SSCI), Index Chemicus, Science Citation Index
176
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
(SCI), Arts & Humanities Citation Index (A&HCI), Current Chemical Reactions, Conference Proceedings
Citation Index: Science and Conference Proceedings Citation Index: Social Science and Humanities.
2. Impact Factor
One of the most important product of ISI is the journal impact factor (IF), which is first designed by Eugene
Garfield. IF is published by ISI as a standardized measure to reveal the mean number of citations to the
published articles in particular journal. Journal Citation Reports (JCR) defines the IF as the frequency of citation
to the “average papers” published in a journal in an especial year. In effect, IF is almost used by research
institutions, universities and governmental agencies to measure the influence and impact of a professional
journal comparing with other journals in the same area of research (Kuo & Rupe, 2007). IF is also applied to
evaluate the individual profession of scientists and researchers. Therefore, editors have a high tendency to
increase the IF of their journals.
In order to calculate the IF of a journal, the number of citations to the published papers in a journal, which is
reported in the JCR year, is divided by the total number of papers published in that journal during two previous
years. Citation may be comes from articles published in the same journal or different proceedings, journals or
books indexed by WOS. As an example, an IF of 3.0 reveals that, on average, the papers published in a particular
journal within two years ago, have been cited three times. Although the IF calculation for journals is based on
articles published and cited in the previous two years, one can calculate the average citation using longer time
periods.
The 5-year journal IF is the number of citations to a particular journal divided by the total number of papers
published in the journal since five years ago. In certain fields of study, the 5-year journal IF is more appropriate
than 2-years because the body of citations in 5-years basis might be large enough to make a reasonable
comparison. In some area of research, it takes a long time, more than two years, to publish a work and response
to previous articles.
The aggregate IF is applied for a subject category rather than a journal only (Khan & Hegde, 2009). The
calculation is the same way as the IF for a journal, but it considers the total number of citations to all journals in
a particular subject group and the number of papers published in these cited journals. The median IF refers to the
median value of IF for all journals in the subject category. The number of citations to a journal divided by the
total papers published in the journal in current year is known as immediacy index. The journal’s half-life in each
year refers to the median age of the article cited in JCR. As an example, if the value of journal’s half-life equals
to six in 2007, it means that the citations from 2002 until 2007 to the journal are 50% of all the citations to the
journal in 2007. The journal ranking table displays the ranking of a journal in related subject category based on
the journal IF (Khan & Hegde, 2009).
3. Praises for Impact Factor
IF as a measure of journals’ influence has several advantages and benefits for researchers as well as librarians,
knowledge managers and information professionals. Some of the advantages are listed below:
1) The primary advantage of impact factor is that it is very easy to measure.
2) Another advantage of IF is to mitigate the absolute citation frequencies. Since large journals provide a
considerable body of citable literature which cites to the small journals, IF discounts the benefits of large
journals in favour of small journals. In addition, there is a tendency to discount the benefits of high frequently
published journals in favour of less frequently published journals and also matured journals over the newly
established journals. In other words, journal IF is a remarkable measure for evaluating the journals since it
offsets the benefits of age and size among the journals.
3) IF is a useful tool to compare different research groups and journals. It is commonly applied to administer
scientific library collections. Librarians, knowledge managers and Information professionals encounter with
limited budgets when they aim to select journals for their institutions and departments. ISI provide them with the
IF as a measure to choose most frequently cited journals (Khan & Hegde, 2009). About 9000 social science and
science journals from 60 countries are indexed by Web of Knowledge.
4) IF for indexed journals are greatly available and easy to understand and use. IF is more acceptable and
popular than other alternative measures (Khan & Hegde, 2009).
4. Criticisms for Impact Factor
However, IF has several inherent shortcomings overcoming the advantages of IF. The following problems refer
to the calculation method of IF:
177
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
1) It should be noted that Thomson Scientific is a commercial institution, which sells its products and
evaluations to research institutions and publishers. Data and measures of IF are not publicly established and it is
not easy for scientists to access these data and measures. Therefore, it is not subject to a peer review process.
Researchers demand a fully transparency over how Thomson Scientific collect data and calculates citation
metrics. Thomson Scientific is unable to clarify the used data to support its published IF and hence its IF is not
reliable. Generally, findings of a scientific article could not be accepted by scientists when the primary data is
not available.
2) The coverage of database is not complete since Thomson Scientific excludes some specific types of sources
from the denominator. As an example, books (as a source for citations) are excluded from the database. Citations
from journals which are not indexed in the ISI are not considered in IF calculation (Khan & Hegde, 2009).
Falagas, Kouranos, Arencibia-Jorge, and Karageorgopoulos (2008) assert that one of the major shortcomings of
journal IF is considering only “citable” papers mainly original papers and reviews.
3) IF could not be an appropriate measure because it is an arithmetic mean of number of citations to each paper
and the arithmetic mean is not an appropriate measure (Joint Committee on Quantitative Assessment of Research
2008). IF does not consider the actual quality of research articles, their impressiveness, or the long-term impact
of journals. It should be noted that IF and citation indexes indicate a specific, and relatively uncertain, type of
performance. They do not provide a direct measure of quality (Moed, 2005). Therefore, Using the IF to derive
the quality of individual articles or their authors seems to be idle (Baum, 2011). In addition, the validity and
appropriate use of the IF as a measure of journal importance is a subject to controversy and another independent
examination might reproduce a different IF (Rossner, Van Epps, & Hill, 2007). IF make more controversy when
it is used to assess the articles published in the journals. IF appraises the reputation of the publishing journal not
the quality of the content in the single papers. In addition, Falagas et al. (2008) assert that IF takes into account
received citations only in a quantitative manner. It is conceivable that the published papers in a journal have a
greater influence on science if these papers are cited by articles with a higher scientific quality.
4) The two-year or five-year window for measuring the IF might be logical for some fields of studies with a fast
moving process of research, while it is not reasonable for some fields of study, which requires a long period for
research or empirical validation and also takes a long time for review. These kinds of research might take a time
longer than two years to be completed and then published. Therefore, citation to the original papers will not be
considered in the IF of publishing journal. In addition, articles, which are published in many years ago and are
still cited, have a significant impact on research area, but unfortunately, citation to these articles will not be
considered due to their old age.
5) Distributing the same value of IF to each article published in a same journal leads to excessive variability in
article citation, and provides the majority of journals and articles with the opportunity to free ride on a few
number of highly cited journals and articles. Only some articles are cited anywhere near the mean journal IF. In
effect, IF highly overstates the impact of most articles, while the impact of rest articles is greatly understated
because of the few highly cited articles in the journal (Baum, 2011). Therefore, journal IF could not portray the
individual impact of single articles and it is not much related to the citedness of individual articles in publishing
journal (Yu & Wang, 2007). Seglan (1997) states that IF is wrongly used to estimate the importance of a single
article based on where it is published. Since IF averages over all published papers, it underestimates the
importance of the highly cited papers whereas overestimating the citations to the average publications.
6) Although, it is feasible to evaluate the IF of the journals in which an individual has published papers, the IF is
not applicable to individual scientists or articles. ISI does not distribute a JCR for the humanities. The relevant
number of citations received by a single paper is a better measure of importance and influence of individual
papers as well as its authors (Khan & Hegde, 2009). Garfield (1998) caution regarding the “misuse in evaluating
individuals” by IF due to a wide variation from paper to paper within an individual journal.
7) The IF is highly depended on different disciplines such as physical sciences and mathematical because of the
speed with which articles are cited in a research area (Van Nierop, 2009). It is not logical to use IF as a measure
to compare journals across different disciplines. The absolute value of IF does not make any sense. For example,
a journal with the value of IF equal to two could not have much impact in fields and disciplines such as
Microbiology, while it would be impressive in Oceanography. Therefore, IF as a measure of comparison
between different fields and disciplines could be considered as invalid. In addition, this comparison has been
widely made not only between the journals, but also between scientists or university departments, while,
absolute IF could not estimate the scientific level of department. In an evaluation program, especially for
doctoral degree granting institutions, reviewers consider the IF of journals in examining the scholarly outputs. It
178
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
makes a bias in which some types of researches are undervalued and falsifies the total contribution of each
scientist (Khan & Hegde, 2009).
8) IF is not relevant in some disciplines such as engineering, where the main scientific outputs are technical
reports, conference proceedings and patents. Regarding the literature, IF is not a relevant measure because books
constitute the most important publications in literature, which cite other books.
9) The number of citations from journals in less universal languages and less-developed countries is understated,
since only ISI database are utilized to define the IF.
It is argued that the simple methodology applied in the calculation of IF provides editors with the opportunity to
use different practices in order to inflate the impact factor of their journals (Garfield, 1996; Hemmingsson,
Skjennald, & Edgren, 2002; The PLoS Medicine Editors, 2006). Reviewers may request from the authors to
expand the article’s citation before publishing the articles. This is usual in peer review process in order to
improve the quality of article. On the other hand, IF indicates the importance of a journal in related subject
category. Therefore, editors have a high tendency to increase their journal IF (Yu, Yang, & He, 2011). There are
several methods for a journal to increase its IF. One policy is to manipulate the IF (Dong, Loh, & Mondry, 2005).
References to a journal can be managed in order to increase the number of citations to a journal and hence raise
the IF (Smith, 1997). Therefore, these journals seem to have additional impact and their quality and quantity will
be increased. However, there is no actual improvement in the journal impact and influence. This manipulation of
IF will only result in artificial evaluation.
1) Self-citation is a common way to manipulate and increase the IF of journals (Miguel & Marti-Bonmati, 2002;
Fassoulaki, Papilas, Paraskeva, & Patris, 2002; Falagas & Kavvadia, 2006; Falagas & Alexiou, 2007; Yu et al.,
2011). Editorial board may force authors to expand the article’s citations not necessarily to improve the quality
of scientific articles but they aim to inflate the IF of their journal in order to artificially raising the journal's
scientific reputation (Wilhite & Fong, 2012). This is a business policy, which is called coercive citation. In effect,
editors require authors to cite papers published in the same journal (self-citation) before the journal accepts to
publish the article even if these citations are not relevant to the research. In their study, (Wilhite & Fong, 2012)
find that 20% of researchers in psychology, sociology, economics and multiple business disciplines has
experienced coercive citation. They find that journals with a lower IF have a high tendency for coercive citation
in order to inflate their IF. Another study documents that coercive citation is common in other scientific
disciplines.
2) Another policy to manipulate the IF is to focus on publishing review articles rather than research articles
because review articles have a large body of literature, which are highly citable (Andersen, Belmont, & Cho,
2006; Falagas et al., 2008; Arnold & Fowler, 2011). Therefore, review articles provide the higher IF for
publishing journals and these journals will have the highest IF in their respective research area. Journals have a
tendency to publish a large number of their articles, which are likely to receive a high citation in near future or
deny accepting papers such as case report in medical journals, which are not expected to be cited.
Considering the above drawbacks of IF, it is suggested that IF as a measure of journal evaluation should have
certain characteristics (Offutt, 2008). It should be independent from the length and number of the papers
published in the journal. In addition, it should consider essential advances in the area over short term or
incremental contributions. IF should have relative stability in a year to year basis. The most important is that IF
should evaluate the impact of a journal during a long period since some articles are still cited after 10, 20, or
even 50 years from the date of publication. In addition, we should consider the warnings by Garfield against
using to compare different IF and scientific field. Beside the IF we can consider other internet databases to
calculate the IF of journals or articles which is collaborating with Elsevier.
The journal IF is an indicator of journal performance, and as such many of the suggestions for performance
indicators could equally be used to the application of journal IF (Greenwood, 2007). In November 2007 the
European Association of Science Editors (EASE) published an official statement suggesting that journal IF can
be applied only and cautiously for assessing and comparing the importance and impact of journals, but not for
evaluating the single articles, and of course not for appraising the research programmes or researchers (EASE,
2007). The International Council for Science (ICSU) Committee on Freedom and Responsibility in the conduct
of Science (CFRS) issued a "Statement on publication practices and indices and the role of peer review in
research assessment" in July 2008, which suggest to consider only a limited number of articles published per
year by each scientist or even give a penalty to scientists due to excessive number of publications in each year
(Smith, 1997). The Deutsche Forschungsgemeinschaft (2010), German Foundation for Science, issued new
strategies to assess only papers and no bibliometric information on candidates in decision making process
179
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
regarding:
"...performance-based funding allocations, postdoctoral qualifications, appointments, or reviewing funding
proposals, [where] increasing importance has been given to numerical indicators such as the h-index and
the impact factor".
The citation style of manipulated journals, especially the number of self-citations in the last two year appears to
be unusual. It can be seen that the self-cited rate of manipulated journals is higher than that of the normal journal.
Therefore, self-citation rate can be used as a measure to specify the manipulated journals and calculate the real
IF by deducting the self-citations from the reported IF. However, it is a difficult task to specify manually each
journal in the JCR database. As discussed, journal might publish a larger number of review articles in order to
inflate their IF. Therefore, the Thomson Scientific website can provide directions for removing review articles
from the calculation (Khan & Hegde, 2009).
5. Conclusion
Although IF is a popular measure of quality for journals, we conclude that journal IF has its own restrictions. We
believe that there should be more examination of whether and how IF evaluates journal quality before it is
broadly accepted as a measure of journal quality. The IF measurement would certainly never assess the peer
review process of journals (Offutt 2008). It means that IF is not an appropriate measure to assess the quality of
published articles. One important factor in assessing a progress case is to consider the number of published
articles but it is an imperfect method because it considers the number of articles only instead of the quality of
publications. However, it is very hard to assign professionals to read articles and give an independent opinion
about the impact of particular scientist on his or her research area.
The implication for researchers and journals is that they should not rely only on this indicator. If we do not
consider the above limitations associated with IF, decisions made based on this measure is potentially misleading.
Generally, a measure will fall into disuse and disrepute among the scientific community members, if it is found
to be invalid, unreliable, or ill-conceived. Although IF has many limitations as discussed in this paper, it does not
lose its reputation and application by the scientific society. Indeed, IF attracts more attention and being used
more frequently by scientists and librarians, knowledge managers and information professionals. Critically,
extensive use of IF may result in destroying editorial and researchers’ behaviour, which could compromise the
quality of scientific articles. Calculation of IF and policies to increase the IF by journals, may push researchers to
consider publication as a business rather than contribution to the area of research. It is not fare that we should
rely on such a non-scientific method which IF appraises the quality of our efforts. It is the time of the timeliness
and importance of a new invention of journal ranking techniques beyond the journal impact factor. There should
be a new research trend with the aim of developing journal rankings that consider not only the raw number of
citations received by published papers, but also the influence or importance of documents which issue these
citations (Palacios-Huerta & Volij, 2004; Bollen, Rodriguez, & van de Sompel, 2006; Bergstrom, 2007; Ma,
Guan, & Zhao, 2008). The new measure should represent scientific impact as a function not of just the quantity
of citations received but of a combination of the quality and the quantity.
References
Andersen, J., Belmont, J., & Cho, C. T. (2006). Journal impact factor in the era of expanding literature. Journal
of Microbiology, Immunology and Infection, 39, 436-443.
Arnold, D. N., & Fowler, K. K. (2011). Nefarious numbers. Notices of the American Mathematical Society, 58(3),
434-437.
Baum, J. A. C. (2011). Free-riding on power laws: Questioning the validity of the impact factor as a measure of
research
quality
in
organization
studies.
Organization,
18(4),
449-466.
http://dx.doi.org/10.1177/1350508411403531
Bergstrom, C. (2007). Eigenfactor: Measuring the value and prestige of scholarly journals, College & Research
Libraries News, 68(5), 314-316.
Bollen, J., Rodriguez, M. A., & Sompel, H. (2006). Journal status, Scientometrics, 69(3), 669-687.
http://dx.doi.org/10.1007/s11192-006-0176-z
Dong, P., Loh, M., & Mondry, A. (2005). The impact factor revisited. Biomedical Digital Libraries, 2.
European Association of Science Editors (EASE). (2010). Statement on inappropriate use of impact factors.
Retrieved from http://www.ease.org.uk/publications/impact-factor-statement
Falagas, M. E., & Alexiou, V. G. (2007). Editors may inappropriately influence authors’ decisions regarding
180
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
selection of references in scientific articles. International Journal of Impotence Research, 19, 443-445.
http://dx.doi.org/10.1038/sj.ijir.3901583
Falagas, M. E., & Kavvadia, P. (2006). Eigenlob: Self-citation in biomedical journals. Federation of American
Societies for Experimental Biology Journal, 20, 1039-1042. http://dx.doi.org/10.1096/fj.06-0603ufm
Falagas, M. E., Kouranos, V. D., Arencibia-Jorge, R., & Karageorgopoulos, D. E. (2008). Comparison of
SCImago journal rank indicator with journal impact factor. The Federation of American Societies for
Experimental Biology Journal, 22(8), 2623-2628.
Fassoulaki, A., Papilas, K., Paraskeva, A., & Patris, K. (2002). Impact factor bias and proposed adjustments for
its
determination.
Acta
Anaesthesiologica
Scandinavica,
46,
902-905.
http://dx.doi.org/10.1034/j.1399-6576.2002.460723.x
Garfield, E. (1996). How can impact factors be improved? British Medical Journal, 313, 411-413.
http://dx.doi.org/10.1136/bmj.313.7054.411
Garfield, E. (1998). Der impact faktor und seine richtige anwendung. Der Unfallchirurg, 101(6), 413-414.
Garner, R., Lunin, L., & Baker, L. (1967). Three Drexel information science research studies. Madison: Drexel
Press.
Giles, C. L., Bollacker, K. D., & Lawrence, S. (1998). An automatic citation indexing system, Digital Libraries
98- Third ACM Conference on Digital Libraries, New York, 89-98.
Greenwood, D. C. (2007). Reliability of journal impact factor rankings. BMC Medical Research Methodology,
7(48), 1-6.
Hemmingsson, A., Mygind, T., Skjennald, A., & Edgren, J. (2002). Manipulation of impact factors by editors of
scientific
journals.
American
Journal
of
Roentgenology,
178(3),
767-767.
http://dx.doi.org/10.2214/ajr.178.3.1780767
Joint Committee on Quantitative Assessment of Research. (2008, June 12). Citation Statistics, 1-26.
Khan, & Hegde, (2009). Is impact factor true evaluation for ranking quality measure? Journal of Library &
Information Technology, 29(3), 55-58.
Kuo, W. & Rupe, J. (2007). R-impact: reliability-based citation impact factor. IEEE Transactions on Reliability
56(3), 366-367. http://dx.doi.org/10.1109/TR.2007.902789
Ma, N., Guan, J., & Zhao, Y. (2008). Bringing PageRank to the citation analysis. Information Processing and
Management, 44(2), 800-810. http://dx.doi.org/10.1016/j.ipm.2007.06.006
Miguel, A., & Marti-Bonmati, L. (2002). Self-citation: comparison between Radiologia, European radiology and
radiology for 1997–1998. European Radiology, 12(1), 248-252. http://dx.doi.org/10.1007/s003300100894
Moed, H. F. (2005). Citation analysis in research evaluation. Information Science and Knowledge Management.
Springer.
Nigam, A., & Nigam, P. K. (2012). Citation index and impact factor. Indian Journal of Dermatology,
Venereology, and Leprology, 78(4), 511-516. http://dx.doi.org/10.4103/0378-6323.98093
Offutt, J. (2008). The journal impact factor. Journal of software testing verification and reliability, 18(1).
Palacios-Huerta, I., & Volij, O. (2004). The measurement of intellectual influence. Econometrica, 72(3), 963-977.
http://dx.doi.org/10.1111/j.1468-0262.2004.00519.x
Rossner, M., Van Epps, H., & Hill, E. (2007). Show me the data. Journal of Cell Biology, 179(6), 1091-1092.
http://dx.doi.org/10.1083/jcb.200711140
Seglen, P. O. (1997). Why the impact factor of journals should not be used for evaluating research. British
Medical Journal, 314(7079), 498-502. http://dx.doi.org/10.1136/bmj.314.7079.497
Small, H. (1973). Co-citation in the scientific literature: A new measure of the relationship between two
documents. Journal of the American Society for Information Science, 24(4), 265-269.
http://dx.doi.org/10.1002/asi.4630240406
Smith, R. (1997). Journal accused of manipulating impact factor. British Medical Journal, 314(15), 461.
http://dx.doi.org/10.1136/bmj.314.7079.461d
The PLoS Medicine Editors. (2006). The impact factor game: It is time to find a better way to assess the
scientific literature. PLoS Med., 3, 291. http://dx.doi.org/10.1371/journal.pmed.0030291
181
www.ccsenet.org/ass
Asian Social Science
Vol. 9, No. 5; 2013
Van Nierop, E. (2009). Why do statistics journals have low impact factors? Statistica Neerlandica, 63(1), 52-62.
http://dx.doi.org/10.1111/j.1467-9574.2008.00408.x
Wilhite, A. W., & Fong, E. A. (2012). Coercive citation in academic publishing. Science, 335(6068), 542-543.
http://dx.doi.org/10.1126/science.1212540
Yu, G., & Wang, L. (2007). The self-cited rate of scientific journals and the manipulation of their impact factors.
Scientometrics, 73(3), 356-366. http://dx.doi.org/10.1007/s11192-007-1779-8
Yu, G., Yang, D. H., & He, H. X. (2011). An automatic recognition method of journal impact factor manipulation.
Journal of Information Science, 37(3), 235-245. http://dx.doi.org/10.1177/0165551511400954
182
Australian Journal of Basic and Applied Sciences, 7(4): 198-202, 2013
ISSN 1991-8178
Does it Matter Which Citation Tool is Used to Compare the h-index of a Group of
Highly Cited Researchers?
1
Hadi Farhadi, 2Hadi Salehi, 3Melor Md Yunus, 4Arezoo Aghaei Chadegani, 4Maryam Farhadi,
4
Masood Fooladi and 5Nader Ale Ebrahim
1
School of Psychology and Human Development, Faculty of Social Sciences and Humanities,
Universiti Kebangsaan Malaysia (UKM), Malaysia
2
Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University, Najafabad,
Isfahan, Iran
3
Faculty of Education, Universiti Kebangsaan Malaysia (UKM), Malaysia
4
Department of Accounting, Mobarakeh Branch, Islamic Azad University, Mobarakeh, Isfahan, Iran
5
Department of Engineering Design and Manufacture, Faculty of Engineering, University of Malaya,
Kuala Lumpur, Malaysia
Abstract: h-index retrieved by citation indexes (Scopus, Google scholar, and Web of Science) is used
to measure the scientific performance and the research impact studies based on the number of
publications and citations of a scientist. It also is easily available and may be used for performance
measures of scientists, and for recruitment decisions. The aim of this study is to investigate the
difference between the outputs and results from these three citation databases namely Scopus, Google
Scholar, and Web of Science based upon the h-index of a group of highly cited researchers (Nobel
Prize winner scientist). The purposive sampling method was adopted to collect the required data. The
results showed that there is a significant difference in the h-index between three citation indexes of
Scopus, Google scholar, and Web of Science; the Google scholar h-index was more than the h-index in
two other databases. It was also concluded that there is a significant positive relationship between hindices based on Google scholar and Scopus. The citation indexes of Scopus, Google scholar, and Web
of Science may be useful for evaluating h-index of scientists but they have some limitations as well.
Key words: h-index, Scopus, Google Scholar, Web of Science, Nobel Prize, Physics, Chemistry,
Economic Sciences.
INTRODUCTION
Nowadays, when citation information is needed there are a few sources to rely on. The most comprehensive
web version of citation indexes sources are Scopus (http://www.scopus.com/), Google Scholar
(http://scholar.google.com/), and the Web of Science (http://portal.isiknowledge.com/). Scopus, Google scholar,
and Web of Science are regarded as the most useful and trustful source for searching. They are valuable tools
for searching which provide the citation searching and ranking by times cited (Mikki, 2009). These three
sources of citation indexes are multidisciplinary and international coverage and used to assess the scientific
output worldwide (Bar‐Ilan, 2008).
Background:
To quantify the research output of a single scientist, Hirsch (2005) introduced the h-index. h-index is
calculated by the number of publications and the number of citations received. As defined by Hirsch (2005), a
scientist has index h if h of his or her Np papers have at least h citations each and the other (Np-h) papers have ≤
h citations each. It is a single number which supports a good representation of the scientific lifetime
achievement (Egghe & Rousseau, 2006; van Raan, 2006). For example, if a scientist has published 15 papers
that each had at least 15 citations, h-index of him/her will be 15. Although this single number is simple to
compute and it takes into account both the quantity and impact of the papers, it is essential to consider which
database is reporting this single number (h-index).
h-index is computed based upon the data from the aforementioned citation indexes to measure the scientific
output of a researcher. The data from these citation indexes are important for scientists and universities in all
over the world. For instance, h-index is one of the impressive and effective factors used by promotion
committees to evaluate research productivity and impact at universities. Although, citing and cited documents
from these citation tools are associated to each other, h-index can be obtained separately from different citation
databases of Scopus, Google Scholar, and Web of Science. That is because each of these citation indexes has a
Corresponding Author: Hadi Farhadi (School of Psychology and Human Development, Faculty of Social Sciences and
Humanities, Universiti Kebangsaan Malaysia (UKM), Malaysia
E-mail: [email protected]
198
Aust. J. Basic & Appl. Sci., 7(4): 198-202, 2013
different collection policy which influences both the publications covered and the number of citations (Bar‐Ilan,
2008). Hence, the outputs and the results from these sources of citation indexes should not be approximately
similar.
This study aims to distinguish the major differences on h-index of scientists between three citation
databases of Scopus, Google Scholar, and Web of Science. Therefore, based upon the literature, this study
attempts to answer the following research questions:
1) How much the results from these sources of citation indexes are different?
2) Can universes and promotion committees use the h-index of a scientist as a tool for synthesizing and
depicting the scientific production in a more exhaustive way?
To answer the above research questions, the current study investigated the differences between the outputs
and the results from these three citation databases (Scopus, Google Scholar, and Web of Science) based upon
the h-index of a group of highly cited researchers (Nobel Prize winner scientists).
Methodology:
There exist three commonly web versions of citation indexes sources to measure the h-index of a scientist.
Hence, this study computed the h-index for 12 Nobel Prize winners based upon data from Scopus, Google
Scholar, and Web of Science for three different fields of study namely physics, chemistry and economic
sciences. Then, we compared the h-index data of these 12 Nobel Prize winners to find out the correlation
between the h-index of three mentioned citation indexes sources. Purposive sampling method was used to
collect the required data. To do so, we used the official web site of the Nobel Prize (http://www.nobelprize.org/)
as our starting point, which lists the all Nobel Prize winners.
Every year, the Nobel Foundation in Stockholm, Sweden, presents an international award called “Nobel
Prize” to those who have made outstanding contributions to their disciplines. The Nobel Prize includes a medal,
a diploma, and cash award given annually to those who during the preceding year have given the greatest benefit
to human beings. Every year, the respective Nobel Committees invite thousands of university professors,
members of academies and scientists from different countries, previous Nobel laureates, and members of
parliamentary assemblies to submit candidates for the Nobel Prizes for the coming year. These nominators are
chosen in such a way that as many countries and universities as possible are represented over time.
For this study, we selected 12 scientists from the official website of the Nobel Prize who have won the
Nobel Prize in three different fields of physics, chemistry, and economic sciences. Then the name of each
scientist was searched in three citation indexes sources of Scopus, Google Scholar, and Web of Science (see
Figures 1, 2 and 3). We brought Serge Haroche (the physics Nobel prize 2012 winner) as an example in
following figures. As can be seen in the following figures, the h-index of Serge Haroche is 34, 35 and 21 based
upon Scopus, Google Scholar, and Web of Science, respectively.
Fig. 1: Serge Haroche’s h-index by Scopus is 34.
199
Aust. J. Basic & Appl. Sci., 7(4): 198-202, 2013
Fig. 2: Serge Haroche’s h-index by Google Scholar is 35.
Fig. 3: Serge Haroche’s h-index by Web of Science is 21.
RESULTS AND DISCUSSION
Table 1 shows the h-index of 12 Nobel Prize winner scientists in three different citation indexes of Scopus,
Google Scholar and Web of Science.
Table 1: h-index of 12 Nobel Prize winner scientists in three different citation indexes.
Noble Prize Winner’s Name
Year
Research field
Web of Science
Serge Haroche
2012
Physics
21
David J. Wineland
2012
Physics
20
Saul Perlmutter
2011
Physics
35
Brian P. Schmidt
2011
Physics
21
Robert J. Lefkowitz
2012
Chemistry
31
Brian K. Kobilka
2012
Chemistry
24
Dan Shechtman
2011
Chemistry
11
Akira Suzuki
2010
Chemistry
85
Alvin E. Roth
2012
Economic Sciences
4
Thomas J. Sargent
2011
Economic Sciences
11
Christopher A. Sims
2011
Economic Sciences
13
Peter A. Diamond
2010
Economic Sciences
6
200
Scopus
34
47
38
46
106
63
5
56
28
21
13
14
Google Scholar
35
23
32
62
167
70
13
79
68
77
64
61
Aust. J. Basic & Appl. Sci., 7(4): 198-202, 2013
Our first step in understanding the differences between Scopus, Google Scholar, and Web of Science based
upon the scientists’ h-index was to examine the means of each citation tool resource. Figure 4 shows the mean
of h-index for Nobel Prize winner scientists based upon data from Google Scholar (mean=62.58), Scopus
(39.25), and Web of Science (23.5). As can be seen in Figure 4, the mean of Google scholar h-index (mean =
62.58) for scientist in this research is more than the mean of Scopus h-index (mean = 39.25) and mean of Web
of Science h-index (mean = 23.5). The results can be reasonable because Google Scholar covers more articles
indexed than other two citation tools and it contains books and conference proceedings which may alter
considerable citation metrics (Vanclay, 2007; Bar‐Ilan, 2008; Mikki, 2009).
Fig. 4: Comparing the h-index of 12 Nobel Prize winners based upon data from Scopus, Google Scholar and
Web of Science.
Table 2 also shows the multiple regression between the h-index of 12 Nobel Prize winners based upon data
from Scopus, Google Scholar and Web of Science. The multiple regression results show that there is a positive
significant relationship between Scopus and Google Scholar h-index (β=.799, p<.05). This indicates that more
h-index in Scopus citation tool corresponds with more Google Scholar h-index. In other words, we can conclude
that the h-index of Scopus can predict the Google scholar h-index. However, this study could not find a
significant relationship between the h-index in Web of Science and Google Scholar h-index.
Table 2: Multiple regression between the h-index of 12 Nobel Prize winners based upon data from Scopus, Google Scholar and Web of
Science.
B
Beta
t
Constant
24.761
1.561
Web of Science
-.300
-.164
-.639
Scopus
1.143
.799
3.122*
R2= 0.54, F= 5.32, p<0.05
Based upon the samples’ field of studies, the results showed that the h-index mean of chemistry (mean =
59.17) is more than the h-index mean of physics (mean = 37.5) and economy (mean = 31.67). It means that the
citation and publication in chemistry area are more than those of physics and economy (see Figure 5).
Fig. 5: Comparing the h-index of 12 Nobel Prize winners based upon their research area.
201
Aust. J. Basic & Appl. Sci., 7(4): 198-202, 2013
Conclusions:
h-index retrieved by citation indexes of Scopus, Google Scholar, and Web of Science is used to measure the
scientific performance and the research impact studies based upon the number of publications and citations of a
scientist. It also is easily available and may be used for performance measures of scientists and for recruitment
decisions. As the results showed for the samples of this study, the difference in the h-index between three
citation indexes of Scopus, Google Scholar and Web of Science is obvious and noticeable. The Google Scholar
h-index is more in comparison with the h-index in two other databases. Moreover, the findings showed that
there is a significant relationship between h-indices based upon Google Scholar and Scopus.
We can conclude that it matters which citation tools are used to compute the h-index of scientists.
Therefore, the universities and promotion committees need to consider which citation tool is used to compute
the h-index of a researcher or a lecturer. In general, when we intend to report the h-index of a scientist, we need
to consider and mention the sources of citation indexes. We cannot compare the h-index of two scientists
without mentioning the sources of citation indexes. The citation indexes of Scopus, Google Scholar and Web of
Science might be useful for evaluating the h-index of scientists but they have their own limitations as well.
REFERENCES
Bar‐Ilan, J., 2008. Informetrics at the beginning of the 21st century: A review. Journal of Informetrics, 2(1):
1‐52.
Bar‐Ilan, J., 2008. Which h‐index? A comparison of WoS, Scopus and Google Scholar. Scientometrics,
74(2): 257‐271.
Egghe, L., R. Rousseau, 2006. An informetric model for the Hirsch-index, Scientometrics, 69(1): 121-129.
Google Scholar. (No Date). About Google Scholar. Retrieved 27 January 2013, from
http://scholar.google.com/.
Hirsch, J.E., 2005. An index to quantify an individual’s scientific research output. Proceedings of the
National Academy of Sciences, 102(46): 16569‐16572.
ISI WoS. Thomson Scientific. (No Date). Web of Science. Retrieved 27 January 2013, from http://
http://portal.isiknowledge.com/.
Mikki, S., 2009. Google Scholar Compared to Web of Science: A Literature Review. Nordic Journal of
Information Literacy in Higher Education, 1(1): 41‐51.
Nobel Prize. (No Date). About Nobel Prize. Retrieved 27 January 2013, from http://www.nobelprize.org/.
Scopus. (No Date). About Scopus. Retrieved 27 January 2013, from http://www.scopus.com/.
Van Raan, A.F.J., 2006. Comparison of the Hirsch-index with standard bibliometric indicators and with
peer judgment for 147 chemistry research groups. Scientometrics, 67(3): 491-502.
Vanclay, J.K., 2007. On the robustness of the h‐index. Journal of the American Society for Information
Science and Technology, 58(10): 1547‐1550.
202
International Journal of Management Science
and Business Administration
Volume 1, Issue 3, February 2015, Pages 6-15
Does a Long Reference List Guarantee More Citations?
Analysis of Malaysian Highly Cited and Review Papers
Nader Ale Ebrahim1, H. Ebrahimian2, Maryam Mousavi3, Farzad Tahriri3
1
Research Support Unit, Centre of Research Services, Institute of Research
Management and Monitoring (IPPP), University of Malaya, Malaysia
2
3
Institute of Mathematical Sciences, Faculty Science, University Malaya.
Centre for Product Design and Manufacturing, Department of Mechanical
Engineering, Faculty of Engineering, University of Malaya, 50603, Kuala Lumpur,
Malaysia
Corresponding author (e-mail):[email protected]
Abstract: Earlier publications have shown that the number of references as well as the number of
received citations are field-dependent. Consequently, a long reference list may lead to more
citations. The purpose of this article is to study the concrete relationship between number of
references and citation counts. This article tries to find an answer for the concrete case of
Malaysian highly cited papers and Malaysian review papers. Malaysian paper is a paper with at least
one Malaysian affilation. A total of 2466 papers consisting of two sets, namely 1966 review papers
and 500 highly-cited articles, are studied. The statistical analysis shows that an increase in the
number of references leads to a slight increase in the number of citations. Yet, this increase is not
statistically significant. Therefore, a researcher should not try to increase the number of received
citations by artificially increasing the number of references.
Key words: H-index, Citation analysis, Bibliometrics, Impact factor, Performance evaluation,
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
Relations between citations and references
1. Introduction
Researchers seeking citation tracking to find the most influential articles for a particular topic and to
see how often their own published papers are cited (Bakkalbasi et al. 2006). On the other hand
universities are looking for citations because of its influence in the university ranking (Ale Ebrahim
et al. 2013, Ioannidis 2010, Bornmann, Leydesdorff, and Wang 2014). A citation count is the
number of times a research work such as a journal article is cited by other works. The citation per
paper meaningfully influence a number of metrics, including total citation counts, citation speed, the
ratio of external to internal cites, diffusion scores and h-index (Carley, Porter, and Youtie 2013).
Citation counts still commonly use for the measure of research papers quality and reputation (Abt
and Garfield 2002). The number of citations that an article receives measured its impact on a
specific field (Lai, Darius, and Lerut 2012). Citation analysis is one of the most important tools to
evaluate research performance (Bornmann et al. 2012). Citation indicator is important for scientists
and universities in all over the world (Farhadi, Salehi, Yunus, et al. 2013). In the early stage, the
relationship between the number of references and the number of the paper citation was
investigated in the 1965 (UZUN 2006, de Solla Price 1965). A long reference list at the end of a
research paper may be the key to ensuring that it is well cited (Corbyn 2010, Ball 2008). Hence,
citation counts are correlated with reference frequencies (Abt and Garfield 2002). Webster,
Jonason, and Schember (2009) raised the question “Does the number of references an article
contains predict its citation count?” and found that reference counts explained 19% of the variance
in the citation counts. Lancho-Barrantes, Guerrero-Bote, and Moya-Anegón (2010) found that not
only the number, but also the citation impact of the cited references correlated with the citation
counts for a paper. The higher the impact of the cited references, the higher the later impact of the
citing paper (Bornmann et al. 2012). Review articles are usually highly cited compare to other types
of papers (Meho 2007).
Review papers represent the existing knowledge in a given field and more likely to be cited
(Alimohammadi and Sajjadi 2009). Several bibliometric studies highlighted that citation counts are a
function of many factors besides the scientific quality (Bornmann et al. 2012), length of paper (Abt
and Garfield 2002), visibility (Ale Ebrahim et al. 2014), optimize scholarly literature for academic
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
search engines (Beel, Gipp, and Wilde 2010), add the name of study in the title of all publications
(Sarli and Holmes 2011), publishing in a journal with higher impact factor (Vanclay 2013), internet
usage (Farhadi, Salehi, Embi, et al. 2013), gross domestic product (GDP) (Gholizadeh et al. 2014),
number of authors (Krause 2009), self-archiving (Gargouri et al. 2010), publish in an open access
journal (Swan 2010), collaborate with international authors (Pislyakov and Shukshina 2012), write
paper with a Nobel laureates (Ball 2011) and many other (Ale Ebrahim et al. 2013) including write a
review paper (Vanclay 2013) and use more references (Corbyn 2010). In this study the relationship
between number of references and citation counts is determined. Webster, Jonason, and Schember
(2009) mentioned “On average, review articles actually showed less of the relationship than
standard articles” (Corbyn 2010). So, in this research both review and standard articles (papers)
were investigated. 2466 articles consist of 1966 Malaysian review and 500 highly cited papers were
selected to examine the relationship between number of references and citation counts in the given
article.
2.Materials and methods
All data were obtained through Web of Science online academic database provided by Thomson
Scientific. This database included the necessary information to examine the relationship between
reference and citation counts for every review and highly cited papers published in Malaysia since
1980 to October 2013. Science Citation Index Expanded, Social Sciences Citation Index and Arts &
Humanities Citation Index, were searched for reviews and highly cited papers. For each paper, all
Bibliometrics data, especially the number of references and the number of times the paper has been
cited during the interval between the year of publication and the year 2013, have been
collected.Two samples set were selected: 1- The sample number one consisted of 1966 review
papers in all disciplines from Malaysia, according to the Web of Knowledge’s classification system.
Citation statistics produced by shorter than three years’ time frame may not be sufficiently stable
(Adams 2005, UZUN 2006). Because, papers appearing in the Web of Science databases over the
last few years, have not had enough time to accumulate a stable number of citations (Webster,
Jonason, and Schember 2009). Therefore, the time span limited from 1980 to November, 2010;
yielding a subsample of 721 publications (37% of the original sample). Publications with zero citation
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
were removed. In order to select the highly cited paper a threshold 10 times cited per year is
considered. The association between the number of references (independent variable) and time
cited per year (dependent variable) of highly cited review papers investigated with linear and nonlinear models. 2- The sample number two comprises 500 highly cited publications from Malaysia.
According to the Web Of Science classification, the results are obtained based on the article type
and exclude the review articles, editorial material, conference papers and book review.
3. Results and discussion
Two sets of data 1- 1966 review papers and 2- 500 high cited papers, were investigated separately.
The results and discussions are coming as follows.
Outliers for sample one (1966 review papers)
Due to the effect of the age of an article, the number of citations cannot be a reference of highly
cited paper. Therefore, the citation per year selected as a reference for highly cited paper. Papers
with 10 times cited per year is considered as highly cited paper. Figure 3-1 shows the number of
times cited per year for 660 review papers. A threshold was visually determined on 50 times cited
per year. Papers with more than 50 times cited yearly is called “extremely high cited paper” and
detected as outliers. Papers with more than 300 listed references also detected as outliers (3-2).
Figure 3-1 Number of times cited per year vs number of review papers references
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
Figure 3-2 Number of times cited per year vs number of references in review paper
Correlation analysis for sample one (1966 review papers)
The correlation between variables was modeled with regression model, linear model
y = α x + β and exponential model, non-linear model y = α eβx. The goodness of both model was
then measured with Spearman's rho
(ρ ) , Kendall's tau (τ ) and Pearson correlation coefficient (r ) .
The result of correlation analysis is summarized in 3-1.
Coefficient Statistic value P-value
Equation
ρ
0.127
0.193 y = 0.0639x + 10
0.0041x
τ
0.112
0.093 y = 10e
0.152
0.120
r
Model
Linear
Non-linear
Table 3-1 The result of correlation analysis of highly-cited review papers
The association between variables is graphically illustrated with scatter plots. The trend of these
associations was drawn with solid lines. Refer to Figure 3 and Figure 4, both linear and non-linear
models are not significantly fitted, trends are positive which support the hypothesis “For a given
review paper, increasing in the number of references may have result of increasing the times cited
per year”.
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
Figure 2-3 Relationship between number of references and citation counts in review papers (linear model)
Figure 3-3 Relationship between number of references and citation counts in review papers (Exponential
model)
Outlier detection for sample two (500 highly cited papers)
Papers with 10 times cited per year is considered as highly cited paper. Papers that cited more than
100 times per year is considered as extremely high cited paper and detected as an outlier. Figure 5
and Figure 6 are showing raw data and filtered data respectively.
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
Figure 3-4 Raw data - Number of times cited per year vs number of references 500 highly cited papers
Figure 3-5 Filtered data - Number of times cited per year vs number of references in 500 highly cited
papers
Correlation analysis for sample two (500 highly cited papers)
The association between the number of references (independent variable) and time cited per year
(dependent variable) of first 500 high cited papers investigated with linear and non-linear model
correlation analysis. The correlation was modeled with regression model, linear model y = α x + β
and exponential model, non-linear model y = α eβx. The goodness of fit was then measured with
Spearman's rho
(ρ ) ,
Kendall's tau
(τ ) and
Pearson correlation coefficient
(r ) .
The result of
correlation analysis is summarized in Table 3-2.
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
Coefficient Statistic value P-value
Equation
ρ
0.08
0.914 y = 0.03 x + 2.4
0.08 x
τ
0.04
0.467 y = 1.1 e
0.06
0.419
r
Model
Linear
Non-linear
Table 3-2 The result of correlation analysis of 500 highly cited papers
The association between variables is graphically illustrated with scatter plots. The trend of these
associations is shown by the solid lines. Figure 3-7 and Figure 3-8 shows, although both linear and
non-linear models are not significantly fitted, positive values of correlation coefficients are still
suggesting a positive trend (positive correlation) on the number of references and the number of
times cited per year.
Figure 3-6 Relationship between number of references and citation counts in 500 highly cited (linear model)
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
Figure 3-7 Relationship between number of references and citation counts
in 500 highly cited (Exponential Model)
4. Conclusion
This study shows that since the trend between the citation count and the number of references is
not statistically significant, we cannot conclude that there is a significant association between the
citation count of Malaysia review papers between the given period and number of references
contained in the paper. The correlation coefficient is not statistically significant. However, r = 0.152
based on the population of 721 articles. Malaysian review papers get more citations than other
types of papers. The number of references in the article has the lowest impact on the citation
compares with review paper. As this study looked only Malaysia review papers and 500 highly-cited
article, it would be necessary to conduct a similar study in the otherworld and types of papers. It
would be important to examine whether in other types of papers the relationship investigated here
have significant correlated or not. The research considered the general definition of citations.
Therefore, future studies may make a diffrentianain between “perfunctory citations” and “organic
citations” citations as Tang and Safer (2008) defined “perfunctory citations” is occurred only once
and in the introduction, “organic citations” as references cited for “conceptual ideas” and
“methodology and data” reasons.
ACKNOWLEDGEMENT
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
Sincere thanks to Dr. Bojan Obrenović and the International Journal of Management Science and
Business Administration’s board members for their useful advices.
References
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Abt, Helmut A., and Eugene Garfield. 2002. "Is the relationship between numbers of references
and paper lengths the same for all sciences?" Journal of the American Society for Information
Science and Technology 53 (13):1106-1112. doi: 10.1002/asi.10151.
Adams, Jonathan. 2005. "Early citation counts correlate with accumulated impact."
Scientometrics 63 (3):567-581. doi: 10.1007/s11192-005-0228-9.
Ale Ebrahim, Nader, Hadi Salehi, Mohamed Amin Embi, Farid Habibi Tanha, Hossein Gholizadeh,
and Seyed Mohammad Motahar. 2014. "Visibility and Citation Impact." International Education
Studies 7 (4):120-125. doi: 10.5539/ies.v7n4p120.
Ale Ebrahim, Nader, Hadi Salehi, Mohamed Amin Embi, Farid Habibi Tanha, Hossein Gholizadeh,
Seyed Mohammad Motahar, and Ali Ordi. 2013. "Effective Strategies for Increasing Citation
Frequency." International Education Studies 6 (11):93-99. doi: 10.5539/ies.v6n11p93.
Alimohammadi, Dariush, and Mahshid Sajjadi. 2009. "Correlation between references and
citations." Webology 6 (2):a71.
Bakkalbasi, Nisa, Kathleen Bauer, Janis Glover, and Lei Wang. 2006. "Three options for citation
tracking: Google Scholar, Scopus and Web of Science." Biomedical Digital Libraries 3 (1):7. doi:
10.1186/1742-5581-3-7.
Ball, Philip. 2008. "A longer paper gathers more citations." Nature 455 (7211):274-275. doi:
10.1038/455274a.
Ball, Philip. 2011. "Are scientific reputations boosted artificially?" Nature, 6 May.
Beel, Jöran, Bela Gipp, and Erik Wilde. 2010. "Academic Search Engine Optimization (ASEO)."
Journal of Scholarly Publishing 41 (2):176-190. doi: 10.3138/jsp.41.2.176.
Bornmann, L., L. Leydesdorff, and J. Wang. 2014. "How to improve the prediction based on
citation impact percentiles for years shortly after the publication date?" Journal of Informetrics
8 (1):175-180. doi: 10.1016/j.joi.2013.11.005.
Bornmann, Lutz, Hermann Schier, Werner Marx, and Hans-Dieter Daniel. 2012. "What factors
determine citation counts of publications in chemistry besides their quality?" Journal of
Informetrics 6 (1):11-18. doi: http://dx.doi.org/10.1016/j.joi.2011.08.004.
Carley, S., A. L. Porter, and J. Youtie. 2013. "Toward a more precise definition of self-citation."
Scientometrics 94 (2):777-780. doi: 10.1007/s11192-012-0745-2.
Corbyn, Zoë. 2010. "An easy way to boost a paper's citations."
Nature 406. doi:
10.1038/news.2010.406
de Solla Price, Derek J. 1965. "Networks of Scientific Papers." Science 149 (3683):510-515.
doi: 10.1126/science.149.3683.510.
Farhadi, Hadi, Hadi Salehi, Melor Md Yunus, Arezoo Aghaei Chadegani, Maryam Farhadi,
Masood Fooladi, and Nader Ale Ebrahim. 2013. "Does it Matter Which Citation Tool is Used to
Compare the h-index of a Group of Highly Cited Researchers?" Australian Journal of Basic and
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Nader Ale Ebrahim, H. Ebrahimian, Maryam Mousavi, Farzad Tahriri
Does a Long Reference List Guarantee More Citations? Analysis of Malaysian Highly Cited and Review Papers
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Applied Sciences 7 (4):198-202. doi: arXiv:1306.0727.
Farhadi, Maryam, Hadi Salehi, Mohamed Amin Embi, Masood Fooladi, Hadi Farhadi, Arezoo
Aghaei Chadegani, and Nader Ale Ebrahim. 2013. "Contribution of Information and
Communication Technology (ICT) in Country’S H-Index." Journal of Theoretical and Applied
Information Technology 57 (1):122-127. doi: 10.5281/zenodo.7715.
Gargouri, Yassine, Chawki Hajjem, Vincent Larivière, Yves Gingras, Les Carr, Tim Brody, and
Stevan Harnad. 2010. "Self-Selected or Mandated, Open Access Increases Citation Impact for
Higher Quality Research." PLoS ONE 5 (10):e13636. doi: 10.1371/journal.pone.0013636.
Gholizadeh, Hossein, Hadi Salehi, Mohamed Amin Embi, Mahmoud Danaee, Seyed Mohammad
Motahar, Nader Ale Ebrahim, Farid Habibi Tanha, and Noor Azuan Abu Osman. 2014.
"Relationship among Economic Growth, Internet Usage and Publication Productivity:
Comparison among ASEAN and World’s Best Countries." Modern Applied Science 8 (2):160170. doi: 10.5539/mas.v8n2p160.
Ioannidis, J. P. A. 2010. "Is there a glass ceiling for highly cited scientists at the top of research
universities?" Faseb Journal 24 (12):4635-4638. doi: 10.1096/fj.10-162974.
Krause, Kate. 2009. Increasing your Article's Citation Rates. Open Access Week. Accessed 28
May 2013.
Lai, Quirino, Tom Darius, and Jan Lerut. 2012. "The one hundred most frequently cited articles
in the field of clinical liver transplantation." Transplant International 25 (6):e76-e77.
Lancho-Barrantes, Bárbara S., Vicente P. Guerrero-Bote, and Félix Moya-Anegón. 2010. "What
lies behind the averages and significance of citation indicators in different disciplines?" Journal
of Information Science 36 (3):371-382. doi: 10.1177/0165551510366077.
Meho, Lokman I. 2007. "The rise and rise of citation analysis." Physics World 20:32–36.
Pislyakov, Vladimir, and Elena Shukshina. 2012. "Measuring Excellence in Russia: Highly Cited
Papers, Leading Institutions, Patterns of National and International Collaboration." Proceedings
of STI 2012, Montréal.
Sarli, Cathy, and Kristi Holmes. 2011. "Strategies for Enhancing the Impact of Research."
Washington University School of Medicine, Accessed 9 May. https://becker.wustl.edu/impactassessment/strategies.
Swan, Alma. 2010.
Tang, Rong, and Martin A. Safer. 2008. "Author-rated importance of cited references in biology
and psychology publications."
Journal of Documentation 64 (2):246-272. doi:
10.1108/00220410810858047.
Uzun, Ali. 2006. "Statistical relationship of some basic bibliometric indicators in scientometrics
research." International Workshop on Webometrics, Informetrics and Scientometrics & Seventh
COLLNET Meeting, Nancy (France), May 10 - 12, 2006.
Vanclay, Jerome K. 2013. "Factors affecting citation rates in environmental science." Journal of
Informetrics 7 (2):265-271. doi: http://dx.doi.org/10.1016/j.joi.2012.11.009.
Webster, Gregory D, Peter K Jonason, and Tatiana Orozco Schember. 2009. "Hot topics and
popular papers in evolutionary psychology: Analyses of title words and citation counts in
Evolution and Human Behavior, 1979–2008." Evolutionary Psychology 7 (3):348-362.
ISSN 1849-5664 (online)
http://researchleap.com/category/international-journal-of-management-science-and-business-administration
ISSN 1849-5419 (print) International Journal of Management Science And Business Administration Vol 1. No 3. February 2015., pp. 6 -15
Modern Applied Science; Vol. 8, No. 5; 2014
ISSN 1913-1844 E-ISSN 1913-1852
Published by Canadian Center of Science and Education
Economic Growth and Internet Usage Impact on Publication
Productivity among ASEAN’s and World’s Best Universities
Hossein Gholizadeh1, Hadi Salehi2, Mohamed Amin Embi3, Mahmoud Danaee4, Ali Ordi5, Farid Habibi Tanha6,
Nader Ale Ebrahim7 & Noor Azuan Abu Osman1
1
Department of Biomedical Engineering, Faculty of Engineering, University of Malaya, 50603 Kuala Lumpur,
Malaysia
2
Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University, Najafabad, Isfahan, Iran
3
Faculty of Education, Universiti Kebangsaan Malaysia, Bangi, 43600, Malaysia
4
Faculty of Agriculture, Roudehen Branch, Islamic Azad University, Roudehen, Iran
5
Universiti Teknologi Malaysia, Advance Informatics School (AIS), Kuala Lumpur, Malaysia
6
Department of Financial Sciences, University of Economic Sciences, Tehran, 1593656311, Iran
7
Research Support Unit, Centre of Research Services, Institute of Research Management and Monitoring (IPPP),
University of Malaya, Malaysia
Correspondence: Hadi Salehi, Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University,
Najafabad, Isfahan, Iran. E-mail: [email protected]
Received: July 3, 2014
doi:10.5539/mas.v8n5p169
Accepted: July 21, 2014
Online Published: September 1, 2014
URL: http://dx.doi.org/10.5539/mas.v8n5p169
Abstract
Measuring the number of papers which are published each year, publication productivity is the factor which
shows the reputation of universities and countries. However, the effect of growing economy and using internet
on the publication productivity in Asian countries has not been discovered yet. The present research is going to
figure out the publication productivity among the elite universities in Asian countries and also ten top
universities around the world in the last twenty years (from 1993 to 2012). Furthermore, the current research is
aimed to study the relationship among publication, gross domestic product (GDP) and internet usage. It is worth
to mention that the publication of the top Ten Malaysian Universities was regarded for the similar period of time.
To get the exact numbers of documents like papers, conference articles, review papers and letters which are
published by the universities in the last twenty years, the writer of the same paper used the Science Direct
database. Moreover, the data for GDP and the number of internet usage was collected through the World Bank
database (World Data Bank).To compare all kinds of publications,one-way ANOVAwas used and to investigate
the impact of economic growth and internet usageon publication productivity, multiple regression analysis was
applied.The results showed that therate of publication growth was 1.9, 20.9, and 65.5 % in top universities in the
world, ASEAN countries and Malaysia, respectively.The results also showed thatthere was a positive and
significant correlationbetween GDP and the number of internet users with the number of publications in ASEAN
and Malaysian universities. Internet usagehad much more influence in comparison withthe GDP in predicting the
number of publicationsamong these groups except for top ten Malaysian universities from 2003 to 2012. In
summary, publication trends in top ten Malaysian and ASEAN universities are promising. However, policy
makers and science managersshouldspend much more percentage of their GDP on Internet facilities and research
studies that their outputs lead to more rapid economic growthand internet usage.
Keywords: ASEAN, publication productivity, documents, internet usage, GDP, Malaysian Universities,
publication trend
1. Introduction
Ten countries in Southeast Asia formed the geo-political and economic Association of Southeast Asian Nations.
It was first formed on 8 August 1967 by Indonesia, Malaysia, the Philippines, Singapore and Thailand. Some
other countries like Brunei, Burma (Myanmar), Cambodia, Laos, and Vietnam became the members of this
group and expanded it. The chief purpose of organizing this group is increasing economic growth (Sarel, 1997).
169
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
Gross domestic product (GDP) which refers to the market price of all officially realized goods and services
which are produced in a country in a specific period of time is the chief tool to measure the economy of a
country. Reviewing the previously published papers, the writer found out that there is a relationship between
economic growth with education and paper publication. The above mentioned result is not fully support by
realistic facts (Jin, 2013; Nelson, 1966; Lucas Jr, 1988; Becker, 2009; Romer, 1990). Positive and significant
impact of education on economic growth has been found by Mankiw et al. (1992) and Barro (1991) (Mankiw et
al., 1992 & Barro, 1991); however, a possibility of reverse relationship between economic growth and education
was shown by Bils and Klenow (2000). Besides, Jin and Jin newly indicated that the effect of publication
productivity on economic growth is not the same in different fields. For example, there is a positive relationship
between engineering and science with economic growth, while the social sciences do not have the same effect on
economic growth (Jin, 2013).
Nowadays, compared with the past twenty years, studying the publication productivity is a main topic for the
researchers and students because the findings of the researches can positively affect the whole community Zain
et al., 2009). According to the recent development rules, the numbers of educated employees have been
enhanced. This matter helps the economy of the countries to grow very fast (Jin, 2013). It has been found out
that those countries which are highly developed like the United Stated and England are among the world’s top
productive research universities. The number of publication shows the research productivity and is employed to
grade the countries and universities (Yazit and Zainab, 2007, Narin and Hamilton, 1996, Toutkoushian et al.,
2003, Liu and Cheng, 2005, Meho and Spurgin). It can also be used to determine author’s productivity or the
publication productivity of research groups (Liu, 2005; Hart, 2000; Uzun, 2002; Gu, 2001; Fox, 1983).
Numerously referring to the previously published papers by the new papers a lot, shows the following
identification and also the effect in the field of study. Those review articles which refer to other articles a lot, can
give us some facts about the major areas of discipline besides, they can emphasize the increase of specific fields.
Moreover, more frequently cited papers are mostly written by famous researchers who can impress future
directions of the field by their ideas (Lefaivre and O’Brien, 2011, Kelly et al., 2010, Ponce and Lozano, 2010,
Joynt and Leonard, 1980).
Several indicators of academic or research performance are used to rank educational institutes and universities.
They includealumni and staff winning Nobel Prizes and Fields Medals, highly cited researchers, papers
published in Nature and Science, papers indexed in major citation indices, and the per capita academic
performance of an institution. The Academic Ranking of World Universities (ARWU) is the first global ranking
of universities to be published. Today, ARWU is regarded to be one of the three most influential and widely
observed international university rankings, along with the QS World University Rankings and the Times Higher
Education World University Rankings. The Academic Ranking of World Universities (ARWU), commonly
known as the Shanghai Ranking, is a publication that was founded and compiled by the Shanghai Jiao Tong
University to rank universities globally. The rankings have been conducted since 2003 and updated annually.
The current study is mainly going to investigate the amount of publication productivity among the best
universities in ASEAN countries and world’s top ten universities from 1993 to 2002 and 2003 to 2012.The study
also aimed to achieve the following objectives:


Studying the relationship among publication productivity, gross domestic product (current US $) and
internet users
Examining the publication direction of ten elite Malaysian Universities in a specific time periods
Since the Science Direct offers about 20% more inclusion than Web of Science, it has been used as the first
full-text theoretical database in this research. The researchers think that there is a positive relationship among the
economic growth, the numbers of people who can use the internet and also the number of publication of papers
in elite Asian universities and also the ten best universities in the whole world.
2. Methodology
ScienceDirect database was used to collect the number of documentsincluding articles, conference papers,
review papers, letters and books published in the last two decades from 1993 to 2002 and 2003to 2012. These
data were collectedto make a comparison among the top university in each ASEAN country, top tenuniversities
in Malaysia andtop tenuniversities in the world. To find the first university in each ASEAN countryand top
tenMalaysian universities, we used the number of publications in ScienceDirect database. Moreover,to determine
the top tenuniversities in the world, the Academic Ranking of World Universities (ARWU) was used.
Furthermore,in each university,the main subject area (overall), number of papers published in Nature and
Science journals, and the most cited paperswere identified.The numbers of citations that eachpaper could receive
170
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
during one week were identified as well.
To compare all kinds of publicationsamongtop ten universities in the world, ASEAN and Malaysia), one-way
ANOVA was applied. As the homogeneity test was not met, the Welch statistic was used to test the equality of
means. Moreover, to evaluate the relationship between publications (articles, conference papers, review papers,
letters) with GDP and Internet usage, Spearman correlation coefficient test was applied.To investigate economic
growth and internet use impact on publication productivity,multiple regression was applied to examine what
extent the proposed multiple liner regression model is supported by the research data. The regression examined
how well the number of publications could be predicted from GDP and internet usage.In the multiple regression
model, GDP and Internet usagewere set as the independent variables and the number of publications was
considered as the dependent variable.
3. Results
According to the Academic Ranking of World Universities (ARWU), the top tenuniversities in the world are
mostly located in the United States (8 universities) or United Kingdom (2 universities) (see Table 1). Moreover,
the first university in each ASEAN countryand top tenuniversities in Malaysia based on the number of
publications in ScienceDirect database were listed in Table 1.The main research areas in world’s best universities
were physics andastronomy (7 universities), medicine (2 universities) and engineering (1 university). In these
universities, the average number of papers published in Nature and Sciencewere 1586 and 1419, respectively
(see Table 1).
Table 1. The average number of published papers
Main
Institution (country)
established
Overall
Main
Subject
papers
papers
(year)
publication*
Subject area
area (%
(Nature)
(Science)
publication)
Harvard University
(US)
Stanford University
(US)
1636
74433
1891
110914
1861
134794
1868
158231
1209
135913
1891
62675
1764
62273
1754
112569
1890
90126
1096
122553
1905
74484
1949
1943
Physics and
most
most
cited
cited
paper
paper
(citation)
(citation)
22/7/2013
29/7/2013
most cited
paper
(publication
date)
14.8
1430
2294
10224
10255
1990
13.6
861
1593
6249
6266
2001
20.1
1563
1860
11678
11732
2000
15.3
1864
2233
18659
18757
1965
14.9
4099
644
7966
7977
1990
26.3
974
1134
8657
8705
1995
20.1
754
945
6123
6136
1998
Medicine
17.9
676
1403
10425
10484
1998
Medicine
21.7
980
1560
11741
11777
1953
15.2
2658
526
10198
10216
2001
Engineering
17.9
71
0
2171
2180
2003
21563
Medicine
14.9
24
0
445
449
2000
20291
Medicine
41.9
0
0
1494
1503
2005
Astronomy
Engineering
Massachusetts
Institute of
Technology (MIT)
Physics and
Astronomy
Top world's universities
(US)
University of
California, Berkeley
(US)
University of
Cambridge (UK)
California Institute
of Technology (US)
Princeton University
(US)
Columbia
University (US)
University of
Chicago (US)
University of
Oxford (UK)
Physics and
Astronomy
Physics and
Astronomy
Physics and
Astronomy
Physics and
Astronomy
Physics and
Astronomy
Universities
Top ASEAN
National University
of Singapore
(Singapore)
University of
Malaya (Malaysia)
Mahidol University
171
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
(Thailand)
Institut Teknologi
Bandung
1959
2979
1960
2955
Engineering
23.6
0
5
330
330
2001
54
15
11
857
861
1972
15.2
0
0
153
153
2000
10.4
0
0
222
224
2003
44.9
0
0
680
687
2009
26.3
0
0
171
171
2003
77
1999
(Indonesia)
International Rice
Research Institute
Agricultural
(Philippines)
Vietnam National
University
1945
1230
1985
864
(Vietnam)
UNIVERSITI
BRUNEI
DARUSSALAM
and
Biological
Computer
Science
Agricultural
and
Biological
(Brunei)
Institut Pasteur du
Cambodge
1953
251
1996
178
Medicine
(Cambodia)
National University
of Laos (Laos)
University of
Yangon (Myanmar)
University of
Malaya
Universiti Sains
Malaysia
Universiti Putra
Top Malaysian Universities
Malaysia
Agricultural
and
Biological
1878
109
Chemistry
13.6
0
1
77
1949
21572
Medicine
14.9
24
0
449
1962
17054
13.3
0
0
808
1969
16322
15.7
0
0
453
Material
Science
452
811
2004
1996
Agricultural
and
Biological
2008
455
Universiti
Kebangsaan
1970
15010
Engineering
15.5
0
0
449
Malaysia
UniversitiTeknologi
Malaysia
UniversitiTeknologi
MARA
International Islamic
University Malaysia
Multimedia
University
UniversitiTeknologi
Petronas
Universiti Malaysia
Perlis
2004
452
1975
10134
Engineering
26.1
0
0
260
1956-1965
6784
Engineering
22.5
0
0
305
1983
3995
Engineering
19.8
0
0
100
1996
3872
Engineering
27
0
0
275
1997
3343
23.5
0
0
77
2001
2321
32.6
0
0
137
Computer
science
Engineering
261
307
101
276
77
137
2004
2010
2007
2001
2008
2007
However, top universities in ASEAN countries could averagely publish 11 papers in Nature Journal and 2 papers
in Science journal. The average numbers of citations for the most cited papers in each university in these three
groups (world, ASEAN and Malaysia) were 10192, 660 and 331, respectively. Furthermore, the resultsshowed
39 citations per week for most cited papers in world's top universities while it was 4 citations per weekin
ASEAN universities (see Table 1, Figure 1).
172
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
Figure 1. Number of citations in one week for the most cited papers in each university (Top); Publication trend
(Down)
University of Singapore with 74484 papers was the first in ASEAN countries and was followed by University of
Malaya with 21563 publications among ASEAN countries. Moreover, the University of Yangon in Myanmar had
the leastpublications (109) among ASEAN countries.Interestingly, publication trends in ASEAN countries are
promising compared to the top ten universities in the world (see Appendixes 1 and 2).The rate of publication
growth between these twodecades was 1.9, 20.9 and 65.5 percentagein top universities in the world, ASEAN and
Malaysia, respectively (see Figure 1).
To compare all kinds of publications, the results of ANOVA showed that there were significant differences
among these three groups. Duncan’s multiple rangetest showed that there wasno significant difference between
ASEAN and Malaysia universities, while both of them were significantly lower than top ten world universities
(see Table 2).
Table 3 shows the correlation between indices (articles, conference papers, review papers, and letters), GDP and
internet users for these three groups. The results showedthat there was a positive and significant
correlationbetween GDP and the number of internet users with the number of publicationsin ASEAN and top ten
Malaysian universities. However, there was a negative correlationbetween GDP and internet users with the
number of letters published from 1993 to 2002 in Malaysia. Moreover, there was a negative and significant
correlationbetween GDP and the number of articles published in world's top universities (see Table 3).
TheR-squared (R2) value presented in Table 4showed some information about the goodness of fit of a model. In
regression, the R2 coefficient of determination is a statistical measure of how well the regression line
approximates the real data points. TheR2 of 1 indicates that the regression line perfectly fits the observeddata.
173
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
TheR2value of 0.584 for ASEAN universities from 1993 to 2002 implies that the two-predictor model explained
about 58.4% of the variance in publications. Table 5 revealed that based on the reported value of the F-statistic,
the model fits the data. This means that the slope of the estimated linear regression model line was not equal to
zero; thus, it confirmedthat there was a linear relationship between publication and the two-predictor variables.
Table 2. A (Welch robust Tests of Equality of Means); B (Pairwise comparison using Tamhane test)
(A)
1993-2002
Welch statistic
2003-2012
df
Sig.
Welch Statistic
df
Sig.
306.521
137.71
P<0.05
222.531
182.888
P<0.05
119.638
135.955
P<0.05
77.143
188.254
P<0.05
Review
159.91
142.948
P<0.05
221.178
160.595
P<0.05
Letter
143.944
168.747
P<0.05
101.268
156.708
P<0.05
Total publication
319.809
137.21
P<0.05
230.36
180.99
P<0.05
Article
Conference
(B)*
2
2
2003-201
1993-200
Decade
Article
Conference
Review
Letter
Total
publication
Top world
2030.73a
264.1a
124.6a
23.93a
2443.36a
ASEAN
222.98b
28.76b
6.55b
1.65b
259.94b
Malaysia
66.58c
8.87b
2.13b
0.78b
78.36c
Top world
3304.72a
776.77a
374.79a
58.3a
4514.58a
ASEAN
574.85b
166.54b
35b
7.04b
783.43b
Malaysia
509.03b
235.81b
22.39b
3.31b
770.54b
Means with same letter are not significantly different in each column.
*data are log transformed.
Table 3. Correlation coefficient between the number of articles, conference papers, review papers and letters with
GDP and internet usage for Top ten universities of the world, ASEAN and Malaysia
1993-2002
WOELD
ASEAN
MALAYSIA
2003-2012
GDP
internet
GDP
internet
Article
-0.042
.229*
-0.004
.531**
Conference
0.178
.232*
0.012
0.158
Review
0.185
.497**
-.443**
-0.062
Letter
-0.128
0.137
-.256*
.324**
Article
.723**
.611**
.779**
.800**
Conference
.775**
.599**
.737**
.739**
Review
.565**
.574**
.518**
.705**
Letter
.416**
.416**
.415**
.567**
Article
0.133
.280**
.624**
.595**
Conference
.249*
.400**
.879**
.876**
Review
0.09
0.171
.530**
.442**
Letter
-0.03
-0.008
.338**
.258*
**. Correlation is significant at the 0.01 level (2-tailed).
*. Correlation is significant at the 0.05 level (2-tailed).
174
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
Table 4. Results of multiple linear regression between GDP and internet usage with total number of publications
for Top ten universities of the world, ASEAN and Malaysia, respectively
1993-2002
Unstandardized Coefficients
group
B
Standardized Coefficients
Std. Error
t
Sig.
R2
0.187
Beta
Top world
GDP*
-1149.03
279.863
-0.408
-4.106
P<0.05
internet
21.03
4.943
0.423
4.254
P<0.05
GDP*
197.843
70.785
0.214
2.795
P<0.05
internet
43.194
4.661
0.708
9.267
P<0.05
GDP*
100.412
164.662
0.066
0.61
0.543
internet
1.771
0.888
0.216
1.994
P<0.05
Standardized Coefficients
t
ASEAN
0.584
Malaysia
0.063
* log transformed.
2003-2012
Unstandardized Coefficients
group
B
Std. Error
Sig.
R2
Beta
Top world
GDP*
-1461.77
446.388
internet
121.544
25.935
GDP*
363.724
161.614
-3.275
P<0.05
0.425
4.687
P<0.05
0.177
2.251
P<0.05
-0.297
0.276
ASEAN
internet
42.706
4.983
0.676
8.57
P<0.05
GDP*
6219.755
1339.622
1.141
4.643
P<0.05
internet
-50.131
23.01
-0.535
-2.179
0.102
0.551
Malaysia
0.421
* log transformed.
Standardized regression coefficients are presented in Table 4 to explain the importance of two predictors in
predicting the number of publications. Independent variable with a high beta coefficient is highly important in
contributing to the prediction of the number of total publications. Based on the beta values obtained, the beta
coefficient in theworld’s top university was -0.408 and -0.297 for GDP and 0.423 and 0.425 for Internet
usagebetween 1993-2002 and 2003-2012, respectively. This means that Internet usagehad a much more
significant influence than the GDP in predicting the number of publications. The results showedthe same issuefor
ASEAN universities meaning thatin these two decades, Internet usagehad much more significant effects than the
GDP in predicting the number of publications. Interestingly, GDP had more power (1.141) than the number of
internetusers (-0.535) from 2003 to 2012 in predicting the number of publications in top ten Malaysian
universities.
4. Discussion and Conclusion
In this study, a comparisonwas made among the top university in each ASEAN country, top ten Malaysian
universities and the world’s top universitiesregarding the relationship between economic growth and internet
usagewithpublication productivity from 1993 to 2002 and 2003 to 2012. Shanghai Ranking (ARWU) was used to
determine the top ten universities in the world as it is the first global ranking of universities to be publishedand
one of the three most influential and widely observed international university rankings. Moreover, the numbers
of publications were used to find the first university in each ASEAN country and top ten Malaysian universities
because most of these universities were not listed in the ARWU, QS, or the Times Higher Education World
175
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
University Rankings.
Publication productivity is an indicator of research output and could be used to rank countries and universities
(Yazit and Zainab, 2007, Narin and Hamilton, 1996, Toutkoushian et al., 2003, Liu and Cheng, 2005, Meho and
Spurgin). It can also be used to determine author’s productivity or the publication productivity of research
groups and to assess the productivity of persons in a particular discipline (Liu and Cheng, 2005, Hart, 2000,
Uzun, 2002, Gu and Zainab, 2001, Fox, 1983, Yi et al., 2008). World's best universities are mostly working on
physics andastronomy. Institutes whichare working in these fields could publish more papers and get more
citations. They could publish 144.2% and 709.5% more papers comparedto ASEAN universities in Nature and
Science journals, respectively (see Table 1). They could alsoreceive 9.8 times more citations per week for their
most cited papers (see Table 1).
Table 5. ANOVA table
1993
ANOVAb
group
world
asean
my
Model
1
1
1
Sum of Squares
df
Mean Square
Regression
1.756E7
2
8781193.216
Residual
6.875E7
97
708732.439
Total
8.631E7
99
Regression
1.656E7
2
8279590.609
Residual
1.126E7
71
158575.023
Total
2.782E7
73
57700.227
2
28850.113
Residual
853014.813
97
8793.967
Total
910715.040
99
Regression
F
Sig.
12.390
.000a
52.212
.000a
3.281
.042a
a. Predictors: (Constant), internet, log GDP.
b. Dependent Variable: total publication.
2003
ANOVAb
group
world
asean
my
Model
1
1
1
Sum of Squares
df
Mean Square
Regression
5.958E7
2
2.979E7
Residual
1.440E8
87
1655372.703
Total
2.036E8
89
Regression
9.489E7
2
4.745E7
Residual
7.379E7
78
946035.347
Total
1.687E8
80
Regression
2.040E7
2
1.020E7
Residual
2.658E7
87
305522.329
Total
4.698E7
89
a. Predictors: (Constant), internet, log GDP.
b. Dependent Variable: total publication.
176
F
Sig.
17.996
.000a
50.153
.000a
33.387
.000a
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
Publication trendfrom 1993 to 2012 in ASEAN and top ten Malaysian universitieswith 20.9 and 65.5% growth
ispromising compared to the world’s top ten universities (1.9%). ASEAN and Malaysian universities havebeen
averagely established over the past 66 and 37 years ago, respectively; while the top universities in the world have
been averagely stabilised 327 years ago. This can be one of the reasons of high trend in publication productivity
in ASEAN and Malaysian universities. In addition, the growth of publications might be overestimated due to the
omitted variables, and a reverse causality from GDP to publications will be another possibility to occur.The
number of publications in Malaysia increased dramatically after 2007 to 2012 (almost 5 times) (see Appendixes
1 and 2). One of the main reasons for increasing the number of paper publications in Malaysia could be greatly
concentrating on enhancing the quality of research in the universities which are specific for researching like
University of Malaya. Specified in the 10th Malaysia Plan, 1% of Malaysia GDP will be spent on the
development and research projects. Furthermore, the other important reason could be the change of dissertations
direction from conventional into research-based. PhD students have enhanced 10 times in the past years (there
were about 4000 students in 2002, while they have increased to 40,000 in 2012). The top ten universities in
Malaysia are shown in Table 1. It should be mentioned that the first five universities shown are ranked as the
research universities and get more government funding.
Thefindings of this study showed a positive and significant correlationbetween GDP and the number of
publications in ASEAN and Malaysian universities. This finding was in line withthe findings of previous
research studies that showedpositive and significant correlationbetween education and economic growth
(Mankiw et al., 1992; Barro, 1991; Bils and Klenow, 2000). Interestingly, there is a negative and significant
correlation between the number of publications (articles and letters) and economic growth in top ten Malaysian
universities from 1993 to 2012. Besides, Jin and Jin showed in their recent study that the publication productivity
affect the economic growth differently in various fields.In top ten universities located in the United States and
United Kingdom, GDP could not have positive effects to increase the number of publications. While the
significant effects of GDP on the number of publications were seen in Malaysian universities especially from
2003 to 2012.
However, it should be noted that only the effects of GDP and internet usageon publications of recent years were
evaluated in this study. Since the educational effects are accomplished over longer horizons, further investigation
of the educational effects using the data that correspond to the stock of publications in earlier years would be
important.
Science managers and policy makers in ASEAN countries shouldspend much more percentage of their GDP on
development and research projects as Malaysia does.Moreover, increasing the number of PhD students and
changing the universityprograms to paper-based dissertations could be another solution to increase the number of
publications. Furthermore, it is necessary to find out different ways to improve the quality and visibility of
theresearch studiesand invest more on research studies that their outputs lead to more rapid economic growth.
References
Barro, R. J. (1991). Economic growth in a cross section of countries. The Quarterly Journal of Economics, 106,
407-443. http://dx.doi.org/10.2307/2937943
Becker, G. S. (2009). Human capital: A theoretical and empirical analysis, with special reference to education.
University of Chicago Press.
Bils, M., & Klenow, P. J. (2000). Does schooling cause growth? American Economic Review, 1160-1183.
http://dx.doi.org/10.1257/aer.90.5.1160
Falagas, M. E., Pitsouni, E. I., Malietzis, G. A., & Pappas, G. (2008). Comparison of PubMed, Scopus, web of
science, and Google scholar: Strengths and weaknesses. The FASEB Journal, 22, 338-342.
http://dx.doi.org/10.1096/fj.07-9492LSF
Fox, M. F. (1983). Publication productivity among scientists: A critical review. Social Studies of Science, 13,
285-305. http://dx.doi.org/10.1177/030631283013002005
Gu, Y., & Zainab, A. (2001). Publication productivity of Malaysian researchers in the field of Computer Science
and Information Technology. Malaysian Journal of Library & Information Science, 6, 1-23.
Hart, R. L. (2000). Co-authorship in the academic library literature: A survey of attitudes and behaviors. The
Journal of Academic Librarianship, 26, 339-345. http://dx.doi.org/10.1016/S0099-1333(00)00140-3
Jin, J. C., & Jin, L. (2013). Research publications and economic growth: Evidence from cross-country
regressions. Applied Economics, 45, 983-990. http://dx.doi.org/10.1080/00036846.2011.613785
177
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
Joynt, R. L., & Leonard, J. A. (1980). Dantrolene sodium suspension in treatment of spastic cerebral palsy.
Developmental
Medicine
&
Child
Neurology,
22,
755-767.
http://dx.doi.org/10.1111/j.1469-8749.1980.tb03742.x
Kelly, J., Glynn, R., O’briain, D., Felle, P., & Mccabe, J. (2010). The 100 classic papers of orthopaedic surgery a
bibliometric analysis. Journal of Bone & Joint Surgery, British Volume, 92, 1338-1343.
http://dx.doi.org/10.1302/0301-620X.92B10.24867
Lefaivre, K. A., & O’brien, P. J. (2011). 100 most cited articles in orthopaedic surgery. Clinical Orthopaedics
and Related Research®, 469, 1487-1497. http://dx.doi.org/10.1007/s11999-010-1604-1
Liu, N. C., & Cheng, Y. (2005). The academic ranking of world universities. Higher education in Europe, 30,
127-136. http://dx.doi.org/10.1080/03797720500260116
Lucas, Jr, R. E. (1988). On the mechanics of economic development. Journal of monetary economics, 22, 3-42.
http://dx.doi.org/10.1016/0304-3932(88)90168-7
Mankiw, N. G., Romer, D., & Weil, D. N. (1992). A contribution to the empirics of economic growth. The
Quarterly Journal of Economics, 107, 407-437. http://dx.doi.org/10.2307/2118477
Meho, L. I., & Spurgin, K. M. (2005). Ranking the research productivity of lis faculty and schools: An
evaluation of data sources and research methods. Journal of the American Society for Information Science
and Technology, 56(12), 1314–1331.
Narin, F., & Hamilton, K. S. (1996). Bibliometric performance measures. Scientometrics, 36, 293-310.
http://dx.doi.org/10.1007/BF02129596
Nelson, R. R., & Phelps, E. S. (1966). Investment in humans, technological diffusion, and economic growth. The
American Economic Review, 56, 69-75.
Ponce, F. A., & Lozano, A. M. (2010). Highly cited works in neurosurgery. Part I: the 100 top-cited papers in
neurosurgical
journals:
A
review.
Journal
of
neurosurgery,
112,
223-232.
http://dx.doi.org/10.3171/2009.12.JNS091599
Romer, P. M. (1990). Endogenous technological change. Journal of political Economy, S71-S102.
http://dx.doi.org/10.1086/261725
Sarel, M. (1997). Growth and productivity in ASEAN countries. International Monetary Fund.
Toutkoushian, R. K., Porter, S. R., Danielson, C., & Hollis, P. R. (2003). Using publications counts to measure
an institution's research productivity. Research in Higher Education, 44, 121-148.
http://dx.doi.org/10.1023/A:1022070227966
Uzun, A. (2002). Productivity ratings of institutions based on publication in Scientometrics, Informetrics, and
Bibliometrics, 1981–2000. Scientometrics, 53, 297-307. http://dx.doi.org/10.1023/A:1014864827441
Yazit, N., & Zainab, A. (2007). Publication productivity of Malaysian authors and institutions in LIS. Malaysian
Journal of Library and Information Science, 12.
Yi, H., Ao, X., & Ho, Y. S. (2008). Use of citation per publication as an indicator to evaluate pentachlorophenol
research. Scientometrics, 75, 67-80. http://dx.doi.org/10.1007/s11192-007-1849-y
Zain, Z. M., Ishak, R., & Ghani, E. K. (2009). The influence of corporate culture on organisational commitment:
A study on a Malaysian listed company. European Journal of Economics, Finance and Administrative
Sciences, 17, 16-26.
Copyrights
Copyright for this article is retained by the author(s), with first publication rights granted to the journal.
This is an open-access article distributed under the terms and conditions of the Creative Commons Attribution
license (http://creativecommons.org/licenses/by/3.0/).
178
International Education Studies; Vol. 6, No. 11; 2013
ISSN 1913-9020
E-ISSN 1913-9039
Published by Canadian Center of Science and Education
Effective Strategies for Increasing Citation Frequency
Nader Ale Ebrahim1, Hadi Salehi2, Mohamed Amin Embi3, Farid Habibi Tanha4, Hossein Gholizadeh5, Seyed
Mohammad Motahar6 & Ali Ordi7
1
Research Support Unit, Centre of Research Services, Institute of Research Management and Monitoring (IPPP),
University of Malaya, Malaysia
2
Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University, Najafabad, Isfahan, Iran
3
Faculty of Education, Universiti Kebangsaan Malaysia, Bangi, 43600, Malaysia
4
Department of Financial Sciences , University of Economic Sciences, Tehran, 1593656311, Iran
5
Department of Biomedical Engineering, Faculty of Engineering, University of Malaya, 50603 Kuala Lumpur,
Malaysia
6
Faculty of Information Science and Technology,Universiti Kebangsaan Malaysia, Bangi, 43600, Malaysia
7
Universiti Teknologi Malaysia, Advance Informatics School (AIS), Kuala Lumpur, Malaysia
Correspondence: Nader Ale Ebrahim, Research Support Unit, Centre of Research Services, Institute of Research
Management and Monitoring (IPPP), University of Malaya, 50603, Kuala Lumpur, Malaysia. Tel:
603-7967-7355. E-mail: [email protected]
Received: September 12, 2013
doi:10.5539/ies.v6n11p93
Accepted: October 12, 2013
Online Published: October 23, 2013
URL: http://dx.doi.org/10.5539/ies.v6n11p93
Abstract
Due to the effect of citation impact on The Higher Education (THE) world university ranking system, most of
the researchers are looking for some helpful techniques to increase their citation record. This paper by reviewing
the relevant articles extracts 33 different ways for increasing the citations possibilities. The results show that the
article visibility has tended to receive more download and citations. This is probably the first study to collect
over 30 different ways to improve the citation record. Further study is needed to explore and expand these
techniques in specific fields of study in order to make the results more precisely.
Keywords: university ranking, improve citation, citation frequency, research impact, open access, h-index
1. Introduction
The research output is an essential part of an institution’s measure and evaluation of research quality. Previously,
the number of publication and journal impact factors were the means to derive research ratings. Recent approach
for rating research quality rely on the number of citations per article. Citation shows that how many times an
article has been cited by other articles (Fooladi et al. 2013). Citations to research publications are a quality
indicator, important for both the author and the affiliated university (Jones and Evans 2013). Most researchers
are evaluated based upon their publications as well as the numbers of citations their publications receive. One of
the key ways to increase citations is to expose the research output to the widest possible audience. If people are
unaware of the research, they won’t be citing it. The more researchers and students in other fields who have
heard about the research, the researcher will receive more citations. Citations to an article might strongly depend
on the visibility, rather than the merit of the article (Marashi et al. 2013). Ale Ebrahim (2012) argues that
publishing a high quality paper in scientific journals will be a halfway of receiving citation in the future. The rest
of the way is advertising and disseminating the publications by using the proper “Research Tools”. Post-print
publishing means to make peer-reviewed, published research articles freely available to anyone with an internet
connection, often greatly increases the citation frequency of articles (LiU E-Press 2007). The following section
introduces 33 ways which may help the researchers to improve the number of their citations.
93
www.ccsenet.org/ies
International Education Studies
Vol. 6, No. 11; 2013
2. 33 Ways for Improving Citations
1. Use A Unique Name Consistently Throughout Academic Careers
Authors are highly advised to use the same variation of their name consistently throughout their academic
careers. If the name is a common name, consider adding your full middle name to distinguish it from other
authors. Consistency enhances retrieval (Sarli and Holmes 2011).
2. Use a standardized institutional affiliation and address, using no abbreviations (Sarli and Holmes 2011).
Standardization of author affiliation is important to make sure work can be attributed to the correct author and
institution (Jones and Evans 2013). Providing accurate contact details are essential so that researchers can
contact directly for queries, further information and discussions about the publication (Wong 2008).
3. Repeat key phrases in the abstract while writing naturally.
Make some key phrases of your study and repeat them in the abstract page of your paper. Since search engines
and citation trackers search the abstract of your article, the normal repetition of key words increases the chance
of your paper to be retrieved more easily (Sarli and Holmes, 2011; Jones and Evans, 2013).
4. Assign keyword terms to the manuscript (Sarli and Holmes 2011).
Using keywords is a vital part of abstract writing, because of the practice of retrieving information electronically:
keywords act as the search term. Use keywords that are specific, and that reflect what is essential about the paper.
Put yourself in the position of someone researching in your field: what would you look for? Consider also
whether you can use any of the current "buzzwords" (Emerald Guide 2012).
5. Make a unique phrase that reflects author's research interest and use it throughout academic life.
Add the name of study in the title of all publications and use the same title/ name consistently (Sarli and Holmes
2011).
6. Publish in journal with high impact factor (Vanclay 2013).
The most effective strategy to increase citation rates is publishing in a journal with higher impact factor (Vanclay
2013). Dhawan and Gupta (2005) studied 1101 papers and found that articles published in high impact factor
journals increase the probability of getting cited.
7. Self-archive articles.
Free online availability increases a paper's impact (Lawrence 2001); therefore, maximize the visibility of your
research by making copies of your articles available online (Jones and Evans 2013). Gargouri et al. (2010) have
made a strong and a declarative link between self-archiving and increased citation performance.
8. Keep your professional web pages and published lists up to date (Jones and Evans 2013).
The advantage of self-archive on the web and make a link between published lists is obvious. Freely accessible
articles increase citations by 50% or more (Harnad 2006).
9. Make your research easy to find, especially for online searchers (Jones and Evans 2013).
Jamali and Nikzad (2011) investigated 2172 articles and found that there is a positive relationship between the
number of downloads and citations. Research shows that there is a correlation between highly cited articles and
the likelihood of it being online (Vaughan and Shaw 2003).
10. Open Access (OA) increases citation rate (MacCallum and Parthasarathy 2006).
Free access increases citation rates, searching online is more efficient and following hyperlinks quickly leads
researchers to their prevailing opinion (Evans 2008). Open Access has a positive impact on growth of citations
(Table 1) (Swan 2010).
11. Deposit paper in Open Access repository (Frost 2009).
For example, Ray Frost is a chemist who publishes prolifically. About three years ago, he began depositing his
articles in the Queensland University of Technology (QUT) repository. So far, he has deposited over 300 of his
articles. Figure 1 (derived data from the Web of Science) shows the patterns of publication and citations to those
publications. When Ray started putting his articles into the QUT repository, the numbers of citations began to
take off. The latest count is 1200 in one year. Even though Ray’s publication rate went up a bit over this period,
the increase in citations is impressive (Frost 2009).
94
www.ccsenet.org/ies
International Education Studies
Vol. 6, No. 11; 2013
Table 1. Effect of Open Access (OA) to increase the level of citations (Swan 2010)
Size of OA citation advantage when found
(and where explicitly stated by discipline)
% increase in citations with Open Access
Physics/astronomy
Mathematics
Biology
Electrical engineering
Computer science
Political science
Philosophy
Medicine
Communication studies (IT)
Agricultural sciences
170 to 580
35 to 91
-5 to 36
51
157
86
45
300 to 450
200
200 to 600
Figure 1. The patterns of citations to Ray Frost’s articles that are deposited in the Queensland University of
Technology (QUT) repository (Frost 2009)
12. Publish with international authors (Pislyakov and Shukshina 2012)
Citation analysis shows that papers with international co-authors are cited up to four times more often than those
without international co-authors (Jones and Evans 2013). Krause (2009) argued that articles published with
multi-countries or multi-institutes collaborations get cited more. Authors who are often involved in international
collaboration received more citations (Aksnes 2003).
13. Team-authored articles get cited more (Krause 2009)
Wuchty et al. (2007) have used 19.9 million papers over 5o years and demonstrated that team-authored articles
typically produce more frequently cited research than individuals. A recent study by Cotropia and Petherbridge
(2013) in law review articles which were published within two decades also demonstrated that team research is
on average more frequently cited than individual research. Typically high cited articles are authored by a large
number of scientists (Aksnes 2003).
14. Use more references
There is a ridiculously strong relationship between the number of citations a paper receives and the number of its
references (Corbyn 2010).
15. Publish a longer paper
A longer paper gathers more citations (Ball 2008; Abt 1998). Hamrick et al. (2010) indicated that longer papers
are associated with more citations.
16. Publish papers with a Nobel laureates (Ball 2011)
95
www.ccsenet.org/ies
International Education Studies
Vol. 6, No. 11; 2013
Some landmark papers of Nobel laureates quite quickly give their authors a sudden boost in citation rate and this
boost extends to the author's earlier papers too, even if they were in unrelated areas (Ball 2011).
17. Contribute to Wikipedia (SAGE 2012)
Try to contribute in Wikipedia. As a good example, one paper (Nader Ale Ebrahim et al. 2009) that was used as
a reference in defining virtual teams in Wikipedia has received significant citations in comparison to the rest of
the articles from the same author.
18. Start blogging (SAGE 2012)
Use blogs and podcasts to leverage on-going researcher discussion on the Internet (Taylor & Francis Group
2012a). Web 2.0 tools such as wikis and blogs can be created to inform, describe and link people’s research
interests and publications (Wong 2008). Authors are encouraged to promote their papers through the addition of
links which web search engines such as Google take particular notice for their page ranks (Smith 2005).
19. Join academic social networking sites (Taylor & Francis Group 2012b)
Increasing the availability of articles through social networking sites broadens dissemination, increases use, and
enhances professional visibility which lead to increased citations and usage. Academica is an online social
reference tool that allows reference sharing among academics and researchers. Alternatively, researchers may
use Citeulike to share their interests in research publications (Wong 2008). Academica, Citeulike, ResearchGate
and Linkedin are just a few examples of knowledge sharing tools to make others aware of research articles that
may be of relevance to authors and hence get cited.
20. Write a review paper
Reviews are more likely to be cited than original research papers. Some types of articles including editorials,
letters to editors, news items, meeting abstracts, and case studies are generally poorly cited (Taylor & Francis
Group 2012a). Authors seeking to be well cited should aim to write comprehensive and substantial review
articles, and submit them to journals with a high impact factor that carry previous articles on the topic (Vanclay
2013).
21. Papers published after having first been rejected elsewhere receive significantly more citations (Ball 2012)
Submission history affected post-publication impact, resubmissions from other journals received significantly
more citations than first-intent submissions (Calcagno et al. 2012).
22. Papers with a larger number of “callouts” be likely to receive a higher number of citations (Hamrick et al.
2010)
A “callout” is a phrase or sentence from the paper that is displayed in a different font, somewhere in the paper.
Generally, callouts are inserted by the editorial staff to call attention to potentially interesting aspects of a paper
(Hamrick et al. 2010).
23. Avoid to select a question type of title
Jamali and Nikzad (2011) investigated 2172 articles and found different types of title effects on the number of
downloads and citations. Especially articles with question type titles tend to be downloaded more but cited less
than the others.
24. Sharing detailed research data
Open data (Publicly-available datasets) are significantly associated with a 69% increase in citations to articles
that accompany the data. This correlation is independent of Journal Impact Factor, country of authors and time
since publication (Piwowar et al. 2007; Swan 2010).
25. Publish across disciplines
Publishing across disciplines has been found to increase citation e.g. chemistry, biological science and physics
(Ortega and Antell 2006).
26. Present a working paper (Taylor & Francis Group 2012a)
Try to go to a prestigious conference and present some parts of your research or publish working paper. Working
papers are freely available before and after the articles are published. Researchers may upload their working
papers into open access repositories including the personal websites or more formal repositories such as arXiv
and SSRN (McCabe 2011).
27. Publish your article in one of the journals everyone in your discipline reads (Taylor & Francis Group 2012a)
96
www.ccsenet.org/ies
International Education Studies
Vol. 6, No. 11; 2013
Choosing a journal that matches with a researcher’s field of study is thus very important because it makes it
more likely that the article receives more citation. A journal which covers a broad range of disciplines may be
the best.
28. Publicize yourself - link your latest published article to your email signature (Taylor & Francis Group 2012a)
A great way to spread researchers’ outputs and get extra attention of email recipient is to add a link to the latest
publication. This little section of contact information that most people ignore, provides a good platform for
publication marketing.
29. Publish your work in a journal with the highest number of abstracting and indexing (Nader Ale Ebrahim
2012)
Abstracting and indexing services generate visibility for the material published in most journals (Garfield and
Merton 1979). Citation potential increases by attributing to the high visibility of scientific materials. Therefore, a
journal with the highest number of abstracting and indexing in different databases can be a good target.
30. Create a podcast describing the research project
Research is not just text and figures. Create a podcast describing the research project and submit the podcast to
YouTube or Vimeo (Sarli and Holmes 2011). Podcasts can describe the research efforts. Video is an increasingly
important way for researchers to communicate their results and welcome submissions of podcasts from authors
and editors (Sarli and Holmes 2011).
31. Make an online CV Like ORCID or ResearcherID
Online CV makes a link between the list of published papers and open access versions of relevant articles (Sahu
2005). Online CV increases researchers’ output visibility to the academic community.
32. Publish tutorials papers
Tutorial paper is “a paper that organizes and introduces work in the field. A tutorial paper assumes its audience
is inexpert; it emphasizes the basic concepts of the field and provides concrete examples that embody these
concepts (ACM 2013)”. Tutorials papers tend to have a higher number of citations (Hamrick et al. 2010).
and finally;
33.
Use
all
“Enhancing
Visibility
and
Impact”
tools
http://www.mindmeister.com/39583892/research-tools-by-nader-ale-ebrahim
which
are
available
on
Familiarity with academic advertisement tools allows the researcher to increase his/her h-index in the short time.
H-index shows the academicians’ influences in the specified field of research (Aghaei Chadegani et al. 2013).
Therefore, a person with higher levels of h-index has higher quality publications with high amount of citations
(Nader Ale Ebrahim 2012). The advertisement section of the above mind map includes the tools which can assist
the researchers to disseminate and increase visibility of their published papers.
3. Conclusion
Publishing research output in high-impact journals is a primary concern of the researchers. The researchers also
need to consider different ways to receive more citations after publishing a paper. When their works are
published, they are concerned about citation which is directly related to the paper’s quality and visibility. The
researchers cannot increase the quality of their published papers; therefore, they can apply some of these 33 key
points to increase the visibility of their published papers.
References
Abt, H. A. (1998). Why some papers have long citation lifetimes. [10.1038/27355]. Nature, 395(6704), 756-757.
http://dx.doi.org/10.1038/27355
ACM. (2013). ACM Computing Surveys. Retrieved May 30, 2013, from http://csur.acm.org/author_info.html
Aghaei Chadegani, A., Salehi, H., Yunus, M. M., Farhadi, H., Fooladi, M., & Farhadi, M., et al. (2013). A
Comparison between Two Main Academic Literature Collections: Web of Science and Scopus Databases.
Asian Social Science, 9(5), 18-26. http://dx.doi.org/10.5539/ass.v9n5p18
Aksnes, D. W. (2003). Characteristics of highly cited papers. Research Evaluation 12(3), 159-170.
http://dx.doi.org/10.3152/147154403781776645
Ale Ebrahim, N. (2012). Publication Marketing Tools “Enhancing Research Visibility and Improving Citations”.
Research Tools in Education Series, 1(2), 1-86.
97
www.ccsenet.org/ies
International Education Studies
Vol. 6, No. 11; 2013
Ale Ebrahim, N., Ahmed, S., & Taha, Z. (2009). Virtual R & D teams in small and medium enterprises: A
literature review. [Review]. Scientific Research and Essay, 4(13), 1575–1590.
Ball,
P. (2008). A longer paper
http://dx.doi.org/10.1038/455274a
gathers
more
citations.
Nature
455(7211),
274-275.
Are scientific reputations boosted artificially? (2011, 6 May). Nature.
Ball, P. (2012, 11 October). Rejection improves eventual impact of manuscripts. Nature.
Calcagno, V., Demoinet, E., Gollner, K., Guidi, L., Ruths, D., & de Mazancourt, C. (2012). Flows of Research
Manuscripts Among Scientific Journals Reveal Hidden Submission Patterns. Science, 338(6110),
1065-1069. http://dx.doi.org/10.1126/science.1227833
Corbyn, Z. (2010). An easy way to boost a paper's citations. Nature. http://dx.doi.org/10.1038/news.2010.406
Cotropia, C. A., & Petherbridge, L. (2013). The Dominance of Teams in the Production of Legal Knowledge.
Loyola-LA Legal Studies.
Dhawan, S., & Gupta, B. (2005). Evaluation of Indian physics research on journal impact factor and citations
count: A comparative study. DESIDOC Journal of Library & Information Technology, 25(3), 3-7.
Emerald Guide. (2012). How to... write an abstract. Retrieved
http://www.emeraldinsight.com/authors/guides/write/abstracts.htm?part=1
May
9,
2013,
from
Evans, J. A. (2008). Electronic Publication and the Narrowing of Science and Scholarship. Science, 321(5887),
395-399. http://dx.doi.org/10.1126/science.1150473
Fooladi, M., Salehi, H., Yunus, M. M., Farhadi, M., Aghaei Chadegani, A., Farhadi, H., et al. (2013). Do
Criticisms Overcome the Praises of Journal Impact Factor? Asian Social Science, 9(5), 176-182,
http://dx.doi.org/10.5539/ass.v9n5p176
Frost, R. (2009). Case study: Open Access visibility and impact of an individual researcher. Retrieved May 9,
2013,
from
http://www.openscholarship.org/jcms/c_6220/case-study-open-access-visibility-and-impact-of-an-individua
l-researcher
Garfield, E., & Merton, R. K. (1979). Perspective on Citation Analysis of Scientists. In Citation indexing: Its
theory and application in science, technology, and humanities (Vol. 8). Wiley New York.
Gargouri, Y., Hajjem, C., Larivière, V., Gingras, Y., Carr, L., Brody, T., et al. (2010). Self-Selected or Mandated,
Open Access Increases Citation Impact for Higher Quality Research. PLoS ONE, 5(10), e13636.
http://dx.doi.org/10.1371/journal.pone.0013636
Hamrick, T. A., Fricker, R. D., & Brown, G. G. (2010). Assessing What Distinguishes Highly Cited from
Less-Cited
Papers
Published
in
Interfaces.
Interfaces
40(6),
454-464.
http://dx.doi.org/10.1287/inte.1100.0527
Harnad, S. (2006). Publish or perish—self-archive to flourish: the green route to open access. ERCIM News, 64.
Jamali, H. R., & Nikzad, M. (2011). Article title type and its relation with the number of downloads and citations.
Scientometrics, 88(2), 653-661. http://dx.doi.org/10.1007/s11192-011-0412-z
Jones, K., & Evans, K. (2013). Good Practices for Improving Citations to your Published Work (pp. 2).
University of BATH.
Krause, K. (2009). Increasing your Article's Citation Rates. Open Access Week.
Lawrence, S. (2001). Free online availability substantially increases a paper's impact. [10.1038/35079151].
Nature, 411(6837), 521-521.
LiU E-Press (2007). One way to increase citation frequency. Retrieved May 9, 2013, from
http://www.ep.liu.se/authorinf/postpubl.en.asp
MacCallum, C. J., & Parthasarathy, H. (2006). Open Access Increases Citation Rate. PLoS Biol, 4(5), e176,
http://dx.doi.org/10.1371/journal.pbio.0040176
Marashi, S.-A., Seyed Mohammad Amin, H.-N., Alishah, K., Hadi, M., Karimi, A., Hosseinian, S., et al. (2013).
Impact
of
Wikipedia
on
citation
trends.
EXCLI
Journal,
12,
15-19.
http://dx.doi.org/citeulike-article-id:12202824
98
www.ccsenet.org/ies
International Education Studies
Vol. 6, No. 11; 2013
McCabe, M. J. (2011). Online Access and the Scientific Journal Market: An Economist’s Perspective. (Vol.
Draft Report for the National Academy of Sciences, pp. 1-36). University of Michigan and SKEMA
Business School.
Ortega, L., & Antell, K. (2006). Tracking Cross-Disciplinary Information Use by Author Affiliation:
Demonstration of a Method. College & Research Libraries, 67(5), 446-462.
Pislyakov, V., & Shukshina, E. (2012). Measuring Excellence in Russia: Highly Cited Papers, Leading
Institutions, Patterns of National and International Collaboration. Paper presented at the Proceedings of
STI 2012, Montréal.
Piwowar, H. A., Day, R. S., & Fridsma, D. B. (2007). Sharing Detailed Research Data Is Associated with
Increased Citation Rate. PLoS ONE, 2(3), 308. http://dx.doi.org/10.1371/journal.pone.0000308
SAGE (2012). 10 Ways to Increase Usage and Citation of your Published Article Using Social Media. Retrieved
May 9, 2013, from http://www.sagepub.com/authors/journal/10ways.sp
Sahu, D. (2005). Open Access: Why India Should Brace it? In (pp. 1-49).
Sarli, C., & Holmes, K. (2011). Strategies for Enhancing the Impact of Research. Retrieved May 9, 2013, from
https://becker.wustl.edu/impact-assessment/strategies
Smith, A. G. (2005). Citations and Links as a Measure of Effectiveness of Online LIS Journals. IFLA Journal,
31(1), 76-84. http://dx.doi.org/10.1177/0340035205052651
Swan, A. (2010). The Open Access citation advantage: Studies and results to date.
Taylor & Francis Group (2012a). Optimize citations. Retrieved
http://journalauthors.tandf.co.uk/beyondpublication/optimizingcitations.asp
Taylor & Francis Group (2012b). Promote your article. Retrieved
http://journalauthors.tandf.co.uk/beyondpublication/promotearticle.asp
May
May
9,
9,
2013,
from
2013,
from
Vanclay, J. K. (2013). Factors affecting citation rates in environmental science. Journal of Informetrics, 7(2),
265-271. http://dx.doi.org/10.1016/j.joi.2012.11.009
Vaughan, L., & Shaw, D. (2003). Bibliographic and Web citations: What is the difference? Journal of the
American
Society
for
Information
Science
and
Technology,
54(14),
1313-1322.
http://dx.doi.org/10.1002/asi.10338
Wong, R. (2008). Ways to Maximise Citations for Researchers (pp. 1-7). University of Sheffield.
Wuchty, S., Jones, B. F., & Uzzi, B. (2007). The Increasing Dominance of Teams in Production of Knowledge.
Science, 316(5827), 1036-1039. http://dx.doi.org/10.1126/science.1136099
Copyrights
Copyright for this article is retained by the author(s), with first publication rights granted to the journal.
This is an open-access article distributed under the terms and conditions of the Creative Commons Attribution
license (http://creativecommons.org/licenses/by/3.0/).
99
Modern Applied Science; Vol. 8, No. 5; 2014
ISSN 1913-1844 E-ISSN 1913-1852
Published by Canadian Center of Science and Education
Equality of Google Scholar with Web of Science Citations: Case of
Malaysian Engineering Highly Cited Papers
Nader Ale Ebrahim1, Hadi Salehi2, Mohamed Amin Embi3, Mahmoud Danaee4, Marjan Mohammadjafari5,
Azam Zavvari6, Masoud Shakiba6 & Masoomeh Shahbazi-Moghadam7
1
Research Support Unit, Centre of Research Services, Institute of Research Management and Monitoring (IPPP),
University of Malaya, Malaysia
2
Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University, Najafabad, Isfahan, Iran
3
Faculty of Education, Universiti Kebangsaan Malaysia (UKM), Bangi, 43600, Malaysia
4
Faculty of Agriculture, Roudehen Branch, Islamic Azad University, Roudehen, Iran
5
Department of Industrial Engineering, Faculty of Engineering, Science and Research Branch, Islamic Azad
University, Kerman, Iran
6
Center for Software Technology and Management, Faculty of Information Science and Technology, Universiti
Kebangsaan Malaysia, 43600 UKM Bangi, Selangor, Malaysia
7
Perdana School of Science, Technology and Innovation Policy, Universiti Teknologi Malaysia
Correspondence: Hadi Salehi, Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University,
Najafabad, Isfahan, Iran. E-mail: [email protected]
Received: June 1, 2014
doi:10.5539/mas.v8n5p63
Accepted: June 23, 2014
Online Published: August 6, 2014
URL: http://dx.doi.org/10.5539/mas.v8n5p63
Abstract
This study uses citation analysis from two citation tracking databases, Google Scholar (GS) and ISI Web of
Science, in order to test the correlation between them and examine the effect of the number of paper versions on
citations. The data were retrieved from the Essential Science Indicators and Google Scholar for 101 highly cited
papers from Malaysia in the field of engineering. An equation for estimating the citation in ISI based on Google
scholar is offered. The results show a significant and positive relationship between both citation in Google
Scholar and ISI Web of Science with the number of versions. This relationship is higher between versions and
ISI citations (r = 0.395, p<0.01) than between versions and Google Scholar citations (r = 0.315, p<0.01). Free
access to data provided by Google Scholar and the correlation to get ISI citation which is costly, allow more
transparency in tenure reviews, funding agency and other science policy, to count citations and analyze scholars’
performance more precisely.
Keywords: bibliometrics, citation analysis, evaluations, equivalence, Google Scholar, High cited, ISI Web of
Science, research tools, H-index
1. Introduction
Citation index as a type of Bibliometrics method traces the references in a published article. It shows that how
many times an article has been cited by other articles (Fooladi et al., 2013). Citations are applied to evaluate the
academic performance and the importance of information contained in an article (Zhang, 2009). This feature
helps researchers get a preliminary idea of the articles and research that make an impact in a field of interest. The
avenues to evaluate citation tracking have greatly increased in the past years (Kear & Colbert-Lewis, 2011).
Citation analysis was monopolized for decades by the system developed by Eugene Garfield at the Institute for
Scientific Information (ISI) now owned by Thomson Reuter Scientific (Bensman, 2011). ISI Web of Science is a
publication and citation database which covers all domains of science and social science for many years (Aghaei
Chadegani et al., 2013). In 2004, two competitors emerged, Scopus and Google Scholar (Bakkalbasi, Bauer,
Glover, & Wang, 2006). Google Inc. released the beta version of ‘Google Scholar’ (GS)
(http://scholar.google.com) in November 2004 (Pauly & Stergiou, 2005). These three tools, ISI from Thomson
Reuters, Google Scholar (GS) from Google Inc. and Scopus from Elsevier are used by academics to track their
citation rates. Access to ISI Web of Science is subscription-based service while GS provides a free alternative to
retrieve the citation counts. Therefore, the researchers need to estimate their citation in ISI by knowing the GS
63
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
citation counts. On the other hand, publishing a research paper in a scholarly journal is necessary but not
sufficient for receiving citations in the future (Nader Ale Ebrahim, 2013). The paper should be visible to the
relevant users and authors in order to get citations. The visibility of the paper is defined by the number of paper
versions which are available in the Google Scholar database. The number of citations will be limited to the
versions of the published article on the web. The literature has shown increased visibility by making research
outputs available through open access repositories, wider access results and higher citation impact (Nader Ale
Ebrahim et al., 2014; Amancio, Oliveira Jr, & da Fontoura Costa, 2012; Antelman, 2004; Ertürk & Şengül, 2012;
Hardy, Oppenheim, Brody, & Hitchcock, 2005). A paper has greater chance of becoming highly cited whenever
it has more visibility (Nader Ale Ebrahim et al., 2013; Egghe, Guns, & Rousseau, 2013).
The objectives of this paper are two-fold. The first objective is to find the correlation between Google Scholar
and ISI citation in the highly cited papers. The second objective is to find a relationship between the paper
availability and the number of citations.
2. Google Scholar & Web of Science Citations
The citation facility of Google Scholar is a potential new tool for Bibliometrics (Kousha & Thelwall, 2007).
Google Scholar, is a free-of-charge by the giant Google search engine, has been suggested as an alternative or
complementary resource to the commercial citation databases like Web of Knowledge (ISI/Thomson) or Scopus
(Elsevier) (Aguillo, 2011). Google Scholar provides Bibliometrics information on a wide range of scholarly
journals, and other published material, such as peer-reviewed papers, theses, books, abstracts and articles, from
academic publishers, professional societies, preprint repositories, universities and other scholarly organizations
(Orduña-Malea & Delgado López-Cózar, 2014). GS also introduced two new services in recent years: Google
Scholar Author Citation Tracker in 2011 and Google Scholar Metrics for Publications in April 2012 (Jacso,
2012). Perhaps some of these documents would not otherwise be indexed by search engines such as Google, so
they would be "invisible" to web searchers, and clearly some would be similarly invisible to Web of Science
users, since it is dominated by academic journals (Kousha & Thelwall, 2007). On the other hand, the Thomson
Reuters/Institute for Scientific Information databases (ISI) or Web of Science database (actually there is
ambiguity between different names of former ISI), include three databases: Science Citation Index/Science
Citation Index Expanded (SCI/SCIE) (SCIE is the online version of SCI), Social Science Citation Index (SSC)
and Arts and Humanities Citation Index (AHCI) (Larsen & von Ins, 2010). Since 1964 the Science Citation
Index (SCI) has been a leading tool in indexing (Garfield, 1972).
Few studies have been done to find a correlation between GS with WoS citations. Cabezas-Clavijo and
Delgado-Lopez-Cozar (2013) found that the average h-index values in Google Scholar are almost 30% higher
than those obtained in ISI Web of Science, and about 15% higher than those collected by Scopus. GS citation
data differed greatly from the findings using citations from the fee-based databases such as ISI Web of Science
(Bornmann et al., 2009). Google Scholar overestimates the number of citable articles (in comparison with formal
citation services such as Scopus and Thomson Reuters) because of the automated way it collects data, including
‘grey’ literature such as theses (Hooper, 2012). The first objective of this study is to find the correlation between
Google Scholar and ISI citation in the highly cited papers.
3. Visibility and Citation Impact
Nader Ale Ebrahim et al. (2014) based on a case study confirmed that the article visibility will greatly improve
the citation impact. The journal visibility has an important influence on the journal citation impact (Yue &
Wilson, 2004). Therefore, greater visibility caused higher citation impact (Zheng et al., 2012). In contrast, lack of
visibility has condensed a significant citation impact (Rotich & Musakali, 2013). Nader Ale Ebrahim et al. (2013)
by reviewing the relevant papers extracts 33 different ways for increasing the citations possibilities. The results
show that the article visibility has tended to receive more download and citations. In order to improve the
visibility of scholars’ works and make them relevant on the academic scene, electronic publishing will be
advisable. This provides the potential to readers to search and locate the articles at minimum time within one
journal or across multiple journals. This includes publishing articles in journals that are reputable and listed in
various databases and peer reviewed (Rotich & Musakali, 2013). Free online availability substantially increases
a paper’s impact (Lawrence, 2001a). Lawrence (2001a, 2001b) demonstrated a correlation between the
likelihood of online availability of the full-text article and the total number of citations. He further showed that
the relative citation counts for articles available online are on average 336% higher than those for articles not
found online (Craig, Plume, McVeigh, Pringle, & Amin, 2007).
However, there are limited resources to explain the relationship between the paper availability and the number of
citations (Craig et al., 2007; Lawrence, 2001b; McCabe & Snyder, 2013; Solomon, Laakso, & Björk, 2013).
64
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
None of them discussed about the relationship between the number of versions, and citation. The number of
“versions” will be shown in any Google Scholar search result. Figure 1 shows 34 different versions of an article
entitled “Virtual Teams: a Literature Review (Nader Ale Ebrahim, Ahmed, & Taha, 2009)” and number of
citations. The second objective of this research is to find a relationship between the paper availability and the
number of citations.
Figure 1. The number of “versions” in the Google Scholar search result
4. Methodology
Highly cited papers from Malaysia in the field of engineering were retrieved from the Essential Science
Indicators (ESI) which is one the Web of Knowledge (WoK) databases. ESI provides access to a comprehensive
compilation of scientists’ performance statistics and science trend data derived from WoK Thomson Reuters
databases. Total citation counts and cites per paper are indicators of influence and impact of each paper. There is
a threshold to select highly cited papers according to the baseline data in ESI. This threshold is different from
one discipline to another one. ESI rankings are determined for the most cited authors, institutions, countries, and
journals (The Thomson Corporation, 2013). The paper must be published within the last 10-year plus four-month
period (January 1, 2003-April 30, 2013) and must be cited above threshold level, in order to be selected.
Essential Science Indicators data used in this research have been updated as of July 1, 2013.
Google Scholar which is a free online database was used for deriving the number of citations and versions of the
ESI highly cited papers. The data from ESI was collected on 29 July 2013 and Google Scholar data was
collected on 31 July 2013. The total numbers of 101 papers were listed in ESI as highly cited papers from
Malaysia in the field of engineering. The lists of 101 papers were retrieved from ESI database and then were
exported to an Excel sheet. A search engine was developed to get the number of citations and versions from
Google Scholar. This gadget assisted the present researchers to collect the data more preciously and faster than
searching for the papers one by one. The Statistical Package for the Social Sciences (SPSS) was used for
analyzing the data. The results are illustrated in the following section.
5. Results and Discussion
The number of citations which were derived from Web of Knowledge platform hereafter are called ISI citation.
To study the relationship among the number of citations in Google scholar and ISI and the number of versions,
correlation coefficients were computed.
Table 1 shows descriptive statistics of the variables.
Table 1. Descriptive statistic of variables
N
Minimum
Maximum
Mean
Std. Deviation
Versions
101
2
28
5.62
3.078
Cited in Google Scholar
101
4
348
80.76
71.718
ISI citation
101
5
189
43.15
36.076
As both numbers of citations in Google scholar and ISI were distributed normally, Pearson correlation coefficient
(r) was used and the results showed a very high positive and significant association (r = 0.932 , P<0.01) between
the number of citations in Google scholar and ISI for the articles that were published during 2006 to 2012 from
Malaysia in the field of engineering. To study the relationship between both citation and the number of versions,
65
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
Spearman Rho was used due to the non-normal distribution of the versions. The results showed a significant and
positive relationship between both citations in Google Scholar and ISI with the number of versions. This
relationship was higher between versions and ISI citations (r = 0.395, p<0.01) than between versions and Google
Scholar citations (r = 0.315, p<0.01). Linear regression was also applied to predict the number of citations in ISI
based on Google Scholar citations. The results showed a very high predictability (R2 = 0.836) for the linear
model (see
Figure 2) which was significant (F = 511.63, p<0.01). Therefore, the final equation for estimating the citation in
ISI based on Google Scholar is:
ISI Citation = 5.961 + 0.460 (Google Scholar citation)
Figure 2. Scatter diagram between ISI citation and Google Scholar citation
To study the effect of the number of versions on both citations in Google Scholar and ISI, simple linear
regression was applied. The results indicated that the number of versions had a significant positive effect on
citations in both databases (see Table 2 and Table 3).
Table 2. Summary of regression analysis results
R Square
0.276
0.272
Model a
Model b
F
39.12**
38.316**
β
0.532
0.528
t
6.255
6.19
p
<0.01
<0.01
Predictors: Versions
a: Dependent Variable: Cited in Google Scholar, b: Dependent Variable: ISI citation
Table 3. Descriptive statistics of variables - Year
Year
N
Versions
Cited in Google Scholar
ISI citation
Mean
SD
Mean
SD
Mean
SD
Before 2009
20
7.75
5.25
152.85
103.741
79.8
46.6
2009
26
6.08
1.695
101.19
38.948
56.96
20.577
2010
18
5.11
2.193
70.17
50.097
41.44
26.86
66
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
2011
16
4.31
1.352
49.25
33.66
21.31
12.015
2012
21
4.48
2.089
19.9
9.518
9.24
3.315
A comparison between Google Scholar and ISI citation for highly cited papers from Malaysia in the field of
engineering (see Figure 3) shows that the citation counts in Google Scholar are always higher than the number of
citations in ISI.
Figure 3. Comparison between Google Scholar and ISI citations
6. Conclusion
The number of publications and the number of citations in ISI Web of Science are used to measure the
researchers’ scientific performance and their research impact. However, these numbers are not freely available.
Therefore, the offered equation can be used as a reference to convert the number of Google Scholar citations to
ISI citations. On the other hand, the number of versions of both systems has a significant positive effect on the
number of citations. This finding supports other researchers’ (Amancio et al., 2012; Antelman, 2004; Egghe et al.,
2013; Ertürk & Şengül, 2012; Hardy et al., 2005) findings related to the paper visibility. The results of this study
indicate that there is a strong correlation between the number of citations in Google Scholar and ISI Web of
Science. Therefore, the researchers can increase the impact of their research by increasing the visibility of their
research papers (or paper versions). Future study is needed to determine the relationship between citation counts
on the other databases such as Microsoft Academic Research, Scopus, SiteSeer index and ISI by considering
journal article and conference papers.
Reference
Aghaei Chadegani, A., Salehi, H., Yunus, M. M., Farhadi, H., Fooladi, M., Farhadi, M., & Ale Ebrahim, N.
(2013). A Comparison between Two Main Academic Literature Collections: Web of Science and Scopus
Databases. Asian Social Science, 9(5), 18-26. http://dx.doi.org/10.5539/ass.v9n5p18
Aguillo, I. F. (2011). Is Google Scholar useful for Bibliometrics? A Webometric Analysis. In E. Noyons, P.
Ngulube, & J. Leta (Eds.), Proceedings of Issi 2011: The 13th Conference of the International Society for
Scientometrics and Informetrics, Vols 1 and 2 (pp. 19-25). Leuven: Int Soc Scientometrics &
Informetrics-Issi.
Ale Ebrahim, N. (2013). Introduction to the Research Tools Mind Map. Research World, 10(4), 1-3.
http://dx.doi.org/10.5281/zenodo.7712
Ale Ebrahim, N., Ahmed, S., & Taha, Z. (2009). Virtual Teams: a Literature Review. Australian Journal of Basic
and Applied Sciences, 3(3), 2653-2669. http://dx.doi.org/10.6084/m9.figshare.1067906
Ale Ebrahim, N., Salehi, H., Embi, M. A., Habibi Tanha, F., Gholizadeh, H., & Motahar, S. M. (2014). Visibility
67
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
and Citation Impact. International Education Studies, 7(4), 120-125. http://dx.doi.org/10.5539/ies.v7n4p120
Ale Ebrahim, N., Salehi, H., Embi, M. A., Habibi Tanha, F., Gholizadeh, H., Motahar, S. M., & Ordi, A. (2013).
Effective Strategies for Increasing Citation Frequency. International Education Studies, 6(11), 93-99.
http://dx.doi.org/10.5539/ies.v6n11p93
Amancio, D. R., Oliveira Jr, O. N., & da Fontoura Costa, L. (2012). Three-feature model to reproduce the
topology of citation networks and the effects from authors’ visibility on their h-index. Journal of
Informetrics, 6(3), 427-434. http://dx.doi.org/10.1016/j.joi.2012.02.005
Antelman, K. (2004). Do open-access articles have a greater research impact? College & Research Libraries,
65(5), 372-382. http://dx.doi.org/10.5860/crl.65.5.372
Bakkalbasi, N., Bauer, K., Glover, J., & Wang, L. (2006). Three options for citation tracking: Google Scholar,
Scopus and Web of Science. Biomedical Digital Libraries, 3(1), 7. http://dx.doi.org/10.1186/1742-5581-3-7
Bensman, S. (2011). Anne-Wil Harzing: The publish or perish book: Your guide to effective and responsible
citation analysis. Scientometrics, 88(1), 339-342. http://dx.doi.org/10.1007/s11192-011-0388-8
Bornmann, L., Marx, W., Schier, H., Rahm, E., Thor, A., & Daniel, H. D. (2009). Convergent validity of
bibliometric Google Scholar data in the field of chemistry-Citation counts for papers that were accepted by
Angewandte Chemie International Edition or rejected but published elsewhere, using Google Scholar,
Science Citation Index, Scopus, and Chemical Abstracts. Journal of Informetrics, 3(1), 27-35.
http://dx.doi.org/10.1016/j.joi.2008.11.r001
Cabezas-Clavijo, A., & Delgado-Lopez-Cozar, E. (2013). Google Scholar and the h-index in biomedicine: The
popularization
of
bibliometric
assessment.
Medicina
Intensiva,
37(5),
343-354.
http://dx.doi.org/10.1016/j.medin.2013.01.008
Craig, I. D., Plume, A. M., McVeigh, M. E., Pringle, J., & Amin, M. (2007). Do open access articles have greater
citation impact?: A critical review of the literature. Journal of Informetrics, 1(3), 239-248.
http://dx.doi.org/10.1016/j.joi.2007.04.001
Egghe, L., Guns, R., & Rousseau, R. (2013). Measuring co-authors' contribution to an article's visibility.
Scientometrics, 95(1), 55-67. http://dx.doi.org/10.1007/s11192-012-0832-4
Ertürk, K., & Şengül, G. (2012). Self Archiving in Atılım University. In S. Kurbanoğlu, U. Al, P. Erdoğan, Y.
Tonta, & N. Uçak (Eds.), E-Science and Information Management (Vol. 317, pp. 79-86): Springer Berlin
Heidelberg. http://dx.doi.org/10.1007/978-3-642-33299-9_11
Fooladi, M., Salehi, H., Yunus, M. M., Farhadi, M., Aghaei Chadegani, A., Farhadi, H., & Ale Ebrahim, N.
(2013). Do Criticisms Overcome the Praises of Journal Impact Factor? Asian Social Science, 9(5), 176-182.
http://dx.doi.org/10.5539/ass.v9n5p176
Garfield, E. (1972). Citation analysis as a tool in journal evaluation. Science, 178, 471-479.
http://dx.doi.org/10.1126/science.178.4060.471
Hardy, R., Oppenheim, C., Brody, T., & Hitchcock, S. (2005). Open Access Citation Information.
Hooper, S. L. (2012). Citations:
http://dx.doi.org/10.1038/483036c
not
all
measures
are
equal.
Nature,
483(7387),
36-36.
Jacso, P. (2012). Google Scholar Metrics for Publications The software and content features of a new open
access
bibliometric
service.
Online
Information
Review,
36(4),
604-619.
http://dx.doi.org/10.1108/14684521211254121
Kear, R., & Colbert-Lewis, D. (2011). Citation searching and bibliometric measures: Resources for ranking and
tracking. College & Research Libraries News, 72(8), 470-474.
Kousha, K., & Thelwall, M. (2007). Google Scholar citations and Google Web-URL citations: A multi-discipline
exploratory analysis. Journal of the American Society for Information Science and Technology, 58(7),
1055-1065. http://dx.doi.org/10.1002/asi.v58:7
Larsen, P. O., & von Ins, M. (2010). The rate of growth in scientific publication and the decline in coverage
provided
by
Science
Citation
Index.
Scientometrics,
84(3),
575-603.
http://dx.doi.org/10.1007/s11192-010-0202-z
Lawrence, S. (2001a). Free online availability substantially increases a paper's impact. Nature, 411(6837),
521-521. http://dx.doi.org/10.1038/35079151
68
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 5; 2014
Lawrence, S. (2001b). Online or invisible. Nature, 411(6837), 521. http://dx.doi.org/10.1038/35079151
McCabe, M. J., & Snyder, C. M. (2013). Does Online Availability Increase Citations? Theory and Evidence from
a Panel of Economics and Business Journals: SSRN working paper.
Orduña-Malea, E., & Delgado López-Cózar, E. (2014). Google Scholar Metrics evolution: an analysis according
to languages. Scientometrics, 98(3), 2353-2367. http://dx.doi.org/10.1007/s11192-013-1164-8
Pauly, D., & Stergiou, K. I. (2005). Equivalence of results from two citation analyses: Thomson ISI’s Citation
Index and Google’s Scholar service. Ethics in Science and Environmental Politics, 5, 33-35.
Rotich, D. C., & Musakali, J. J. (2013). Publish or Perish: Remaining Academically Relevant and Visible In the
Global Academic Scene through Scholarly Publishing. Paper presented at the Conference and Programme
Chairs, South Africa.
Solomon, D. J., Laakso, M., & Björk, B.-C. (2013). A longitudinal comparison of citation rates and growth
among
open
access
journals.
Journal
of
Informetrics,
7(3),
642-650.
http://dx.doi.org/10.1007/s11192-013-1164-8
The Thomson Corporation. (2013). Essential Science Indicators, Product Overview. Retrieved from
http://esi.webofknowledge.com/help//h_whatis.htm
Yue, W. P., & Wilson, C. S. (2004). Measuring the citation impact of research journals in clinical neurology: A
structural
equation
modelling
analysis.
Scientometrics,
60(3),
317-332.
http://dx.doi.org/10.1023/B:SCIE.0000034377.93437.18
Zhang, C.-T. (2009). The e-Index, Complementing the h-Index for Excess Citations. PLoS ONE, 4(5), e5429.
http://dx.doi.org/10.1371/journal.pone.0005429
Zheng, J., Zhao, Z. Y., Zhang, X., Chen, D. Z., Huang, M. H., Lei, X. P., … Zhao, Y. H. (2012). International
scientific and technological collaboration of China from 2004 to 2008: a perspective from paper and patent
analysis. Scientometrics 91(1), 65-80. http://dx.doi.org/10.1007/s11192-011-0529-0
Copyrights
Copyright for this article is retained by the author(s), with first publication rights granted to the journal.
This is an open-access article distributed under the terms and conditions of the Creative Commons Attribution
license (http://creativecommons.org/licenses/by/3.0/).
69
International Education Studies; Vol. 7, No. 7; 2014
ISSN 1913-9020 E-ISSN 1913-9039
Published by Canadian Center of Science and Education
Ethical and Unethical Methods of Plagiarism Prevention in Academic
Writing
Kaveh Bakhtiyari1,2, Hadi Salehi3, Mohamed Amin Embi4, Masoud Shakiba2,5, Azam Zavvari2,5, Masoomeh
Shahbazi-Moghadam6, Nader Ale Ebrahim7 & Marjan Mohammadjafari8
1
Department of Computer & Cognitive Science, Faculty of Engineering, University of Duisburg-Essen,
Duisburg, Germany
2
Department of Electrical, Electronics and Systems Engineering, Faculty of Engineering and Built Environment,
Universiti Kebangsaan Malaysia, Bangi, Malaysia
3
Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University, Najafabad, Isfahan, Iran
4
Faculty of Education, Universiti Kebangsaan Malaysia (UKM), Bangi, Malaysia
5
Industrial Computing Group, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia,
Bangi, Selangor, Malaysia
6
Perdana School of Science, Technology and Innovation Policy, Universiti Teknologi Malaysia, Malaysia
7
Research Support Unit, Centre of Research Services, Institute of Research Management and Monitoring (IPPP),
University of Malaya, Malaysia
8
Department of Industrial Engineering, Faculty of Engineering, Science and Research Branch, Islamic Azad
University, Kerman, Iran
Correspondence: Kaveh Bakhtiyari, Department of Electrical, Electronics and Systems Engineering, Faculty of
Engineering and Built Environment, Universiti Kebangsaan Malaysia, 43600 UKM Bangi, Selangor Darul Ehsan,
Malaysia. E-mail: [email protected]
Received: February 10, 2014
doi:10.5539/ies.v7n7p52
Accepted: May 13, 2014
Online Published: June 19, 2014
URL: http://dx.doi.org/10.5539/ies.v7n7p52
Abstract
This paper discusses plagiarism origins, and the ethical solutions to prevent it. It also reviews some unethical
approaches, which may be used to decrease the plagiarism rate in academic writings. We propose eight ethical
techniques to avoid unconscious and accidental plagiarism in manuscripts without using online systems such as
Turnitin and/or iThenticate for cross checking and plagiarism detection. The efficiency of the proposed
techniques is evaluated on five different texts using students individually. After application of the techniques on
the texts, they were checked by Turnitin to produce the plagiarism and similarity report. At the end, the
“effective factor” of each method has been compared with each other; and the best result went to a hybrid
combination of all techniques to avoid plagiarism. The hybrid of ethical methods decreased the plagiarism rate
reported by Turnitin from nearly 100% to the average of 8.4% on 5 manuscripts.
Keywords: plagiarism, plagiarism prevention, academic cheating, ethical approaches, unethical approaches
1. Introduction
Academic publishing is a key factor for any person in academia. All research students, lecturers, professors and
researchers in general tend to publish their works and the results of their studies in academic conference
proceedings, books, journals or magazines (Hockenos, 2013). Even though getting a paper published is very
important, the originality and ethic of the context of the published paper is very important as well. The most
common problem in academic manuscripts is “Plagiarism” (Carrol, 2002; Hawley, 1984; Patel, Bakhtiyari, &
Taghavi, 2011).
Plagiarism has been defined by different resources. But one of the best definitions of plagiarism is represented
on Dictionary.com web site:
“an act or instance of using or closely imitating the language and thoughts of another author without
authorization and the representation of that author's work as one's own, as by not crediting the original
52
www.ccsenet.org/ies
International Education Studies
Vol. 7, No. 7; 2014
author.” (Dictionary.com, 2013)
In Etymonline, it is also defined as “kidnapper, seducer, plunderer” which is originality since 1597. In American
Heritage, it is called as “Literary theft”, and in Encyclopedia Britannica is known as “fraudulence, forgery,
piracy-practices” (Encyclopedia-Britannica, 2013). Therefore, plagiarism is the act of copying someone else text,
idea or language and publish it as it is his/her own work. It is very clear from the above definitions that
plagiarism is against the law, and the authors should avoid plagiarism in their own manuscripts.
Nowadays, plagiarism is growing along with increase of publications and academic papers. One of the most
important criteria to evaluate the academics is their publications. Therefore, it is very important to help them to
increase their publication quality, and make them aware of plagiarism consequences. This paper gives a specific
attention to the authors, who plagiarize unconsciously without being aware of the plagiarism. Therefore, the
present study is to propose some ethical techniques which can prevent or at least decrease the plagiarism. Beside
the ethical techniques, we also present the unethical methods of masking plagiarism, which are employed by the
authors who consciously plagiarize. However, in order to decrease the rate of plagiarism, we do not recommend
using unethical methods which are not accepted in academia. But it is worth mentioning them to drag librarian
and academic lecturers’ attention. To evaluate the efficiency of proposed ethical techniques, they are applied in
five different texts using 16 students. The results had shown promising efficiency in plagiarism.
2. Plagiarism: Literature & Causes
At the beginning, we discuss the causes of plagiarism, and analyze how an author might commit plagiarism.
There are two types of authors who have plagiarism in their manuscripts. The first type is the direct plagiarizing
that the author would plagiarize by duplicating another author’s text exactly as it is, and pretending as his/her
own work. It is commonly known as copy-pasting in word processing applications. The second type is about
authors who plagiarize unconsciously. There are many factors which may cause these problems. Some of these
factors are as follows:
1) Uncited Ideas & Concepts: Scholars are used to read many articles to improve other methodologies in order
to deliver a contribution in science. Using other researcher’s idea is considered as plagiarism unless his/her
research article would be cited properly in the manuscript. Even though authors cite the other scholars’
research, they should avoid using the same words and language structures of the original text.
2) Many Authors: In many manuscripts, there might be many authors involved. However, not every single
author is aware whether the others have been honest and ethical in their writings or not. In a collaborative
work, each author is responsible for what he/she writes, but when it gets published all the authors are
partially responsible for the whole published material. Usually all authors trust each other in publications.
However, always the authors may not be honest. There is a famous quotation which says: “Sometimes too
much trust kills you”.
3) Accidental Similarity: The next factor is when the text of the author has been closely similar to another
author’s by chance. When an idea or a sentence is as usual as everybody uses in their own text, it might be
possible that a sentence or a concept would be recognized as plagiarism in plagiarism detection systems.
4) Fixed Definitions: At almost all sciences, there are fixed definitions which authors write and cite them, and
modification would make them a change or twist the meaning.
5) Cross-Text Plagiarism (Text Recycling): A piece of writing consists of several parts such as Introduction,
Body and Conclusion. These parts are dependent on the manuscript goals, applications and styles. However,
they might have more sections including Abstract, Evaluation, etc. Sometimes, authors need to repeat some
concepts in various parts of the text. For example, in the abstract they have to talk about what they are
presenting and discussing in the whole paper. Meanwhile, at the same time, each part of the abstract has
been discussed in details in the body of the manuscript. Another example, in methodologies, evaluations, or
even conclusion; the authors would definitely address the problems and issues which have been discussed
and elaborated in the other sections (e.g. Introduction) (IEEE, 2011). These texts with the same meaning
occurring in various places of the text might cause language and word similarity inside the manuscript. To
elaborate, a sentence in introduction with the definition of the problem statement might be duplicated
somewhere else in the paper (e.g. in Conclusion) with the exact language and sentence structure
(iThenticate.com, 2011). This problem is a very common issue with plagiarism. The ability of writing a
sentence with a same meaning in various styles and structures is an art, and it needs practice and authority in
both language and science. Co-authoring and having the manuscript proofread help significantly to
overcome this problem.
53
www.ccsenet.org/ies
International Education Studies
Vol. 7, No. 7; 2014
6) Self-Plagiarism (Similarity): Researchers might publish more than one article from their own research. Even
though they are different, they have many similarities in different sections of the articles. These papers may
fall into plagiarism by the following three cases:
a.
Redundancy and Duplication: This is also known as dual publication. It refers to the publication of a
single research article in more than one journal. Schein has done a research that 93 out of 660 studies
(more than 14%) from 3 surgical major journals were in case of dual publication (Schein & Paladugu,
2001).
Dual publication is not always considered as plagiarism. In two cases, dual publication is accepted and
ethical.
b.
i)
Summaries and abstracts which have been presented and published in a conference
proceedings can get published in a journal with some extension on the original paper. This
extension usually should be minimum 30%.
ii)
A published article can be republished in another language. But in the second publication, the
existence of the original (first article) should be mentioned and addressed.
Salami Slicing: Sometimes when the research has different aspects, the authors generate various articles
on different sections of the research such as literature review and methodology. They may also split up
the results (data fragmentation) in order to generate different research papers. In all these papers, the
basics such as the Introduction and Problem Statements are very similar to each other. Therefore,
writing the same concepts in different words is usually difficult. Some authors only copy the text from
their own previous published papers.
Besides data fragmentation, there is another issue called data augmentation. Data augmentation refers to
the publication of the studies which have been already published, and subsequently, the authors would
collect new data to improve the result and the contribution of the research. Having both results mixed
together and get it published might easily mislead the readers (malpractice).
c.
Copyright Infringement: The act of copying a text from the author’s previous published manuscripts is
another case of plagiarism. Although the authors have written the original text, it was published under
the copyright of a publisher. According to the copyright rules and regulations, making a copy of the
published text in whole section or a part is considered as plagiarism (or self-plagiarism).
7) Metadata: At the time of producing a document, the word processors (such as Microsoft Office Word) make
some metadata inside the document, which are hidden from the eyes and print. The metadata consist of
many useful information, and sometimes they cause headaches to the author (Hockenos, 2013; Mehta,
Meyer, & Murgai, 2013). For example, in a Microsoft Word document file, the generated metadata contains
the amount of times in minutes which took to complete the file, the last 10 authors of the file (last 10
computer names which edited the file), the company which has registered Microsoft Office product, and in
some cases with the track changes turned on, the deleted texts could also be found and extracted. Dennis
Rader was prosecuted because of murder, torture and bind. He was found and caught by police because his
word processor left his company information in his sent document’s metadata (Anderson, 2013).
The plagiarism as the problem, and its causes have been discussed. However the main challenge is to provide a
solution on “How can we avoid the plagiarism?”
3. Plagiarism Prevention Methods and Techniques
In the previous section, the driving factors of plagiarism in a manuscript were discussed. Here, few solutions are
presented to avoid unwanted plagiarism. As it has been discussed earlier, there are two groups of authors with
plagiarism. The first group of authors who are aware of plagiarism and what they do. For this group the only
solution is providing them sufficient education on consequences of ethical and unethical writings. The second
group includes the authors who plagiarize unconsciously without any intention to do so. Accordingly, there are
two major methods to cover up plagiarism by these two groups of authors: 1) Unethical methods, 2) Ethical
methods.
Unethical methods provide the solutions to decreasing the plagiarism by bypassing the plagiarism detection
software. In fact, the plagiarism has not been solved by these methods, but the software would report very low or
even 0% of plagiarism and similarity. Ethical methods talk about how to change the text properly in order to
decrease and to avoid the real amount of the plagiarism. The methods discussed in this paper are considered as
the authors have followed the ethical behavior in their research and have cited all references correctly and
54
www.ccsenet.org/ies
International Education Studies
Vol. 7, No. 7; 2014
properly in their manuscripts. In the follow, the solutions, methods and techniques are discussed in both of
unethical and ethical approaches.
3.1 Unethical Methods
Unethical methods refer to the techniques in order to decrease the plagiarism rate without any effort to change
the original text. These techniques aim to bypass plagiarism detection algorithms and to get a reasonable result of
plagiarism on a highly plagiarized context (Gillam, Marinuzzi, & Ioannou, 2010). However, they are not decent
and not recommended by any librarian. In this section, these methods are discussed and elaborated to inform
authors, lecturers, librarians and students to avoid and to be mindful of these methods. Also to make sure that
these techniques have not been applied to the produced manuscripts to be published.
Besides the unethical methods, there is a new concept in publishing and writing industry called “Ghost writing”.
Ghost writing means the action of getting other people to write on behalf of someone else. Although it is legal
and common among politicians and famous people, it is prohibited in the academic world. There are ghost
writers who write articles, books, academic assignments, or even university theses on behalf of a researcher for
money. Even though using ghost writing service seems legal, easy and common in many areas, but it is
considered as an unethical behavior in academia.
3.1.1 Mojibake
Mojibake is a Japanese word means transformation, and here it literally means character transformation. Using
Mojibake is only available in multi-layer documents such as PDF. A PDF document has a minimum of two
layers. The first layer (top-layer) is visual, and it is visible to the reader and printer. This layer is similar to an
image, and its text cannot be rendered. The second layer (bottom-layer) is a textual layer, and it contains the text
of the visual layer. When a text is selected, copied or searched; the textual layer is used and it is shown on the
visual layer (ISO, 2008).
In PDF documents, it is possible to keep the visual layer, which is image based, and change the textual layer to
the non-meaningful expressions. Figure 1 shows a sample of a PDF document which its textual layer is changed
to Mojibake (Gillam et al., 2010).
Figure 1. A sample of a text and its Mojibake in a PDF document
55
www.ccsenet.org/ies
International Education Studies
Vol. 7, No. 7; 2014
By having such PDF document submitted to plagiarism detection software, the system reads the textual layer and
checks its plagiarism rate which would be equal to 0%. This method is useful for the copyright protection in
order to prevent PDF documents’ texts from being copied and reused in other documents.
3.1.2 Cyrillic Letters
Coincidentally, some of the letters in Latin and Cyrillic look the same, but they are actually different. There are
three letters of “a”, “o” and “e” which they exist in both Latin and Cyrillic and they all look pretty much the
same in both alphabetical systems. These letters have also different values in ASCII and UNICODE for
computer systems. It means that they would be recognized as different characters in computer applications.
As an unethical action, the letters of “a”, “o” and “e” can be replaced with their similar Cyrillic character in a
whole text and those letters with the replaced characters will be invisible and unsearchable for plagiarism
detection software (Gillam et al., 2010).
3.1.3 Space Replacement
In a text, every word is separated from the adjacent words with a simple space. In fact, spaces define the words.
If there is no space in a text, it means that there is only one word even with hundreds of characters. This method
tries to eliminate the spaces of a text for plagiarism detection software, but on the other hand, make it readable to
the humans. If the spaces would be replaced with dots or some Chinese characters and then recolored to white,
they will not be visible in rich-text area. But in the plagiarism detection software, it would be considered as only
one word as it only processes the plain text and eliminates the rich-text’s styles. In the following, there is the
main text and its space-replaced version (Patel et al., 2011).
Original Text:
“Before Unicode, it was necessary to match text encoding with a font using the same encoding system. Failure
to do this produced unreadable gibberish whose specific appearance varied depending on the exact
combination of text encoding and font encoding.”
Text with spaces replaced with a Chinese character:
Before 画 Unicode,画 it 画 was 画 necessary 画 to 画 match 画 text 画 encoding 画 with 画 a 画 font 画 using
画 the 画 same 画 encoding 画 system.画 Failure 画 to 画 do 画 this 画 produced 画 unreadable 画 gibberish
画 whose 画 specific 画 appearance 画 varied 画 depending 画 on 画 the 画 exact 画 combination 画 of 画 text
画 encoding 画 and 画 font 画 encoding.
The second text with 39 words is considered as a single word for anti-plagiarism software with the same number
of characters as there is no space in the whole text.
3.1.4 Automatic Translation
System Translation (SysTran) is a translation of a text from one language into another by using computer
systems (Google, 2011; SysTranSoft.com, 2011). These systems have been improving every day and more users
are using such systems. Even though these systems cannot translate the text accurately as a human does, they are
very fast and their results are reliable. Using SysTran for an academic manuscript is a double-sided blade, which
has ethical and unethical uses.
There are lots of research studies available in other languages, and they have not been translated into English.
Using SysTran to translate the research and publish as their own work is considered as an unethical approach.
Because what it has been claimed is not the author’s original work.
3.1.5 Synonyms Replacement
Most of word processors are equipped with a dictionary or a thesaurus to suggest synonyms and antonyms for
the words. This feature is very useful for the author to replace the similar words in a paragraph with their
synonyms. Although this feature helps, it can be used to reduce plagiarism from a copied text. If some words in a
copied text would be replaced with their synonyms, the plagiarism detection systems usually fail to detect them.
However, the final text has been changed, the original structure of the context is still the same, and it is
considered as an unethical method to reduce the plagiarism.
56
www.ccsenet.org/ies
International Education Studies
Vol. 7, No. 7; 2014
3.1.6 Text Image
All of the available plagiarism detection systems only process the text of a manuscript, and they simply ignore
the images. If the authors put the text in an image, they would be skipped because the system cannot recognize
the text in the images of a manuscript (Patel et al., 2011). This detection is not only difficult for the system, but
also having a text image can be quite tough for humans to be recognized manually. For example, a text image in
a Word document is pretty clear to be identified, but in a PDF document, it is not easy to be found.
3.2 Ethical Methods
Ethical methods focus on the text words and structures to change them properly and to decrease the similarity
percentage as much as possible. These techniques should be used and applied with consideration of citing the
references correctly and properly in the manuscripts. These methods do not rely on the weaknesses of the
plagiarism detection algorithms.
3.2.1 Reading from Many Sources
When a researcher intends to read about a topic, usually there are loads of references to be read. Researchers
might get the point by reading only one reference, and reading the other similar articles might look time wasting
in some points.
There is a general psychological aspect of reading only one single reference. When a person reads a text, he/she
would respond by using the same structure when he/she is questioned. For example, imagine a person who has
watched a movie or read a book. Now he is retelling it to the others. He might use the same or very close
dialogues of the actors unconsciously to describe what has happened in the movie. Although they do not want to
tell the exact words and they have not memorized them; they repeat what they have heard (or read).
But when a person reads the same concept in many resources, his brain would make an imaginative conceptual
structure in his mind. While he decides to write, his brain discusses what exists as a structural model of the
concept in his brain and not the words and language structures. Reading from many resources will cause the
reader to have the concept of context fixed in his mind. However, reading a single text for many times would
have the words of context fixed in mind, and not the concept. People usually do this (reading a single text for
many times) to memorize a text.
3.2.2 Writing after a Few Days of Reading
It is quite common to write exactly what you have just read. Writing instantly after reading has the benefit of
being accurate and not missing any important part, but it also increases the chance of facing with plagiarism in
the text. According to the forgetting curve of human memory by Hermann Ebbinghaus, what it has been read
will be forgotten a day after by 20% (Ebbinghaus, 1913).
Figure 2. Forgetting curve of human brain by Hermann Ebbinghaus
57
www.ccsenet.org/ies
International Education Studies
Vol. 7, No. 7; 2014
Figure 2 shows the forgetting curve that after a day, 20% of what we have learnt is lost. If we do not review it in
3 days, we have almost forgotten 40%, and in almost a week, it is completely gone. But if after a day, we review
it for a few minutes, we can get it back to 100% again. This time the slope of the forgetting curve is less than the
previous time. Therefore, after 2 days of the review, we might lose 20% which was possible previously in only 1
day. This chart shows that after 4 times of periodic revision, it would be fixed in our minds.
By using this curve, it can be concluded that it is preferable to write a few days after you have read a reference.
In this way, the exact words and language structure will not be in your mind anymore, and your brain will make
up the new sentences automatically.
3.2.3 Using Thesaurus/Dictionary
A plagiarized sentence is an expression with the duplicated words. So, replacing the words with synonyms and
antonyms can help the authors to decrease the chance and amount of similarity in the context. Using dictionaries
and thesauruses are good tools for this purpose. Beside the paper-based dictionaries, almost most of the word
processing software is featured with an integrated dictionary to provide word replacements. Using these tools not
only decrease the plagiarism rate, but also it beautifies and makes fluent texts by replacing the repeated words in
a paragraph.
This technique is not ethical by itself, because replacing the words do not change the sentence structures. But
having the other approaches to avoid plagiarism next to this method is ethical.
3.2.4 As Many Proofread as Possible
Many authors would just have their own work submitted for publication just after being finished with writing. Of
course, they might be very good in writing or confident in the science, but having the paper proofread as many as
possible would also make some minor changes and improvements in the language structure of the paper. These
changes would definitely eliminate the plagiarism and similarity rate.
There are three issues that the authors should be aware of them. Firstly, they primarily should get people with the
same research discipline to proofread the text. Therefore, they only would modify the context as the meaning
remains the same. Asking various people with various research areas can just make amphibologies and opacities
in the context.
The second issue is that spreading the paper to many people can increase the risk of research piracy. Some of the
colleagues may publish the manuscript before the main author publishes the paper. Therefore, the authors should
find the people who are trustworthy and reliable.
The third issue is to remember to review the paper every time after the proof readings. Even if the authors trust
the knowledge of colleagues who have proofread the paper, still they should check to see everything has been
done correctly. It is very possible to make mistakes or twist the meanings when they edit a manuscript.
3.2.5 Using Machine Translators
Some people find it very tricky to write a sentence with a same meaning in several structures (paraphrase);
sometimes they cannot find any other solution rather than copying from the original location. Finding people to
paraphrase them is not always possible, or it is time consuming. Using machine translators is a very easy but not
a fast solution.
Computer translators use Natural Language Processing (NLP) techniques. NLP is a research area and it is not
going to be discussed in this paper. But here we get to know how it works in multi lingual translators. They
choose a language (usually English) as a main language, and they translate every other language from and into
this language. For example, if a user would like to translate a text from French into German, the system firstly
translates French into English, and then English into German. The reason of this action is based on the graph
mathematics. Making the translation rules of each language into another one in a multi-lingual translator will
make a mesh network of relations of languages and rules, which makes a really huge set of rules. A multi-lingual
translator with the support of 10 languages would only need 9 sets of rules, which are between 9 languages and
the one as the main language. But making a mesh network of sets for 10 languages is equal to 45 sets of rules as
following Equation 1.
(1)
According to a research in 2011 (Patel et al., 2011), the result of translated text from a western language into
another western, back to the source language will not make many changes. So, it is suggested to translate it
58
www.ccsenet.org/ies
International Education Studies
Vol. 7, No. 7; 2014
across the eastern and western languages and at the end back to the source language. For example, consider that
there is a text in English. It is suggested to translate it into an eastern language (e.g. Chinese), and then from the
eastern language (e.g. Chinese) into a western language (e.g. French), then repeating this procedure by
translating from western into eastern languages, and finally back into the original language which is English in
this example. The resulted text would be quite different with the original text.
Despite being an easy to use method, there are a few cautions that we should be aware of them. Firstly, the
generated language might be grammatically incorrect and it should be edited. Sometimes editing a sentence is
easier than paraphrasing the whole sentence or paragraph. Secondly, editing incorrect sentences takes a long time,
and it is not as easy and fast as it looks like, but it helps a lot.
3.2.6 Periodic Self-Reading
Once the paper is written, the authors should not leave or submit it. If they read what they have just written, they
can only find the grammatical mistakes or if a word is missing in the text. Further failures will not be recovered
in that stage. But if the authors read the text a few days later, they can cover more failures such as dim sections,
or even having more new sentences written in order to extend the paper. Self-reading the paper periodically can
fix the mistakes, improve and increase the quality of the writing. In spite of the benefits of having a periodic
self-reading, over-reading a paper can cause obsessions.
3.2.7 Quoting
There are some cases that the authors find themselves inevitable to write the exact words as the original text. For
example, it is a famous saying, a scientific abbreviation or a known expression. In these cases, using quotation is
a solution. Quoting someone else words is always ignored and not being considered as plagiarism. Quoting can
be from a few words to hundreds. This technique is simply putting the text into two quotation marks and
referencing the source at the end of the expression. For example, “A person who never made a mistake never
tried anything new”—Albert Einstein.
Besides the ethical solution of quoting, opening a quote in the beginning of the manuscript, and closing it at the
end is an unethical and ineligible approach.
3.2.8 Realistic & Imaginative Context
A solution (methodology) plays one of the most important roles for the whole journey of the research. The right
solution drives the author to the destination, and the wrong one can lead him/her to nowhere. Imagination is the
first place where the new ideas are born.
A researcher tries to use his imaginations and to lead them to be realistic, and respect them by writing, feeding,
raising, improving, and at the end publishing them.
Although an imagination might be funny and stupid at the first sight, it has the potential of opening a new door
toward the problem solving. Writing of imagination is usually and basically plagiarism-free, because they are
purely the authors’ properties.
4. Efficiency Evaluation of Plagiarism Prevention Techniques
In this section, the above ethical techniques are evaluated and the results are compared to check how effective
they are in decrement of plagiarism and similarity at academic writings. A group of 16 individual students from
the National University of Malaysia (Bangi, Malaysia) and Erican Language Centre (Kajang, Malaysia) has been
selected. Five texts consist of 400-500 words were chosen from Wikipedia to have a primary 95% - 100%
similarity index in Turnitin software. In the first stage, each of those above ethical techniques was taught to two
students, and they were requested to apply that specific technique on all those 5 texts from Wikipedia. Also the
students were allowed to edit the text as long as the meaning is the same. At the end of the first stage, each of
those ethical techniques was applied to all 5 texts by two students. And 80 texts were produced from a nearly 100%
plagiarized texts (copied from Wikipedia). These 80 texts were checked for plagiarism by Turnitin system.
59
Plagiarism
www.ccsenet.org/ies
International Education Studies
100
90
80
70
60
50
40
30
20
10
0
Vol. 7, No. 7; 2014
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
Text 1 40
51
20
25
65
59
75
81
34
43
70
75
85
89
45
51
Text 2 37
39
25
30
70
64
75
85
30
35
71
76
88
86
48
48
Text 3 47
43
29
32
75
63
79
87
28
41
65
70
87
81
41
51
Text 4 51
49
35
28
72
61
83
79
25
42
72
70
82
85
56
53
Text 5 42
55
21
29
68
62
78
80
35
45
75
73
80
79
42
50
Student No.
Text 1
Text 2
Text 3
Text 4
Text 5
Figure 3. Similarity index of each text by each student
Plagiarism
Each 2 students are trained with an ethical method. Students 1 & 2 are trained with method 1 (3.2.1), students 3
& 4 with method 2 (3.2.2) and respectively, until the last 2 students of 15 & 16 with method 8 (3.2.8). Figure 3
shows 80 values in percentage resulted by 16 students on 5 different texts. It shows that methods number 2
(Writing after a Few Days of Reading), 5 (Using Machine Translators) and 8 (Realistic and Imaginative Context)
are the top 3 methods to avoid plagiarism in a text. To clarify the above result, each of the 2 values of each
method in a paper is averaged, and it is illustrated in Figure 4.
100
90
80
70
60
50
40
30
20
10
0
1
2
3
4
5
6
7
8
Text 1
45.5
22.5
62
78
38.5
72.5
87
48
Text 2
38
27.5
67
80
32.5
73.5
87
48
Text 3
45
30.5
69
83
34.5
67.5
84
46
Text 4
50
31.5
66.5
81
33.5
71
83.5
54.5
Text 5
48.5
25
65
79
40
74
79.5
46
Methods
Text 1
Text 2
Text 3
Text 4
Text 5
Figure 4. Average of each method on every single text
Figure 5 illustrates the average of the similarity index after applying each method on all 5 texts. Lower values
show less similarity and plagiarism rate of each method on the texts.
60
www.ccsenet.org/ies
International Education Studies
Vol. 7, No. 7; 2014
100
90
80
6, 71.7
3, 65.9
70
Plagiarism
7, 84.2
4, 80.2
60
8, 48.5
1, 45.4
50
5, 35.8
40
2, 27.4
30
20
10
0
1
2
3
4
5
6
7
8
Method #
Figure 5. Average of similarity index of each method on all texts
In Figures 4 and 5, the effectiveness of each method is illustrated. As it can be shown, the best method to avoid
plagiarism went to method 2 (Writing after a Few Days of Reading). We also decided to measure how these
techniques might affect on the texts by applying all together.
Then, for the final stage, all 16 students were requested to apply their learnt methods together on the texts again.
This time, they were applying their methods collaboratively at the same time on 5 texts. And at the end, we had 5
texts with 8 methods applied on by 16 students. These 5 papers were evaluated to check the similarity index.
Table 1 shows the plagiarism rate of each text after applying all 8 methods. The similarity indices of these texts
were reduced from nearly 100% to the average of 8.4%, which is a brilliant similarity index by using of 8
methods and different students.
Table 1. Plagiarism rates of all 5 texts by applying 8 methods
Text #
Plagiarism
1
7%
2
5%
3
9%
4
13%
5
8%
Average
8.4%
5. Conclusion
This paper started with the explanations of plagiarism issues and literary piracy in academic manuscripts. It also
presented few unethical methods to warn the authors, students and librarians about them. Then it proposed
ethical methods to avoid the plagiarism and to decrease the similarity rate in the academic writings. These ethical
methods were all evaluated on 5 manuscripts by 16 students and checked by Turnitin. The results show that all
those methods can reduce the similarity index significantly. The most effective method was the second approach,
“Writing after a Few Days of Reading”. The hybrid of all these methods can make an almost plagiarism-free
context down to the average of 8.4% of similarity index. This research can be used by researchers to improve the
quality and originality of their academic writings.
References
Anderson,
J.
(2013).
Breaking
the
BTK
Killer's
61
Cipher.
Cryptologia,
37(3),
204-209.
www.ccsenet.org/ies
International Education Studies
Vol. 7, No. 7; 2014
http://dx.doi.org/10.1080/01611194.2013.797047
Carrol, J. (2002). A handbook for detecting plagiarism in higher education. London: Oxford: The Oxford Centre
for Staff and Learning Development, Oxford Brookes University.
Dictionary.com.
(2013).
Plagiarism.
Retrieved
http://dictionary.reference.com/browse/Plagiarism
January
8,
2013,
from
Ebbinghaus, H. (1913). Memory: A contribution to experimental psychology. Teachers college, Columbia
University.
Encyclopedia-Britannica.
(2013).
Plagiarism.
http://dictionary.reference.com/browse/Plagiarism
Retrieved
January
8,
2013,
from
Gillam, L., Marinuzzi, J., & Ioannou, P. (2010). TurnItOff: Defeating plagiarism detection systems. Paper
presented at the 11th Higher Education Academy-ICS Annual Conference, University of Durham, UK.
Google. (2011). Google Translator. Retrieved April 20, 2011, from http://translate.google.com/
Hawley, C. S. (1984). The thieves of academe: Plagiarism in the university system. Improving college and
university teaching, 32(1), 35-39. http://dx.doi.org/10.1080/00193089.1984.10533838.
Hockenos, P. (2013). High-Profile Plagiarism Prompts Soul-Searching in German Universities. Chronicle of
Higher Education.
IEEE.
(2011).
Cross
Check
User
Guide.
Retrieved
http://www.ieee.org/documents/CrossCheck_User_Guide.pdf
April
20,
2011,
from
ISO. (2008). Document management–Portable document format Part 1: PDF 1.7 (Vol. ISO 32000-1:2008):
International Organiztion for Standardization.
iThenticate.com.
(2011,
08/04/2011).
CrossCheck.
http://www.crossref.org/crosscheck/index.html
Retrieved
April
20,
2011,
from
Mehta, S., Meyer, D., & Murgai, P. (2013). Identifying and ranking networked biographies and referral paths
corresponding to selected qualifications: US Patent 20,130,018,871.
Patel, A., Bakhtiyari, K., & Taghavi, M. (2011). Evaluation of Cheating Detection Methods in Academic
Writings. Library Hi Tech, 29(4), 623-640. http://dx.doi.org/10.1108/07378831111189732.
Schein, M., & Paladugu, R. (2001). Redundant surgical publications: Tip of the iceberg? Surgery, 129(6),
655-661. http://dx.doi.org/10.1067/msy.2001.114549.
SysTranSoft.com. (2011). SYSTRAN–Online Translation Software. Retrieved April 20, 2011, from
http://www.systransoft.com/
Copyrights
Copyright for this article is retained by the author(s), with first publication rights granted to the journal.
This is an open-access article distributed under the terms and conditions of the Creative Commons Attribution
license (http://creativecommons.org/licenses/by/3.0/).
62
Cite as: N. Ale Ebrahim, “How to Promote Your Article,” University of Malaya Research Bulletin,
vol. 1, no. 1, 23 June, 2014. doi: 10.6084/m9.figshare.1083502
How to Promote Your Article
Nader Ale Ebrahim
Research Support Unit, Centre of Research Services, Institute of Research Management and Monitoring
(IPPP), University of Malaya, Kuala Lumpur, Malaysia
Abstract:
Writing an article for online distribution in a way that maximized the chances of citation hits, is
different from preparing one for print journals in some small, but important, respects. To be
cited, articles have to be visible in an electronic environment. Therefore, publishing a high
quality paper in scientific journals will be a halfway of receiving citation in the future. The rest
of the way is advertising and disseminating the publications by using the proper “Research
Tools”. Familiarity with the tools allows the researcher to increase his/her h-index in the short
time.
Keywords: University ranking, Improve citation, Citation frequency, Research impact, Open
access, h-index, Increase citations, Research tools
Inroduction:
This article provides a list of simple yet effective ways to promote your publications [1-3].
1.
2.
3.
4.
5.
6.
Use a unique name consistently throughout academic careers.
Use a standardized institutional affiliation and address, using no abbreviations
Repeat key phrases in the abstract while writing naturally.
Assign keyword terms to the manuscript
Optimize your article for Google Scholar and other academic search engines
Make a unique phrase that reflects author's research interest and use it throughout
academic life
7. The article metadata should be correct (especially author and title).
8. Publish in journal with high impact factor
9. Self-archive articles
10. Keep your professional web pages and published lists up to date (Make an attractive
homepage that is clear about what you are working on.)
11. Make your research easy to find, especially for online searchers
12. Publish in an Open Access (OA) journal
13. Deposit paper in Open Access repository
14. Publish with international authors
15. Write article collaboratively (Team-authored articles get cited more)
Cite as: N. Ale Ebrahim, “How to Promote Your Article,” University of Malaya Research Bulletin,
vol. 1, no. 1, 23 June, 2014. doi: 10.6084/m9.figshare.1083502
16. Use more references
17. Cite relevant thought influencers in your article (To be the best, cite the best).
18. Publish a longer paper
19. Publish papers with a Nobel laureates
20. Contribute to Wikipedia
21. Start blogging and share your blog post with target customers
22. Share your article on all your social media platforms
23. Interact with your peer connections through academic social media.
24. Keep track of all your international contacts
25. Follow the conversation on academic social media
26. When your paper is finally published, individually email the pdf to selected collaborators
27. Respond to comments.
28. Write a review paper
29. Avoid to select a question type of title
30. Sharing detailed research data
31. Publish across disciplines
32. Present a working paper
33. Publish your article in one of the journals everyone in your discipline reads
34. Add your latest published article’s link to your email signature
35. Take 50 photocopies of your best 1-2 papers to conferences, and leave them on the
brochure desk as a handout.
36. Increase number of publications in peer-reviewed journals
37. After a conference takes the paper, correct it, extend it, and submit it to a journal.
38. Publish your work in a journal with the highest number of abstracting and indexing
39. Create a podcast describing the research project
40. Make an online CV Like ORCID or ResearcherID.
41. Publish tutorials papers
42. Follow-up the papers which cited your article
43. Use all
“Enhancing Visibility and Impact” tools which are available on
http://www.mindmeister.com/39583892/research-tools-by-nader-ale-ebrahim.
Cite as: N. Ale Ebrahim, “How to Promote Your Article,” University of Malaya Research Bulletin,
vol. 1, no. 1, 23 June, 2014. doi: 10.6084/m9.figshare.1083502
Source: http://contentmarketinginstitute.com/2011/03/blog-post-to-dos/
References
[1]
N. Ale Ebrahim, H. Salehi, M. A. Embi, F. Habibi Tanha, H. Gholizadeh, S. M. Motahar,
and A. Ordi, “Effective Strategies for Increasing Citation Frequency,” International
Education Studies, vol. 6, no. 11, pp. 93-99, October 23, 2013.
[2]
N. Ale Ebrahim, H. Salehi, M. A. Embi, F. Habibi Tanha, H. Gholizadeh, and S. M.
Motahar, “Visibility and Citation Impact,” International Education Studies, vol. 7, no. 4,
pp. 120-125, March 30, 2014.
[3]
N. Ale Ebrahim, “Introduction to the Research Tools Mind Map,” Research World, vol.
10, no. 4, pp. 1-3, June 14, 2013.
Journal of Theoretical and Applied Information Technology
10th November 2013. Vol. 57 No.1
© 2005 - 2013 JATIT & LLS. All rights reserved.
ISSN: 1992-8645
www.jatit.org
E-ISSN: 1817-3195
CONTRIBUTION OF INFORMATION AND
COMMUNICATION TECHNOLOGY (ICT) IN COUNTRY’S HINDEX
1
MARYAM FARHADI1*, HADI SALEHI2,3, MOHAMED AMIN EMBI2, MASOOD FOOLADI1,
HADI FARHADI4, AREZOO AGHAEI CHADEGANI1, and NADER ALE EBRAHIM5
1
Department of Accounting, Mobarakeh Branch, Islamic Azad University, Mobarakeh, Isfahan, Iran
2
Faculty of Education, Universiti Kebangsaan Malaysia (UKM), Bangi, 43600, Malaysia
3
Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University, Najafabad, Isfahan,
Iran
4
School of Psychology and Human Development, Faculty of Social Sciences and Humanities, Universiti
Kebangsaan Malaysia (UKM), Malaysia
5
Department of Engineering Design and Manufacture, Faculty of Engineering, University of Malaya, Kuala
Lumpur, Malaysia
*
Corresponding author: Maryam Farhadi, Email: [email protected]
ABSTRACT
The aim of this study is to examine the effect of Information and Communication Technology (ICT)
development on country’s scientific ranking as measured by H-index. Moreover, this study applies ICT
development sub-indices including ICT Use, ICT Access and ICT skill to find the distinct effect of these
sub-indices on country’s H-index. To this purpose, required data for the panel of 14 Middle East countries
over the period 1995 to 2009 is collected. Findings of the current study show that ICT development
increases the H-index of the sample countries. The results also indicate that ICT Use and ICT Skill subindices positively contribute to higher H-index but the effect of ICT access on country’s H-index is not
clear.
Keywords: Information and Communication Technology (ICT) development, H-index, Middle East
However, ICT revolution and use of internet
caused many structural changes [3]. As an
example, the United States productivity has been
revived during the late 1990s and the early 2000s
because of the ICT development [4]. There are
many studies introducing ICT as an important
element to facilitate economic growth in both
developing and developed countries [5, 6]. In
addition, ICT has the potential to increase health
systems through the designing new methods to
prevent and detect disease [7]. Other studies
confirm the dominant role of ICT in improving
and modernizing the learning methods and
educational systems [8, 9]. The fast progress in
technology in the last two decades has reduced
the cost of ICTs. In this situation, it is easier for
more people to use and enjoy the advantages of
ICTs [10].
1. INTRODUCTION
In the last decades, technology has improved
the life style and made easier the human
activities through a new civilization. A new
revolution
named
Information
and
Communication Technology (ICT) has occurred
in the present world [1]. ICT is defined as a
concept including information equipment as well
as computers and software, auxiliary equipment
that is connected to a computer (such as
calculators, cash registers, photocopiers),
communications equipment, and instruments [2].
In other words, ICT is a technology with the aim
of
connecting
computers
and
other
communication equipment to gather, produce,
process, classify, manage, create, and distribute
information. After the invention of machines in
1970s, industrial revolution has been started and
machines have been used instead of handworks.
122
Journal of Theoretical and Applied Information Technology
10th November 2013. Vol. 57 No.1
© 2005 - 2013 JATIT & LLS. All rights reserved.
ISSN: 1992-8645
www.jatit.org
Although, the key role of ICT is approved in
different areas such as engineering [11],
education [12], health care [13], and economy
[6], its effect on the degree of scientific ranking
of a country is yet unexplored. Therefore, the
aim of this study is to explore the impact of ICT
development as measured by ICT Development
Index (IDI) on H-index as a proxy of scientific
progress of a country.
E-ISSN: 1817-3195
East countries, this study applies the regression
techniques with countries’ H-index as dependent
variable and IDI as independent variable as
follows:
H it =
α 0 + α1IDIit + ε it
(1)
where α 's are parameters to be estimated, H
represents H-index, IDI stands for ICT
development index and subscripts i and t show
the ith country in tth year, respectively. ε is the
error term which is assumed to be independent of
IDI. Then, for more inspections, instead of IDI,
which captures the composite effect of ICT
development, the current study uses three ICT
sub-indices including ICT Access, ICT Use and
ICT Skill as explanatory variables. In this
approach, we can test the distinct impact of each
sub-index
on
the
country’s
H-index.
Consequently, in this step of the analysis,
following equation (Equation 2) is used:
In addition, in order to shed more lights on
this area, present study applies three components
of ICT development namely, ICT Skill, ICT Use
and ICT Access to examine the distinct impact of
these components on H-index as a proxy of
scientific progress of a country. To this purpose,
this study applies a multiple regression analysis
on a panel of 14 Middle East countries over the
period 1995 to 2009.
The current study suggests the use of H-index
provided by SCImago database [14] to measure
the scientific ranking of country. Based on
Hirsch [15], a scientist or a group of scientists
has H-index of h when h of their papers has at
least h citations. In other words, H-index is the
number of articles with citation number higher or
equal to h [15]. The H-index, which has been
invented by Hirsch in 2005, proposes a simple
and apparently strong measure to compare the
scientific productivity and visibility of countries
and institutions [16].
H it =
µ0 + µ1Accessit + µ2 Useit + µ3Skillit + δ it
(2)
where µ 's are parameters to be estimated, H
shows H-index and Access, Use and Skill stand
for ICT development sub-indices. Moreover,
subscripts i and t indicate the ith country in tth
year, respectively, and δ is the error term which
is assumed to be independent of Access, Use and
Skill variables.
H-index can measure the quantity and quality
of papers simultaneously since the total number
of published articles and the number of citations
received by that articles are considered in the
calculation of H-index. In addition, there are
many studies, which reveal the benefits of Hindex as an appropriate and fair tool for
evaluating and ranking scientists, laboratories,
scientific journals, universities, institutions and
countries [17, 18, 19, 20, 21].
2.2. Data and Measurement
2.2.1.
ICT development
The IDI, as an indicator of ICT development
is computed based on Principal Component
Analysis (PCA), which combines various
indicators in order to make single value. For the
first time, IDI is introduced in 2009 by
Measuring the Information Society [23] and
considers the degree of progress in ICTs in more
than 150 countries. Composition of this indicator
includes ICT Access showing infrastructure and
access, ICT Skill referring to ICT skill or
capability and ICT Use displaying the extent of
using ICT. Required ICT data to compute the
IDI is extracted from United Nations Educational
Scientific and Cultural Organization (UNESCO)
and ITU. Nowadays, ICT is considered as an
important factor in development of countries,
which are moving towards knowledge or
Hirsch [15] proposes that H-index can be
used to appraise not only the scientists but also
countries and institutions. Hence, following
Jacso [22], current study also uses H-index as a
proxy to measure scientific ranking of countries.
2. METHODOLOGY AND DATA
2.1. Empirical Model
In order to examine the effect of ICT
development on scientific degree of 14 Middle
123
Journal of Theoretical and Applied Information Technology
10th November 2013. Vol. 57 No.1
© 2005 - 2013 JATIT & LLS. All rights reserved.
ISSN: 1992-8645
www.jatit.org
information-based societies. It is assumed that
these countries have experienced the process of
ICT development to turn into an information
society. Therefore, the ITU [23] suggests the
three-phase model including ICT readiness
(showing the degree of networked infrastructure
and ICT access), ICT intensity (showing the
degree of ICT use by the society) and ICT
impact (showing the result of effective and
efficient ICT use). IDI is a composite indicator
including ICT access, ICT skill and ICT use.
Each sub-index is composed of a number of
indicators explained as follow:
E-ISSN: 1817-3195
Figure 1. Three Stages In The Evolution Towards An
Information Society (Source: ITU [25])
1. Access Sub-Index shows the availability of
ICT infrastructure and individuals’ access to
basic ICTs. This sub-index has five indicators
including fixed telephone lines per 100
inhabitants, mobile cellular telephone
subscriptions
per
100
inhabitants,
international Internet bandwidth (bit/s) per
Internet user, proportion of households with a
computer and proportion of households with
Internet access at home.
2.2.2.
h-index
Different sources report the H-index. Some of
these sources are subscription-based sources
such as Scopus and Web of Science (WoS) and
some of them are free access sources such as
Harzing's Publish or Perish program on the basis
of Google Scholar entries. It should be noted that
these sources report different H-index for the
same institution, country and scholar because of
their different coverage. For example, WoS
covers a high range of published journals while it
does not cover high impact conferences.
Although, the coverage of publications in Scopus
is poor prior to 1996, it covers conferences
properly. Documents in Google Scholar receive
more citations in comparison with those in
Scopus and WoS. It is argued that Google
Scholar has the most coverage of journals and
conferences particularly those published after
1990 [26].
2. Use Sub-Index shows the real degree of ICTs
strength of use based on the available
infrastructure
and
access
indicators.
Considering the limitation on data for ICT
Use at the global scale and results of the
PCA, this subgroup has three indicators
including Internet users per 100 inhabitants,
fixed broadband Internet subscribers per 100
inhabitants
and
mobile
broadband
subscriptions per 100 inhabitants.
3. Skill Sub-Index captures the degree of ICT
skill in different countries. Since the required
data for many developing countries are not
collected, an appropriate indicator can be the
degree of literacy and education. Especially
in developing countries, the poor level of
education is a major obstacle to the effective
use of internet and computers. With the
inclusion of ICT in school courses, school
attendance may offer an appropriate indicator
for students’ exposure to the Internet or
computers. Therefore, this subgroup has three
indicators including, adult literacy rate,
secondary gross enrolment ratio and tertiary
gross enrolment ratio. The required data for
this subgroup are extracted from the
UNESCO Institute for Statistics (UIS) [24].
The current study collects the required data
on H-index for the panel of 14 Middle East
countries over the period 1995 to 2009, from
SCImago database [14]. The SCImago Journal
and Country Rank website presents the journal
and country scientific indicators based on the
information included in the Scopus database.
3. FINDING AND DISCUSSION
In this section, we first estimate Equation 1,
which explores the impact of ICT development
on H-index as a proxy of country’s scientific
ranking. Then, applying the three ICT subindices, this study presents the estimation results
for Equation 2, which is corresponding to the
effect of ICT development sub-indices on Hindex.
124
Journal of Theoretical and Applied Information Technology
10th November 2013. Vol. 57 No.1
© 2005 - 2013 JATIT & LLS. All rights reserved.
ISSN: 1992-8645
www.jatit.org
E-ISSN: 1817-3195
Table 1. Regression Results For The Impact Of ICT Development On H-Index
Dependent Variable: H-index
Variables
Coefficient
Standard Error
IDI
31.69
4.67***
Constant
24.16
13.35*
Observation = 210
F(1, 208) = 46.07***
R-squared = 0.1813
Adjusted R-squared = 0.1774
Note:
***, ** and * denote statistically significant at 1%, 5% and 10%, respectively.
Standard errors are heteroskedasticity consistent.
sample countries. The estimated coefficients
show a positive and significant effect of ICT Use
and ICT Skill on country’s scientific ranking.
This finding asserts that increasing the
application of ICTs and improving the ICT skills
might significantly increase the scientific degree
of a country. The results also indicate that if a
country increases the ICT Use and ICT Skill
indices by one unit, its H-index will raise by 133
and 407 units, respectively. As can be seen in
Table 2, the coefficient explaining the effect of
ICT Access on H-index is negative but not
significant. Finally, based on the adjusted Rsquared, ICT sub indices can explain 27.11% of
variations in country’s H-index.
Table 1 shows the effect of ICT development
on each country’s H-index. The estimation
results indicate a positive and significant effect
of ICT development on country’s scientific
ranking as proxied by H-index. This finding
suggests that suitable policies aim at improving
the level of ICT development can significantly
increase the scientific rankings of a country. The
results also assert that if a country raises its ICT
development index by one unit, the H-index will
increase by 31 units. As can be seen in Table 1,
the adjusted R-squared is equal to 0.1774, which
means that ICT development can explain 17.74%
of variations in country’s H-index.
Table 2 exhibits the impact of ICT
development sub-indices on H-index of 14
Table 2. Regression Results For The Impact Of ICT Development Sub-Indices On H-Index
Dependent Variable: H-index
Variables
T
6.79
1.81
Coefficient
Standard Error
t
ICT Access
-25.46
49.80
-0.51
ICT Use
133.23
64.64**
2.06
ICT Skill
407.28
64.86***
6.28
Constant
-151.10
35.82***
-4.22
Observation = 210
F(3, 206) = 26.91***
R-squared = 0.2816
Adjusted R-squared = 0.2711
Note:
***, ** and * denote statistically significant at 1%, 5% and 10%, respectively.
Standard errors are heteroskedasticity consistent.
125
Journal of Theoretical and Applied Information Technology
10th November 2013. Vol. 57 No.1
© 2005 - 2013 JATIT & LLS. All rights reserved.
ISSN: 1992-8645
www.jatit.org
E-ISSN: 1817-3195
[7] Haridasan, K., Rangarajan, S. and Pirio, G.
“Re-Inventing Health Care Training In The
Developing World: The Case For Satellite
Applications In Rural Environments”, Online
Journal of Space Communication, 14(Winter),
2009.
[8] Sharma, A., Gandhar, K., Sharma, S. and
Seema, S. “Role of ICT in the Process of
Teaching and Learning”, Journal of Education
and Practice, Vol. 2, No. 5, 2011, pp. 1-6.
[9] Sangrà, A. and González-Sanmamed, M. “The
role of information and communication
technologies in improving teaching and
learning processes in primary and secondary
schools”, Journal of Asynchronous Learning
Networks, Vol. 15, No. 4, 2011, pp. 47-59.
[10]
International Telecommunication Union,
“Measuring the information society: The ICT
Development Index”, Place des Nations CH1211 Geneva Switzerland, 2011.
[11]
C.C. Aldea, A.D. Popescu, A. Draghici
and G. Draghici, “ICT Tools Functionalities
Analysis for the Decision Making Process of
Their Implementation in Virtual Engineering
Teams”, Procedia Technology, Vol. 5, 2012,
pp. 649-658.
[12]
F. Scheuermann and F. Pedró,
“Assessing the effects of ICT in education:
Indicators, criteria and benchmarks for
international comparisons”, Luxembourg:
Publications Office of the European Union,
2009.
[13]
A.R. Bakker, “Health care and ICT,
partnership is a must”, International Journal of
Medical Informatics, Vol. 66, No. 1-3, 2002,
pp. 51-57.
[14] SCImago, SJR — SCImago Journal &
Country Rank, 2007, Retrieved March 11,
2013, from http://www.scimagojr.com
[15] J.E. Hirsch, “An index to quantify an
individual’s scientific research output”,
Proceedings of the National Academy of
Science, Vol. 102, No. 46, 2005, pp. 1656916572.
[16] J.F. Molinari and A. Molinari, “A new
methodology
for
ranking
scientific
institutions”, Scientometrics, Vol. 75, No. 1,
2008, pp. 163-174.
[17] P. Ball, “Index aims for fair ranking of
scientists”, Nature, Vol. 436, No. 900, 2005.
doi:10.1038/436900a.
[18] G. Saad, “Exploring the h-index at the
author and journal levels using bibliometric
data of productive consumer scholars and
business-related
journals
respectively”,
4. CONCLUSION AND IMPLICATIONS
This paper focuses on examining the effect of
ICT development on the scientific degree of 14
Middle East countries as proxied by H-index. The
results of the regression analysis show that ICT
development has a positive and significant effect
on the H-index of these countries. Moreover,
applying the ICT development sub-indices, this
study finds that ICT Use and ICT Skill might have
positive effects on H-index but the effect of ICT
Access is not clear since its corresponding
coefficient is not statistically significant.
Therefore, we can conclude that ICT induces
outcomes that leads to higher H-index values and
raises the scientific level of sample countries. In
this situation, policy makers should aim
increasing the level of ICT development through
increasing its indicators including fixed and
mobile telephone lines, international Internet
bandwidth, proportion of households with a
computer and Internet access, Internet users, fixed
and mobile broadband Internet subscribers, adult
literacy rate, secondary and tertiary gross
enrolment ratios.
REFERENCES:
[1] R.J. Gordon, “Does the ‘New Economy’
measure up to the great inventions of the
past?”, Journal of Economic Perspectives,
Vol. 14, No. 4, 2000, pp. 49-74.
[2] S.D. Oliner and D.E. Sichel, “Computers and
output growth revisited: How big is the
puzzle?”, Brookings Papers on Economic
Activity, Vol. 2, 1994, pp. 273-334.
[3] S.S. Cohen, J.B. Delong and J. Zysman,
“Tools for thought: What is new and important
about the E-economy?”, BRIE Working Paper,
No. 138, 2000.
[4] D.W. Jorgenson and K.J. Stiroh, “Information
technology and growth”, American Economic
review Papers and Proceedings, Vol. 89, No.
2, 1999, pp. 109-115.
[5] Gruber, H. and Koutroumpis, P. “Mobile
telecommunications and the impact on
economic development”, Economic Policy,
Vol. 26, No. 67, 2011, pp. 387-426.
[6] Farhadi, M., Ismail, R. and Fooladi, M.
“Information and Communication Technology
Use and Economic Growth”, PLoS ONE, Vol.
7,
No.
11,
2012,
e48903.
doi:10.1371/journal.pone.0048903
126
Journal of Theoretical and Applied Information Technology
10th November 2013. Vol. 57 No.1
© 2005 - 2013 JATIT & LLS. All rights reserved.
ISSN: 1992-8645
www.jatit.org
Scientometrics, Vol. 69, No. 1, 2006, pp. 117120.
[19] W. Glanzel, “On the h-index - A
mathematical approach to a new measure of
publication activity and citation impact”,
Scientometrics, Vol. 67, No. 2, 2006, pp. 315321.
[20] T. Braun, W. Glanzel and A. Schubert, “A
Hirsch-type index for journals Scientometrics,
Vol. 69, No. 1, 2006, pp. 169-173.
[21] K.M. Mclntyre, L. Hawkes, A. WaretSzkuta, S. Morand and M. Baylis, “The HIndex as a Quantitative Indicator of the
Relative Impact of Human Diseases”, PLoS
ONE, Vol. 6, No. 5, 2011. e19558.
doi:10.1371/journal.pone.0019558.
[22] P. Jacso, “The h-index for countries in Web
of Science and Scopus”, Online Information
Review, Vol. 33, No. 4, 2009, pp. 831-837.
[23] International Telecommunication Union,
“Measuring the information society: The ICT
Development Index”, Place des Nations CH1211 Geneva Switzerland, 2009.
[24] UNESCO Institute for Statistics (UIS),
2011. http://www.uis.unesco.org.
[25] International Telecommunication Union,
“Measuring the information society: The ICT
Development Index”, Place des Nations CH1211 Geneva Switzerland, 2010.
[26] L.I. Meho and K. Yang, “Impact of data
sources on citation counts and rankings of LIS
faculty: Web of science versus scopus and
google scholar”, Journal of the American
Society for Information Science and
Technology, Vol. 58, No. 13, 2007, pp. 2105–
2125.
127
E-ISSN: 1817-3195
From the SelectedWorks of Nader Ale Ebrahim
December 2014
Optimize Your Article for Search Engine
Contact
Author
Start Your Own
SelectedWorks
Available at: http://works.bepress.com/aleebrahim/93
Notify Me
of New Work
Preprint version of: N. Ale Ebrahim, “Optimize Your Article for Search Engine,” University of Malaya Research
Bulletin, vol. 2, no. 1, 23 December, 2014.
Optimize Your Article for Search Engine
Nader Ale Ebrahim
Research Support Unit, Centre of Research Services, IPPP, University of Malaya
Abstract
This article provides guidelines on how to optimize scholarly literature for academic search
engines like Google Scholar, in order to increase the article visibility and citations.
Introduction
Academic search engine optimization (ASEO) is a debatable topic [1]. Some researchers believe
“Designing scientific articles just for SEO would undermine the credibility and definitively the
quality of the articles and the research presented” [2]. On the other hand, the inclusion of the
articles in the index improves the ability to make their articles available to the academic
community [3]. “If you want your paper be "accepted" by a search engine you have to accept the
search engines style guidelines” [2]. The literature indicates a significant and positive
relationship between both citation in Google Scholar and ISI Web of Science with the number of
versions [4]. Increasing the visibility means rises the number of versions of an article and vice
versa [5]. The number of “versions” will be shown in a Google Scholar search result. Figure 1
shows 33 different versions of an article entitled “Virtual Teams: a Literature Review [6]” and
number of citations.
Figure 1 The number of “versions” in the Google Scholar search result
Discussion
There are several techniques that you can optimize your article for search engines:
1. Usage of strong keywords and synonyms; Once the keywords are chosen, they need to be
mentioned in the right places: in the title, and as often as possible in the abstract and the
body of the text (but, of course, not so often as to annoy readers).[1],
Preprint version of: N. Ale Ebrahim, “Optimize Your Article for Search Engine,” University of Malaya Research
Bulletin, vol. 2, no. 1, 23 December, 2014.
2. The completeness of documents metadata; When documents are converted to PDF, all
metadata should be correct (especially author and title). Some search engines use PDF
metadata to identify the file or to display information about the article on the search
results page. It may also be beneficial to give a meaningful file name to each article [1].
3. Write a good and short title [7];
4. Consistent spelling of authors’ first and last names;
5. Refer to reliable resources with high citations;
6. Select a proper journal; and
7. Deposit your paper in different repository; are some examples [8].
Google Scholar as one of the academic search engine indexes scholarly material from proprietary
sources such as subscription and open access journals, conference proceedings and white papers.
Google Scholar weighted heavily the title, the Journal name and author names also multiple
versions of an article, (e.g. on your home page and in UM Research Repository) [1] (please refer
to http://authorservices.wiley.com/bauthor/seo.asp [7] for more detail).
Beel, Gipp and Wilde [1] believe that academic search engine optimization (ASEO) should be a
common procedure for researchers, similar to, selecting an appropriate journal for publication.
ASEO should not be seen as a guide on how to cheat academic search engines. Rather, it is about
helping academic search engines to understand the content of research papers and, thus, about
how making this content more widely and easily available.
References
[1]
J. Beel, B. Gipp, and E. Wilde, “Academic Search Engine Optimization (ASEO),”
Journal of Scholarly Publishing, vol. 41, no. 2, pp. 176-190, 01/01/, 2010.
[2]
http://blog.webometrics.org.uk. "Academic Search Engine Optimization: An inevitable
evil?," 19 September, 2014; http://blog.webometrics.org.uk/2010/03/academic-searchengine-optimization-an-inevitable-evil/.
[3]
J. Beel, and B. Gipp, “On the robustness of google scholar against spam,” in Proceedings
of the 21st ACM conference on Hypertext and hypermedia (HT’10), Toronto, Ontario,
Canada, 2010, pp. 297-298.
[4]
N. Ale Ebrahim, H. Salehi, M. A. Embi, M. Danaee, M. Mohammadjafari, A. Zavvari,
M. Shakiba, and M. Shahbazi-Moghadam, “Equality of Google Scholar with Web of
Science Citations: Case of Malaysian Engineering Highly Cited Papers,” Modern Applied
Science, vol. 8, no. 5, pp. 63-69, August 6, 2014.
[5]
N. Ale Ebrahim, H. Salehi, M. A. Embi, F. Habibi Tanha, H. Gholizadeh, and S. M.
Motahar, “Visibility and Citation Impact,” International Education Studies, vol. 7, no. 4,
pp. 120-125, March 30, 2014.
Preprint version of: N. Ale Ebrahim, “Optimize Your Article for Search Engine,” University of Malaya Research
Bulletin, vol. 2, no. 1, 23 December, 2014.
[6]
N. Ale Ebrahim, S. Ahmed, and Z. Taha, “Virtual Teams: a Literature Review,”
Australian Journal of Basic and Applied Sciences, vol. 3, no. 3, pp. 2653-2669, 2009.
[7]
"Optimizing
Your
Article
for
Search
http://authorservices.wiley.com/bauthor/seo.asp.
[8]
N. Ale Ebrahim, H. Salehi, M. A. Embi, F. Habibi Tanha, H. Gholizadeh, S. M. Motahar,
and A. Ordi, “Effective Strategies for Increasing Citation Frequency,” International
Education Studies, vol. 6, no. 11, pp. 93-99, October 23, 2013.
Engines,"
23
September;
Modern Applied Science; Vol. 8, No. 2; 2014
ISSN 1913-1844
E-ISSN 1913-1852
Published by Canadian Center of Science and Education
Relationship among Economic Growth, Internet Usage and
Publication Productivity: Comparison among ASEAN and World’s
Best Countries
Hossein Gholizadeh1, Hadi Salehi2, Mohamed Amin Embi3, Mahmoud Danaee4, Seyed Mohammad Motahar5,
Nader Ale Ebrahim6, Farid Habibi Tanha7 & Noor Azuan Abu Osman1
1
Department of Biomedical Engineering, Faculty of Engineering, University of Malaya, 50603 Kuala Lumpur,
Malaysia
2
Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University, Najafabad, Isfahan, Iran
3
Faculty of Education, Universiti Kebangsaan Malaysia, Bangi, 43600, Malaysia
4
Faculty of Agriculture, Roudehen Branch, Islamic Azad University, Roudehen, Iran
5
Faculty of Information Science and Technology,Universiti Kebangsaan Malaysia, Bangi, 43600, Malaysia
6
Research Support Unit, Centre of Research Services, Institute of Research Management and Monitoring (IPPP),
University of Malaya, Malaysia
7
Department of Financial Sciences, University of Economic Sciences, Tehran, 1593656311, Iran
Correspondence: Hadi Salehi, Faculty of Literature and Humanities, Najafabad Branch, Islamic Azad University,
Najafabad, Isfahan, Iran. Tel: 98-913-104-1419. E-mail: [email protected]
Received: January 5, 2014
doi:10.5539/mas.v8n2p160
Accepted: February 7, 2014
Online Published: March 14, 2014
URL: http://dx.doi.org/10.5539/mas.v8n2p160
Abstract
Publication productivity, as measured by the number of papers, has been regarded as one of the main indicators
of reputation of countries and institutions. Nevertheless, the relationship among research publications, economic
growth and World Wide Web in ASEAN countries is still unclear. The main intention of this study was to
identify publication productivity among ASEAN and the world’s top ten countries in the last 16 years
(1996-2011). This study also aimed at finding the relationship among publication, gross domestic product (GDP)
and internet usage. Furthermore, the publication trend in the 10 first Malaysian universities was evaluated for the
same periods. Scopus database was used to find the overall documents, overall citations, citations per document
and international collaboration from 1996 to 2011 for each country. The World Bank database (World Data Bank)
was used to collect the data for GDP and the number of internet users. Moreover, to evaluate 10 top Malaysian
universities, the number of published articles, conferences, reviews, and letters for the same periods was
collected. The results of this study showed significant differences among ASEAN and top 10 countries regarding
publication productivity. Moreover, a positive and significant relationship was observed between indices, GDP
and internet usage for these countries. Surprisingly, international collaboration had a significant and negative
relationship with economic growth. Malaysia had fewer citations per document (7.64) and international
collaboration (36.9%) among ASEAN countries. In conclusion, international collaboration between academic
institutes and researchers is influenced by economic growth and access to internet in the countries. Furthermore,
publication trends in ASEAN countries are promising. However, policy makers and science managers should try
to find different ways to increase the quality of the research publication and to raise citation per document.
Keywords: ASEAN countries, publication productivity, citation, internet user, collaboration, economic growth
1. Introduction
The Association of Southeast Asian Nations, a geo-political and economic organization including ten countries
located in Southeast Asia, was formed on 8 August 1967 by Indonesia, Malaysia, the Philippines, Singapore and
Thailand. Since then, the membership has expanded to include Brunei, Burma (Myanmar), Cambodia, Laos, and
Vietnam. One of the main aims of this association is accelerating economic growth (Sarel, 1997). The main tool
for measuring a country's economy is Gross domestic product (GDP) that is the market value of all officially
recognized final goods and services produced within a country in a given period of time.
160
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
Based on the literature, there is a relationship between economic growth and education and research publication.
However, this relationship has not been well supported by realistic evidence (Jin & Jin, 2013; Nelson & Phelps,
1966; Lucas, 1988; Becker, 2009; Romer, 1990). The Researches by Mankiw et al. (1992) and Barro (1991)
showed positive and significant effects of education on economic growth (Mankive et al., 1992; Barro, 1991).
While the study by Bils and Klenow (2000) showed a possibility of reverse relationship between economic
growth and education (Bils & Klenow, 2000). Moreover, in a recent research study, Jin and Jin (2013) indicated
that publication productivity in different fields has dissimilar effects on economic growth. For instance,
engineering and science have positive influences on economic growth in comparison with publication output in
the field of social sciences (J. Jin & L. Jin, 2013). These days, publication productivity is a main task for
researchers and students than in the past decades and the results of research should benefit the community (Zain
et al., 2009). Based on new developmental principles, the number of knowledgeable employees could be
increased by education (J. Jin & L. Jin, 2013). In addition, educated human resource would help the economy of
the countries to grow faster. Moreover, it is widely observed that the world’s top productive research universities
are in the highly developed countries such as the United States, the United Kingdom and Canada et al.
Publication count is an indicator of research productivity and used to rank countries and universities (Liu &
Cheng, 2005; Meho & Spurgin, 2005; Narin & Hamilton, 1996; Toutkoushian et al., 2003; Yazit & Zainab, 2007).
It can also be used to determine authors’ productivity or the publication productivity of research groups (Liu &
Cheng, 2005; Hart, 2000; Uzun, 2002; Gu & Zainab, 2001; Fox, 1983). The number of citations of previously
published works is an indicator of its subsequent recognition and impact in a field of study. Reviewing articles
that are frequently cited can provide information about the dominant areas of a discipline and also highlight the
growth of particular fields. Furthermore, top-cited articles are often written by recognized experts who can offer
insights into future directions of the discipline (Joynt & Leonard, 1980; Kelly et al., 2010; Lefaivre & O’Brien,
2011; Ponce & Lozano, 2010).
Science Citation Index appeared in 1961 (following that, the Social Sciences Index was formulated in 1966).
Using this index, it has been possible to answer questions such as “who publishes more articles” and “whose
articles are being cited more often?” (Bar-Ilan, 2008; Heeks & Bailur, 2007; Meho & Yang, 2007). Hirsch
developed the H-index (proposed by Jorge E. Hirsch, a physicist at UCSD, in 2005) as an alternative to
traditional bibliometric measures. The H-index is a single bibliometric indicator that combines measures of
productivity and impact into one. Hirsch argued that this combination reflects the impact of a publication on the
corresponding scientific community (Braun et al., 2006; Schubert & Glänzel, 2007). H-index retrieved by
citation indexes (Scopus, Google scholar, and Web of Science) is used to measure the scientific performance and
the research impact studies based on the number of publications and citations of a scientist. It is also easily
available and may be used for performance measures of scientists and for recruitment decisions (Aghaei et al.,
2013; Ale Ebrahim et al., 2013; Brown & Gardner, 1985; Fooladi et al., 2013).
In this study, we used the human development index (HDI) to find the top ten countries around the world. HDI is
a composite statistic used to rank countries according to their development levels from “very high” to “low”.
Countries are placed based on life expectancy, education, living standard, child welfare, health care, economic
welfare, and population’s happiness.
The main purpose of this study was to compare publication productivity among ASEAN and the world’s top ten
countries from 1996 to 2011. In addition, we intended to find the relationship among publication productivity,
gross domestic product (current US$), and internet usage. Finally, we aimed to investigate the publication trend
of the 10 first Malaysian universities in the same periods. In this study, Scopus database
(WWW.SCIMAGOJR.COM) was used since it is a primary full-text scientific database and offers about 20%
more coverage than Web of Science (Anad & Sen, 1994).
We hypothesized publications of academic papers as the research outputs have positive relationship with
economic growth and the number of people with access to the worldwide network. Also, there is a significant
difference between ASEAN publication productivity and the world’s top ten countries.
2. Method
To identify publication productivity among ASEAN and the world’s top ten countries, Scopus database
(SCImago) was used to compare overall documents and citations, cited documents, un-cited documents,
self-citations, citations per document and international collaboration from 1996 to 2011. Moreover, the World
Bank database (World Data Bank) was used to collect data for GDP and the number of internet users in these
countries.
To find the top ten countries around the world, we used the human development index (HDI) (Anad & Sen,
161
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
1994). Moreover, to find the 10 first universities in Malaysia, we used the number of publications in Scopus
database. Furthermore, to evaluate these universities, the number of published articles, conferences, reviews, and
letters for the same periods was collected (1996-2011).
2.1 Data Analysis
Prior to data analysis, normality test was done for the obtained data. The normality test revealed that all variables
were not distributed normally. So, to answer the research questions, non parametric test was used. To evaluate
the relationship among related items of publication, GDP and Internet usage, Spearman correlation coefficient
test was applied. Moreover, to compare ASEAN countries with the top ten countries, Mann–Whitney U test was
used. For evaluating the differences in number of documents, citations, self-citations, and citations per document
among ASEAN countries, Kruskal-Wallis test was applied. The Kruskal-Wallis test is a nonparametric test that
compares three or more unmatched groups and it is an extension of the Mann-Whitney U test to allow the
comparison of more than two independent groups.
3. Results
Based on human development index (HDI), the top ten countries around the world are listed as follow: Norway,
Australia, USA, The Netherlands, Germany, New Zealand, Ireland, Sweden, Switzerland, and Japan.
Table 1. Comparison between ASEAN countries and the 10 first countries around the world (publication
productivity and country ranking)
ASEAN
TOP
TEN
Country ranking*
Overall Documents
(1996_2011) in Scopus
publication
H-Index
Brunei
1064
133
140
Cambodia
1296
128
112
Indonesia
16139
63
57
Laos
853
141
135
Malaysia
75530
42
54
Myanmar
906
139
146
Philippines
11326
70
56
Singapore
126881
32
30
Thailand
69637
43
39
Vietnam
13172
67
60
Norway
141143
31
21
Australia
592533
11
11
USA
6149455
1
1
Netherland
487784
14
8
Germany
1581429
5
3
New Zealand
114495
34
26
Ireland
91125
39
27
Sweden
337135
18
10
Switzerland
350253
17
9
Japan
1604017
4
6
Country
* based on 238 countries in the world.
162
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
Table 2. Kruskal Wallis test
ASEAN countries
Brunei
Cambodia
Indonesia
Laos
Malaysia
Myanmar
Philippines
Singapore
Thailand
Vietnam
ASEAN countries
Brunei
Cambodia
Indonesia
Laos
Malaysia
Myanmar
Philippines
Singapore
Thailand
Vietnam
ASEAN countries
Brunei
Cambodia
Indonesia
Laos
Malaysia
Myanmar
Philippines
Singapore
Thailand
Vietnam
ASEAN countries
Brunei
Cambodia
Indonesia
Laos
Malaysia
Myanmar
Philippines
Singapore
Thailand
Vietnam
Documents (Mean)
66.5
81
1008.688
53.3125
4720.625
56.625
707.875
7930.063
4352.313
823.25
Overall Citation (Mean)
484.0625
801.1875
7406.875
478.1875
17601.19
428.0625
7586.875
83167.75
33060.56
6725.25
Self-citation (Mean)
43.4375
88.0625
857.4375
64.875
4345.5625
33.625
866.0625
11939.563
5831.5
789.25
Citation per document
(Mean)
9.5081
12.2075
10.8456
11.8269
7.64
11.5994
12.9463
12.5063
10.6606
11.755
SD
32.26
64.46
663.05
45.95
5443.05
41.21
283.60
3940.70
2963.70
609.77
SD
194.40
614.26
2823.18
400.11
5917.94
245.58
2320.56
32859.46
14704.18
2789.02
SD
19.69
72.82
223.05
50.88
2483.06
18.62
240.79
4584.45
2407.33
394.57
Chi square*
p.value
138
<0.01
Chi square*
p.value
140.628
<0.01
Chi square*
p.value
142.822
<0.01
SD
Chi square*
p.value
5.37
6.05
5.18
5.86
3.98
6.61
6.20
5.02
4.97
5.75
13.099
<0.01
To investigate the differences among ASEAN countries, Kruskal Wallis test was applied. The results for all these
indices showed that there were significant differences among these countries.
163
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
As Tables 1 and 2 show, Singapore had the highest publications (126881 documents) and Laos had the lowest
publications (853 documents) among ASEAN countries, while Malaysia ranking (with 75530 documents) was
the second and the forty second among ASEAN and all the world countries, respectively (based on 238 countries
in the world). Laos with 853 documents, and the publication ranking of 141 from 1996 to 2011 was the last
country in this region. The USA was the first country in the world with the highest number of publications
(6149455), H-index (1035) and overall citation (114546415) (see Table 1).
Interestingly, Malaysia had the fewest citations per document (overall 7.64 cite per document) and the least
collaboration with the world (37%) among ASEAN countries, whereas Laos had the most international
collaboration (91%) in this region from 1996 to 2011 (see Figure 1 and Table 2). Furthermore, Malaysia and the
United States had the highest self-citations with 24 % and 48% of all the citations that were relieved during that
period among ASEAN and the top ten countries in the world (see Table 2 and Figure 1). Moreover, Malaysia had
the least collaboration with the world in the last 16 years among ASEAN countries (1996-2011). As Table 2
shows, there are significant differences among the number of documents (H(2) = 138, p < 0.01), citations (H(2)
= 140.62, p<0.01), self-citations (H(2) = 142.82, p<0.01), and citations per document (H(2) = 13.09, p<0.01)
among ASEAN countries.
Publication trends in ASEAN countries are promising in comparison with the top ten countries in the world (see
Figure 2 and Table 3). There was a significant difference (p<0.01) between ASEAN and developed countries in
number of documents, citable documents, citations (overall), self-citations, cited documents and research
collaboration with the world. The rate of growth (Figure 2) in scientific publication was 24% in Cambodia (the
highest in ASEAN), while Japan had 1.4% of growth in publication among these countries.
To evaluate the relationship among related items of publication productivity and GDP and Internet usage,
Spearman correlation coefficient was applied. Table 4 shows the correlation between indices and GDP and
internet usage for ASEAN and the top ten countries separately. The results showed that there was a positive and
significant relationship between GDP and the number of publications, citable documents, citation, self-citation
and cited documents in ASEAN and the top ten countries in the world. However, one variable of international
collaboration had a significant and negative relationship with GDP.
Regarding the relationship between internet usage and all the variables, we found a significant and positive
relationship in ASEAN countries except for the variable of international collaboration. However, there was a
positive and significant (p<0.01) relationship between international collaboration and internet usage in the top
ten countries.
Prior to data analysis, normality test was conducted. The results revealed that all variables were not distributed
normally so that we used Kruskal Wallis test, a non-parametric test, to examine publication productivity among
the top ten Malaysian universities. Table 5 shows the mean and standard deviation for all types of publications
among the top ten universities in Malaysia. The results of Kruskal Wallis test confirmed a significant difference
among these universities. The institution with the first ranking in Malaysia was University of Malaya with 13105
publications from 1996 to 2011. The main field of research in this university was physics and astronomy with
15 % of all publications. et al.,Moreover, the result of Spearman correlation coefficient test showed a positive
and significant relationship among the number of articles, conference papers, reviews and letters published
during these years. Except the number of conference papers and letters, all other coefficients showed a strong
relationship.
Table 3. Comparison between ASEAN and the top 10 countries for publication
REGION
ASEAN
TOP TEN COUNTRIES
Mean
SD
Mean
SD
Z
P value
Documents
1980.0
3453.7
71558.6
112595.0
-14.5
<0.01
Citable Documents
1859.9
3263.2
67369.3
104758.0
-14.518
<0.01
Citation (overall)
15774.0
27026.4
1190700.0
2256220.0
-14.9
<0.01
Self-citation
2485.9
4098.7
446135.2
1079040.0
-14.86
<0.01
Cited Documents
1277.8
2021.3
53836.4
84267.2
-14.748
<0.01
International
Collaboration
63.2
19.4
39.8
11.2
-10.206
<0.01
To compare ASEAN countries and the top ten countries, Mann–Whitney U test was applied. The results showed
significant differences for all indices
164
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
Table 4. Spearman correlation coefficient
REGION
Index
GDP
Internet users
Documents
.853**
.681**
Citable Documents
.854**
.669**
Citation (overall)
.699**
.511**
Self-citation
.737**
.534**
Cited Documents
.823**
.646**
% International Collaboration
-.502**
-.595**
Documents
.972**
.201*
Citable Documents
.973**
.199*
Citation (overall)
.757**
-.224**
Self-citation
.780**
-.165*
Cited Documents
.940**
0.121
% International Collaboration
**. Correlation is significant at the 0.01 level (2-tailed).
-.407**
.557**
ASEAN
TOP TEN
*. Correlation is significant at the 0.05 level (2-tailed).
Table 5. Kruskl Wallis test was applied to examine publication productivity among the top ten Malaysian
universities
Article
Conference
Review
Letter
university
Mean
SD
Mean
SD
Mean
SD
Mean
SD
University of Malaya
638
618
125
152
32
34
9
6
Universiti Sains Malaysia
570
609
131
160
19
24
4
4
Universiti Putra Malaysia
570
645
114
143
20
24
1
2
Universiti Kebangsaan Malaysia
416
509
188
287
14
20
2
3
Universiti Teknologi Malaysia
180
262
210
282
6
12
0
0
Universiti Teknologi MARA
103
153
150
281
4
7
0
1
International Islamic University
Malaysia
98
131
57
92
4
4
1
2
Multimedia University
111
100
89
87
4
5
0
0
Universiti Teknologi Petronas
51
98
85
155
2
4
0
0
Universiti Malaysia Perlis
32
58
43
70
1
2
0
0
Chi square
84
22
73
90
P value
<0.01
<0.01
<0.01
<0.01
Article
Conference
Review
Conference
.817**
Review
.894**
.717**
Letter
.746**
.488**
Spearman’s rho Correlation Coefficients
165
.741**
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
A
B
C
D
Figure 1. Comparison among ASEAN and World's Best Countries regarding self citation (A); Citation per
document (B); Un-cited documents (C); and International collaboration (D)
166
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
A
B
C
Figure 2. Publication trend in ASEAN countries and World's Best Countries (A); Publication trend in Malaysia
(B); Citation per document in Malaysia (C)
167
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
4. Discussion and Conclusion
In the current study, we made a comparison in Scopus database (SCImago) between ASEAN and the top 10
countries in the world regarding the relationship among economic growth, internet usage and publication
productivity from 1996 to 2011. In addition, we made a comparison among the first 10 Malaysian universities.
We used SJR, or SCImago Journal and Country Rank, as a measure of the scientific prestige of scholarly sources
derived from the Google Page Rank™ algorithm. It attempts to even out the differences in citation practice in
different subject fields and facilitate direct comparisons of sources. Furthermore, the SCImago Journal and
Country Rank portal displays journal and country scientific indicators developed from the information contained
in the Scopus® database. These indicators can be used to assess and analyze scientific domains.
Publication productivity could be a good indicator of research output and used to rank countries, research
institutes, or researchers in different fields (Liu & Cheng, 2005; Meho & Spurgin, 2005; Narin & Hamilton,
1996; Toutkoushian et al., 2003; Yazit & Zainab, 2007). The impact of a publication is assessed in terms of the
number of citations that it has received in relation to other outputs in the journal (Yi et al., 2008). However, the
number of times cited for an article is highly correlated with the length of time since its publication (Yi et al.,
2008). There was a significant difference (p < 0.01) between citation per document (Yi et al., 2008; Calver &
Bradley 2009) in ASEAN region (11.1 citation per document) and developed countries (17.7) (see Table 3).
Malaysia got 31.5 % and 57.1 % less citation per document in comparison with the average of citation per
document in ASEAN and top ten developed countries respectively. However, Malaysia had the highest
self-citation in this region with almost 24 % of all citations which were received during this period. The United
States had the highest number of publications as well as the highest self-citation (48%) in the world. Malaysia
had the least collaboration with the world in comparison with other ASEAN countries and it could be one of the
reasons for receiving less citation per document (Ale Ebrahim et al., 2013). Based on the literature, the mean
value of citation per publication of collaborative papers was higher than that of single country publications.
Publication trends in ASEAN countries with 10.1% growth are promising in comparison with the developed
countries studied in our research (2.3%). Figure 2 represents the publication trend and the number of
publications in Malaysia between 1996 and 2011. The number of publications increased dramatically after 2007
with 4783 to 18875 documents in 2011 (almost 4 times). However, the number of citations decreased rapidly
from 5.09 to 0.47 citations per document during these years (11times). One of the main reasons for the increase
in the number of publications in Malaysia could be focusing heavily on increasing the quality of research in its
research universities such as University of Malaya. Malaysia spends 1% of its gross domestic product on
development and research projects as determined in the 10th Malaysian Plan. Moreover, increasing the number
of PhD students and changing their program from conventional dissertations to paper-based dissertations (Its
requirement is publication in high peer-reviewed journals) could be another reason. The number of PhD students
in Malaysia has increased 10 times in recent years (from about 4,000 students in 2002 to almost 40,000 in 2012).
Table 5 shows the 10 first universities in Malaysia. The first top five universities have been chosen as research
universities and received additional government funding. Two third (68.8 %) of all Malaysian publications have
been produced by these research universities from 1996 to 2011.
Chemistry, Physics, and Astronomy are the first fields regarding the number of publications and citations.
Institutes that are working in these fields could publish more papers and get more citations. Malaysian
universities mostly work on Engineering and Medicine in comparison with the top universities located in the top
countries such as the United States.
Our findings in this study showed a positive and significant relationship between GDP and the number of
publications. This finding was similar to the previous researches that found positive and significant relationship
between education and economic growth (Mankiw et al., 1992; Barro, 1991; Bils & Klenow, 2000). Moreover, a
recent research study by Jin and Jin (2013) indicated that publication productivity in different fields has
dissimilar effects on economic growth. Surprisingly, for one variable i.e., international collaboration, we found
that there is a significant and negative relationship between international collaboration and the GDP in ASEAN
countries, while the number of people with access to the World Wide Web in developed countries had significant
and positive relationship with research collaboration projects. The countries with more economic growth like the
United States had less research collaboration with other counties.
Publication trends in ASEAN countries and especially in Malaysia (20.14 times increased) are encouraging in
comparison with the top ten developed countries. Nevertheless, these publications could not get reasonable
citations. Researchers, science managers, as well as policy makers should try to find different ways to increase
citation per document (Ale Ebrahim et al., 2013) such as improving the quality and visibility of their researches
168
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
and invest more on research studies whose outputs lead to more rapid economic growth.
Acknowledgements
The authors wish to extend their best gratitude to Mrs. Elham Sadat Yahyavi, and Mrs. Rezvan ...... for their kind
assistance in this project.
References
Aghaei Chadegani, A., Salehi, H., Md Yunus, M., Farhadi, H., Fooladi, M., Farhadi, M., & Ale Ebrahim, N.
(2013). A comparison between two main academic literature collections: Web of Science and Scopus
databases. Asian Social Science, 9(5), 18-26. http://dx.doi.org/10.5539/ass.v9n5p18
Ale Ebrahim, N., Salehi, H., Embi, M. A., Habibi, T. F., Gholizadeh, H., & Motahar, S. M. (2013). Effective
Strategies for Increasing Citation Frequency. International Education Studies, 6, 93-99.
http://dx.doi.org/10.5539/ies.v6n11p93
Anand, S., & Sen, A. (1994). Human development index: methodology and measurement. Human Development
Report Office (HDRO), United Nations Development Programme (UNDP).
Bar-Ilan, J. (2008). Which h-index?—A comparison of WoS, Scopus and Google Scholar. Scientometrics, 74(2),
257-271. http://dx.doi.org/10.1007/s11192-008-0216-y
Barro, R. J. (1991). Economic growth in a cross section of countries. The quarterly journal of economics, 106(2),
407-443. http://dx.doi.org/10.2307/2937943
Becker, G. S. (2009). Human capital: A theoretical and empirical analysis, with special reference to education.
University of Chicago Press.
Bils, M., & Klenow, P. J. (2000). Does schooling cause growth? American economic review, 1160-1183.
http://dx.doi.org/10.1257/aer.90.5.1160
Braun, T., Glänzel, W., & Schubert, A. (2006). A Hirsch-type index for journals. Scientometrics, 69(1), 169-173.
http://dx.doi.org/10.1007/s11192-006-0147-4
Brown, L. D., & Gardner, J. C. (1985). Using citation analysis to assess the impact of journals and articles on
contemporary accounting research (CAR). Journal of Accounting Research, 23(1), 84-109.
http://dx.doi.org/10.2307/2490908
Calver, M. C., & Bradley, J. S. (2009). Should we use the mean citations per paper to summarise a journal’s
impact or to rank journals in the same field? Scientometrics, 81(3), 611-615.
http://dx.doi.org/10.1007/s11192-008-2229-y
Falagas, M. E., Pitsouni, E. I., Malietzis, G. A., & Pappas, G. (2008). Comparison of PubMed, Scopus, web of
science, and Google scholar: strengths and weaknesses. The FASEB Journal, 22(2), 338-342.
http://dx.doi.org/10.1096/fj.07-9492LSF
Fooladi, M., Salehi, H., Md Yunus, M., Farhadi, M., Aghaei Chadegani, A., Farhadi, H., & Ale Ebrahim, N.
(2013). Does Criticisms Overcome the Praises of Journal Impact Factor? Asian Social Science, 9(5),
176-182. http://dx.doi.org/10.5539/ass.v9n5p176
Fox, M. F. (1983). Publication productivity among scientists: A critical review. Social Studies of Science, 13(2),
285-305. http://dx.doi.org/10.1177/030631283013002005
Gu, Y., & Zainab, A. N. (2001). Publication productivity of Malaysian researchers in the field of Computer
Science and Information Technology. Malaysian Journal of Library & Information Science, 6(1), 1-23.
Hart, R. L. (2000). Co-authorship in the academic library literature: a survey of attitudes and behaviors. The
Journal of Academic Librarianship, 26(5), 339-345. http://dx.doi.org/10.1016/S0099-1333(00)00140-3
Heeks, R., & Bailur, S. (2007). Analyzing e-government research: Perspectives, philosophies, theories, methods,
and
practice.
Government
Information
Quarterly,
24(2),
243-265.
http://dx.doi.org/10.1016/j.giq.2006.06.005
Jin, J. C., & Jin, L. (2013). Research publications and economic growth: evidence from cross-country
regressions. Applied Economics, 45(8), 983-990. http://dx.doi.org/10.1080/00036846.2011.613785
Joynt, R. L., & Leonard, J. A. (1980). Dantrolene sodium suspension in treatment of spastic cerebral palsy.
Developmental
Medicine
&
Child
Neurology,
22(6),
755-767.
http://dx.doi.org/10.1111/j.1469-8749.1980.tb03742.x
169
www.ccsenet.org/mas
Modern Applied Science
Vol. 8, No. 2; 2014
Kelly, J. C., Glynn, R. W., O’Briain, D. E., Felle, P., & McCabe, J. P. (2010). The 100 classic papers of
orthopaedic surgery a bibliometric analysis. Journal of Bone & Joint Surgery, British Volume, 92(10),
1338-1343. http://dx.doi.org/10.1302/0301-620X.92B10.24867
Lefaivre, K. A., & O’Brien, P. J. (2011). 100 most cited articles in orthopaedic surgery. Clinical Orthopaedics
and Related Research®, 469(5), 1487-1497. http://dx.doi.org/10.1007/s11999-010-1604-1
Liu, N. C., & Cheng, Y. (2005). The academic ranking of world universities. Higher education in Europe, 30(2),
127-136. http://dx.doi.org/10.1080/03797720500260116
Lucas Jr, R. E. (1988). On the mechanics of economic development. Journal of monetary economics, 22(1), 3-42.
http://dx.doi.org/10.1016/0304-3932(88)90168-7
Mankiw, N. G., Romer, D., & Weil, D. N. (1992). A contribution to the empirics of economic growth. The
quarterly journal of economics, 107(2), 407-437. http://dx.doi.org/10.2307/2118477
Meho, L. I., & Spurgin, K. M. (2005). Ranking the research productivity of library and information science
faculty and schools: An evaluation of data sources and research methods. Journal of the American Society
for Information Science and Technology, 56(12), 1314-1331. http://dx.doi.org/10.1002/asi.20227
Meho, L. I., & Yang, K. (2007). Impact of data sources on citation counts and rankings of LIS faculty: Web of
Science versus Scopus and Google Scholar. Journal of the American Society for Information Science and
Technology, 58(13), 2105-2125. http://dx.doi.org/10.1002/asi.20677
Narin, F., & Hamilton, K. S. (1996). Bibliometric performance measures. Scientometrics, 36(3), 293-310.
http://dx.doi.org/10.1007/BF02129596
Nelson, R. R., & Phelps, E. S. (1966). Investment in humans, technological diffusion, and economic growth. The
American Economic Review, 69-75.
Ponce, F. A., & Lozano, A. M. (2010). Highly cited works in neurosurgery. Part I: the 100 top-cited papers in
neurosurgical
journals:
A
review.
Journal
of
Neurosurgery,
112(2),
223-232.
http://dx.doi.org/10.3171/2009.12.JNS091599
Romer, P. M. (1990). Endogenous technological change. Journal of Political Economy, S71-S102.
http://dx.doi.org/10.1086/261725
Sarel, M. (1997). Growth and productivity in ASEAN countries: International Monetary Fund.
Schubert, A., & Glänzel, W. (2007). A systematic analysis of Hirsch-type indices for journals. Journal of
Informetrics, 1(3), 179-184. http://dx.doi.org/10.1016/j.joi.2006.12.002
Toutkoushian, R. K., Porter, S. R., Danielson, C., & Hollis, P. R. (2003). Using publications counts to measure
an institution's research productivity. Research in Higher Education, 44(2), 121-148.
http://dx.doi.org/10.1023/A:1022070227966
Uzun, A. (2002). Productivity ratings of institutions based on publication in Scientometrics, Informetrics, and
Bibliometrics, 1981-2000. Scientometrics, 53(3), 297-307. http://dx.doi.org/10.1023/A:1014864827441
Yazit, N., & Zainab, A. N. (2007). Publication productivity of Malaysian authors and institutions in LIS.
Malaysian Journal of Library and Information Science, 12.
Yi, H., Ao, X., & Ho, Y. S. (2008). Use of citation per publication as an indicator to evaluate pentachlorophenol
research. Scientometrics, 75(1), 67-80. http://dx.doi.org/10.1007/s11192-007-1849-y
Zain, Z. M., Ishak, R., & Ghani, E. K. (2009). The influence of corporate culture on organisational commitment:
A study on a Malaysian listed company. European Journal of Economics, Finance and Administrative
Sciences, 17(17), 16-26.
Copyrights
Copyright for this article is retained by the author(s), with first publication rights granted to the journal.
This is an open-access article distributed under the terms and conditions of the Creative Commons Attribution
license (http://creativecommons.org/licenses/by/3.0/).
170
HOME | CURRENT | ARCHIVES | FORUM
Research World, Volume 10, 2013
Online Version
Article A10.4
Introduction to the Research Tools Mind Map
Nader Ale Ebrahim
Research Support Unit, Centre of Research Services
Institute of Research Management and Monitoring (IPPP)
University of Malaya, Malaysia
aleebrahim[at]siswa.um.edu.my
Published Online: June 14, 2013
With the increasing use of information and communications technology (ICT), researchers are able to use
computer software tools to find, organise, manipulate, analyse, and share relevant information. However,
there are hundreds of such tools to select from, for various research-related uses. I have collected over
700 tools that can help researchers do their work efficiently. It is assembled as an interactive Web-based
mind map, titled Research Tools, which is updated periodically. Created using the MindMeister software
platform, Research Tools provides a convenient interface with zoom in/out, screen drag, and other userfriendly features.
Being a mind map, Research Tools consists of a hierarchical set of nodes. It has four main nodes: (1)
Searching the literature, (2) Writing a paper, (3) Targeting suitable journals, and (4) Enhancing visibility
and impact, and six auxiliary nodes. In the mind map, each parent node has several child nodes
associated with it, which can be accessed by clicking on the plus (+) sign. Several free tools can be found
in the child nodes. Some paid tools are also included.
The four main nodes are described below, listing some tools as examples. In the mind map, a green tick
sign indicates my preference for some tools. Users can access the tools by clicking on the arrow sign that
appears next to the tool title.
1. Searching the Literature
Tools under this node would assist first-year research students and other early-stage researchers in
expanding their knowledge base and developing a research focus (Ale Ebrahim, 2013). A sample of
these tools is given below, grouped under the specific purposes for which these tools might be used:
(a) Find keywords: Google AdWords, Google Trends, Springer Realtime, KwMap
(b) Develop a new keyword/phrase: Microsoft Adcenter Lab, keyword density analyser, Free Keyword
Tool
(c) Find research materials: Web of Science, Scopus®, ProQuest, Research Papers in Economics,
Munich Personal RePEc Archive (MPRA)
(d) Evaluate a paper: Publish or Perish, Readermete, Microsoft Academic Research
(e) Access databases: Social Science Research Network, Refdoc, iSEEK Education, Database of
Electrical Engineering & Computer Science
(f) Compare resources: Top 30 Web Tools to Search, 100 Useful Tips and Tools to Research the Deep
Web
(g) Search visually: Kartoo, Redz
2. Writing a Paper
Several tools on research writing have been assembled under this node. These tools help researchers in
various aspects of scientific and scholarly writing. Here is a sample of such tools:
(a) Write and edit: Ginger, Spell Checker, Plagiarisma, Springer Exemplar
(b) Analyse and summarise texts: Smmry, Tools 4 Noobs, Free Summarizer
(c) Search desktop: DtSearch, Copernic Desktop Search
(d) Write collaboratively: Google Drive, Adobe acrobat professional
(e) Synthesise and visualise information: MindMeister, MindMaple, Text2mindmap
(f) Manage bibliographic data: Mendeley, EndNote, Zotero
(g) Build and share article collections: Bibliogo, WizFolio
(h) Detect and avoid plagiarism: Viper, The Plagiarism Checker, Turnitin, Paper Rater
3. Targeting Suitable Journals
Sometimes, researchers are not sure about the journal they should target for publishing their research
papers. They would like to evaluate a journal’s visibility and credibility before submitting their papers.
Researchers can use a range of tools to identify, assess, and select appropriate journals with a view to
enhancing their publication profile (Ale Ebrahim, 2010).
(a) Identify potential journals: Elsevier journal finder, JANE (from The Biosemantics Group), Springer
journal advisor
(b) Detect relevant journals: Google Alerts, My Citation Alerts (from Web of Knowledge)
(c) Evaluate journal quality (impact factor and journal ranking): Journal Citation Report (Thomson
Reuters), SJR (Scimago Lab), Eigenfactor® (University of Washington)
4. Enhancing Visibility and Impact
Publishing a research paper in a scholarly journals is necessary but not sufficient for receiving citations in
future. We need to ensure that the paper is visible to the relevant users and authors. There are tools that
help in enhancing the visibility and readership of research papers. Effective use of these tools can result
in increased citations and, thus, improve the h-index of the author (h-index shows the author’s influence
as a researcher, see Ale Ebrahim, 2012; Aghaei Chadegani et al., 2013). Here is a sample of tools to
increase the visibility of one’s published papers:
(a) Create an online CV: ResearcherID, Publications List, ORCID, Google Scholar My Citations, A visual
resume
(b) Create an online repository: CiteuLike, Getcited, PubZone, Social Science Open Access Repository,
arXiv, HAL
(c) Share bookmarks: Diigo, Delicious, BibSonomy, ScienceWISE
(d) Publish and share information: Scribd, Docstoc®, figshare, Udemy
(e) Network with other researchers: ResearchGate, Academia, LinkedIn
5. Auxiliary Nodes
Research tools which could not be categorised under the above four main nodes have been placed under
six auxiliary nodes. The auxiliary nodes are: Survey, Links, Keeping up-to-date (Alert service), h-index,
General, and Download. The tools under these nodes can assist researchers in keeping track of their
research interest and activities in various ways.
(a) Stay current with research: Science direct alerts, MIT press, PubMed, Conference Alerts, WikiCFP
(b) Collect data: SurveyMonkey, Formsite, Survey share, Google Docs, eSurveysPro, Free Online
Survey, KwikSurvey
(c) h-index: Prediction scientific h-index, Finding h-index in Web of Science, Optimizing your article for
search engines, 10 Ways to Increase Usage and Citation of your Article Using Social Media
6. Conclusion
We can save plenty of research time by using specific tools for specific purposes. This can facilitate quick
progress and early completion of research projects. In my experience, research students who use these
tools are able to complete their literature review in less than 3 months.
Acknowledgement
Sincere thanks to Professor D. P. Dash for his kind encouragement to write this article and his useful
advice on its organisation.
References
Aghaei Chadegani, A., Salehi, H., Yunus, M. M., Farhadi, H., Fooladi, M., Farhadi, M., & Ale Ebrahim, N.
(2013). A comparison between two main academic literature collections: Web of Science and Scopus
databases. Asian Social Science, 9(5), 18-26. Retrieved from http://dx.doi.org/10.5539/ass.v9n5p18
Ale Ebrahim, N. (2010). Target ISI journals: How to write/publish ISI papers [Presentation material].
Retrieved from http://works.bepress.com/aleebrahim/1/
Ale Ebrahim, N. (2012). Publication marketing tools: Enhancing research visibility and improving citations
[Presentation material]. Retrieved from http://works.bepress.com/aleebrahim/64/
Ale Ebrahim, N. (2013). The effective use of research tools and resources. [Presentation material].
Retrieved from http://works.bepress.com/aleebrahim/73/
Suggested Citation: Ale Ebrahim, N. (2013). Introduction to the Research Tools mind map. Research
World, 10, Article A10.4. Retrieved from http://www1.ximb.ac.in/RW.nsf/pages/A10.4
Copyleft
The article may be used freely, for a noncommercial purpose, as long as the original source is
properly acknowledged.
Xavier Institute of Management, Xavier Square, Bhubaneswar 751013, India
Research World (ISSN 0974-2379) http://www1.ximb.ac.in/RW.nsf/pages/Home
American J. of Engineering and Applied Sciences 5 (1): 9-14, 2012
ISSN 1941-7020
© 2012 Science Publications
Technology Use in the Virtual R&D Teams
1
Nader Ale Ebrahim, 2Shamsuddin Ahmed,
Salwa Hanim Abdul Rashid and 2Zahari Taha
1
Department of Engineering Design and Manufacture,
Faculty of Engineering, University of Malaya Kuala Lumpur, Malaysia
2
Faculty of Manufacturing Engineering and Management Technology,
University Malaysia Pahang, 26300 Gambang, Pahang, Malaysia
1
Abstract: Problem statement: Although, literature proves the importance of the technology role in
the effectiveness of virtual Research and Development (R&D) teams for new product development.
However, the factors that make technology construct in a virtual R&D team are still ambiguous. The
manager of virtual R&D teams for new product development does not know which type of technology
should be used. Approach: To address the gap and answer the question, the study presents a set of
factors that make a technology construct. The proposed construct modified by finding of the field
survey (N = 240). We empirically examine the relationship between construct and its factors by
employing the Structural Equation Modeling (SEM). A measurement model built base on the 19
preliminary factors that extracted from literature review. The result shows 10 factors out of 19 factors
maintaining to make technology construct. Results: These 10 technology factors can be grouped into two
constructs namely Web base communication and Web base data sharing. The findings can help new product
development managers of enterprises to concentrate in the main factors for leading an effective virtual R&D
team. In addition, it provides a guideline for software developers as well. Conclusion: The second and third
generation technologies are now more suitable for developing new products through virtual R&D teams.
Key words: Collaboration teams, questionnaires performance, cross-functional teams, product
development, structural equation modeling, measurement model, literature review
even around the world. Therefore, enterprises have no
choice but to disperse their new product units to access
such dispersed knowledge and skills (Kratzer et al.,
2005). As a result, enterprises are finding that internal
development of all technology needed for new products
and processes are difficult or impossible. They must
increasingly receive technology from external sources
(Stock and Tatikonda, 2004).
Virtualization in NPD has recently started to make
serious headway due to developments in technologyvirtuality in NPD which is now technically possible
(Leenders et al., 2003). As product development
becomes the more complex, supply chain, also have to
collaborate more closely than in the past. These kinds
of collaborations almost always involve individuals
from different locations, so virtual team working
supported by Information Technology (IT), offers
notable potential benefits (Anderson et al., 2007).
Although the use of the internet in NPD has received
notable attention in the literature, little is written about
collaborative tool and effective virtual teams for NPD
(Ebrahim et al., 2009a). In addition, literature shows the
INTRODUCTION
Virtual teams are defined as “small temporary
groups of geographically, organizationally and/or time
dispersed knowledge workers who coordinate their
work, mainly with electronic information and
communication technologies to carry out one or more
organization tasks” (Ebrahim et al., 2009b). Virtual
R&D team is a form of a virtual team, which includes
the features of virtual teams and concentrates on R&D
activities (Ebrahim et al., 2011). The members of a
virtual R&D team use different degrees of
communication technology to complete the research
without space, time and organizational boundaries
(Nader et al., 2010a, Husain and Yong, 2009). “We are
becoming more virtual all the time!” is heard in many
global corporations today (Chudoba et al., 2005). On
the other hand, New Product Development (NPD) is
widely recognized as a key to corporate prosperity
(Lam et al., 2007). The specialized skills and talents
needed for developing new products often remain
locally in pockets of excellence around the company or
Corresponding Author: Nader Ale Ebrahim, Department of Engineering Design and Manufacture, Faculty of Engineering,
University of Malaya Kuala Lumpur, Malaysia
9
Am. J. Engg. & Applied Sci., 5 (1): 9-14, 2012
factors that make technology construct in a virtual R&D
team are still ambiguous. I this study we try to fill the
gap in the literature.
This study is structured as follows. First, base on
prior research we extract the 19 factors of technology
construct in the virtual R&D teams. Next, Structural
Equation Modeling (SEM) is used as the analytical tool
for testing the estimating and testing the technology
construct measurement models. Then adjust the
preliminary technology construct the model by fitting
the model according to the SEM fitness indices and
made a final measurement model. The study infers with
a discussion and future guidelines.
Factors that make technology construct in a virtual
R&D team are still ambiguous. We extracted 19importance factors related to the technology construct,
base on a comprehensive review on technology view in
the virtual R&D team working. Table 1 summarized the
factors and their supported references. E-mails and
conference calls are generally known as first generation
technologies while online discussion boards, power
point presentations, video tools and online meeting
tools are second-generation technologies. Third
generation technology refers typically to web-enabled
shared workspaces with the intranet or internet (LeeKelley and Sankey, 2008).
Literature review: Virtual teams use digital
communications, video and audio links, electronic
whiteboards, e-mail, instant messaging, websites, chat
rooms, as substitutes for physical collocation of the
team members (Baskerville and Nandhakumar, 2007,
Pauleen and Yoong, 2001). Simple transmission of
information from point A to point B is not enough; the
virtual environment presents significant challenges to
effective communication (Walvoord et al., 2008). Being
equipped with even the most advanced technologies are
not enough to make a virtual team effective, since the
internal group dynamics and external support
mechanisms must also be present for a team to succeed in
the virtual world (Lurey and Raisinghani, 2001). Virtual
teams are technology-mediated groups of people from
different discipline that work on common tasks (Dekker et
al., 2008) so the way the technology is implemented seems
to make the virtual teams outcome more or less likely
(Anderson et al., 2007). Virtual R&D team instructor
should choose the suitable technology based on the
purpose of the team (Ebrahim et al., 2009c).
Research method: To build a measurement model of
technology construct in virtual R&D teams for new
product development, we conducted a web-based
survey mainly in Malaysian and Iranian manufacturing
enterprises, in a random sample of small and medium
enterprises. Web-based survey method is selected
because; it is a cost-effective and quick result to get
feedback from the belief of the respondent. A Likert
scale from one to five was used. This set up gave
respondents a series of attitude dimensions. For each
factor, the respondent was asked whether, the factor is
not important or extremely important by using a Likert
scale rating. The questionnaire was emailed to the
managing director, R&D manager, the new product
development manager, project and design manager and
suitable people who were most familiar with the R&D
activities in the firm. The rapid expansion of Internet
users has given web-based surveys the potential to
become a powerful tool in survey research (Sills and
Song, 2002, Nader et al., 2010b).
Table 1: Summary of the factors related to the technology construct in the virtual teams
Factor name Factor descriptions
References
Tech1
Use internet and electronic mail
(Redoli et al., 2008, Pauleen and Yoong, 2001,
Lee-Kelley and Sankey, 2008, Thissen et al., 2007)
Tech2
Online meeting on need basis
(Chen et al., 2007; Lee-Kelley and Sankey, 2008;
Pena-Mora et al., 2000; Thissen et al., 2007)
Tech3
Web conferencing
(Coleman and Levine, 2007; Thissen et al., 2007,
Zemliansky and Amant, 2008; Ebrahim et al., 2009c)
Tech4
Seminar on the Web
(Zemliansky and Amant, 2008)
Tech5
Shared work spaces
(Lee-Kelley and Sankey, 2008)
Tech6
Video conferencing
(Chen et al., 2007; Zemliansky and Amant, 2008)
Tech7
Audio conferencing
(Chen et al., 2007; Lee-Kelley and Sankey, 2008; Zemliansky
Tech8
Online presentations
(Lee-Kelley and Sankey, 2008) and Amant, 2008)
Tech9
Share documents (off-line)
(Coleman and Levine, 2007; Ebrahim et al., 2009c)
Tech10
Share what’s on your computer desktop with people in
other locations (Remote access and control)
(Thissen et al., 2007; Ale et al., 2009)
Tech11
Do not install engineering software
(Coleman and Levine, 2007; Kotelnikov, 2007,
(get service through web browser)
Vasileva, 2009)
Tech12
Access service from any computer (in Network)
(Thissen et al., 2007; Vasileva, 2009)
Tech13
Standard phone service and hybrid services
(Thissen et al., 2007; Ebrahim et al., 2009c)
Tech14
Access shared files anytime, from any computer
(Lee-Kelley and Sankey, 2008)
Tech15
Web database
(Coleman and Levine, 2007; Zemliansky and Amant, 2008;
Ebrahim et al., 2009c)
Tech16
Provide instant collaboration
(Coleman and Levine, 2007; Thissen et al., 2007)
Tech17
Software as a service (canceling the need to install and run
the application on the own computer)
(Coleman and Levine, 2007; Thissen et al., 2007)
Tech18
Virtual research center for product development
(Zemliansky and Amant, 2008)
Tech19
Can be integrated/compatible with the other tools and systems
(Coleman and Levine, 2007; Kotelnikov, 2007)
10
Am. J. Engg. & Applied Sci., 5 (1): 9-14, 2012
Invitation e-mails were sent to each respondent,
reaching 972 valid email accounts, with reminders
following every two weeks up to three months. 240
enterprises completed the questionnaire, for an overall
response rate of 24.7% Table 2.
Convergent validity was established using a
calculation of the factor loading, Average Variance
Extracted (AVE) and Composite Reliability (CR). The
factors that have standardized loadings exceeded 0.50,
were maintained (Dibrell et al., 2008). The initial
measurement model was consisting of 19 factors (Tech1
to Tech19). After revising the measurement model by
deleting Tech1, Tech10, Tech11 and Tech13, the AVE
and CR were calculated. AVE larger than 0.5 is the
threshold (McNamara et al., 2008). CR is calculated by
squaring the sum of loadings, then dividing it by the sum
of squared loadings, plus the sum of the measurement
error (Lin et al., 2008). CR should be greater than 0.6
(Huang, 2009). The measurement model had acceptable
convergent validity since the calculated CR and AVE
were 0.930 and 0.613 respectively.
For discriminant validity, we performed AMOS
software using Maximum Likelihood method (ML).
The fitting indices checked with their respective
acceptance values Table 4. We run the AMOS for the
model Ver1 (technology construct with 15 factors) and
found a nonsignificant chi-square per degrees of
freedom (CMIN/DF = 7.232). Most of the rest of fit
indices was not in the acceptable range.
RESULTS
Anderson and Gerbing (1988) suggested using
Confirmatory Factor Analysis (CFA) for scale
development because it affords stricter interpretation of
unidimensionality than what is provided by more
traditional approaches, such as coefficient alpha, itemtotal correlations and exploratory factor analysis. The
evidence that the measures were one-dimensional,
where a set of indicators (factors) shares only a single
underlying construct, was assessed using CFA
(Anderson and Gerbing, 1988). After data collection,
the measures purification procedures should be used to
assess their reliability, unidimensionality, discriminate
validity and convergent validity (Anderson and
Gerbing, 1988).
For reliability analysis, Cronbach’s Alpha
(Cronbach, 1951) was employed to each factor. As
shown in Table 3, all the items with Cronbach’s α
greater than threshold 0.6 were included in the analysis
and the rest omitted from analysis. So, the factors
Tech1, Tech10, Tech11 and Tech13 freed from further
analysis. In general, the reliability of the questionnaire’s
instruments displayed a good reliability across samples.
Structural Equation Modeling (SEM) using AMOS
18 was employed for validation of the measurement
model. This statistical analysis are estimated
simultaneously for both the measurement and structural
models (Dibrell et al., 2008). To ensure the factors
make a right construct, the measurement model
examined for model fit. Given this, the model assessed
for the convergent and discriminant validity.
Table 2: Summarized online survey data collection
Numbers of emails sent enterprises
Total responses (Click the online web page)
Total responses / received questionnaire (%)
Total completed
Total completed / received questionnaire (%)
Table 3: Summary of the final measures and reliabilities
Corrected itemCronbach’s alpha
Factor name
total correlation
if Item deleted
Tech1
0.525
0.943
Tech2
0.755
0.939
Tech3
0.777
0.939
Tech4
0.717
0.940
Tech5
0.759
0.939
Tech6
0.722
0.940
Tech7
0.731
0.939
Tech8
0.780
0.939
Tech9
0.610
0.942
Tech10
0.576
0.942
Tech11
0.571
0.943
Tech12
0.686
0.940
Tech13
0.519
0.943
Tech14
0.624
0.941
Tech15
0.696
0.940
Tech16
0.642
0.941
Tech17
0.678
0.940
Tech18
0.649
0.941
Tech19
0.615
0.942
3625
972.0
26.8
240.0
24.7
Table 4: Fitting indices (adopted from (Byrne, 2001)
Fit Indices
Desired range
χ2 /degrees of freedom (CMIN/DF) ≤ 2.00
IFI (Incremental Fit Index)
≥ 0.90
Comparative Fit Index (CFI)
Coefficient values range from zero to 1.00, with values close to 0.95 showing superior fit
RMSEA (Root Mean Squire
values less than .05 show good fit and values as high as .08 represent reasonable fit,
Error of Approximation)
from 0.08-0.10 show mediocre fit and those greater than 0.10 show poor fit
Root mean square residual (RMR) ≤ 0.08
Goodness-of-Fit Index (GFI)
≥ 0.90
Normed Fit Index (NFI)
Coefficient values range from zero to 1.00, with values close to 0.95 showing superior fit
Relative Fit Index (RFI)
Coefficient values range from zero to 1.00, with values close to 0.95 showing superior fit
Tucker-Lewis Index (TLI)
Values ranging from zero to 1.00, with values close to 0.95 (for large samples) being indicative of good fit
11
Am. J. Engg. & Applied Sci., 5 (1): 9-14, 2012
Fig. 2: Final measurement model
Fig. 1: Measurement model Ver2
The results of the final measurement model of
technology construct in virtual R&D team for
developing a new product, shows the share of two main
contrasts, which are strongly correlated to each other:
Thus, refer to the AMOS Modification Indices (MI)
some of the factors that had the lowest factor loading or
the same effect of remaining factor, were deleted. With
this modification, the measurement model Ver2 had a
significant chi-square per degrees of freedom
(CMIN/DF = 4.767); other fit indices, RMSEA, RMR
and GFI also were in the acceptable range. Therefore,
the best fitting model was the measurement model Ver2
Fig. 1 and it used for further analysis.
•
•
DISCUSSION
The final measurement developed made base on the
measurement model ver2 by classifying the factors in
two groups according to their relevant factor loading
with the threshold 0.83. The proper name for each
group can be web base; communications and data
sharing respectively. As displayed in Fig. 2 each factor
loading was above 0.62 and significant. Overall, the
final measurement model produced good fit indices
(CMIN/DF = 2.889, RMR = .04, GFI = 0.929, RFI =
0.929, NFI = 0.949, TLI = 0.952, CFI = 0.966 IFI =
0.964, RMSEA = 0.089).
While fitting the technology construct the
measurement model the factors Tech14 (access shared
files anytime, from any computer), Tech15 (web
database), Tech16 (provide instant collaboration),
Tech17 (software as a service (eliminating the need to
install and run the application on the own computer))
and Tech19 (can be integrated/compatible with the
other tools and systems) were dropped. Modification
indices (MI) base on regression weights shows Tech17,
Tech 18 and Tech19 are highly correlated, so one
representative (Tech18) from this group is enough.
Tech14 to Tech16 are strongly correlated with Tech12,
so the remaining factor represents the deleted ones.
Web base communications consists of online
meeting on needed basis, web conferencing,
seminar on the web, video conferencing, audio
conferencing and online presentations
Web base data sharing consists of shared work
spaces, share documents (off-line), access service
from any computer (in network) and virtual
research center for product development
According to Lee-Kelley and Sankey (2008) these
two constructs belong to the second and third
generation of technology. Equip virtual R&D team
members with the suitable technology make the teams
more effective. Therefore, the manager of NPD should
provide the facilities and infrastructures for the virtual
R&D teams to achieve the higher level of team
effectiveness.
CONCLUSION
Research so far has explored the 19 factors for
working together virtually; however, us still less know
about the factors which has main contributions in the
technology construct of the virtual R&D teams for New
product development. The findings of this study extend
the literatures and help to build a foundation for further
understanding of the technology elements in the virtual
R&D teams for new product development. The
measurement model shows ten factors that make the
technology constructs. These ten factors can be sorted
by their factor loading which are reflecting the factor
weight. Therefore, the software developer or the
managers of the NPD are able to provide a better
12
Am. J. Engg. & Applied Sci., 5 (1): 9-14, 2012
platform for virtual team working by concentrating on
the main factors. The second and third generation of
technology (refer to definition of Lee-Kelley and
Sankey (2008) is now more suitable for developing a
new product through virtual R&D teams.
Future research is needed to examine the effects of
each factor to perform the virtual R&D teams while the
others constructs of virtual teams such as process and
people are present. A new SEM is needed to
demonstrative the relationship between factorsconstructs and constructs-constructs which is not
investigated yet in the literature.
Dekker, D.M., C.G. Rutte and V.D.P.T. Berg, 2008.
Cultural differences in the perception of critical
interaction behaviors in global virtual teams. Int. J.
Intercu.
Relations,
32:
441-452.
DOI:
10.1016/j.ijintrel.2008.06.003
Dibrell, C., P.S. Davis and J. Craig, 2008. Fueling
Innovation through Information Technology in
SMEs. J. Small Bus. Manage., 46: 203-218. DOI:
10.1111/j.1540-627X.2008.00240.x
Ebrahim, N.A., S. Ahmed, Z. Taha, 2009a. Modified
stage-gate: A conceptual model of virtual product
development process. Afr. J. Marke. Manag, 1:
211-219.
Ebrahim, N.A., S. Ahmed, Z. Taha, 2009b. Virtual
R&D teams in small and medium enterprises: A
literature review. Sci. Res. Essay, 4: 1575-1590.
DOI: 10.2139/ssrn.1530904
Ebrahim, N.A., S. Ahmed, Z. Taha, 2009c. Virtual
teams: A literature review. Australian J. Basic
Applied Sci., 3: 2653-2669.
Ebrahim, N.A., S.H.A. Rashid, S. Ahmed, Z. Taha,
2011. The Effectiveness of Virtual R&D Teams in
SMEs: Experiences of Malaysian SMEs. Indu.
Eng. Manag. Syst., 10: 109-114.
Huang, C.C., 2009. Knowledge sharing and group
cohesiveness on performance: An empirical study
of technology R&D teams in Taiwan.
Technovation,
29:
786-797.
DOI:
1016/j.technovation.2009.04.003
Husain, A.J.A. and C.H. Yong, 2009. Utility-based
policy management system for virtual organization.
J.
Comput.
Sci.,
5:
635-645.
DOI:
10.3844/jcssp.2009.635.645
Kotelnikov, V., 2007. Kim Hak-Su Asia-Pacific
Development Information Programme e-Primers
for the Information Economy, Society and Polity.
1st Edn., ISBN: 9789748283937, pp: 27.
Kratzer, J., R.T.A.J. Leenders and J.M.L.V. Engelen,
2005. Keeping virtual R&D teams creative. CBS
Interactive.
Lam, P.K., K.S. Chin, J.B. Yang and W. Liang, 2007.
Self-assessment of conflict management in clientsupplier collaborative new product development.
Indus. Manag. Data Syst., 107: 688-714. DOI:
10.1108/02635570710750435
Lee-Kelley, L. and T. Sankey, 2008. Global virtual
teams for value creation and project success: A
case study. Int. J. Project Manag. 26: 51-62. DOI:
10.1016/j.ijproman.2007.08.010
Leenders, R.T.A.J., J.M.L.V. Engelen and J. Kratzer,
2003. Virtuality, communication and new product
team creativity: A social network perspective. J.
Eng. Technol. Manag., 20: 69-92. DOI:
10.1016/S0923-4748(03)00005-5
REFERENCES
Ale, E., N. Ahme and Z. Taha, 2009. Innovation and
Technology Facilitator University of Tehran.
Anderson, A.H., R. Mcewan, J. Bal and J. Carletta,
2007. Virtual team meetings: An analysis of
communication and context. Comput. Human Behav.,
23: 2558-2580. DOI: 10.1016/j.chb.2007.01.001
Anderson, J.C. and D.W. Gerbing, 1988. Structural
equation modeling in practice: A review and
recommended two-step approach. Psycholo. Bull.,
103: 411-423. DOI: 10.1037/0033-2909.103.3.411
Baskerville, R. and J. Nandhakumar, 2007. Activating
and perpetuating virtual teams: Now that we’re
mobile, where do we go? IEEE Trans. Profess.
Commun.,
50:
17-34.
DOI:
10.1109/TPC.2006.890849
Byrne, B.M., 2001. Structural equation modeling with
AMOS: Basic Concepts, Applications and
Programming. 1st Edn., Routledge, USA., ISBN10: 0805841040 pp: 352.
Chen, M., Y. Liou, C.W. Wang, Y.W. Fan and Y.P.J. Chi,
2007. TeamSpirit: Design, implementation and
evaluation of a Web-based group decision support
system. Decision Support Syst., 43: 1186-1202.
DOI: 10.1016/j.dss.2005.07.008
Chudoba, K.M., E. Wynn, M. Lu, B.M. WatsonManheim, 2005. How virtual are we? Measuring
virtuality and understanding its impact in a global
organization. Inform. Syst. J., 15: 279-306. DOI:
10.1111/j.1365-2575.2005.00200.x
Coleman, D. and S. Levine, 2007. Collaboration 2.0 :
technology and best practices for successful
collaboration in a web 2.0 world. 1st Edn., Happy
About, California, USA, ISBN-10: 1600050719
pp: 320.
Cronbach, L.J., 1951. Coefficient alpha and the internal
structure of tests. Psychometrika, 16: 297-334.
DOI: 10.1007/BF02310555
13
Am. J. Engg. & Applied Sci., 5 (1): 9-14, 2012
Sills, S.J. and C. Song, 2002. Innovations in survey
research an application of web-based surveys. Soc.
Sci.
Comp.
Rev.,
20:
22-30.
DOI:
10.1177/089443930202000103
Stock, G.N. and M.V. Tatikonda, 2004. External
technology integration in product and process
development. Int. J. Oper. Production Manage., 24:
642-665. DOI: 10.1108/01443570410541975
Thissen, M.R., J.M. Page, M.C. Bharathi and T.L.
Austin, 2007. Communication tools for distributed
software development teams. Proceedings of the
2007 ACM SIGMIS CPR Conference on Computer
Personnel Research: The Global Information
Technology Workforce. New York, DOI:
10.1145/1235000.1235007
Vasileva, S.E., 2009. Authority-based and bottom-up
diffusion
of
collaboration
information
technologies: Constraints and enablements. mba
(mgmt
information
systems).
Universitat
Koblenz-Landau.
Walvoord, A.A.G., E.R. Redden, L.R. Elliott and M.D.
Coovert, 2008. Empowering followers in virtual
teams: Guiding principles from theory and practice.
Comput. Human Behav., 24: 1884-1906. DOI:
10.1016/j.chb.2008.02.006
Zemliansky, P. and K.S. Amant, 2008. Handbook of
Research on Virtual Workplaces and the New
Nature of Business Practices. 1st Edn., IGI
Global Snippet, Hershey PA, ISBN-10:
159904871X, pp: 733.
Lin, C., C. Standing and Y.C. Liu, 2008. A model to
develop effective virtual teams. Decision Support
Syst., 45: 1031-1045. DOI: 10.1016/j.dss.2008.04.002
Lurey, J.S. and M.S. Raisinghani, 2001. An empirical
study of best practices in virtual teams. Inform.
Manage., 38: 523-544. DOI: 10.1016/S03787206(01)00074-X
Mcnamara, K., A.R. Dennis and T.A. Carte, 2008. It's
the thought that counts: The mediating effects of
information processing in virtual team decision
making. Inform. Syst. Manage., 25: 20-32. DOI:
10.1080/10580530701777123
Nader, A.E., A. Shamsuddin and T. Zahari, 2010b.
Virtual R&D teams and SMEs growth: A
comparative study between Iranian and Malaysian
SMEs. Afr. J. Bus. Manage., 4: 2368-2379.
Nader, A.E., A. Shamsuddin, A. Rashid, S. Hanim and
T., Zahari, 2010a. The Effectiveness of Virtual
R&D Teams in SMEs: Experiences of Malaysian
SMEs. Proceedings of the 11th Asia Pacific
Industrial Engineering and Management Systems
Conference Dec. 9-9, Melaka, Malaysia. Kuala
Lumpur, Malaysia: University of Malaya Press, pp:
1-6.
Pauleen, D.J. and P. Yoong, 2001. Facilitating virtual
team relationships via Internet and conventional
communication channels. Int. Res., 11: 190-202.
DOI: 10.1108/10662240110396450
Pena-Mora, F., K. Hussein, S. Vadhavkar and K.
Benjamin, 2000. CAIRO: A concurrent engineering
meeting environment for virtual design teams.
Artif. Intell. Eng., 14: 203-219. DOI:
10.1016/S0954-1810(00)00016-9
Redoli, J., R. Mompo, J. Garcia-Diez and M. LopezCoronado, 2008. A model for the assessment and
development of Internet-based information and
communication services in small and medium
enterprises. Technovation, 28: 424-435. DOI:
10.1016/j.technovation.2008.02.008
14