Using Sentence-Selection Heuristics to Rank Text
Segments in TXTRACTOR
Daniel McDonald and Hsinchun Chen
Artificial Intelligence Lab
Management Information Systems Department
University of Arizona
Tucson, AZ 85721, USA
520-621-2748
{dmm, hchen}@eller.arizona.edu
Indicative text summarization systems support the user in
deciding which documents to view in their totality and which to
ignore. Some summarization techniques use measures of query
relevance to tailor the summary to a specific query [22] [5].
Providing tools for users to sift through query results can
potentially ease the burden of information overload.
Using document summaries can also potentially improve the
results of queries on digital libraries. Relevance feedback methods
usually select terms from entire documents in order to expand
queries. Lam-Adesina and Jones found query-expansion using
document summaries to be considerably more effective than
query-expansion using full-documents [13]. Other summarization
research explores the processing of summaries instead of full
documents in information retrieval tasks [18, 21]. Using
summaries instead of full documents in a digital library has the
potential to speed query processing and facilitate greater post
retrieval analysis, again potentially easing the burden of
information overload.
ABSTRACT
TXTRACTOR is a tool that uses established sentence-selection
heuristics to rank text segments, producing summaries that
contain a user-defined number of sentences. The purpose of
identifying text segments is to maximize topic diversity, which is
an adaptation of the Maximal Marginal Relevance criterion used
by Carbonell and Goldstein [5]. Sentence selection heuristics are
then used to rank the segments. We hypothesize that ranking text
segments via traditional sentence-selection heuristics produces a
balanced summary with more useful information than one
produced by using segmentation alone. The proposed summary is
created in a three-step process, which includes 1) sentence
evaluation 2) segment identification and 3) segment ranking. As
the required length of the summary changes, low-ranking
segments can then be dropped from (or higher ranking segments
added to) the summary. We compare the output of TXTRACTOR
to the output of a segmentation tool based on the TextTiling
algorithm to validate the approach.
1.2 Background
Categories and Subject Descriptors
Approaches to text summarization vary greatly. A distinction
frequently is made between summaries generated by text
extraction and those that generate text abstracts. Text extraction is
widely used [10], utilizing sentences from a document to create a
summary. Early examples of summarization techniques utilized
text extraction [16]. Text abstraction programs, on the other hand,
produce grammatical sentences that summarize a document’s
concepts. The concepts in an abstract are often thought of as
having been compressed. While the formation of an abstract may
better fit the idea of a summary, its creation involves greater
complexity and difficulty [10]. Producing abstracts usually
involves several stages such as topic fusion and text generation
that are not required for text extracts. Recent summarization
research has largely focused on text extraction with renewed
interest in sentence-selection summarization methods in particular
[17]. An extracted summary remains closer to the original
document, by using sentences from the text, thus limiting the bias
that might otherwise appear in a summary [16]. TXTRACTOR
continues this trend by utilizing text extraction methods to
produce summaries.
The goals of text summarizers can be categorized by their intent,
focus, and coverage [7]. Intent refers to the potential use of the
summary. Firmin and Chrzanowski divide a summary’s intent into
three main categories: indicative, informative, and evaluative.
Indicative summaries give an indication of the central topic of the
original text or enough information to judge the text’s relevancy.
I.2.7 Natural Language Processing - Language parsing and
understanding, Text analysis
General Terms:
Algorithms
Keywords
Text summarization, text segmentation, Information Retrieval,
text extraction
1. INTRODUCTION
1.1 Digital Libraries
Automatic text summarization offers potential benefits to the
operation and design of digital libraries. As digital libraries grow
in size, so does the user’s need for information filtering tools.
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies
are not made or distributed for profit or commercial advantage and that
copies bear this notice and the full citation on the first page. To copy
otherwise, or republish, to post on servers or to redistribute to lists,
requires prior specific permission and/or a fee.
JCDL’02, July 13-17, 2002, Portland, Oregon, USA.
Copyright 2002 ACM 1-58113-513-0/02/0007…$5.00.
28
Informative summaries can serve as substitutes for the full
documents and evaluative summaries express the point of view of
the author on a given topic. Focus refers to the summary’s scope,
whether generic or query-relevant. A generic summary is based on
the original text, while a query-relevant summary is based on a
topic selected by the user. Finally, coverage refers to the number
of documents that contribute to the summary, whether the
summary is based on a single document or multiple documents.
TXTRACTOR uses a text extraction approach to produce
summaries that are categorized as indicative, generic, and based
only on single documents.
conclusion” are more likely to appear in scientific literature than
in newspaper articles [10]. Position-based methods are also
domain-dependent. The first sentence in a paragraph contains the
topic sentence in some domains, whereas it is the last sentence
elsewhere. Combined with other techniques, however, these
extraction methods can still contribute to the quality of a
summary.
2.2 Document Segmentation
Document segmentation is an Information Retrieval (IR) approach
to summarization. Narrowing the scope from a collection of
documents, the IR approach views a single document as a
collection of words and phrases from which topic boundaries must
be identified [10]. Recent research in this field, particularly the
TextTiling algorithm [9], seems to show that a document’s topic
boundaries can be identified with a fair amount of success. Once a
document’s segments have been identified, sentences from within
the segments are typically extracted using word-based rules in
order to turn a document’s segments into a summary. Breaking a
document into segments identifies the document’s topic
boundaries. Segmentation is a nice way to make sure that a
document’s topics are adequately represented in a summary.
The IR approach to extraction does have some weaknesses.
Having a word-level focus “prevents researchers from employing
reasoning at the non-word level” [10]. While the IR technique
successfully segments single documents into topic areas [9], the
selection of sentences to extract from within those topic areas
could be improved by using many different heuristics, both wordbased and those that utilize language knowledge. In addition, once
a document is segmented, there is no way to know which of the
segments is the most salient to the overall document. Some
mechanism is required to rank segments so that the most pertinent
topic information either gets extra coverage in the summary, or is
covered first in the summary. A practical problem is also
addressed by ranking segments. When the required number of
sentences in a summary is less than the number of identified
segments, there must be an intelligent way to decide which
segments will not be covered. A possible solution perhaps is to
force a document to have a certain number of segments that match
the number of sentences allowed in the summary. Presetting the
number of acceptable topic areas, however, seems to defeat the
process of true segment identification. Segment boundaries would
seem arbitrary if there were a limit on their number. The process
of finding a document’s topic areas is separate from that of
selecting representative sentences to appear in the summary. A
ranking of the segments, however, would allow a summary to
grow and shrink, while extracting sentences from as many of the
highest ranked topic areas as possible. While this approach would
not be suited to an informative summary, ranking segments and
controlling the number of sentences in a summary are acceptable
for an indicative summary.
2. RELATED RESEARCH
TXTRACTOR is most strongly related to the research by
Carbonell and Goldstein [5] that strives to reduce the redundancy
of information in a query-focused summary. Carbonell and
Goldstein introduce the concept of Maximal Marginal Relevance
(MMR), where each sentence is ranked based on a combination of
a relevance and diversity measure. The consideration of diversity
in TXTRACTOR is achieved by segmenting a document using the
TextTiling algorithm [9]. Sentences coming from different text
segments are considered adequately diverse. All text segments
must be represented in a summary before additional sentences
from an already represented segment can be added. Nomoto [18]
and Radev [19] also present different ways to implement diversity
calculations for summary creation. Different from the
summarization work done by Carbonell and Goldstein, however,
TXRACTOR is not query-focused, but rather uses sentenceselection heuristics, instead of query relevance, to rank a
document’s sentences.
2.1 Sentence Selection
Much research has been done on techniques to identify sentences
that effectively summarize a document. Luhn in 1958 first utilized
word-frequency-based rules to identify sentences for summaries
[16]. Edmundson (1969) added three rules in addition to word
frequencies for selecting sentences to extract, including cue
phrases (e.g., “significant,” “impossible,” “hardly”), title and
heading words, and sentence location (words starting a paragraph
were more heavily weighted) [6]. The ideas behind these older
approaches are still referenced in modern text extraction research.
Sentence-selection methods assist in finding the salient sentences
in a document. By salient, we mean sentences a user would
include in a summary. There has been much review of sentenceselection methods in research. Teufel and Moens found the use of
cue phrases to be the best individual method [24]. Kupiec et al.,
on the other hand, found the position-based method to be the best
[12]. Regarding the combining of sentence-selection heuristics,
research conducted by Kupiec, Pedersen, and Chen found that the
best mix of extraction methods included position, cue phrase, and
sentence length. C. Aone et al. tested several different variations
of tf*idf and the using or suppressing of proper names in their
system DimSum [3]. Goldstein et al. found that summary
sentences had 90 percent more proper nouns per sentence [8].
When deciding which combination of extraction methods to use in
TXTRACTOR, we assume each method is independent and its
impact can be aggregated into the total sentence score. As we
conduct additional summarization experimentation, we will
further refine our use of sentence-selection methods and add
additional promising methods.
Despite the usefulness of sentence extraction methods in finding
salient sentences, they cannot alone produce the highest-quality
extracts. Sentence-selection techniques are often domain
dependent. For example, the words “Abstract” and “in
2.3 Combination Proposal
TXTRACTOR attempts to capture the benefits of sentenceselection summarization and document segmentation while
overcoming many of their deficiencies. The document
segmentation algorithm identifies the document’s main topic
areas, while sentence-selection heuristics identify the salient
sentences of a summary. The topic areas are used as the
foundation for the summary and the salient sentences are used as a
compass guiding the inclusion of certain topic areas. Document
segmentation provides a thorough domain-independent analysis of
the entire document, created in a bottom-up manner. Sentenceselection heuristics provide saliency information in a structured
29
top-down manner. In addition, we have included many sentenceselection techniques in order to reduce the domain-dependency
effect of any one heuristic. We hypothesize that ranking a
document’s segments on the basis of their containing one or many
of the document’s salient sentences will produce summaries that
are more information rich than those produced by the
segmentation-only approach.
sentence. Thus, a high tf*idf score for a sentence is normalized for
sentence length. The resulting score is then added to the
sentence’s score.
3.1.4 Sentence position in a paragraph
As the sentences are extracted from the original document, new
lines and carriage returns signal the beginning of new paragraphs.
The beginning sentence of a document and the beginning sentence
in a paragraph are given additional weight due to their greater
summarizing potential.
3. TXTRACTOR IMPLEMENTATION
TXTRACTOR is a summarizer based on text extraction written in
Java. Its major components include sentence-selection rules and a
segmentation algorithm. The summarization process takes place in
three main steps 1) sentence evaluation 2) segmentation or topic
boundary identification and 3) segment ranking and extraction.
3.1.5 The sentence length
The length of a sentence can provide clues as to its usefulness in a
summary [12] [3]. Before adding sentence length summarization
heuristic, we tried to achieve the same effect by simply not
averaging tf*idf scores over the number of words in the sentence.
Longer sentences would naturally be scored higher because they
would contain more non-stop word terms. This approach overly
weighted long sentences to the point where scores from the tf*idf
equation would overpower the scores in other areas. Normalizing
that score would mute the value of a concentrated sentence with
many document-wide terms. To solve this problem, we made
sentence length its own rule. The length of a sentence is calculated
and its impact added to the sentence’s weight.
Because each of the five sentence-selection rules is calculated
differently, each score had to be normalized so the impact of each
rule would be comparable. For example, the impact of an extra
proper noun in a sentence is not the same as that of a sentence
occurring first in a paragraph or of a sentence being very long.
The current normalization factor for each heuristic was
determined through experimentation.
TXTRACTOR has a configuration option that allows the user to
adjust the impact of each sentence-selection heuristic, without
having to recompile the program. Because each extraction
heuristic was normalized, a user can change the weighting of a
particular heuristic and immediately judge its impact on the
summary. Including the configuration capability facilitates
experimentation with different heuristic weights. In addition,
while the summary-generation logic of TXTRACTOR was
designed to be reasonably domain-independent, a user can still
change the weighting given to different sentence-selection rules
through the configuration option, thus customizing the
summarizer to different domains and uses. For example, a user
may want to see as many proper nouns as possible in the
summary. Increasing the weight of the proper nouns rule will
cause sentences with proper nouns to move to the top of the
sentence ranking and thus appear more often in the summary.
Once each sentence is scored based on the above five heuristics,
the sentences are then ranked according to their summarizing
value. Unlike segmentation-only approaches, sentences are ranked
against other sentences from the entire document, not only those
sentences within the same topic area. An example of sentence
scoring is shown in Figure 1. The three sentences listed are the top
three sentences extracted from a document entitled “May the
Source be With You” from Wired Magazine [15]. A nearly
complete copy of the article is found in Figure 4. Each sentence in
Figure 1 comes from a different topic segment. The highestscoring sentence greatly benefits from being the first sentence in
the article (+20). This position heuristic has been shown to be the
most effective of all sentence-selection heuristics [12]. All three
sentences begin a paragraph (+10). The second sentence contains
the cue phase “thus”, adding 10 points. The cue phrase allows it to
outrank the third sentence despite the third sentence having the
highest values of the three for tf*idf and sentence length.
3.1 Sentence Evaluation:
The summarization process begins by parsing the sentences of the
original text using a program that recognizes 60 abbreviations and
various punctuation anomalies. The original orders of the
sentences are preserved so they can be added to the summary in
the intended order. Once the sentences are identified,
TXTRACTOR begins ranking each sentence. We use five
sentence-selection heuristics to evaluate the document’s
sentences. The following different ranking methods have an
impact on the corresponding scores of each of the sentences.
3.1.1 Presence of cue phrases
Currently, each sentence is checked for the existence of ten
different cue phrases (e.g. “in summary,” “in conclusion,” “in
short,” “therefore”). Cue phrases are words that signal to a reader
that the author is going to summarize his or her idea or subject.
The cue phrases are loaded out of a text file so that additional
words can be easily added as more experimentation is done in this
area.
3.1.2 Proper nouns in the sentence
A TXTRACTOR-generated summary is meant to provide enough
information for a user to be able to decide whether she or he
wants to read the original document in its entirety. Important to
this decision is the existence of certain proper names and places.
Currently, TXTRACTOR simply reads each sentence and counts
the capitalized words, not including the opening word in the
sentence. This is meant as a temporary implementation until a full
entity-extraction algorithm can be implemented. The total number
of capitalized words in each sentence is then averaged over the
number of words in the sentence. Shorter sentences are thus not
penalized for having fewer proper nouns than longer sentences.
The average number of proper nouns is then normalized and
added to the sentence’s score.
3.1.3 TF*IDF
Tf*idf measures how frequently the words in a sentence occur
relative to their occurrence in the entire document. Sentences that
have document words in common are scored higher. To calculate
tf*idf, the occurrence of every word in a sentence and the word’s
total occurrences in the document are totaled. Before the terms are
totaled, however, each word is made lower-case and stemmed
using the Porter stemmer. The Porter stemmer is one of the most
widely used stemming algorithms [11] and can be thought of as a
lexicon-free stemmer because it uses cascaded rewrite rules that
can be run very quickly and do not require the use of a lexicon.
Stemming is performed so that words with the same stem but
different affixes may be treated as the same word when
calculating the frequency of a particular term. The tf*idf
calculation is computed and then averaged over the length of the
30
(topic: 0) (sentence 0) (score: 95) “The laws protecting software code are stifling creativity,
destroying knowledge, and betraying the public trust.” (First document sentence: +30, first sentence of
paragraph + 20, 0 for proper nouns, +34 for tf*idf, +11 for sentence length = score of 95)
(topic: 5) (sentence 52) (score: 85) “Thus, I would dramatically reduce the safeguards for software from the ordinary term of 95 years to an initial term of 5 years, renewable once.” (First sentence of
paragraph + 20, +10 for cue phrase “thus,” +0 for proper nouns, +41 for tf*idf, +14 for sentence
length = score of 85)
(topic: 1) (sentence 23) (score: 85) “Finally, while control is needed, and perfectly warranted, our bias
should be clear up front: Monopolies are not justified by theory; they should be permitted only when
justified by facts.” (First sentence of paragraph + 20, +1 for proper nouns, +45 for tf*idf, +18 for
sentence length = score of 85)
Figure 1- the weighting of individual sentences
example of the segment ranking is shown in Figure 2. Highranking sentences are added first to the summary. Two sentences
from the same segment are not included in the summary
(regardless of their ranking) until a sentence from each segment
has been included. Once all segments are represented in the
summary, then the process starts over adding one sentence from
all segments. Remaining ranked-sentences are added by segment
3.2 Segmentation:
The segmentation algorithm used is based on the TextTiling
algorithm developed by Marti Hearst [9]. The TextTiling
algorithm analyzes a document and determines where the topic
boundaries are located. A topic boundary can be thought of as the
point at which the author of the document changes subjects or
themes. The first step in the TextTiling algorithm is to divide the
text into token-sequences, removing any words that appear on the
stop list. We have used a token-sequence length of 20 and the
same stop word list used by Marti Hearst in TextTiling. Tokensequences are then combined to form blocks. Blocks are
compared using a similarity algorithm. The comparison between
blocks functions like a sliding window. The first block contains
the first token-sequence plus k token-sequences before it. The
second block contains the second token-sequence and the k tokensequences after it. The value for k used in our summarizer is 10,
also the same one used by Marti Hearst in TextTiling. The blocks
are then compared using an algorithm that returns the similarity as
a percentage, which is derived from the number of times the same
terms appear in the two blocks being compared. The Jaccard
coefficient is used for the similarity equation, which differs
slightly from the normalized inner product equation used by
Hearst. We did not consider the impact of using different
similarity equations to be significant. The Jaccard coefficient is as
follows:
L
∑ (w
ik
Si, j =
L
k =1
L
∑w +∑w
k =1
2
ik
k =1
2
jk
2. Segment Ranking
1. Text Segmentation
with ranked sentences
1.
2.
5.
15.
18.
Sentence Ordering
A two-sentence
summary would
include sentences
ranked 1 & 3
A
five-sentence
summary
would
include
sentences
ranked 1, 2, 15, 3, &
4
3.
w jk )
4.
L
8.
− ∑ wik w jk
k =1
Figure 2- text segmentation and sentence-selection combined
Si,j is the similarity between the two blocks of grouped tokensequences i and j, wik is the number of occurrences of term k in
the block, and L is the total number of blocks.
Once the topic boundaries have been identified, TXTRACTOR
then assigns each sentence to a document segment. After all
sentences have been given weights and assigned to segments, then
segment ranking and sentence extraction can operate.
until the summary-length requirement is met. Once the length
requirement is met, the sentences are then sorted by the order in
which they appeared in the original document and displayed on
the screen. Figure 3 shows some pseudo code for the segment
ranking routine. Document segmentation, therefore, provides the
topic structure for a document within which sentence selection
can be utilized to identify the salient topic areas. It is of practical
advantage to rank segments so that a user can easily change the
desired length of the summary while the ranking routine identifies
3.3 Segment Ranking:
Once a document is segmented into its main topic areas,
TXTRACTOR ranks the document segments based on the scores
given to sentences by the sentence-selection heuristics. An
31
4.3 An Example Document
Rank segments (Array of ranked sentences)
while(summary length not achieved)
for( each ranked sentence in array )
if(sentence segment not already used)
if(summary length achieved)
break;
add sentence to summary
end if
else
add to temp array for recursive call
end else
end for
Rank segments( temp array )
end while
end Rank segments
for( all summary sentences)
rank sentences by original document order
end for
While not included in the summaries evaluated in the user studies,
the article in Figure 4 is a good example of the differences
between the TXTRACTOR (reference by “TXT#” in the figure)
summaries and the summaries generated by the segmentation-only
approach (referenced by “SEG#” in the figure) [15]. Large
asterisks and segment numbers highlight the breaks in the
document segments. The first sentence selected by TXTRACTOR
is the first sentence in the document, despite the fact that it had a
10-point lower tf*idf score than the first segmentation sentence.
The first sentence, however, is a very good summarizing sentence.
The segmentation approach then selects a second sentence from
the first topic area. The two summaries then select the same
sentence from segment two to add to their summaries. Later,
TXTRACTOR skips over the third topic area, while the
segmentation algorithm adds its final two sentences from that
topic area. A segmentation summary tries to include sentences
from every segment. TXTRACTOR had ranked the two sentences
added in the segmentation summary as 50th and 62nd respectively.
The sentences had low scores for sentence length and somewhat
low scores for tf*idf. Sentences three and four for the
TXTRACTOR summary are scored highly due to the included cue
phrase, “thus”. The final sentence selected by TXTRACTOR is
not rated in the top five best sentences (it is sixth), but because
two sentences in the top-five come from the first topic area, room
in the summary is preserved for a segment not already
represented. Thus, the ranking routine includes a sentence from
the seventh segment in the summary, instead of duplicating
sentences in a segment.
Figure 3 – Pseudo code for segment ranking
which segments, represented by their sentences, to add to or drop
from the summary.
4. PRELIMINARY TESTING
As a preliminary test of the performance of the TXTRACTOR
summarizer, subjects compared summaries produced by
segmentation alone with summaries produced by TXTRACTOR.
A length limit of five sentences was imposed on all the
summaries.
4.4 Document Selection
Five documents were deliberately selected from various different
subject domains. Document subjects ranged from psychology and
sports to arts and science. The subjects of the documents were
varied in order to see whether the TXTRACTOR approach had
limitations in certain subject domains. Effort was also made to
vary the length of the documents. The numbers of words in the
documents ranged from 537 up to 13,293. Different lengths of
documents were selected so that varied numbers of segments
would be created. By including long articles, we hoped to get
preliminary clues as to which summarizer prioritized a
document’s segments best and whether prioritizing segments led
to improved summaries. In this experiment, we did not ask the
subjects to judge the cohesiveness of the summaries. We tried to
focus the user on the information content of the summary, not its
cohesiveness.
We selected the following five documents to be summarized:
1. Turning Snooping Into Art, by Noah Shachtman, 773
words [23]
2. No. 4 Virginia suffers first loss of season, Game Day
Recap, 537 words [20]
3. Nanotech Fine Tuning by Mark K. Anderson, 654
words [2]
4. Ann Landers by Ann Landers, 650 words [14]
5. A Primer on Narcissism by Sam Vaknin, Ph.D, 13,293
words [25]
4.1 Segmented Summaries
The summaries produced by the segmentation-only approach used
the same segmenting code as that used in TXTRACTOR. After
segments were identified, every word in each segment, except for
those on the stop list, was scored based on the tf*idf equation. The
two highest-ranking terms from the segment were identified along
with the first occurrence of each of the terms in the segment. The
sentence(s) where the first occurrences were identified was then
added to the summary. Each segment produced one or two
sentences, depending on whether one sentence contained the top
two keywords or not. The same procedure for every segment was
carried out until there was at least one sentence from every
segment in the summary. In cases where there were more than
five sentences, only the first five sentences were included in the
summary for comparison purposes.
4.2 TXTRACTOR Configuration
The configuration settings of the document-segmenting algorithm
were kept constant between the TXTRACTOR system and the
segmentation-only system. Token-sequences of 20 characters
were used and 10 token-sequences were added together to form a
block for the similarity comparisons. Blocks were allowed to
cross sentence boundaries and no stemming or noun phrasing was
applied in identifying the document segments. Stemming,
however, was used in calculating tf*idf in the sentence-selection
portion of TXTRACTOR. The segmenting code was allowed to
determine how many topic areas the document had instead of
being forced to generate the boundaries for a predetermined
number of topics.
4.5 Experiment Participants
Five subjects were chosen to compare the TXTRACTOR
summary with the summary generated by the segmentation–only
approach. All subjects were above the age of 20 and were either
completing or had already obtained a bachelor’s degree. The
subjects were emailed the ten summaries, grouped by original
document. Participants were directed to choose the summary that
32
1**[ The laws protecting software code are stifling creativity, destroying knowledge, and betraying the public trust.
TXT1
Legal heavy Lawrence Lessig argues it's time to bust the copyright monopoly.
In the early 1970s, RCA was experimenting with a new technology for distributing film on magnetic tape - what we would come to call video.
SEG1
Researchers were keen not only to find a means for reproducing celluloid with high fidelity but also to discover a way to control the use of the
technology. Their aim was a method that could restrict the use of a film distributed on video, allowing the studio to maximize the film's return from
distribution.
The technology eventually chosen was relatively simple. A video would play once, and when finished, the cassette would lock into place. If a customer
wanted to play the tape again, she would have to return it to the video store and have it unlocked.….
They were horrified. They would "never," Feely reported, permit their content to be distributed in that form, because the content - however clever the
self-locking tape was - was still insufficiently controlled. How could they know, one of the Disney execs asked Feely, "how many people were going to
be sitting there watching" a film? What's to stop someone else from coming in and watching for free?
SEG2
We live in a world with "free" content, and this freedom is not an imperfection. We listen to the radio without paying for the songs we hear;
we hear friends humming tunes that they have not licensed. We tell jokes that reference movie plots without the permission of the directors. We read
our children books, borrowed from a library, without paying the original copyright holder for the performance rights. The fact that content at a particular
time may be free tells us nothing about whether using that content is theft. Similarly, in arguing for increasing content owners' control over content
users, it's not sufficient to say "They didn't pay for this use."
Second, the reason perfect control has not been our tradition's aim is that creation always involves building upon something else. There is no art that
]**2**[
Monopoly controls have been the exception in free
doesn't reuse. And there will be less art if every reuse is taxed by the appropriator.
societies; they have been the rule in closed societies.
Finally, while control is needed, and perfectly warranted, our bias should be clear up front: Monopolies are not justified by theory;
TXT2
SEG3
they should be permitted only when justified by facts. If there is no solid basis for extending a certain monopoly protection, then we should not extend
that protection. This does not mean that every copyright must prove its value initially. That would be a far too cumbersome system of control. But it
does mean that every system or category of copyright or patent should prove its worth. Before the monopoly should be permitted, there must be reason
to believe it will do some good - for society, and not just for monopoly holders.
]**3**[
One example of this expansion of control is in the realm of software.
Like authors and publishers, coders (or more likely, the
SEG4
companies they work for) enjoy decades of copyright protection. Yet the public gets very little in return. The current term of protection for software is
the life of an author plus 70 years, or, if it's work-for-hire, a total of 95 years. This is a bastardization of the Constitution's requirement that copyright be
for "limited times." By the time Apple's Macintosh operating system finally falls into the public domain, there will be no machine that could possibly
run it. The term of copyright for software is effectively unlimited.
Worse, the copyright system safeguards software without creating any new knowledge in return. When the system protects Hemingway, we at
SEG5
]**4**[
least get to see how Hemingway writes.
We get to learn about his style and the tricks he uses to make his work succeed. We can see this
because it is the nature of creative writing that the writing is public. There is no such thing as language that conveys meaning while not simultaneously
transmitting its words. Software is different: Software gets compiled, and the compiled code is essentially unreadable; but in order to copyright
TXT3
software, the author need not reveal the source code. Thus, while the English department gets to analyze Virginia Woolf's novels to train
its students in better writing, the computer science department doesn't get to examine Apple's operating system to train its students in better coding.
]**5**[
The harm that comes from this system of protecting creativity is greater than the loss experienced by computer science education.
While
the creative works from the 16th century can still be accessed and used by others, the data in some software programs from the 1990s is already
inaccessible. Once a company that produces a certain product goes out of business, it has no simple way to uncover how its product encoded data. The
code is thus lost, and the software is inaccessible. Knowledge has been destroyed.
Copyright law doesn't require the release of source code because it is believed that software would become unprotectable. The open source movement
might throw that view into doubt, but even if one believes it, the remedy (no source code) is worse than the disease. There are plenty of ways for
software to be secured without the safeguards of law. Copy-protection systems, for example, give the copyright holder plenty of control over how and
when the software is copied.
]**6**[
And one
If society is to give software producers more protection than they would otherwise take, then we should get something in return.
thing we could get would be access to the source code after the copyright expires.
TXT4
Thus, I would dramatically reduce the safeguards for software - from the ordinary term of 95 years to an initial term of 5 years, renewable
once. And I would extend that government-backed protection only if the author submitted a duplicate of the source code to be held in escrow while the
work was protected. Once the copyright expired, that escrowed version would be publicly available from the copyright office.
Most programmers should like this change. No code lives for 10 years, and getting access to the source code of even orphaned software projects would
]**7**[
Software
benefit all. More important, it would unlock the knowledge built into this protected code for others to build upon as they see fit.
would thus be like every other creative work - open for others to see and to learn from.
There are other ways that the government could help free up resources for innovation. …
One context in particular where this could do some good is in orphaned software. Companies often decide that the costs of developing or maintaining
software outweigh the benefits. They therefore "orphan" the software by neither selling it nor supporting it. They have little reason, however, to make
the software's source code available to others. The code simply disappears, and the products become useless.
Software gets 95 years of copyright protection. By the time the Mac OS finally falls into the public domain, no machine will be able to run it.
TXT5
But if Congress created an incentive for these companies to donate their code to a conservancy, then others could build on the earlier work
and produce updated or altered versions. This in turn could improve the software available by preserving the knowledge that was built into the original
code. Orphans could be adopted by others who saw their special benefit.
]**8**[
The problems with software are just examples of the problems found generally with creativity.
Our trend in copyright law has been to
enclose as much as we can; the consequence of this enclosure is a stifling of creativity and innovation. If the Internet teaches us anything, it is that great
value comes from leaving core resources in a commons, where they're free for people to build upon as they see fit. An Innovation Commons was the
essence - the core - of the Internet. We are now corrupting this core, and this corruption will in turn destroy the opportunity for creativity that the
Internet built.
]**
Figure 4 - original text showing topics and sentences extracted via TXTRACTOR and a segmentation-only approach
33
provided the most pertinent information and seemed to be the
most useful in general.
7. REFERENCES
[1]
4.6 Results
[2]
Despite the small size of the experiment, we were able to observe
some encouraging responses. Of the 25 comparisons that were
made between summaries, the TXTRACTOR summary was
preferred 14 times, the segmentation-only summary was preferred
8 times, and 3 times the summaries were judged to be more or less
the same. The subjects therefore preferred the TXTRACTOR
summaries 7:4 over the summaries generated by segmentation
only. After submitting their responses, the subjects were told
which summarizer produced which summary. The participants
then volunteered explanations for why they choose the summary
they did. A common sentiment was that the TXTRACTOR
summary contained more information, but the sentences
sometimes did not flow well. When the sentences flowed well, as
judged by the participants, the TXTRACTOR-produced summary
was usually preferred. An interesting note is that even though we
had not instructed the subjects to assess the readability of the
summary, users did not ignore the summary’s cohesiveness. It
seems, even with indicative summaries,
that poor readability can distract a subject from information
content.
[3]
[4]
[5]
[6]
[7]
5. CONCLUSION & FUTURE DIRECTION
Based on our tests, the TXTRACTOR summarizer outperformed
the summarizer based solely on segmentation. The hypothesis that
ranking segments through the use of established sentenceselection heuristics leads to better text-extracted summaries
appears to be promising. There is much that can be done,
however, to improve the performance of the summarizer. Future
improvements to TXTRACTOR include implementing the local
salience method of cohesion analysis [4]. The local salience
method is based on the assumption that relevant words and
phrases are revealed by a “combination of grammatical, syntactic,
and contextual parameters”. The original document is parsed to
identify a sentence’s subjects and predicates. Different weights
are then given to sentences based on the part-of-speech containing
the term being analyzed. Experimentation will be conducted on
how many parts-of-speech to parse out of each sentence.
Additional research is needed to tune the weights of the sentenceselection methods being used. Much research that has been done
in this area could be incorporated into our work. In addition,
analyzing the discourse context of the sentences should help
improve the cohesiveness of the summaries.
We are currently planning on conducting more complete
experiments and user studies on our combined segmentation and
sentence-selection approach to summarization. We are looking to
test our summarization approach on a larger scale, similar to that
done at the May 1998 SUMMAC conference [1]. Finally, we
would like to implement and test TXTRACTOR in different
digital library domains such as medical libraries and web pages.
[8]
[9]
[10]
[11]
[12]
[13]
[14]
6. ACKNOWLEDGMENTS
We would like to express our gratitude to NSF Digital Library
Initiative-2, “High-performance Digital Library Systems: From
Information Retrieval to Knowledge Management,” IIS-9817473,
April 1999 – March 2002. We also would like to thank William
Oliver for his implementation of the TextTiling algorithm and
Karina McDonald for her feedback on the summaries.
[15]
[16]
34
in TIPSTER Text Phase III 18-Month Workshop, (Fairfax,
VA, 1998).
Anderson, M.K. Nanotech Fine Tuning
http://www.wired.com/news/technology/0,1282,494472,00.html.
Aone, C., Okurowski, M.E., Gorlinsky, J. and Larsen, B. A
Trainable Summarizer with Knowledge Acquired from
Robust NLP Techniques. in Maybury, M.T. ed. Advances in
Automatic Text Summarization, The MIT Press, Cambridge,
1999, 71-80.
Boguraev, B. and Kennedy, C., Salience-based Content
Characterization of Text Documents. in Proceedings of the
Workshop on Intelligent Scalable Text Summarization at the
ACL/EACL Conference, (Madrid, Spain, 1997), 2-9.
Carbonell, J. and Goldstein, J., The Use of MMR,
Diversity-Based Reranking for Reordering Documents and
Producing Summaries. in SIGIR, (Melbourne, Austrailia,
1998), 335-336.
Edmundson, H.P. New Methods in Automatic Extracting. in
Maybury, M.T. ed. Advances in Automatic Text
Summarization, The MIT Press, Cambridge, 1969, 23-42.
Firmin, T. and Chrzanowski, M.J. An Evaluation of
Automatic Text Summarization Systems. in Maybury, M.T.
ed. Advances in Automatic Text Summarization, The MIT
Press, Cambridge, 1999.
Goldstein, J., Kantrowitz, M., Mittal, V. and Carbonell, J.,
Summarizing Text Documents: Sentence Selection and
Evaluation Metrics. in 22nd International Conference on
Research and Development in Information Retrieval,
(1999).
Hearst, M.A. Segmenting Text into Multi-Paragraph
Subtopic Passages. Computational Linguistics (23(1)). 3364.
Hovy, E. and Lin, C.-Y. Automated Text Summarization in
SUMMARIST. in Maybury, M.T. ed. Advances in
Automatic Text Summarization, The MIT Press, Cambridge,
1999, 81-94.
Jurafsky, D. and Martin, J.H. Speech and Language
Processing: An Introduction to Natural Language
Processing, Computational Linguistics, and Speech
Recognition. Prentice Hall, Upper Saddle River, 2000.
Kupiec, J., Pedersen, J. and Chen, F., A Trainable
Document Summarizer. in Proceedings of the 18th ACMSIGIR Conference, (1995), 68-73.
Lam-Adesina, A.M. and Jones, G.J.F., Applying
Summarization Techniques for Term Selection in
Relevance Feedback. in SIGIR, (new Orleans, Louisiana,
USA, 2001), 1-9.
Landers, A. Ann Landers
http://www.washingtonpost.com/wp-dyn/articles/A628232002Jan4.html.
Lessig, L. May the Source Be With You. Wired Magazine,
9.12 (December).
http://www.wired.com/wired/archive/9.12/lessig.html.
Luhn, H.P. The Automatic Creation of Literature Abstracts.
in Maybury, M.T. ed. Advances in Automatic Text
Summarization, The MIT Press, Cambridge, 1958, 15-22.
[17] Mani, I. and Maybury, M.T. (eds.). Advances in Automatic
Text Summarization. The MIT Press, Cambridge, 1999.
[18] Nomoto, T. and Matsumoto, Y., A New Approach to
Unsupervised Text Summarization. in SIGIR, (New
Orleans, LA, USA, 2001), 26-34.
[19] Radev, D.R., Jing, H. and Budzikowska, M., Centroidbased summarization of mulitiple documents: sentence
extraction, utility-based evaluation, and user studies. in
ACL/NAAL Workshop on Summarization, (Seatle, WA.,
2000).
[20] Recap. No. 4 Virginia suffers first loss of season
http://sports.espn.go.com/ncaa/mbasketball/recap?gameId
=220050189, 2002.
[21] Sakai, T. and Jones, K.S., Generic Summaries for Indexing
in Information Retrieval. in SIGIR, (New Orleans,
Louisiana, USA, 2001), 190-198.
[22] Sanderson, M., Accurate user directed summarization from
existing tools. in Conference on Information and
Knowledge Management, (Bethesda, MD, USA, 1998), 4551.
[23] Shachtman, N. Turning Snooping Into Art
http://www.wired.com/news/culture/0,1284,49439,00.html.
[24] Teufel, S. and Moens, M., Sentence Extraction as a
Classification Task. in Workshop on Intelligent Scalable
Summarization ACL/EACL Conference, (Madrid, Spain,
1999), 58-65.
[25] Vaknin, S. A Primer on Narcissism
http://www.mentalhelp.net/poc/view_doc.php/type/doc/id/41
9.
35
© Copyright 2026 Paperzz