Integrated text and line-art extraction from a topographic map

IJDAR (2000) 2: 177–185
Integrated text and line-art extraction from a topographic map
Luyang Li1 , George Nagy2 , Ashok Samal3 , Sharad Seth3 , Yihong Xu4
1
2
3
4
Panasonic Information and Networking Laboratories, Princeton, NY, USA
Rensselaer Polytechnic Institute, Troy, NY, USA
University of Nebraska – Lincoln, Department of Computer Science, Lincoln, NE 68588, USA
Hewlett-Packard Laboratories, Palo Alto, CA, USA
Received January 1, 2000 / Revised January 21, 2000
Abstract. Our proposed approach to text and line-art
extraction requires accurately locating a text-string box
and identifying external line vectors incident on the box.
The results of extrapolating these vectors inside the box
are passed to an experimental single-font optical character reader (OCR) program, specifically trained for the
font used for street labels. In the first evaluation experiment, automated techniques are used to identify the
boxes and the line vectors. In the second, more comprehensive, experiment an operator marks these using a
graphical user interface. OCR results on 544 instances
of overlapped street-name boxes show the following improvements due to the integrated processing: the error
rate is reduced from 4.1% to 2.0% for characters and
from 11.8% to 6.4% for words.
Key words: Graphics recognition – Map processing –
Text-graphics segmentation – Geographic information
systems – Character recognition – Cooperative processing – Interactive processing – Adaptation – Form processing
1 Introduction
Both government and business organizations must frequently convert existing paper maps of cities and larger
regions into a computer-readable form that can be interfaced with existing geographical information systems
(GIS). Often, the task involves a batch of maps of the
same kind, but each covering a different geographic area.
Examples might be a set of topographic maps or road
maps of a foreign country issued by the same agency.
Until recently, coordinate digitizing tables were used
for such conversion. Currently, the most efficient method
is to raster-scan the maps, display them in segments on
a computer screen, and digitize them using interactive
CAD (e.g., Autocad) or GIS (e.g., ARC/INFO) tools.
Parts of the process, such as tracing a single curve, might
Correspondence to: S. Seth
be automated. Completely automatic conversion, while
still the objective of many academic research projects,
remains elusive. However, when a large amount of similar
data needs to be converted at the same time, it is highly
desirable to reduce the amount of manual intervention
required as the task progresses.
In 1997, the United States National Imagery and
Mapping Agency (NIMA) sponsored a series of research
activities as part of the Intelligent Map Understanding
Project. The project was designed to fill both in-house
and commercial needs. The University of NebraskaLincoln and Rensselaer Polytechnic Institute jointly participated in these activities. Our primary contribution in
this work was to identify the bottlenecks in large-volume
conversions using a general-purpose system and develop
a framework under which the operator time for these
tasks could be progressively reduced.
Our specific task was to develop methods to extract
the street network and street labels from the bitmap representation of a United States Geological Survey (USGS)
topographic map, and convert them to a GIS format similar to that of the TIGER database developed by the Bureau of Census [USD97]. Here the streets are polylines,
and the associated street names are in ASCII. (We did
not attempt to duplicate the TIGER city-block structure.) The main research objective was to demonstrate
that iterated processing, feedback from downstream processes, and judicious use of operator interaction improve
the results.
In the USGS map, street lines and names are all
printed in black. Thus separating the black layer with
a color filter provides a starting point for processing the
image. However, major challenges remain even with this
simple and effective step of information reduction. The
text and line layers frequently overlap; text strings appear in many orientations; the line layer contains not
only streets but also grid lines and political boundaries.
Similarly, the black text layer contains not just the street
names but also elevation values and names of neighborhoods, political districts, schools, hospitals, other builtup areas, physiographic and topographic features. These
challenges are well illustrated in the black layer of a 3-
178
L. Li et al.: Integrated text and line-art extraction from a topographic map
Fig. 1. The black layer of a section of USGS map
inch square chip from the Washington D.C. East map in
Fig. 1.
Because map conversion is a niche market, with government agencies serving as the primary customer base,
it has not received as much attention in the commercial
world as the conversion of forms and legal documents.
However, with the spread of the Web and GPS technologies, navigational databases are finding use in invehicle navigation, intelligent traffic and transportation
systems, and such online applications as routing and city
guides. Street layer extraction is the first step towards
building detailed navigational databases. As such, any
significant improvement in this process will help alleviate the sparse coverage of the existing databases. For
example, in the Navtech database [Nav], detailed area
maps are currently available for only four cities in the
entire states of Iowa, Kansas, Missouri, Nebraska, North
Dakota, and South Dakota.
Preliminary results were presented at a session organized by NIMA at the Conference on Geographic and
Land Information Systems (GIS/LIS) [NSS+ 97], at the
Workshop on Graphics Recognition of the International
Association of Pattern Recognition [NSS+ 98], and at
the International Conference on Document Analysis and
Recognition (ICDAR) [LNS+ 99]. The results also appear in four theses [Kal97,Li98,Siv98,Xu98]. This paper presents new results obtained since the conclusion of
the NIMA project but it draws on methods discussed in
detail in these references.
Briefly, in our past work we accomplished the following steps on a single map. We isolated the black layer
of the map, which contains both the street lines and the
street labels, by appropriate thresholding of the RGB
color components viewed as an HSV (hue-saturationvalue) representation. We separated the 18, 848 × 23, 631
pixel black layer into 32,000 connected components (this
took about 4 min ‘system time’ on a Sun 4M SPARCstation). Then we carried out a preliminary classification
of each component, according to geometric features and
black pixel density, into line art, characters, and icons.
The line art was vectorized using ArcScan [ESRI97],
and refined (‘beautified’ [PW85]) using line-width, linelength, and line-pairing criteria, and finally the constraint
was applied that the street casings must form city blocks.
The double-line street casings were then reduced to a
street centerline representation. The characters were
grouped into ‘word boxes’, rotated to horizontal, and
submitted to an experimental OCR program that did not
require character-level separation and which was specifically trained for the font used for street labels. (We also
tried a commercial omnifont system but found that it
produced extremely poor results on the map text.)
The labels were analyzed according to the direction
and spacing of nearby street lines. The specific and
generic components of the labels (like ‘George Washington’ and ‘Ave’) were assembled, and the street lines
were traced as far as possible and associated with the
complete (and occasionally repeated) labels.
Other important parts of the project were the implementation on top of ARC/INFO of a display of partial
or complete results to allow interactive correction of errors, of a logging system that tracks the time spent by
the operator on various tasks, and of a cost model of
the complete process for estimating the conversion time
required for new tasks in terms of map and operatorrelated parameters.
The novel aspect documented here is the following.
We demonstrate the effect of cooperative processing of
text and line art. This is necessary because the text and
line art often overlap and recognition results on one can
help recognition of the other. We use information from
the street lines to locate and orient label boxes and to
remove overlaying line segments, and use the output of
the character recognition system to refine the street-line
sublayer. We show that with cooperative processing the
error rate of our word-based OCR algorithm can be reduced by a factor of two on noisy samples.
We briefly review relevant sources of information.
The definitive authority on text-graphics separation is
Kasturi. His work with his students stretches over a
decade [FK88, KA88, TK98]. It usually gives much more
emphasis to image processing than to OCR. A good
source of other techniques devised in the last 20 years
for isolating text from illustrations in printed documents,
and for line-text separation in engineering drawings, is
[OK95]. Brief reviews appear in [NSS+ 97] and
[Li98].
There are many, many papers on map conversion in
general, with a particularly valuable early paper by a
practitioner, Rhind [Rhi74]. Good examples of recent
projects are [StK95, Har95], while surveys of the required
tools appear in [EAK95, JV97]. Since we use commercial
software, we don’t review the even larger field of thinning
and vectorization, but a recent survey is [Doe98].
There is not a mass of literature on associating streets
with street labels on scanned maps. The most relevant
paper we found is by G. Myers and his colleagues at SRI
L. Li et al.: Integrated text and line-art extraction from a topographic map
Fig. 2. A street label and street casings from a USGS map
(enlarged)
[MMC+ 96]. They propose a verification-based approach,
but compared to ours the line art and the text are processed essentially separately.
Finally, we note that the problem of separating text
from lines arises in forms processing, which is a major
industry. Although most of the successful techniques are
proprietary, some recent results appear in [NSY96].
2 Data
The source map for all of the following experiments was
WASHINGTON EAST (D.C.) QUADRANGLE, based
on aerial photography taken in 1955. It was revised in
1965, photorevised (as shown in purple) and edited, but
not field checked, in 1979. The map has a scale of
1:24,000 and elevation contour intervals of 10 feet. It is a
polyconic projection based on the 1927 North American
Datum. It shows the Maryland and Virginia state coordinate systems as well as the 1000-m Universal Transverse
Mercator grid.
The size of the map (46 cm × 58 cm) dictates special
processing considerations. The map was color-digitized
by NIMA at 1000 dpi ( 40 lpm) with 8 bits per pixel
onto a CD-ROM. The uncompressed image is 455 MB,
which reduces to 71 MB after lossless compression. Since
such a large image can neither fit in typical workstation
memory nor can be displayed at full resolution, we implemented the retrieval of rectangular ‘chips’ of arbitrary
size and location using either pixel or latitude-longitude
coordinates. The map can thus be processed as a mosaic
of adjacent chips.
The USGS 7.5-minute series of topographic maps is
typical of the best in traditional cartography and packs
an enormous amount of information for diverse uses. The
extraction of data from such a map is correspondingly
more difficult than from specialized maps like cadastral and road maps, or from map separates (overlays).
High-resolution digitization is desirable because at lower
sample-spacing fine lines break, character shapes are distorted, it is even more difficult to segment glyphs from
street lines, and the regularity of the halftone pattern is
destroyed by aliasing.
179
The focus here is on street lines and labels in the
black layer (Fig. 2). In addition to street casings, 0.1 mm
wide and separated by 0.5 mm, the black layer graphics include solid blocks showing buildings in non-builtup areas like parks - some with flag (school) or cross
(church) icons attached; black outlines for running tracks
and other surface features; thick railroad lines; dashed
political boundary lines; and straight grid lines slightly
thinner than the street casings.
There are half-a-dozen fonts and sizes of text. The
street labels are slanted sans-serif caps 1.25 mm high or,
for principal streets, 1.4 mm high. Text in other fonts
denotes districts, monuments, schools, elevation benchmarks, etc. The street labels are typically oriented from
left to right for E-W streets, and from bottom to top
for N-S streets for viewing from the right. These conventions conflict for streets that run NW to SE: there
is a 10-degree region where the orientation is inconsistent. Street labels frequently overlay intersecting street
casings, and occasionally touch icons.
3 Methodology
The interaction between the line art and text processing at several points in the data flow is shown in Fig. 3.
The initial processing examines every pixel only once to
recover the black layer and extract its connected components. Then the candidate constituents of street names
and street casings are identified. The initial classification
is, however, error prone and the resulting errors affect
subsequent OCR and vectorization. The results can be
improved by detailed examination of the vicinity of identified character candidates (Fig. 4). The local processing
consists of two interlaced steps: character grouping and
line structure recovery.
Character grouping
Character grouping is based on the character neighborhood graph, which represents the spatial relations between connected components in a manner similar to that
of O’Gorman’s DocStrum [O’G93]. Characters are assembled into word boxes according to the constraint that
they lie on a straight line and their centroids are separated by about 1.4 times their average width.
In order to extract only street name boxes, the neighboring black pixels are searched for parallel line configurations with the appropriate spacing to establish the
dominant direction of the street-name string. This is particularly useful when nearly all the characters in a string
touch street lines. Combining the information from the
neighboring street casings and the character neighborhood graph, characters are grouped into word boxes to
form the longest possible aligned string. Strings are recursively merged as long as they don’t violate orientation
and word separation constraints. The resulting strings
are subjected to an additional uniformity-of-height constraint to distinguish street words from other text labels.
An example of the resolution of groups of adjacent characters into street-label word boxes is shown in Fig. 5.
180
L. Li et al.: Integrated text and line-art extraction from a topographic map
Color
Map Image
Color Separation
Black Layer
Initial Layer Separation
Text Layer
Line Layer
Character
Grouping
Vectorization
Line Vectors
Text Bounding Boxes
Line Structure
Recovery
Marked Text Image
Character
Template
Matching
Character Templates
Cooperative Processing
Line
Image
Fig. 5. Street-label character grouping. In the label ‘UPSHUR’, three of the six characters are touched by overlapping
street lines, yet the word is found
Text
Image
Fig. 3. Sublayer separation
Fig. 6. Line structure recovery based on graphics extrapolation. From top to bottom: original label box; segmented text;
and segmented street lines
Fig. 4. Local analysis. Above, vectors in the vicinity of a
label box. Below, paired vertical vectors identified and extrapolated as street-lines
The above character-grouping scheme is not yet robust enough to fully recover street labels with a curved
baseline that follow curving streets. Such labels are rare
in Washington East (D.C.) but might be more common
in some suburbs. The scheme also has other shortcomings. It cannot reliably distinguish street-name font from
other text or graphics.
Line structure recovery
The objective of this step is to recover street-line casings
that are interrupted because they overlap text. A black
pixel in a word-box can belong to a character, to a casing,
or to both.
All the vectors in the vicinity of each word box are analyzed. Short vectors which are not aligned with longer
vectors are typically parts of characters. Long vectors
that intersect the word bounding box are extended and
connected, and assigned to the street-line layer. The output of the character recognition routine is used to mark
the location of text pixels. The resulting sublayers, segmented text and segmented street lines, are both subsets
of the original black layer. An example is shown in Fig. 6.
Character template matching
Our OCR is based on template matching which, in addition to converting the word-box bitmaps to symbolic
alphanumeric ASCII labels, allows further improvement
of the line art and text segmentation.
Prototypes for each character class are first extracted
from bitmaps of operator-labeled training data. We note
that labeling a hundred or so word boxes is quite rapid,
because the operator need not segment the characters.
L. Li et al.: Integrated text and line-art extraction from a topographic map
181
Fig. 9. Templates constructed from the training data (first
experiment)
Fig. 7. Label box, matching templates, and residue
Fig. 10. Examples of test data for street labels, rotated to
horizontal
Fig. 8. Examples of training data for street labels, rotated
to horizontal
The segmentation-free prototype extraction and recognition algorithms are described in [NX97,XN99].
The extracted prototypes are matched against the
word boxes. The classification is based on the best fit
for the whole word box, which is found using a levelbuilding algorithm. The matching algorithm can ignore
pixels that are flagged as belonging to line art on grounds
that we cannot know the color of the underlying character pixel. Thus the recognition result itself can discriminate between these two possibilities, as the best-fitting
template should overlay the character pixels.
The residual line art overlaying some character can
introduce a mistake only if it distorts the overlaid character into another recognizable class, which is relatively
rare. (This type of error could be corrected by reference
to a street-name gazetteer.)
The important point is that the best-fitting templates
usually form a clean street label, and the residue that is
not covered by them (Fig. 7) can be analyzed, as mentioned above, for segmentation. The recognized characters could in turn be used to improve the initial templates which are based on too few operator-identified
character samples. All the word boxes could then be reclassified with the improved templates.
4 Experiments
The first experiment shows the performance of our custom character-recognition system on unsegmented label
boxes. The 32 templates extracted from the 92 streetname boxes used as the training set (examples in Fig. 8)
are shown in Fig. 9. Darker shades indicate higher probability of black. A few examples of the 733 automaticallyextracted street name boxes used for testing are displayed in Fig. 10.
After determining the best-fitting combination of
templates that fits a label, the character recognition system may still reject some characters if they are not
Table 1. Results of street label recognition
Street Labels
Other Text
Graphics
Correct
Wrong
Rejected
1285
151
0
57
57
46
94
275
573
matched with high enough confidence. Characters that
are part of street labels are recognized with an accuracy of 95% at a reject rate of 8%. (The error rate was
determined by a string-comparison algorithm.) Most of
the graphics fragments are rejected, but some lines perpendicular to the street box are recognized as ‘I’ or ‘L’
(here the line pixels were not marked). Most alphabetic
characters that do not originate from street labels are
rejected, but when they are not, they are often misrecognized. The complete character-level results are presented
in Table 1.
In our earlier work, the street-line layer was evaluated
on an 8” × 8” section of the map [NSS98]. Comparison
against the digital line graph (DLG) indicated that the
vectorization was 97.5% accurate before operator correction, but only 36% of the street lines were extracted.
The positional accuracy of the intersections was 12 m,
which is just within the National Map Accuracy Standards. An interactive session raised the completeness of
the vectorization to 97%, with an accuracy of 8 m. Of
the 3% residual error 1% is due to some missed streets,
and 2% to some of the corrected streets that ran outside
the chip boundary, where there was no DLG data.
The DLG does not contain street names, so these
could not be verified independently. The geometrical accuracy of the TIGER database, which does contain the
street names, was too low to be used for comparison.
Because of the aforementioned limitations of the current box-finding algorithm (see Fig. 10) and the vectorizer, a comprehensive evaluation of our integrated approach to text and line-art extraction was not possible
with automatically extracted boxes and vectors. Therefore, a custom graphic user interface was designed to
182
L. Li et al.: Integrated text and line-art extraction from a topographic map
Fig. 11. Left: an example street-word box (rectified) with
incident vectors. Right: the result of extrapolation of line vectors
allow an operator to draw the word boxes and the incident vectors. Additionally, the operator could transcribe
the word in ASCII to serve as ground truth for OCR.
The interface was instrumented to log the time spent on
individual steps. The black layer for the whole map was
divided into 12 overlapping chips to simplify data entry. Pause and resume buttons allowed the operator to
break the task of capturing all the street-name boxes on
the map into several sessions.
The left part of Fig. 11 shows an example box with
incident line vectors. The time log revealed that it took
the operator 7.6 h to enter 966 boxes. Of this total time,
60% was spent in drawing boxes, 24% in drawing vectors,
and only 16% in entering text. To simplify the dataentry task, the boxes were initially marked with color
markers on hard copies of chip images. This took about
4 h. Note that these operations were necessary only for
the evaluation of the new OCR algorithm and will not
be required in an automated system.
The operator was asked to ignore the 64 street labels with curved baselines (0.2% of the total number of
boxes). The data files containing the box and the incident vector information were post-processed to extrapolate the vectors within the box and mark the corresponding pixels for OCR with the street-line information (see
the right part of Fig. 11).
Of the 966 boxes, 422 were clean (had no line overlap). Ignoring the many duplicate generic names (‘ST’,
‘AVE’, etc.) from among these, a total of 185 were used
as the training set for the OCR program. The 544 boxes
with line overlap were used for evaluation. Figure 12
shows the set of templates constructed from the clean
glyphs. Note that because these are based on a clean and
much larger sample compared to those shown in Fig. 9,
better OCR results can be expected.
The results of OCR with and without vector information are shown in Table 2. The chips were 7-inch square
(with one inch overlap horizontally and vertically with
the adjoining chips) except for the right-most chips (second index = 2) and the bottom-most chips (first index
= 3) which were smaller. The second and fifth columns
show, respectively, the total number of characters and
words in each chip. The third and fourth columns show
the number of character errors for OCR without and
with vector information, respectively. The corresponding word-error data is shown in the sixth and seventh
Fig. 12. Templates constructed from the training data (second experiment)
(a)
(d)
(b)
(e)
(c)
(f)
Fig. 13. A sample of OCR results with and without consideration of lines. In each case, from top to bottom: original
box image, lines, OCR with lines, and OCR without lines
columns. The cumulative results in the last row show
that with the vector information the character error was
reduced from 4.1% to 2.0% and the word error was reduced from 11.6% to 6.4%. The error count does not include ‘1-I’ confusions, as those are identical glyphs on the
map and cannot be discriminated without context. The
count does include nine character errors due to the lack
of a ‘Q’ template because of which all the Q’s were recognized as O’s. We emphasize that only the overlapped
boxes were considered in this test. The error rate would
be lower if a representative sample of clean boxes were
included in the test.
The six sample results in Fig. 13 illustrate text and
line-art extraction with and without vector information.
In each case, the first row shows the original box image with crossing lines, the second row the extrapolated
lines, the third row the OCR result without vector information, and the fourth row the OCR result with vector
information.
In case a, the letter ‘U’ is recognized as the pair
‘6I’ because of the line crossing this letter in the center. When the line pixels are ignored, the ‘U’ is correctly
recognized. In case b, the line next to ‘L’ causes the best
match to occur for the template ‘E’ instead of ‘L’; when
the line pixels are ignored, this error disappears. Case c
illustrates the rare situation when ignoring the line pixels causes an error. Here OCR makes different errors in
L. Li et al.: Integrated text and line-art extraction from a topographic map
183
Table 2. OCR performance with and without vector information
Chip ID
# chars
Char Errors
w/o vector
with vector
information information
# Words
Word Errors
w/o vector
with vector
information information
s00
s01
s02
s10
s11
s12
s20
s21
s22
s30
s31
s32
250
251
98
303
368
216
167
146
171
108
194
71
10
9
10
6
21
7
11
4
1
4
12
1
3
5
4
8
11
5
2
2
1
2
3
1
60
49
19
63
75
49
43
38
39
31
53
25
5
6
5
4
14
7
8
3
2
2
8
1
3
3
4
4
7
5
2
2
1
2
1
1
Total
2343
96(4.1%)
47(2.0%)
544
64(11.8%)
35(6.4%)
the two cases. With the line pixels considered, the bestmatching template is ‘8’. Without the ggggg the bottom
stroke of ‘E’ disappears and because of the residual line
pixels at the top the best match is for ‘T’. In cases d
and e, the crossing lines add enough black pixels to cause
erroneous recognition of the letters ‘E’ and ‘P’, respectively; these errors are corrected when the line pixels are
ignored.
Finally, case f shows one of the three words that were
correctly recognized without the line information but incorrectly recognized with this information. The line pixels deleted enough of the stem pixels of ‘I’ to cause a
better match for ‘T’. What is also clear from these examples (and from Table 2) is that the OCR algorithm
is able to tolerate a fair amount of noise but its performance is further improved with the line information.
5 Conclusion
Beginning with the connected components of the black
layer of the map, we have shown that almost all streetlabel boxes can be located and extracted using character alignment and size, and the graphic context of the
location of labels adjacent to street lines. Most boxes
are correctly oriented, but sizing these boxes precisely
is more difficult. Even after extensive processing, many
boxes contain some graphics fragments, and some text
characters are missing. This is an area we are currently
pursuing for further improvement.
We have presented a template-based character recognition method that does not require character-level segmentation because it recognizes an entire string of characters. This recognizer tolerates overlapping graphics
segments much better than commercial omnifont OCR
and is able to take advantage of the overlapping line
information, thereby further reducing its error rate by
half. The resulting accuracy appears sufficient to invoke
context-based processing (i.e., a gazetteer) to advantage.
We have shown that the integrated approach also
helps eliminate character fragments (within the boxes)
from the vector layer. In future research we intend to
exploit this to improve vectorization.
The percentage of correctly extracted street vectors,
and of correctly extracted and recognized labels, is evidently far too low to be useable without extensive operator interaction. In previous work, we estimated that the
automated processing that we had developed reduced the
time required to digitize the street network from about
24 h to 16 h. With some of the additional improvements
mentioned here, especially close interaction between text
and graphics processing, it appears feasible to reduce the
overall time to about 10 h.
According to the operator log resulting from the correction of the 8” × 8” chip, the most time-consuming
aspects of operator intervention are vectorization, identification of the street label locations, and association
of street labels with street lines. Our methods are quite
effective for the last two of these three tasks, label box
positioning and association. As shown earlier, few labels
are missed entirely, and given correct street vectors and
street labels, the association is practically foolproof.
It therefore appears desirable to concentrate further
effort on more complete vectorization, perhaps even by
relaxing the constraints that now guarantee high accuracy. This will have the greatest direct impact on reducing operator time, as well as a beneficial secondary
effect on label box location. Further improving the accuracy of character recognition will have only a minor
effect, because typing-in the correct labels of the rotated
label boxes, without the need to point-click the location
or association, is very fast.
Wholly automatic conversion of documents of the
complexity of topographic maps is still well over the horizon, but tools can be developed for reducing the cost of
conversion. We are continuing our research with this objective in mind.
Acknowledgements. We gratefully acknowledge the hours of
careful work by Chakravarthy Terlapu in capturing the
street-label box information. This work directly followed the
184
L. Li et al.: Integrated text and line-art extraction from a topographic map
recent project supported by the National Imagery and Mapping Agency as a part of the Intelligent Map Understanding
Project. Support was also provided by the Central Research
Laboratory, Hitachi, Inc., and the NRI Geospatial Decision
Support Grant of the University of Nebraska. Part of this
work was carried out in the New York State Center for Automation Technologies (CAT) at Rensselaer Polytechnic Institute. The CAT is partially funded by a block grant from
the New York State Science and Technology Foundation.
References
[Doe98]
D. Doermann. An introduction to vectorization
and segmentation. Graphics Recognition: Algorithms and Systems, LNCS 1389, Berlin Heidelberg New York: Springer, pp. 1–8, 1998
[EAK95] L. Eikvil, K. Aas, H. Koren. Tools for interactive
map conversion and vectorization. IEEE Conf.
on Comput. Vision and Pattern Recognition, pp.
937–930, 1995
[ESRI97] Environmental Systems Research Institute Inc.
ARC/INFO Software. Redlands, CA, 1997
[FK88]
L.A. Fletcher, R. Kasturi. A robust algorithm for
text string separation from mixed text-graphics
images. IEEE Trans. on Pattern Anal. and Mach.
Intell., 10(6): 910–918 (1988)
[Har95]
J. Hartog. A framework for knowledge-based map
interpretation. Doctoral Dissertation, Technical
University of Delft, 1995
[JV97]
R. Jansen, A. Vossepoel. Adaptive vectorization
of line drawing images. Comput. Vision and Image Understanding, 65(1): 38–56 (1997)
[KA88]
R. Kasturi, L. Alemany. Information extraction
from images of paper-based maps. IEEE. Trans.
SE, 14(5): 671–675 (1988)
[Kal97]
K. Kalafala. Identification of street lines and
intersections in a scanned urban topographic
map. MS Thesis, Rensselaer Polytechnic Institute, Troy, 1997
[Li98]
L. Li. Adaptive text/line separation in document
images based on vectorization and OCR. MS Thesis, University of Nebraska - Lincoln, 1998
[LNS+ 99] L. Li, G. Nagy, A. Samal, S. Seth, Y. Xu. Cooperative text and line-art extraction from a topographic map. Proc. 5th Int. Conf. on Document
Anal. and Recognition (ICDAR ’99), pp. 467–470,
1999
[MMC+ 96] G.K. Myers, P. Mulgaonkar, C. Chen, J. DeCurtins, E. Chen. Verification based approach
for automated text and feature extraction
from raster-scanned maps. Graphics Recognition:
Methods and Applications, LNCS 1072, Berlin
Heidelberg New York: Springer, pp. 190–203,
1996
[Nav]
Navtech. Navigational Technologies, home page:
http://www.navtech.com/
[NSS+ 97] G. Nagy, A. Samal, S. Seth, T. Fisher, E. Guthmann, K. Kalafala, L. Li, S. Sivasubraniam,
Y. Xu. Reading street names from maps - technical challenges. Proc. GIS/LIS Conf., Cincinnati,
OH, pp. 89–97, 1997
[NSS+ 98] G. Nagy, A. Samal, S. Seth, T. Fisher, E. Guthmann, K. Kalafala, L. Li, S. Sivasubraniam,
Y. Xu. A prototype for adaptive association of
street names on maps. Graphics Recognition: Algorithms and Systems, K. Tombre,A. Chhabra,
editors, LNCS 1389, Berlin Heidelberg New York:
Springer, pp. 302–313, 1998
[NSY96] S. Naoi, M. Suwa, M. Yabuki. Quantitative evaluation of improved global interpolation in the segmentation of handwritten numbers overlapping a
border. IEICE Trans. Inf. Syst., E79-D(5), May
1996
[NX97]
G. Nagy, Y. Xu. Automatic prototype extraction
for OCR. Proc. Int. Conf. on Document Anal. and
Recognition, ICDAR’97, pp. 278–282, 1997
[O’G93]
L. O’Gorman. The document spectrum for page
layout analysis. IEEE Trans. PAMI, 15(11):
1162–1173 (1993)
[OK95]
L. O’Gorman, R. Kasturi. Document Image Analysis. IEEE Computer Society Press, New York,
1995
[PW85]
T. Pavlidis, C.J. Van Wyk. An automatic beautifier for drawings and illustrations. Procs. ACM
SIGGRAPH, pp. 225–234, 1985
[Rhi74]
D. Rhind. An introduction to the digitizing and
editing of mapped data. Automation in Cartography. In: P. Dale (ed.), British Cartographic Society, 1974
[Siv98]
S. Sivasubramaniam. Street name association in
digitized maps. MS Thesis, Rensselaer Polytechnic Institute, Troy, 1998
[StK95]
A. Smeulders, T. ten Kate. Software systems design for paper map conversion. Graphics Recognition Methods and Appl., pp. 204–211, 1995
[TK98]
P. Tofani, R. Kasturi. Segmentation of text from
color map images. Technical Report CSE-98-007,
Department of CSE, Penn State University, 1998
[USD97] USDC. Tiger tales: a presentation of the automated geographic support system. US Department of Commerce, Bureau of Census, 1997
[XN99]
Y. Xu, G. Nagy. Prototype extraction and adaptive OCR. IEEE Trans. PAMI, 21(12): 1280–1296
(1999)
[Xu98]
Y. Xu. Prototype extraction and OCR. PhD Thesis, Rensselaer Polytechnic Institute, Troy, 1998
L. Li et al.: Integrated text and line-art extraction from a topographic map
185
Luyang Li
Sharad Seth
George Nagy
Yihong Xu
Ashok Samal