AI and Cultural Heritage Semantic Classification of Byzantine Icons Paraskevi Tzouveli, Nikos Simou, Giorgios Stamou, and Stefanos Kollias, National Technical University of Athens R ecent advances in Web technologies that ensure quick, effective information distribution have created the ideal terrain for spreading vari- ous cultures through cultural-content presentation. Nowadays, we are witThis system uses nessing the publication of a huge amount of cultural information on the Web. fuzzy description In the past, people gathered cultural information from physical objects (such as books, sculptures, statues, and paintings). Now, digital collections have become common practice for most cultural-content providers. However, as the amount of the Web’s cultural content grows, search and retrieval procedures for that content become increasingly difficult. Moreover, Web users need more efficient ways to access huge amounts of content. So, researchers have proposed sophisticated browsing and viewing technologies, raising the need for detailed metadata that effectively describes the cultural content. Several annotation standards have been developed and implemented, and Semantic Web technologies provide a solution for the semantic description of collections on the Web.1 Unfortunately, the semantic annotation of cultural content is time consuming and expensive, making it one of the main difficulties of cultural-content publication. Therefore, the need for automatic or semi automatic analysis and classification of cultural assets has emerged. logics and patterns to automatically determine the sacred figure depicted in an icon. March/April 2009 To meet this need, researchers have proposed using image analysis methods (for some examples, see the sidebar on the next page). However, these methods use domain knowledge for only low-level analysis. Also, two main difficulties have arisen in using image analysis to automate annotation and classification of cultural digital assets. The first is the failure of semantic image segmentation and image analysis algorithms in some real-world conditions. This is due to the high variability of content and environmental parameters (luminance and so on), which makes the problem complex. The second difficulty is the extensive and vague nature of domain knowledge (at least for some cultural content), which complicates formal knowledge representation and reasoning. However, some cultural domains are appropriate for automatic analysis and classification methods. Byzantine icon art is one of them. The predefined image content and the low variability of the image characteristics support the successful application of image analysis methods. Consequently, we’ve developed a system that exploits traditional 1541-1672/09/$25.00 © 2009 IEEE Published by the IEEE Computer Society 35 AI and Cultural Heritage Related Work on Image Analysis for Cultural Heritage T he July 2008 IEEE Signal Processing Magazine presented some particularly interesting applications of image analysis to the cultural-heritage domain. Alejandro Ribes and his colleagues described problems concerning the design of multispectral cameras as well as the analysis of the Mona Lisa in the context of the digitization of paintings.1 Anna Pelagotti and her colleagues proposed a novel multispectral digital-imaging technique for clustering image regions in a set of images containing similar characteristics when exposed to electromagnetic radiation. 2 Their technique provides material localization and identification over a painting’s entire surface. C. Richard Johnson and his colleagues described how three research groups developed systems that identify artists on the basis iconographic guidelines to automatically analyze icons. Our system detects a set of concepts and properties depicted in the icon, according to which it recognizes a sacred figure’s face, thus providing semantic classification. Byzantine Iconography Byzantine art refers to the art of Eastern Orthodox states that were concurrent with the Byzantine Empire. Certain artistic traditions born in the Byzantine Empire, particularly regarding icon painting and church architecture, have continued in Greece, Bulgaria, Russia, and other Eastern Orthodox countries up to the present. The icons usually depict sacred figures from the Christian faith, such as Jesus, the Virgin Mary, the apostles, and the saints. At the beginning of the 17th century, iconography manuals represented the most important source of inspiration for icon painters. The 16th-century monk Dionysios from Fourna is credited as the author of the prototypical manual, Interpretation of the Byzantine Art. 2 In it, he defines the rules and iconographic patterns that painters still use to create icons. Byzantine iconography follows a unique convention of painting. The artistic language of Byzantine painters is characterized by apparent simplicity, overemphasized flatness, unreal and symbolic colors, lack of perspective, and strange proportions. The sacred figures are set beyond real time and real space through the use of gold backgrounds. The most important figure in an icon is depicted frontally, focusing on the eyes, facial expression, and hands. Clarity is the rule, not only in describing shapes but also in arranging figures in balanced compositions so that the narrative’s actions are clear. The figures have large eyes, enlarged ears, long, thin noses, and small mouths, each painted following specific rules. When depicting various figures, the artist is obliged to observe the manual’s rules regarding posture, hair and beard style, apparel, and other attributes. Specific sacred 36 of brushwork characteristics that art historians use for authentication. 3 Finally, Howard Leung and his colleagues reported on the preservation of ancient Chinese calligraphy. 4 References 1. A. Ribes et al., “Studying That Smile,” IEEE Signal Processing Magazine, vol. 25, no. 4, 2008, pp. 14–26. 2. A. Pelagotti et al., “Multispectral Imaging of Paintings,” IEEE Signal Processing Magazine, vol. 25, no. 4, 2008, pp. 27–37. 3. C.R. Johnson et al., “Image Processing for Artist Identification,” IEEE Signal Processing Magazine, vol. 25, no. 4, 2008, pp. 37–48. 4. H. Leung, S.T.S. Wong, and H.H.S. Ip, “In the Name of Art,” IEEE Signal Processing Magazine, vol. 25, no. 4, 2008, pp. 49–54. figures of the Christian faith are defined by specific facial characteristics. For example, according to Dionysios, Jesus is illustrated as a young figure with long, straight, thick, black hair and a short, straight, thick, black beard. According to these rules, the area in which the sacred figure will be painted is separated into equal segments. The first segment contains the head, focusing on the eyes and facial expression. This segment contains the sacred figure’s most distinguishable features—for example, hair and beard style. The remaining segments contain the sacred figure’s body, focusing on the hands and arms. The body can be depicted in different poses—for example, standing, sitting, from the head to the waist, and from the head to the thorax. The head should be elliptical. The major axis is along the vertical part of the head and is equal to four times the height of the nose (H). (According to Dionysios, the nose serves as the metric for producing harmonious facial and body proportions.) The minor axis is along the horizontal and is equal to 3H. The head segment can be separated into four equal vertical sections of height H. The hair is in the first part, the forehead is in the second, the nose is in the third, and the area between the mustache and the chin is in the fourth. A Byzantine painter applies four base colors in layers on the face segment. The darking is a dark color for the base of the face segment and the darkest shadows. The proplasmos is a less-dark color for lighter shadows. The sarkoma is for flesh, and the lightening is for highlights, containing a good amount of white and being light enough to show off against the base color. The absolute value of these colors differs from icon to icon. From Art Manual to Semantic Representation Although the knowledge in Dionysios’s manual concerns vague concepts such as “long hair,” “young face,” and so on, it’s quite strict and formally described. Consequently, www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS Sacred-figure we can create an ontological represenrecognition Feature tation of this knowledge using OWL. extraction In this way, the ontology’s axiomatic Semantic skeleton will provide the terminology segmentation and restrictions for Byzantine icons. Semantic interpretation The only problem is that even a Byzantine icon well-tuned image analysis algorithm analysis can’t produce perfect results for a Byzantine wide set of data with different charicon acteristics. Moreover, many Byzandatabase tine icons might be misinterpreted Assertional owing to their age, further affecting knowledge the results. Finally, the vague nature Reasoning of terms such as distance, color, and engine length introduces fuzziness in the catTerminological knowledge egorization of objects. For example, sometimes a sacred figure represents Knowledge representation a young face, but only to a certain and reasoning degree. In such cases, standard description logics (DLs) are insufficient because they can’t deal with vague, uncertain information. To handle Figure 1. The Byzantine icon classification system’s architecture. The analysis such problems, researchers have pro- module extracts information about the icon, and the knowledge representation and posed fuzzy DLs, based on fuzzy sets reasoning module uses this information to infer implicit knowledge that categorizes and fuzzy logic, which are suitable for the icon. handling imprecise information. 3 To classify the important facial features of Byzantine a large set of heuristics that facilitated the design of the icons and detect the sacred figures having specific charac- subsystem modules and helped us select the most apteristics, we combine image analysis algorithms with the propriate semantic segmentation and feature extraction expressive power of fuzzy DLs. Using the rules and pat- methods. terns described in Dionysios’s manual, we constructed a fuzzy knowledge base that’s populated by the results of im- Semantic segmentation. This module divides the icon into age analysis of the digitized paintings. Using fuzzy reason- regions visualizing the icon’s physical objects and their ing and on the basis of the defined terminology, our system parts. The first step of semantic segmentation is face detection. can classify the figures in Byzantine icons. For this, we adopted a fast method based on support vector decision functions.4 This method first uses integral imThe Icon Classification System Figure 1 shows the general architecture of our Byzantine aging for very fast feature evaluation based on Haar basis icon classification system. It consists of the Byzantine icon functions. Then, it uses a simple, efficient classifier based analysis subsystem and the knowledge representation and on the AdaBoost learning algorithm to select a small number of critical visual features from a very large set of poreasoning subsystem. tential features. Finally, this method combines classifiers in a “cascade,” quickly discarding the image’s background Byzantine Icon Analysis This subsystem performs the image processing and analy- regions while concentrating computation on promising sis tasks necessary to detect a set of primitive concepts and face-like regions. An example of this method’s results is properties (such as face and long hair) and form assertional the green rectangle in the first image of the face detection knowledge that semantically describes the icon’s content. It box in Figure 2 (see next page). The next step is localization of the eyes. Beconsists of modules for semantic segmentation, feature excause the nose’s height should be equal to the distraction, and semantic interpretation. The first step of system development was the interpre- tance of one retina from the other, that distance is a tation, with the aid of Byzantine iconography experts, critical parameter for segmentation. The semanticof Dionysios’s manual. This procedure provided us with segmentation module performs eye localization with the March/april 2009 www.computer.org/intelligent 37 AI and Cultural Heritage Face detection extracts the face and its basic components (nose and eyes). Face component detection detects the most important components of the face (beard, moustache, cheek, forehead, and hair, respectively). Base color analysis extracts the base color layers (darking, proplasmos. sarkoma, and lightening). s1 Semantic interpretation gives meaning to specific features, properties, and relationships of face components with the aid of fuzzy partitions. Membership Feature extraction automatically extracts features such as distance, color, and length for each face component and for some segments of the face components. 1 0.7 0.3 0 s2 s3 s4 s5 BelowEarsHair AboveEarsHair 2.8 5 Hairlength Fuzzy partitioning Image description s: Face s1: Hair s2: Forehead s3: Cheek s4: Mustache s5: Beard (s, s1): hasComponent ... Length(s1) = 2.8*H ... darking(s1) = 88% ... Formal assertions s: Face s1: Hair s2: Forehead s3: Cheek s4: Mustache s5: Beard (s, s1): hasComponent ... s1: AboveEarsHair = 0.7 s1: BelowEarsHair = 0.3 ... Figure 2. Image analysis. First, the algorithm detects the sacred figure’s face region, eyes, and nose. Then, it extracts the hair, forehead, cheek, mustache, and beard parts together with the face’s base color layers. Further analysis of the extracted parts provides information about characteristic features. Finally, the algorithm produces a semantic interpretation for each of these features, together with formal assertions. aid of geometric information from the eyes and the surrounding area. 5 We use a training set of sacred-figure eyes to create vector maps, and we use principal components analysis to derive eigenvectors for these maps. For a candidate face segment, the algorithm 5 projects the length and its angle maps on the spaces spanned by the eigenvectors that were evaluated during training. In that way, we use the similarity of the projection 38 weights and model eyes to determine an eye’s presence and position (indicated by the red and green dots in the second image in the face detection box of Figure 2). After eye localization, the module detects the nose. Let H denote the distance between the eyes and consequently the nose’s length. According to Dionysios, the nose starts on a small horizontal line whose center is a distance of H/5 below the midpoint between the eyes. We use a Hough transform to find this line. Therefore, we determine the nose’s ending line by applying a Hough transform at a distance H down vertically from the starting line. The computation of H leads to the detection of the face’s other main components—the hair, forehead, cheek, mustache, and beard—because they are well defined in Dionysios’s manual. The module detects the four base colors by applying an Otsu segmentation algorithm 6 to a restricted area of the face segment (see the base color analysis box in Figure 2), determining a range of intensity values for the pixels. To segment each part of the face, we use H together with Dionysios’s proportion guidelines to define the seeds for graph-cut segmentation.7 The cost functions used as soft constraints incorporate boundary and part information for each of the five segmented parts. To achieve segmentation of each part, we compute the minimum cut that’s the global optimum among all the segmentations satisfying the constraints (see the face-componentdetection box in Figure 2). Feature extraction. This module automatically extracts features from the segmented parts, 8 providing additional information regarding each part’s length, color, shape, and texture. This information constitutes the characteristic values of the features, providing an image description set (see the image-description section of the feature extraction box in Figure 2). Dionysios assigns four features to hair: www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS • color (dark, gray, or white), • density (thin or thick), • length (above the ears or below the ears, and whether the figure has a braid on the left, right, or both shoulders), and • form (straight, wild, or curly). The feature extraction module categorizes hair color (see Figure 2, feature extraction, segment s1) using the intervals of Otsu segmentation. It estimates the density and form by combining a Canny edge detector and a morphological fi lter and by evaluating the mean value of the curvatures of the detected edges. It fi nds the hair’s length using H together with Dionysios’s proportion guidelines. Next, the module determines information about the sacred figure’s age by computing the number of wrinkles on the forehead (s2). Using the Hough transform, the module defi nes the number of vertical and horizontal lines in the forehead, which represent the wrinkles. A young sacred figure has zero or few wrinkles, whereas an older sacred figure has many. The cheek (s3) provides additional information about age. According to Dionysios, a small number of wrinkles on the larger (left or right) part of the cheek indicates a young sacred figure. Conversely, an older sacred figure’s cheek will have a greater number of wrinkles (and therefore a greater proportion of the colors used for shadows). Finally, the module defi nes the feature sets for the mustache (s4) and beard (s5). Using techniques similar to those for hair analysis, it analyzes the mustache’s color and shape. It also extracts information on the beard’s length (short, not very long, long, down to the waist, or down to the knees), color (fair or dark), form (straight, wild, or curly) and shape (sharp, circle-like, circle, wide, or fluffy). can express an assertion in a fuzzy manner, with the aid of fuzzy set theory. This procedure is critical for the knowledge representation system. Suppose that a specific feature’s values lie in the interval X. A fuzzy set A of X associates each value x of X to a real number in the interval [0, 1], with the value of fA(X) at x representing the grade of membership of x in A. So, we defi ne an appropriate fuzzy partition (a set of fuzzy sets) of X and connect each fuzzy set of the partition with a subconcept of the terminology, defi ned by that fuzzy set’s property. For example, having extracted the value 2.8H for the length of hair (s1), we determine that s1 has a membership value of 0.7 for AboveEarsHair and 0.3 for BelowEarsHair (see the graph in the semanticinterpretation box in Figure 2). The defi nition of the fuzzy partitions is based on the heuristics extracted from Dionysios’s manual. the module determines information about Knowledge representation and reasoning This subsystem consists of terminological and assertional knowledge and a reasoning engine. These types of knowledge are the basic components of a knowledge-based system based on DLs, 9 a structured knowledge-representation formalism with decidable-reasoning algorithms. DLs have become popular, especially because of their use in the Semantic Web (as in OWL DL, for example). DLs represent a domain’s important notions as concept and role descriptions. To do this, DLs use a set of concept and role constructors on the basic elements of a domain-specific alphabet. This alphabet consists of a set of individuals (objects) constituting the domain, a set of atomic concepts describing the individuals, and a set of atomic roles that relate the individuals. The concept and role constructors that are used indicate the expressive power and the name of the specific DL. Here, we use SHIN, an expressive subset of OWL DL that employs concept negation, intersection, and union; existential and universal quantifiers; transitive and inverse roles; role hierarchy; and number restrictions. the sacred figure’s age by computing the number of wrinkles on the forehead. Semantic interpretation. This module represents the extracted information in terms of the sacred-figure ontology. Representing the face and its main components is straightforward: we simply write assertions such as s1: Hair (see the semantic-interpretation box in Figure 2). However, each component’s properties are vague. For example, you can’t always directly determine on the basis of length whether the sacred figure has long hair. This difficulty is due to the vagueness of the defi nition of length and the imprecision introduced by feature extraction (see, for example, the result of the segmentation of the hair in the feature extraction box in Figure 2). Nevertheless, we March/april 2009 Terminological and assertional knowledge. The terminological knowledge is an ontological representation of the knowledge in Dionysios’s manual. The assertional knowledge is a formal set of assertions describing a specific icon in terms of the terminological knowledge. For example, the terminology contains axiomatic knowledge such as “a Jesus Face is depicted as a young face with long, straight, thick, www.computer.org/intelligent 39 AI and Cultural Heritage black hair and a short, straight, thick, black beard,” whereas “the face depicted in the icon is a face with long hair in some degree” is a relative assertion. The knowledge base has two main components. The terminological component (TBox) describes the relevant notions of the application domain by stating properties of concepts and roles and their interrelations. TBox is actually a set of concept inclusion axioms of the form C v D, where D is a superconcept of C, and concept equivalence axioms of the form C ≡ D, where C is equivalent to D. The assertional component (ABox) describes a concrete world by stating properties of individuals and their specific interrelations. To deal with the vagueness introduced in our case, we use f-SHIN, a fuzzy extension of SHIN. 3 In f-SHIN, ABox is a finite set of fuzzy assertions of the form 〈a : C⋈n〉, 〈(a, b) : R⋈n〉, where ⋈ stands for ≥, >, ≤, or <, for a, b ∈ I. Intuitively, a fuzzy assertion of the form 〈a : C ≥ n〉 means that the membership degree of a to concept C is at least equal to n. (A formal definition of the syntax and semantics of fuzzy DLs appears elsewhere. 3,10) In the Byzantine-icon-analysis domain, the individuals are the segments that the semantic-segmentation module extracted. The DL concepts classifying the individuals are determined by the properties measured as features during feature extraction (for example, BlackHair and AboveEarsHair). The DL roles mainly describe partonomic (mereological) hierarchies and spatial relations (for example, the axiom isLeftOf– = isRightOf indicates that the role isLeftOf is the inverse of isRightOf, and the axioms Trans(isLeftOf) and Trans(isAboveOf) indicate that isLeftOf and isRightOF are transitive). The terminology’s main concepts are Figure, Face, and FaceComponent. They’re connected with the partonomic role hasPart and its subroles hasSegment and hasComponent, to indicate whether a segment is a subsegment of another segment and so on. Using these roles, we describe the partonomic hierarchy as follows: Face ≡ ∀hasComponent.FaceComponent Figure ≡ ∃hasSegment.Face FaceComponent v Beard t Cheek t Forehead t Hair t Mustache We describe relevant assertions as follows: 〈s:Face ≥ 1.0〉 〈s1:Hair ≥ 1.0〉 〈(s, s1):hasSegment ≥ 1.0〉 40 … The taxonomy is constructed using the properties measured as features during feature extraction. Here are some examples of terminological axioms defining the concept hierarchy: AboveEarsHair v Hair OldCheek v Cheek OldForehead v Forehead HairyBeard v Beard Restrictions concerning the taxonomy mainly describe disjointness and exhaustiveness, forming axioms such as AboveEarsHair v ¬BelowEarsHair Hair v AboveEarsHair t BelowEarsHair Using these concepts, we define more expressive descriptions that specify some specific face characteristics on the basis of Dionysios’s manual. For example, we define a young man’s face as YoungMansFace ≡ ∃hasComponent.BlackHair u ∃hasComponent.(NotVeryLongBeard u BlackBeard) u ∃hasComponent.YoungForehead u ∃hasPart. YoungCheek u ∃hasPart.BlackMustache We similarly define concepts specifying the characteristics of some popular sacred figures. For example, we define JesusFace as JesusFace ≡ YoungMansFace u ∃hasComponent. (BraidOnTheLeftShoulder u HairyHair u StraightHair) u ∃hasComponent. (CircleLikeBeard u StraightBeard) The reasoning engine. This module uses the terminological knowledge and the assertions to recognize the sacred figure. Our system uses FiRE, the Fuzzy Reasoning Engine (www.image.ece.ntua.gr/~nsimou/FiRE), which fully supports f-SHIN. Figure 3 illustrates how the engine performs reasoning and classification for recognizing Jesus as the sacred figure. We use FiRE’s greatest lower bound (GLB) reasoning service, which evaluates an individual’s greatest possible degree of participation in a concept. The concepts of interest in this case, YoungMansFace and JesusFace, only include conjunc- www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS YoungMansFace � �hasComponent. BlackHair � hasComponent. �.(NotVeryLongBeard � BlackBeard) � �hasComponent.YoungForehead � �hasPart.YoungCheek � �hasPart. BlackMoustache s1 s2 s3 s4 s5 JesusFace � YoungMansFace � �hasComponent.(BraidOnTheLeftShoulder � HairyHair � StraightHair) � �hasComponent.(CircleLikeBeard � StraightBeard) JesusFace TBox Reasoning engine s: Face s1: Hair … (s, s1): hasComponent … s1: BraidOnTheLeftShoulder = 0.78 s1: HairyHair = 0.96 s1: StraightHair = 0.86 s1: BlackHair = 0.88 s2: YoungForehead = 0.82 s3: YoungCheek = 0.87 s4: StraightMoustache = 1.0 s4: BlackMoustache = 0.83 s5: CircleLikeBeard = 0.79 s5: StraightBeard = 0.92 s5: NotVeryLongBeard = 1.0 s5: BlackBeard = 0.84 ABox Knowledge representation and reasoning Figure 3. A reasoning example. The extracted information from image analysis constitutes the assertional component (ABox) of the knowledge base; the terminological component (TBox) is defined on the basis of Fourna’s specification. These components form the input to the fuzzy reasoning engine, which infers information about the icon. tions, so we use the fuzzy intersection operator for the minimum. So, the GLB of segment s (the face) in YoungMansFace is 0.82, whereas the GLB of segment s in JesusFace is 0.78. In other words, the extracted features indicate that the face depicted is YoungMansFace to a degree of 0.82 and is JesusFace to a degree of 0.78. Results We evaluated our system on a database, provided by the Mount Sinai Foundation in Greece, containing 2,000 digitized Byzantine icons dating back to the 13th century. The icons depict 50 different characters; according to Dionysios, each character has specific facial features that makes him or her distinguishable. Evaluation of the Byzantine-icon-analysis subsystem produced promising results. The subsystem’s mean response time was approximately 15 seconds on a typical PC. In the semantic-segmentation module, the face detection submodule reached 80 percent accuracy. In most cases, the failure occurred in icons with a destroyed face area. If the submodule detected the face, it almost always detected the March/april 2009 eyes and nose. The base-color-analysis submodule defined the color model used by the icon’s artist, which the feature extraction module later used. Using the nose’s height, which can be easily determined, and taking into account the manual’s rules, the face-component-detection submodule estimated the positions of the background and foreground pixels for every part of the face. These positions constituted the input to the graph-cut algorithm. This submodule achieved 96 percent accuracy. The feature extraction module extracted features for every icon segment. Then, using the fuzzy partitions, the semantic-interpretation module interpreted each segment’s specific features and properties and the relationship of face parts. Thus, that module determined each feature’s degree of membership in a specific class, thereby creating an image description. To evaluate overall system performance, we used precision and recall. Table 1 (on the next page) presents results for 20 of the sacred-figure classes. The more distinctive facial features the sacred figure in a class contained, the better our method performed (for www.computer.org/intelligent 41 AI and Cultural Heritage Table 1. Evaluation of the icon classification system. Class No. of icons in class No. of images classified No. of images correctly classified Precision Recall Jesus 70 67 58 0.87 0.83 Virgin Mary 60 54 43 0.80 0.72 Peter 50 44 38 0.86 0.76 Paul 50 44 39 0.89 0.78 Katerina 40 35 30 0.86 0.75 Ioannis 40 37 29 0.78 0.73 Luke 40 33 28 0.85 0.70 Andrew 40 33 29 0.88 0.73 Stefanos 30 27 24 0.89 0.80 Konstantinos 40 36 30 0.83 0.75 Dimitrios 40 36 31 0.86 0.78 Georgios 40 37 32 0.86 0.80 Eleni 40 37 31 0.84 0.78 Pelegia 20 17 15 0.88 0.75 Nikolaos 40 38 33 0.87 0.83 Basileios 40 37 31 0.84 0.78 Antonios 30 27 20 0.74 0.67 Eythimios 25 23 18 0.78 0.72 Thomas 35 32 26 0.81 0.74 Minas 20 17 14 0.82 0.70 example, for the Jesus class). Conversely, our method performed worse on figures with similar facial features, such as women. sify paintings on the basis of unmixed colors and dense textures. Acknowledgments W e’re developing a version of our system that takes into account information about clothes associated with specific types of figures (such as apostles, kings, angels, and monks) and clothing accessories (kerchiefs and so on) and nonclothing accessories (books, crosses, and so on) associated with specific sacred figures. For example, the system could incorporate the knowledge that icons of the Virgin Mary depict her wearing a kerchief with three stars. Such information might help in classifying figures with similar facial features. Furthermore, an integrated version of our system could eventually date icons and attribute them to a particular artist or artistic school, on the basis of features specific to each artistic period. Such information can help classify an icon even if the specific sacred figure is difficult to recognize. We’re also investigating applying our method to other art movements such as impressionism, trying to clas- 42 Our research has been partly supported by the MORFES (extraction and modeling of figure in Byzantine paintings of Mount Sinai monuments using measures) project, funded by the EU and the Greek Secretariat for Research and Development. We thank the Mount Sinai Foundation for providing access to the rich data set and especially archeologist Dimitris Kalomirakis for his participation in knowledge extraction and his continuing support during this research. Finally, we thank the anonymous reviewers for their constructive comments. References 1.T. Berners-Lee, J. Hendler, and O. Lassila, “The Semantic Web,” Scientific American, May 2001, pp. 34–43. 2.Dionysios tou ek Phourna, Hermeneia tis zografikis technis (interpretation of the Byzantine art), B. Kirschbaum, 1909. 3.G. Stoilos et al., “Reasoning with Very Expressive Fuzzy De scription Logics,” J. Artificial Intelligence Research, vol. 30, no. 5, 2007, pp. 273–320. 4.P. Viola and M.J. Jones, “Robust Real-Time Face Detection,” Int’l J. Computer Vision, vol. 57, no. 2, 2004, pp. 137–152. 5.S. Asteriadis et al., “An Eye Detection Algorithm Using Pixel to Edge Information,” Proc. 2nd IEEE-EURASIP Int’l Symp. www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS Th e A u t h o r s Control, Communications, and Sig nal Processing (ISCCSP 06), 2006; www.eurasip.org/Proceedings/Ext/ ISCCSP2006/defevent/papers/cr1124. pdf. 6.N.A. Otsu, “A Threshold Selection Method from Gray-Level Histograms,” IEEE Trans. Systems, Man, and Cybernetics, vol. 1, no. 9, 1979, pp. 62–66. 7.Y. Boykov and M.-P. Jolly, “Interactive Graph Cuts for Optimal Boundary and Region Segmentation of Objects in N-D Images,” Proc. Int’l Conf. Computer Vision (ICCV 01), vol. 1, IEEE Press, 2001, pp. 105–112. 8.R.C. Gonzalez and R.E. Woods, Digital Image Processing, 3rd ed., Prentice Hall, 2008. 9.F. Baader et al., The Description Logic Handbook: Theory, Implementation and Applications, Cambridge Univ. Press, 2002. 10.G. Stoilos et al., “Uncertainty and the Semantic Web,” IEEE Intelligent Systems, vol. 21, no. 5, 2006, pp. 84–87. For more information on this or any other computing topic, please visit our Digital Library at www.computer.org/csdl. Paraskevi Tzouveli is pursuing her PhD in image and video analysis at the National Technical University of Athens. Her research interests include image and video analysis, information retrieval, knowledge manipulation, e-learning, and digital cultural heritage systems. Tzouveli received her Diploma in electrical and computer engineering from the National Technical University of Athens. Contact her at [email protected]. Nikos Simou is a PhD student in the National Technical University of Athens Depart- ment of Electrical and Computer Engineering and a member of the Image, Video, and Multimedia Systems Laboratory’s Knowledge Technologies Group. His research interests include knowledge representation under certain and uncertain or imprecise knowledge based on crisp and fuzzy description logics, ontologies, and the implementation and optimization of reasoning systems and applications for the Semantic Web. Contact him at [email protected]. Giorgios Stamou is a lecturer in the National Technical University of Athens School of Electrical and Computer Engineering and a member of the university’s Image, Video, and Multimedia Systems Laboratory, heading the lab’s Knowledge Technologies Group. His main research interests include knowledge representation and reasoning, theory of uncertainty, machine learning, semantic interoperability, and digital archives. Contact him at [email protected]. Stefanos Kollias is a professor in the National Technical University of Athens (NTUA) School of Electrical and Computer Engineering and the director of the Image, Video, and Multimedia Systems Laboratory. His research interests include image and video processing, analysis, retrieval, multimedia systems, computer graphics and virtual reality, artificial intelligence, neural networks, human-computer interaction, and digital libraries. Kollias received his PhD in signal processing from NTUA. He’s a member of the European Commission expert group in digital libraries and of the EC committee on semantic interoper ability, leading NTUA participation in DL projects, EDLnet, VideoActive, Europeana Connect, Europeana Network 1.0, and EU Screen. Contact him at [email protected]. AdvertiSER Information March/April 2009 • IEEE Intelligent Systems Advertising Personnel Marion Delaney IEEE Media, Advertising Dir. Phone: +1 415 863 4717 Email: md.ieeemedia@ ieee.org Marian Anderson Sr. Advertising Coordinator Phone: +1 714 821 8380 Fax: +1 714 821 4010 Email: manderson@ computer.org Sandy Brown Sr. Business Development Mgr. Phone: +1 714 821 8380 Fax: +1 714 821 4010 Email: sb.ieeemedia@ ieee.org March/april 2009 Advertising Sales Representatives Product: Recruitment: Midwest/Southwest Darcy Giovingo Phone: +1 847 498 4520 Fax: +1 847 498 5911 Email: [email protected] Mid Atlantic Lisa Rinaldo Phone: +1 732 772 0160 Fax: +1 732 772 0164 Email: lr.ieeemedia@ ieee.org Northwest/Southern CA Tim Matteson Phone: +1 310 836 4064 Fax: +1 310 836 4067 Email: tm.ieeemedia@ ieee.org New England John Restchack Phone: +1 212 419 7578 Fax: +1 212 419 7589 Email: [email protected] Japan Tim Matteson Phone: +1 310 836 4064 Fax: +1 310 836 4067 Email: [email protected] US West Lynne Stickrod Phone: +1 415 931 9782 Fax: +1 415 931 9782 Email: [email protected] Southeast Thomas M. Flynn Phone: +1 770 645 2944 Fax: +1 770 993 4423 Email: flynntom@ mindspring.com Europe Hilary Turnbull Phone: +44 1875 825700 Fax: +44 1875 825701 Email: impress@ impressmedia.com Europe Sven Anacker Phone: +49 202 27169 11 Fax: +49 202 27169 20 Email: sanacker@ intermediapartners.de www.computer.org/intelligent US East Joseph M. Donnelly Phone: +1 732 526 7119 Email: [email protected] US Central Darcy Giovingo Phone: +1 847 498 4520 Fax: +1 847 498 5911 Email: [email protected] 43
© Copyright 2026 Paperzz