Anda di halaman 1dari 9

AI and Cultural Heritage

Semantic
Classification
of Byzantine Icons
Paraskevi Tzouveli, Nikos Simou, Giorgios Stamou, and Stefanos Kollias,
National Technical University of Athens

R
ecent advances in Web technologies that ensure quick, effective infor-

mation distribution have created the ideal terrain for spreading vari-

ous cultures through cultural-content presentation. Nowadays, we are wit-

This system uses nessing the publication of a huge amount of cultural information on the Web.

fuzzy description In the past, people gathered cultural in- To meet this need, researchers have pro-
formation from physical objects (such as posed using image analysis methods (for
logics and patterns books, sculptures, statues, and paintings). some examples, see the sidebar on the next
Now, digital collections have become com- page). However, these methods use domain
to automatically mon practice for most cultural-content knowledge for only low-level analysis. Also,
providers. two main difficulties have arisen in using
determine the sacred However, as the amount of the Web’s cul- image analysis to automate annotation and
tural content grows, search and retrieval classification of cultural digital assets. The
figure depicted procedures for that content become increas- first is the failure of semantic image seg-
ingly difficult. Moreover, Web users need mentation and image analysis algorithms in
in an icon. more efficient ways to access huge amounts some real-world conditions. This is due to
of content. So, researchers have proposed so- the high variability of content and environ-
phisticated browsing and viewing technolo- mental parameters (luminance and so on),
gies, raising the need for detailed metadata which makes the problem complex. The sec-
that effectively describes the cultural con- ond difficulty is the extensive and vague na-
tent. Several annotation standards have been ture of domain knowledge (at least for some
developed and implemented, and Semantic cultural content), which complicates formal
Web technologies provide a solution for the knowledge representation and reasoning.
semantic description of collections on the However, some cultural domains are ap-
Web.1 Unfortunately, the semantic annota- propriate for automatic analysis and classi-
tion of cultural content is time consuming fication methods. Byzantine icon art is one
and expensive, making it one of the main of them. The predefined image content and
difficulties of cultural-content publication. the low variability of the image characteris-
Therefore, the need for automatic or semi­ tics support the successful application of im-
automatic analysis and classification of cul- age analysis methods. Consequently, we’ve
tural assets has emerged. developed a system that exploits traditional

March/April 2009 1541-1672/09/$25.00 © 2009 IEEE 35


Published by the IEEE Computer Society
AI and Cultural Heritage

Related Work on Image Analysis for Cultural Heritage

T
he July 2008 IEEE Signal Processing Magazine pre- of brushwork characteristics that art historians use for au-
sented some particularly interesting applications of thentication. 3 Finally, Howard Leung and his colleagues re-
image analysis to the cultural-heritage domain. Ale- ported on the preservation of ancient Chinese calligraphy. 4
jandro Ribes and his colleagues described problems con-
cerning the design of multispectral cameras as well as the
analysis of the Mona Lisa in the context of the digitization
References
of paintings.1 Anna Pelagotti and her colleagues proposed 1. A. Ribes et al., “Studying That Smile,” IEEE Signal Processing
a novel multispectral digital-imaging technique for clus- Magazine, vol. 25, no. 4, 2008, pp. 14–26.
2. A. Pelagotti et al., “Multispectral Imaging of Paintings,” IEEE
tering image regions in a set of images containing similar
Signal Processing Magazine, vol. 25, no. 4, 2008, pp. 27–37.
characteristics when exposed to electromagnetic radia- 3. C.R. Johnson et al., “Image Processing for Artist Identifica-
tion. 2 Their technique provides material localization and tion,” IEEE Signal Processing Magazine, vol. 25, no. 4, 2008,
identification over a painting’s entire surface. C. Richard pp. 37–48.
Johnson and his colleagues described how three research 4. H. Leung, S.T.S. Wong, and H.H.S. Ip, “In the Name of Art,” IEEE
groups developed systems that identify artists on the basis Signal Processing Magazine, vol. 25, no. 4, 2008, pp. 49–54.

iconographic guidelines to automatically analyze icons. figures of the Christian faith are defined by specific facial
Our system detects a set of concepts and properties de- characteristics. For example, according to Dionysios, Jesus
picted in the icon, according to which it recognizes a sa- is illustrated as a young figure with long, straight, thick,
cred figure’s face, thus providing semantic classification. black hair and a short, straight, thick, black beard.
According to these rules, the area in which the sacred
Byzantine Iconography figure will be painted is separated into equal segments. The
Byzantine art refers to the art of Eastern Orthodox states first segment contains the head, focusing on the eyes and
that were concurrent with the Byzantine Empire. Certain facial expression. This segment contains the sacred fig-
artistic traditions born in the Byzantine Empire, particu- ure’s most distinguishable features—for example, hair and
larly regarding icon painting and church architecture, have beard style. The remaining segments contain the sacred fig-
continued in Greece, Bulgaria, Russia, and other Eastern ure’s body, focusing on the hands and arms. The body can
Orthodox countries up to the present. The icons usually be depicted in different poses—for example, standing, sit-
depict sacred figures from the Christian faith, such as Je- ting, from the head to the waist, and from the head to the
sus, the Virgin Mary, the apostles, and the saints. thorax.
At the beginning of the 17th century, iconography man- The head should be elliptical. The major axis is along
uals represented the most important source of inspiration the vertical part of the head and is equal to four times the
for icon painters. The 16th-century monk Dionysios from height of the nose (H). (According to Dionysios, the nose
Fourna is credited as the author of the prototypical man- serves as the metric for producing harmonious facial and
ual, Interpretation of the Byzantine Art. 2 In it, he defines body proportions.) The minor axis is along the horizontal
the rules and iconographic patterns that painters still use and is equal to 3H. The head segment can be separated into
to create icons. four equal vertical sections of height H. The hair is in the
Byzantine iconography follows a unique convention of first part, the forehead is in the second, the nose is in the
painting. The artistic language of Byzantine painters is third, and the area between the mustache and the chin is in
characterized by apparent simplicity, overemphasized flat- the fourth.
ness, unreal and symbolic colors, lack of perspective, and A Byzantine painter applies four base colors in layers on
strange proportions. The sacred figures are set beyond real the face segment. The darking is a dark color for the base
time and real space through the use of gold backgrounds. of the face segment and the darkest shadows. The proplas-
The most important figure in an icon is depicted frontally, mos is a less-dark color for lighter shadows. The sarkoma
focusing on the eyes, facial expression, and hands. Clarity is for flesh, and the lightening is for highlights, containing
is the rule, not only in describing shapes but also in arrang- a good amount of white and being light enough to show off
ing figures in balanced compositions so that the narrative’s against the base color. The absolute value of these colors
actions are clear. The figures have large eyes, enlarged ears, differs from icon to icon.
long, thin noses, and small mouths, each painted following
specific rules. From Art Manual to Semantic Representation
When depicting various figures, the artist is obliged to Although the knowledge in Dionysios’s manual concerns
observe the manual’s rules regarding posture, hair and vague concepts such as “long hair,” “young face,” and so
beard style, apparel, and other attributes. Specific sacred on, it’s quite strict and formally described. Consequently,

36 www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS


we can create an ontological represen- Sacred-figure
Feature recognition
tation of this knowledge using OWL. extraction
In this way, the ontology’s axiomatic Semantic
skeleton will provide the terminology segmentation
and restrictions for Byzantine icons. Semantic
The only problem is that even a interpretation
well-tuned image analysis algorithm Byzantine icon
can’t produce perfect results for a analysis
wide set of data with different char- Byzantine
icon
acteristics. Moreover, many Byzan- database
tine icons might be misinterpreted
owing to their age, further affecting Assertional
knowledge
the results. Finally, the vague nature
Reasoning
of terms such as distance, color, and engine
length introduces fuzziness in the cat- Terminological
egorization of objects. For example, knowledge
sometimes a sacred figure represents Knowledge representation
a young face, but only to a certain and reasoning
degree. In such cases, standard de-
scription logics (DLs) are insufficient
because they can’t deal with vague,
uncertain information. To handle Figure 1. The Byzantine icon classification system’s architecture. The analysis
such problems, researchers have pro- module extracts information about the icon, and the knowledge representation and
posed fuzzy DLs, based on fuzzy sets reasoning module uses this information to infer implicit knowledge that categorizes
and fuzzy logic, which are suitable for the icon.
handling imprecise information. 3
To classify the important facial features of Byzantine a large set of heuristics that facilitated the design of the
icons and detect the sacred figures having specific charac- subsystem modules and helped us select the most ap-
teristics, we combine image analysis algorithms with the propriate semantic segmentation and feature extraction
expressive power of fuzzy DLs. Using the rules and pat- methods.
terns described in Dionysios’s manual, we constructed a
fuzzy knowledge base that’s populated by the results of im- Semantic segmentation. This module divides the icon into
age analysis of the digitized paintings. Using fuzzy reason- regions visualizing the icon’s physical objects and their
ing and on the basis of the defined terminology, our system parts.
can classify the figures in Byzantine icons. The first step of semantic segmentation is face detection.
For this, we adopted a fast method based on support vec-
The Icon Classification System tor decision functions.4 This method first uses integral im-
Figure 1 shows the general architecture of our Byzantine aging for very fast feature evaluation based on Haar basis
icon classification system. It consists of the Byzantine icon functions. Then, it uses a simple, efficient classifier based
analysis subsystem and the knowledge representation and on the AdaBoost learning algorithm to select a small num-
reasoning subsystem. ber of critical visual features from a very large set of po-
tential features. Finally, this method combines classifiers
Byzantine Icon Analysis in a “cascade,” quickly discarding the image’s background
This subsystem performs the image processing and analy- regions while concentrating computation on promising
sis tasks necessary to detect a set of primitive concepts and face-like regions. An example of this method’s results is
properties (such as face and long hair) and form assertional the green rectangle in the first image of the face detection
knowledge that semantically describes the icon’s content. It box in Figure 2 (see next page).
consists of modules for semantic segmentation, feature ex- The next step is localization of the eyes. Be-
traction, and semantic interpretation. cause the nose’s height should be equal to the dis-
The first step of system development was the interpre- tance of one retina from the other, that distance is a
tation, with the aid of Byzantine iconography experts, critical parameter for segmentation. The semantic-
of Dionysios’s manual. This procedure provided us with segmentation module performs eye localization with the

March/april 2009 www.computer.org/intelligent 37


AI and Cultural Heritage

Face detection
extracts the face
and its basic weights and model eyes to deter-
components mine an eye’s presence and position
(nose and eyes). (indicated by the red and green dots
in the second image in the face de-
tection box of Figure 2).
Face component Base color After eye localization, the mod-
detection analysis extracts
detects the most the base color
ule detects the nose. Let H denote the
important components layers (darking, distance between the eyes and con-
of the face (beard, proplasmos. sequently the nose’s length. Accord-
moustache, cheek, sarkoma, and ing to Dionysios, the nose starts on a
forehead, and hair, lightening).
respectively). small horizontal line whose center is
a distance of H/5 below the midpoint
between the eyes. We use a Hough
s1 Image description transform to find this line. Therefore,
Feature extraction s: Face we determine the nose’s ending line
automatically extracts s2 s1: Hair
features such as
by applying a Hough transform at a
s2: Forehead
distance, color, s3 s3: Cheek distance H down vertically from the
and length for s4: Mustache starting line.
each face s5: Beard The computation of H leads to the
component and s4 (s, s1): hasComponent
for some ...
detection of the face’s other main com-
segments s5 Length(s1) = 2.8*H ponents—the hair, forehead, cheek,
of the face ... mustache, and beard—because they are
components. darking(s1) = 88% well defined in Dionysios’s manual.
...
The module detects the four base
colors by applying an Otsu segmen-
Semantic Formal assertions tation algorithm 6 to a restricted area
interpretation s: Face of the face segment (see the base
gives meaning to BelowEarsHair AboveEarsHair s1: Hair color analysis box in Figure 2), de-
Membership

specific features, 1 s2: Forehead


termining a range of intensity values
properties, and 0.7 s3: Cheek
relationships of s4: Mustache for the pixels.
0.3
face components s5: Beard To segment each part of the face, we
with the aid (s, s1): hasComponent use H together with Dionysios’s pro-
of fuzzy 0 2.8 5 ...
partitions. Hairlength s1: AboveEarsHair = 0.7 portion guidelines to define the seeds
Fuzzy partitioning s1: BelowEarsHair = 0.3 for graph-cut segmentation.7 The cost
... functions used as soft constraints in-
corporate boundary and part infor-
mation for each of the five segmented
Figure 2. Image analysis. First, the algorithm detects the sacred figure’s face region, parts. To achieve segmentation of
eyes, and nose. Then, it extracts the hair, forehead, cheek, mustache, and beard each part, we compute the minimum
parts together with the face’s base color layers. Further analysis of the extracted cut that’s the global optimum among
parts provides information about characteristic features. Finally, the algorithm
all the segmentations satisfying the
produces a semantic interpretation for each of these features, together with formal
assertions.
constraints (see the face-component-
detection box in Figure 2).

aid of geometric information from the eyes and the sur- Feature extraction. This module automatically ex-
rounding area. 5 We use a training set of sacred-figure tracts features from the segmented parts, 8 providing
eyes to create vector maps, and we use principal com- additional information regarding each part’s length,
ponents analysis to derive eigenvectors for these maps. color, shape, and texture. This information constitutes
For a candidate face segment, the algorithm 5 projects the characteristic values of the features, providing an
the length and its angle maps on the spaces spanned image description set (see the image-description sec-
by the eigenvectors that were evaluated during train- tion of the feature extraction box in Figure 2).
ing. In that way, we use the similarity of the projection Dionysios assigns four features to hair:

38 www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS


• color (dark, gray, or white), can express an assertion in a fuzzy manner, with the aid of
• density (thin or thick), fuzzy set theory. This procedure is critical for the knowl-
• length (above the ears or below the ears, and whether the edge representation system.
figure has a braid on the left, right, or both shoulders), and Suppose that a specific feature’s values lie in the inter-
• form (straight, wild, or curly). val X. A fuzzy set A of X associates each value x of X
to a real number in the interval [0, 1], with the value of
The feature extraction module categorizes hair color (see fA(X) at x representing the grade of membership of x in
Figure 2, feature extraction, segment s1) using the intervals A. So, we defi ne an appropriate fuzzy partition (a set of
of Otsu segmentation. It estimates the density and form by fuzzy sets) of X and connect each fuzzy set of the parti-
combining a Canny edge detector and a morphological fi l- tion with a subconcept of the terminology, defi ned by that
ter and by evaluating the mean value of the curvatures of fuzzy set’s property. For example, having extracted the
the detected edges. It fi nds the hair’s length using H to- value 2.8H for the length of hair (s1), we determine that
gether with Dionysios’s proportion guidelines. s1 has a membership value of 0.7 for AboveEarsHair and
Next, the module determines information about the sa- 0.3 for BelowEarsHair (see the graph in the semantic-
cred figure’s age by computing the interpretation box in Figure 2). The
number of wrinkles on the forehead defi nition of the fuzzy partitions is
(s2). Using the Hough transform, the the module determines based on the heuristics extracted from
module defi nes the number of verti- Dionysios’s manual.
cal and horizontal lines in the fore- information about
head, which represent the wrinkles. Knowledge representation
A young sacred figure has zero or few
wrinkles, whereas an older sacred fig-
the sacred figure’s age and reasoning
This subsystem consists of terminolog-
ure has many.
The cheek (s3) provides additional
by computing the ical and assertional knowledge and a
reasoning engine.
information about age. According to
Dionysios, a small number of wrin-
number of wrinkles These types of knowledge are
the basic components of a knowl-
kles on the larger (left or right) part edge-based system based on DLs, 9 a
of the cheek indicates a young sacred on the forehead. structured knowledge-representation
figure. Conversely, an older sacred fig- formalism with decidable-reasoning
ure’s cheek will have a greater number algorithms. DLs have become popular,
of wrinkles (and therefore a greater proportion of the colors especially because of their use in the Semantic Web (as in
used for shadows). OWL DL, for example). DLs represent a domain’s impor-
Finally, the module defi nes the feature sets for the mus- tant notions as concept and role descriptions. To do this,
tache (s4) and beard (s5). Using techniques similar to those DLs use a set of concept and role constructors on the ba-
for hair analysis, it analyzes the mustache’s color and shape. sic elements of a domain-specific alphabet. This alphabet
It also extracts information on the beard’s length (short, not consists of a set of individuals (objects) constituting the do-
very long, long, down to the waist, or down to the knees), main, a set of atomic concepts describing the individuals,
color (fair or dark), form (straight, wild, or curly) and shape and a set of atomic roles that relate the individuals. The
(sharp, circle-like, circle, wide, or fluffy). concept and role constructors that are used indicate the ex-
pressive power and the name of the specific DL. Here, we
Semantic interpretation. This module represents the ex- use SHIN, an expressive subset of OWL DL that employs
tracted information in terms of the sacred-figure ontology. concept negation, intersection, and union; existential and
Representing the face and its main components is universal quantifiers; transitive and inverse roles; role hier-
straightforward: we simply write assertions such as s1: archy; and number restrictions.
Hair (see the semantic-interpretation box in Figure 2).
However, each component’s properties are vague. For ex- Terminological and assertional knowledge. The termino-
ample, you can’t always directly determine on the basis of logical knowledge is an ontological representation of the
length whether the sacred figure has long hair. This dif- knowledge in Dionysios’s manual. The assertional knowl-
ficulty is due to the vagueness of the defi nition of length edge is a formal set of assertions describing a specific icon
and the imprecision introduced by feature extraction (see, in terms of the terminological knowledge. For example, the
for example, the result of the segmentation of the hair in terminology contains axiomatic knowledge such as “a Jesus
the feature extraction box in Figure 2). Nevertheless, we Face is depicted as a young face with long, straight, thick,

March/april 2009 www.computer.org/intelligent 39


AI and Cultural Heritage

black hair and a short, straight, thick, black beard,” whereas …


“the face depicted in the icon is a face with long hair in some
degree” is a relative assertion. The taxonomy is constructed using the properties mea-
The knowledge base has two main components. The sured as features during feature extraction. Here are some
terminological component (TBox) describes the relevant examples of terminological axioms defining the concept
notions of the application domain by stating properties of hierarchy:
concepts and roles and their interrelations. TBox is actu-
ally a set of concept inclusion axioms of the form C v D, AboveEarsHair v Hair
where D is a superconcept of C, and concept equivalence
axioms of the form C ≡ D, where C is equivalent to D. OldCheek v Cheek
The assertional component (ABox) describes a concrete
world by stating properties of individuals and their spe- OldForehead v Forehead
cific interrelations. To deal with the vagueness introduced
in our case, we use f-SHIN, a fuzzy extension of SHIN. 3 In HairyBeard v Beard
f-SHIN, ABox is a finite set of fuzzy assertions of the form
〈a : C⋈n〉, 〈(a, b) : R⋈n〉, where ⋈ stands for ≥, >, ≤, or Restrictions concerning the taxonomy mainly describe dis-
<, for a, b ∈ I. Intuitively, a fuzzy assertion of the form 〈a jointness and exhaustiveness, forming axioms such as
: C ≥ n〉 means that the membership degree of a to concept
C is at least equal to n. (A formal definition of the syntax AboveEarsHair v ¬BelowEarsHair
and semantics of fuzzy DLs appears elsewhere. 3,10)
In the Byzantine-icon-analysis domain, the individuals Hair v AboveEarsHair t BelowEarsHair
are the segments that the semantic-segmentation module
extracted. The DL concepts classifying the individuals are Using these concepts, we define more expressive descrip-
determined by the properties measured as features during tions that specify some specific face characteristics on the
feature extraction (for example, BlackHair and AboveE- basis of Dionysios’s manual. For example, we define a
arsHair). The DL roles mainly describe partonomic young man’s face as
(mereological) hierarchies and spatial relations (for ex-
ample, the axiom isLeftOf– = isRightOf indicates that YoungMansFace ≡ ∃hasComponent.BlackHair u
the role isLeftOf is the inverse of isRightOf, and the ∃hasComponent.(NotVeryLongBeard u BlackBeard)
axioms Trans(isLeftOf) and Trans(isAboveOf) indicate u ∃hasComponent.YoungForehead u ∃hasPart.
that isLeftOf and isRightOF are transitive). YoungCheek u ∃hasPart.BlackMustache
The terminology’s main concepts are Figure, Face, and
FaceComponent. They’re connected with the partonomic We similarly define concepts specifying the characteris-
role hasPart and its subroles hasSegment and hasCom- tics of some popular sacred figures. For example, we define
ponent, to indicate whether a segment is a subsegment of JesusFace as
another segment and so on. Using these roles, we describe
the partonomic hierarchy as follows: JesusFace ≡ YoungMansFace u ∃hasComponent.
(BraidOnTheLeftShoulder u HairyHair
Face ≡ ∀hasComponent.FaceComponent u StraightHair) u ∃hasComponent.
(CircleLikeBeard u StraightBeard)
Figure ≡ ∃hasSegment.Face
The reasoning engine. This module uses the terminologi-
FaceComponent v Beard t Cheek t Forehead t Hair cal knowledge and the assertions to recognize the sacred
t Mustache figure. Our system uses FiRE, the Fuzzy Reasoning Engine
(www.image.ece.ntua.gr/~nsimou/FiRE), which fully sup-
We describe relevant assertions as follows: ports f-SHIN.
Figure 3 illustrates how the engine performs reasoning
〈s:Face ≥ 1.0〉 and classification for recognizing Jesus as the sacred figure.
We use FiRE’s greatest lower bound (GLB) reasoning ser-
〈s1:Hair ≥ 1.0〉 vice, which evaluates an individual’s greatest possible degree
of participation in a concept. The concepts of interest in this
〈(s, s1):hasSegment ≥ 1.0〉 case, YoungMansFace and JesusFace, only include conjunc-

40 www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS


YoungMansFace � �hasComponent.
BlackHair � hasComponent.
�.(NotVeryLongBeard � BlackBeard)
� �hasComponent.YoungForehead
� �hasPart.YoungCheek � �hasPart.
BlackMoustache

JesusFace � YoungMansFace � JesusFace


s1 �hasComponent.(BraidOnTheLeftShoulder
s2 � HairyHair � StraightHair)
� �hasComponent.(CircleLikeBeard
s3 � StraightBeard)
TBox
Reasoning
s4 engine
s: Face
s5 s1: Hair …
(s, s1): hasComponent …

s1: BraidOnTheLeftShoulder = 0.78


s1: HairyHair = 0.96
s1: StraightHair = 0.86
s1: BlackHair = 0.88
s2: YoungForehead = 0.82
s3: YoungCheek = 0.87
s4: StraightMoustache = 1.0
s4: BlackMoustache = 0.83
s5: CircleLikeBeard = 0.79
s5: StraightBeard = 0.92
s5: NotVeryLongBeard = 1.0 Knowledge
s5: BlackBeard = 0.84 ABox representation
and reasoning

Figure 3. A reasoning example. The extracted information from image analysis constitutes the assertional component (ABox) of
the knowledge base; the terminological component (TBox) is defined on the basis of Fourna’s specification. These components
form the input to the fuzzy reasoning engine, which infers information about the icon.

tions, so we use the fuzzy intersection operator for the mini- eyes and nose. The base-color-analysis submodule defined
mum. So, the GLB of segment s (the face) in YoungMansFace the color model used by the icon’s artist, which the fea-
is 0.82, whereas the GLB of segment s in JesusFace is 0.78. ture extraction module later used. Using the nose’s height,
In other words, the extracted features indicate that the face which can be easily determined, and taking into account
depicted is YoungMansFace to a degree of 0.82 and is Je- the manual’s rules, the face-component-detection submod-
susFace to a degree of 0.78. ule estimated the positions of the background and fore-
ground pixels for every part of the face. These positions
Results constituted the input to the graph-cut algorithm. This sub-
We evaluated our system on a database, provided by the module achieved 96 percent accuracy.
Mount Sinai Foundation in Greece, containing 2,000 digi- The feature extraction module extracted features for
tized Byzantine icons dating back to the 13th century. The every icon segment. Then, using the fuzzy partitions, the
icons depict 50 different characters; according to Diony- semantic-interpretation module interpreted each segment’s
sios, each character has specific facial features that makes specific features and properties and the relationship of face
him or her distinguishable. parts. Thus, that module determined each feature’s degree
Evaluation of the Byzantine-icon-analysis subsystem pro- of membership in a specific class, thereby creating an image
duced promising results. The subsystem’s mean response description.
time was approximately 15 seconds on a typical PC. To evaluate overall system performance, we used
In the semantic-segmentation module, the face detection precision and recall. Table 1 (on the next page) pres-
submodule reached 80 percent accuracy. In most cases, the ents results for 20 of the sacred-figure classes. The
failure occurred in icons with a destroyed face area. If the more distinctive facial features the sacred figure in a
submodule detected the face, it almost always detected the class contained, the better our method performed (for

March/april 2009 www.computer.org/intelligent 41


AI and Cultural Heritage

Table 1. Evaluation of the icon classification system.

No. of images
Class No. of icons in class No. of images classified correctly classified Precision Recall

Jesus 70 67 58 0.87 0.83


Virgin Mary 60 54 43 0.80 0.72
Peter 50 44 38 0.86 0.76
Paul 50 44 39 0.89 0.78
Katerina 40 35 30 0.86 0.75
Ioannis 40 37 29 0.78 0.73
Luke 40 33 28 0.85 0.70
Andrew 40 33 29 0.88 0.73
Stefanos 30 27 24 0.89 0.80
Konstantinos 40 36 30 0.83 0.75
Dimitrios 40 36 31 0.86 0.78
Georgios 40 37 32 0.86 0.80
Eleni 40 37 31 0.84 0.78
Pelegia 20 17 15 0.88 0.75
Nikolaos 40 38 33 0.87 0.83
Basileios 40 37 31 0.84 0.78
Antonios 30 27 20 0.74 0.67
Eythimios 25 23 18 0.78 0.72
Thomas 35 32 26 0.81 0.74
Minas 20 17 14 0.82 0.70

example, for the Jesus class). Conversely, our method sify paintings on the basis of unmixed colors and dense
performed worse on figures with similar facial features, textures.
such as women.
Acknowledgments
Our research has been partly supported by the MORFES (extraction
and modeling of figure in Byzantine paintings of Mount Sinai monu-

W e’re developing a version of our system that takes


into account information about clothes associated
with specific types of figures (such as apostles, kings, an-
ments using measures) project, funded by the EU and the Greek Sec-
retariat for Research and Development. We thank the Mount Sinai
Foundation for providing access to the rich data set and especially
archeologist Dimitris Kalomirakis for his participation in knowl-
gels, and monks) and clothing accessories (kerchiefs and edge extraction and his continuing support during this research.
so on) and nonclothing accessories (books, crosses, and so Finally, we thank the anonymous reviewers for their constructive
comments.
on) associated with specific sacred figures. For example,
the system could incorporate the knowledge that icons of
the Virgin Mary depict her wearing a kerchief with three References
stars. Such information might help in classifying figures 1. T. Berners-Lee, J. Hendler, and O. Lassila, “The Semantic
Web,” Scientific American, May 2001, pp. 34–43.
with similar facial features.
2. Dionysios tou ek Phourna, Hermeneia tis zografikis technis
Furthermore, an integrated version of our system could (interpretation of the Byzantine art), B. Kirschbaum, 1909.
eventually date icons and attribute them to a particular 3. G. Stoilos et al., “Reasoning with Very Expressive Fuzzy De­­
artist or artistic school, on the basis of features specific scription Logics,” J. Artificial Intelligence Research, vol. 30,
to each artistic period. Such information can help classify no. 5, 2007, pp. 273–320.
4. P. Viola and M.J. Jones, “Robust Real-Time Face Detection,”
an icon even if the specific sacred figure is difficult to rec- Int’l J. Computer Vision, vol. 57, no. 2, 2004, pp. 137–152.
ognize. We’re also investigating applying our method to 5. S. Asteriadis et al., “An Eye Detection Algorithm Using Pixel
other art movements such as impressionism, trying to clas- to Edge Information,” Proc. 2nd IEEE-EURASIP Int’l Symp.

42 www.computer.org/intelligent IEEE INTELLIGENT SYSTEMS


Th e A u t h o r s
Paraskevi Tzouveli is pursuing her PhD in image and video analysis at the National Tech-
Control, Communications, and Sig­ nical University of Athens. Her research interests include image and video analysis, in-
nal Processing (ISCCSP 06), 2006; formation retrieval, knowledge manipulation, e-learning, and digital cultural heritage
www.eurasip.org/Proceedings/Ext/ systems. Tzouveli received her Diploma in electrical and computer engineering from the
ISCCSP2006/defevent/papers/cr1124. National Technical University of Athens. Contact her at tpar@image.ntua.gr.
pdf.
6. N.A. Otsu, “A Threshold Selection Nikos Simou is a PhD student in the National Technical University of Athens Depart-
Method from Gray-Level Histograms,” ment of Electrical and Computer Engineering and a member of the Image, Video, and
IEEE Trans. Systems, Man, and Cyber- Multimedia Systems Laboratory’s Knowledge Technologies Group. His research interests
netics, vol. 1, no. 9, 1979, pp. 62–66. include knowledge representation under certain and uncertain or imprecise knowledge
7. Y. Boykov and M.-P. Jolly, “Interactive based on crisp and fuzzy description logics, ontologies, and the implementation and op-
Graph Cuts for Optimal Boundary and timization of reasoning systems and applications for the Semantic Web. Contact him at
Region Segmentation of Objects in N-D nsimou@image.ntua.gr.
Images,” Proc. Int’l Conf. Computer
Vision (ICCV 01), vol. 1, IEEE Press, Giorgios Stamou is a lecturer in the National Technical University of Athens School of
2001, pp. 105–112. Electrical and Computer Engineering and a member of the university’s Image, Video, and
8. R.C. Gonzalez and R.E. Woods, Digi- Multimedia Systems Laboratory, heading the lab’s Knowledge Technologies Group. His
tal Image Processing, 3rd ed., Prentice main research interests include knowledge representation and reasoning, theory of uncer-
Hall, 2008. tainty, machine learning, semantic interoperability, and digital archives. Contact him at
9. F. Baader et al., The Description Logic gstam@cs.ntua.gr.
Handbook: Theory, Implementation
and Applications, Cambridge Univ. Stefanos Kollias is a professor in the National Technical University of Athens (NTUA)
Press, 2002. School of Electrical and Computer Engineering and the director of the Image, Video, and
10. G. Stoilos et al., “Uncertainty and the Multimedia Systems Laboratory. His research interests include image and video process-
Semantic Web,” IEEE Intelligent Sys- ing, analysis, retrieval, multimedia systems, computer graphics and virtual reality, artificial
tems, vol. 21, no. 5, 2006, pp. 84–87. intelligence, neural networks, human-computer interaction, and digital libraries. Kollias
received his PhD in signal processing from NTUA. He’s a member of the European Com-
mission expert group in digital libraries and of the EC committee on semantic interoper­
For more information on this or any other ability, leading NTUA participation in DL projects, EDLnet, VideoActive, Europeana
computing topic, please visit our Digital Li- Connect, Europeana Network 1.0, and EU Screen. Contact him at stefanos@cs.ntua.gr.
brary at www.computer.org/csdl.

AdvertiSER Information
March/April 2009 • IEEE Intelligent Systems
Advertising Personnel Advertising Sales Midwest/Southwest Product:
Marion Delaney Representatives Darcy Giovingo
Phone: +1 847 498 4520 US East
IEEE Media,
Fax: +1 847 498 5911 Joseph M. Donnelly
Advertising Dir.
Email: dg.ieeemedia@ieee.org Phone: +1 732 526 7119
Phone: +1 415 863 4717 Recruitment:
Email: jmd.ieeemedia@ieee.org
Email: md.ieeemedia@
Mid Atlantic Northwest/Southern CA
ieee.org
Lisa Rinaldo Tim Matteson US Central
Phone: +1 732 772 0160 Phone: +1 310 836 4064 Darcy Giovingo
Marian Anderson
Fax: +1 732 772 0164 Fax: +1 310 836 4067 Phone: +1 847 498 4520
Sr. Advertising
Email: lr.ieeemedia@ Email: tm.ieeemedia@ Fax: +1 847 498 5911
Coordinator
ieee.org ieee.org Email: dg.ieeemedia@ieee.org
Phone: +1 714 821 8380
Fax: +1 714 821 4010 New England Japan US West
Email: manderson@ John Restchack Tim Matteson Lynne Stickrod
computer.org Phone: +1 212 419 7578 Phone: +1 310 836 4064 Phone: +1 415 931 9782
Fax: +1 212 419 7589 Fax: +1 310 836 4067 Fax: +1 415 931 9782
Sandy Brown Email: j.restchack@ieee.org Email: tm.ieeemedia@ieee.org Email: ls.ieeemedia@ieee.org
Sr. Business
Development Mgr. Southeast Europe Europe
Phone: +1 714 821 8380 Thomas M. Flynn Hilary Turnbull Sven Anacker
Fax: +1 714 821 4010 Phone: +1 770 645 2944 Phone: +44 1875 825700 Phone: +49 202 27169 11
Email: sb.ieeemedia@ Fax: +1 770 993 4423 Fax: +44 1875 825701 Fax: +49 202 27169 20
ieee.org Email: flynntom@ Email: impress@ Email: sanacker@
mindspring.com impressmedia.com intermediapartners.de

March/april 2009 www.computer.org/intelligent 43

Anda mungkin juga menyukai