as a PDF - International Journal of Interoperability in

IBIS – Interoperability in Business Information Systems
IBIS
International Journal of
Interoperability in Business
Information Systems
Issue 2 (2), 2007
ISSN: 1862-6378
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
Publisher:
University of Oldenburg
Department of Business Information Systems
D-26111 Oldenburg
Germany
Tel.: +49 441 798 4480
Fax: +49 441 798 4472
[email protected]
http://www.ibis-journal.net
ISSN: 1862-6378
Editors (in alphabetical order):
Sven Abels, Axel Hahn, John Krogstie, Kai Mertins,
Andreas L. Opdahl, Mathias Uslar
License:
All our articles are published under the Digital Peer Publishing License. This ensures
free distribution and it also guarantees the authors' rights that no content is
modified or reused without citing the names of authors and holders of rights and
the bibliographical information used. Additionally, the rights to use in physical
form, particularly the rights to distribute the work in printed form or on storage
media, are retained by the authors or other rights holders and are not covered by
this license.
The full license text may be found at http://www.ibis-journal.net
Scope:
The capability to efficiently interact, collaborate and exchange information with
business partners and within a company is one of the most important challenges of
each enterprise, especially forced by the global markets and the resulting
competition. Today, many software systems are completely isolated and not
integrated into a homogeneous structure. This makes it hard to exchange
information and to keep business information in sync. Interoperability can be
defined as the ability of enterprise software and applications to interact. In Europe
between 30-40% of total IT budgets is spent on issues tied to Interoperability.
This journal aims in exchanging and presenting research activities in the area of
creating interoperability in business information systems. Ambition of this journal
is to get an overview of current research activities as well as to offer a broad
discussion in selected areas of the interoperability of heterogeneous information
systems. It is proposed to connect research experts from this domain and to
exchange ideas and approaches. It is our goal to connect latest research results
with real-world scenarios in order to increase interoperability in business
information systems.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Review Board:
Submitted articles of this journal are reviewed by the members of the IBIS
review board. At the date of this issue, the review board contains the
following persons in alphabetical order:
Sven Abels, TIE Nederland B.V., Germany
Antonia Albani, Delft University of Technology, Netherlands
Bernhard Bauer, University of Augsburg, Germany
Arne-J. Berre, SINTEF, Distributed Information Systems, Norway
Jean Bézivin, University of Nantes, Brittany, France
Flavio Bonfatti, University of Modena, Italy
Frank-Dieter Dorloff, University of Duisburg-Essen, Germany
Peter Fettke, Institute for Information Systems, DFKI, Germany
Michael Goedicke, University of Duisburg-Essen, Germany
Jan Goossenaerts, University of Eindhoven, Netherlands
Norbert Gronau, University of Potsdam, Germany
Axel Hahn, University of Oldenburg, Germany
Wilhelm Hasselbring, University of Oldenburg, Germany
Martin Hepp, Digital Enterprise Research Institute (DERI), Germany
Willem-Jan van den Heuvel, Tilburg University, Netherlands
Paul Johannesson, Royal Institute of Technology, Sweden
John Krogstie, Trondheim, Norges Teknisk-Naturvitenskapelige
Universitet, Norwegen
Kai-Uwe Kühnberger, University of Osnabrueck, Germany
Kai Mertins, Fraunhofer Institut für Produktionsanlagen und
Konstruktionstechnik, Germany
Michele Missikoff, IASI, National Research Council (CNR), Italy
Andreas Lothe Opdahl University of Bergen, Norway
Frank Teuteberg, University of Osnabrueck, Germany
Mathias Uslar, OFFIS - Institut für Informatik, Germany
Martin Zelm, CIMOSA Association e.V., Germany
The editors thank all reviewers for their help! This IBIS issue wouldn’t be possible
without the support of our reviewers that help us to ensure a high quality.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Content
Editorial………………………………………………………………………………………………………………………7
PLASMA: A Platform for Schema Matching and Management…………………………………….9
Using Ontologies to Model and Understand Product Development………………………….21
An Enterprise Application Integration Architecture Supporting
Ontology Based Approach for B2B Collaboration………………………………………………………39
IESA-Call…………………………………………………………………………………………………………………….65
Call for Articles…………………………………………………………………………………………………………71
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Editorial
Dear reader,
we would like to welcome you to the second issue of the international journal of
Interoperability in Business Information Systems (IBIS).
By writing this editorial, we can announce that due to much request and some good
input, we managed to bring up a new journal system which should better meet
your and our needs towards Open Access publishing. Furthermore, we would like to
welcome our new co-editors, Kai Mertins, John Krogstie and Andreas L.Opdahl.
Professor Kai Mertins (class 1947) is a director of the IPK at Fraunhofer-Institut for
production plants and construction technology. Since 1998 he is additionally been a
honorary professor at the technical University of Berlin. After electromecanician
studies with emphasis measuring and control engineering, he received a engineer
(degree) in Hamburg. Subsequently, he went on to the industrial engineer studies
at the technical University of Berlin (Dipl. engineer). It closed its graduation with
Professor Dr. h. C mult. Dr. Ing. G. Trace over "Steuerung computer-guided
Fertigungssysteme" off. His focal tasks today are: Enterprise modelling, enterprise
management, production organization, knowledge management, factory planning,
job controlling, manufacturing control systems, coworker qualification. Professor
Mertins has comprehensive experiences as a manager of international projects with
European and asiatic consortia.
John Krogstie (http://www.idi.ntnu.no/~krogstie) has a PhD (1995) and a MSc
(1991) in Information Systems, both from the Norwegian University of Science and
Technology (NTNU). He is Professor in Information, Systems at IDI, NTNU,
Trondheim, Norway. He is also a senior advisor at SINTEF. He was employed as a
manager in Accenture 1991-2000. John Krogstie is the Norwegian Representative
for IFIP TC8 and vice-chair of IFIP WG 8.1 on information systems design and
evaluation, where he is the initiator and leader of the task group for Mobile
Information Systems. He was general chair of the CAiSE'07 the conference and has
published around 90 refereed papers in journals, books and archival proceedings
since 1991.
Andreas Lothe Opdahl is Professor of Information Systems Development and Head
of the Information Systems (IS) Group at the Department of Information Science
and Media Studies at the University of Bergen, Norway.
His main research interest is development, maintenance and use of enterprise
models which support IS development and operations, in particular enterprise
models for interoperability between enterprises and their information systems. He
focuses on models and languages that integrate the various perspectives of
stakeholders tightly and that may embed multiple partial models for different uses
and purposes.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 1 (1), 2007
Selected articles for this issue
For this issue, we accepted about 15% of all submissions.The following titles are
included in issue 2 of our second volume:
•
PLASMA: A Platform for Schema Matching and Management
•
Using Ontologies to Model and Understand Product Development
•
An Enterprise Application Integration Architecture Supporting Ontology
Based Approach for B2B Collaboration
We hope you will appreciate the journal and the articles we selected for this issue.
Best regards,
your IBIS Editorial Board
(Axel Hahn, Sven Abels, Mathias Uslar,
Kai Mertins, John Krogstie and Andreas L. Opdahl)
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
PLASMA: A Platform for Schema Matching and Management
Aïcha-Nabila Benharkat
LIRIS-INSA de Lyon
National Institute of Applied Sciences of Lyon
Villeurbanne, FRANCE
[email protected]
Rami Rifaieh
San Diego Supercomputer Center
University of California
San Diego, USA
[email protected]
Sana Sellami, Mohamed Boukhebouze, Youssef Amghar
LIRIS-INSA de Lyon
National Institute of Applied Sciences of Lyon
Villeurbanne, FRANCE
[email protected] [email protected],
[email protected]
Abstract: This paper introduces an XML Schema management platform that promotes
the use of matching techniques to fulfill the requirements of data integration and data
exchange. The existing platforms, in the market, deal only with graphical but not
automatic matching. Several matching algorithms were suggested, by different
researchers, to automate the correspondences discovery between XML Schemas. These
algorithms use a wide range of approaches and matching techniques with
systematically many parameters and weights to be adjusted manually. In our platform
(PLASMA) we suggest a matching algorithm that uses constraints and a combination of
structural and description matching. In addition to our innovative matching algorithm,
we adopt an automating tuning for the various structural parameters used and propose
a Mapping Expression model along with a collection of transformation operators.
Hence, PLASMA enables a semantic matching process followed by a visual mapping
process to optimize the translation process between exchanged documents or
integrated schema.
Introduction
A growing number of applications are giving enterprises new opportunities to
implement cost-effective translation and integration solutions. However, the
problem resides in simplifying, automating, and reusing theses solutions. Building
manual processes or add-hoc scripts are time consuming that fail to provide the
conversions and transformations required between applications. In order to handle
these challenges, a framework for document and data translation should be
provided.
For data integration and data exchange, we should distinguish between two
processes at the schema level: matching and mapping. Firstly, schema matching
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN: 1862-6378
IBIS – Issue 2 (2), 2007
helps to identify the correspondent elements in two schemas. Afterward schema
mapping with expressions, functions, conditions, etc., is required to define how
the matched elements can be mapped [RIF03]. The input of the mapping process is
the list of matched elements and the output is the individual mapping expression of
target elements.
We suggest avoiding the manual process by generating the transformation
automatically on the basis of a set of correspondences generated by a matching
process called EXSMAL and then visually editing these matched with expressions
and mapping rules. EXSMAL is a matching algorithm that we have initially studied
for matching components of two EDI messages [CHU05]. We use this algorithm in
PLASMA, which covers the matching, mapping and code generation processes.
PLASMA aims at creating a new generation of tools with goals in mind to simplify
the task of users working on document exchange. It can be considered as semiautomatic tool for identifying matching between schemas and generating
automatically the data transformation scripts.
The paper is organized as follows: the section 2 reviews related works in the
domain of data exchange. We give special attention to the effort in the domain of
schema matching and schema mapping. After explaining different related works,
we present in section 3 the PLASMA platform and its four layers (Matching,
Filtering, Mapping Transformation and Integration). We start by presenting EXSMAL
algorithm then we show how to filter the matching set produced by EXSMAL in
order to eliminate irrelevant correspondences. We explain then the principles of
ATA: Algorithm of Tuning Automation that leads to reduce human efforts when
adjusting parameters. The following section describes the third layer (Mapper),
where we use the previous filtered set and a dynamic ontology to generate usable
transformation rules. This section ends with a brief discussion of the
transformation layer. Finally, the section 4 concludes the paper.
Related works
XML is increasingly becoming the used format for B2B and B2C with ebXML,
semantic web techniques, and web Services with SOAP. B2B integration servers
include tools for XML transformation starting from a graphical but manual
definition of correspondences. For instance Visual XML Transformation Tool (IBM)
defines graphically XSLT stylesheets to make easier the transformation of XML
documents. However, many visual tools rely on domain knowledge expertise and
technical skills for performing the transformation task.
In this context, schema matching has attracted recently considerable attention in
research. The XML schema matching can be automatic and used for building
complex transformation of XML documents. This process of matching is unavoidably
required for building a platform of document exchange.
In schema matching, we can identify algorithms working with Meta data schemas
(DTD, database schema, XML Schema) such as Cupid [MAD01]. Similarity Flooding is
a structural matching algorithm used for XML Schema, SQL DDL, RDF schema, UML
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
and OEM [MEL02]. This graph based algorithm uses the idea of node influence by
adjacent node [MEL03]. XClust in [MON02] is an integration strategy based on (DTDs
or XML, schema) Clustering. In the ARTEMIS project [BER99] schema integration is
dealt with attributes clustering. The algorithm operates on a hybrid model
(relational and object oriented). The affinity between names is based on a
linguistic thesaurus (synonymy, hypernymy). The affinity between types is based on
a generic table of data type’s similarity. Finally, COMA [DO 02] and COMA++
[AUM05] are frameworks for schema matching. COMA allows using many algorithms
with different techniques for results rendering. The schemas are translated into
acyclic graphs that can be performed simultaneously by different matching module.
Intermediary results can be reused with possibilities of selection and aggregation.
This fully-personalized feature makes it difficult to average user to identify the
best combination between different modules.
On the other hand, schema mapping is studied separately than schema matching.
Clio [MIL01], is presented as a semi automatic tool allowing mappings among
relational schemas. It takes into account the constraints between the source
attributes and the global, schema ones. The correspondences are supposed to be
already identified and expressed in a first order logics language. Protoplasm
[BER04] inspired by COMA's approach, aims at incorporating different existing
approaches in a single platform for customizable schema mapping. The latter uses
the SMM (Schema Matching Model Graph) as an internal representation in which are
translated Meta models as (SQL DDL, ODMG, and XSD). SCIA [WAN04] is yet another
platform for matching, mapping and code generation. It offers a generic platform
to deal with meta-data (XML Schema, DTD) from scientific domain.
We can distinguish in the preceding review that not every matching algorithm has
created a mapping platform with user interface and case studies and none of these
approaches is applied to document exchange in E-Commerce and EDI. We aim in
this work to study an integrated approach that deals essentially with this domain of
interest. We showed in [RIF03] the drawback of not using a dedicated approach to
tackle the problem of document translation and document exchange for EDI. The
similarity between our approach and the cited ones is obvious in term of content.
However, the difference is shown through the quality of matching algorithm used
for EDI document and the usefulness of suggested mapping platform to cover the
state of the art needs for document exchange in E-Commerce (EDI, B2B, etc.).
The PLASMA Framework
PLASMA provides a general framework to perform data exchange with minimal
effort on the user side having limited technical skills. In order to build a complete
framework, we focused firstly on related studies including mapping expression
model and matching algorithm. We studied mapping expressions that are needed
between two schemas and defined a generic model to cover these expressions. We
suggested the EXSMAL algorithm to simplify finding the matches between
schemas.In the framework PLASMA (Figure.1), two XML Schemas entries are
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN: 1862-6378
IBIS – Issue 2 (2), 2007
transformed into trees will be treated through multiple steps, that can be
described as follow:
The matching process yields a set of (1-1, 1-n, n-m) correspondences. A filtering
process is then performed in order to eliminate irrelevant correspondences. This
filtering provides an extension of EXSMAL Algorithm with two types of similarities
(path and internal) and aims at refining matching to 1-1 cardinalities in the case of
irrelevant similarities. Afterward, a mapping process is applied with the user
intervention to clean the resulted matched values. This process includes as well
the definition of mapping rules with respect to mapping expression model. Finally a
transformation process (i.e. code generation) is carried by the system in order to
offer the executable mapping to apply on data instances. In the following, we
describe briefly these steps.
Figure 1: PLASMA Architecture
The matching Layer
The input of our framework is a couple of XML schemas (e.g. Schema 1 and Schema
2). The matching algorithm calculates a set of correspondences EC1. We can
consider as examples the two schemas shown in Figure.1. EC1 is represented by the
set of lines joining source schema S1 on the left side and the target Schema T on
the right side. We assume in this step that user has identified the best set of
coefficient to be used in the matching algorithm.
EXSMAL
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
We have introduced in [CHU05] a new matcing algorithm, taking into consideration
the specific characteristics of EDI message's branching diagram expressed with XML
Schema. Our choice for the XML Schema was motivated by its potential to define
the structure and semantics of an EDI message. EXSMAL is fully described in [RIF05]
with a case study. It realizes matching of XML schemas taking into account textual
description, data types and context description of elements. The similarity
computation is based on a pairwise element similarity (basic and structural) and a
filtering process eliminating the correspondences with a value below the threshold
thraccept given by the user (Figure.2).
The Basic similarity basicSim is computed as the weighted sum of the textual
description (using information retrieval technique) and data type similarity(using a
static matrix defining the XML schema primitive data type affinity).
The Structural similarity structSim is computed by using two modules: the
structural neighbors computing and the aggregation function agg. The structural
neighbors of an element e is a quadruplet <ancestor(e), sibling(e),
immediateChild(e), leaf(e)>. The structural similarity value of two elements s
(source schema) and t (target schema) depends on the similarity value resulting
from the comparison of each pair of structural neighbor’s items. The similarity
value of each structural neighbors items’ pair is computed by using the function
agg(M, thr) which take a matrix M and a threshold value thr ∈ [0, 100] as input. It
returns the aggregated value of the input matrix M.
Figure.2: Snapshot of the matching generated by EXSMAL
The Pair-wise element similarity is then computed as the weighted sum of the
basic similarity value and the structural similarity value. It’s proposed as the final
similarity value for a pair of elements in our approach. The pair wise element
similarity of s and t is computed by the following formula:
Similarity(s,t)= basicSim(s, t)*coeff_base + structSim(s, t)*coeff_struct
Finally, a filtering can be done according to a threshold thraccept.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN: 1862-6378
IBIS – Issue 2 (2), 2007
Constraints Management
We have proposed in [SEL06] an algorithm completing the matching step by using
the constraints in XML Schema. We have identified the different constraints that
can be applied on the elements (sources and destinations) of schemas and
established a classification of these constraints (intrinsic and process). We have
also proposed an algorithm which takes into consideration the intrinsic constraints
and returns the penalty of violation of these constraints. The algorithm was
integrated with EX-SMAL to determine more efficiently the degree of similarity
between elements of XML Schemas.
Tuning of Matching Algorithm
In our work [BOU07], we have tried to automate the tuning of parameters existing
within the matching systems. This will help in reducing the human intervention,
from developers and users, and also to reach better matching quality (precision,
recall, and overall). Indeed, our approach consists in finding the best parameters
to use according to the studied schemas and their characteristics.
In particular, we believe that schema topology (number of nodes, number of paths,
depth, etc.) is directly related to the quality of the matching with respect to
structural parameters. Indeed, matching algorithm uses structural parameters (e.g.
for COMA [DO 02]: P_Parents, P_Children, P_Leaves, etc.) that can influence the
matching results. In order to discover some of the rules that exist between the
schema topology and the values of these parameters, we have proposed the
following approach:
Firstly, we have started by applying series of experiments, hereafter called
benchmark, with three selected algorithms (COMA [DO 02], EXSMAL [RIF05], SCIA
[WAN04]). We have collected schemas and created manual correspondences
between them. Afterward, we have repeatedly applied the automatic matching on
the schemas in the aim of finding a set of consistent mathematical model (or
relations). The consistency of the relations with a topological pattern can help us
refine the values of structural parameters.
Secondly, we have created a data store that includes all the rules empirically
deduced form the benchmark. Then, we have defined a tuning algorithm that uses
these stored statistical relations and apply them on new schemas. In other words,
the algorithm will seek in the entered schemas some topological patterns similar to
those stored in the data store and suggest the best set of parameters’ values for a
specific algorithm.
The two steps of our approach are depicted in Figure.3. The input of the first step
is a set of XML Schema and a set of tools. At first, the benchmark identifies the
structural parameters to study with each tool. Then it tries to apply a common
methodology using heavily data analysis and statistical methods to verify and test
models. The output of the first step is a set of acceptable models also called rules
or relations. The second part of our approach aims at suggesting values to be used
for a specific matching task.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Figure 3 : Steps of the Automatic Tuning Approach
To adjust the various structural parameters used within the studied matching
systems, we proposed in [BOU07] an algorithm for automatic tuning that uses the
results of benchmark and offers a better precision for the matching task.
Automatic Tuning Algorithm (ATA) has as input two schemas and the tool that we
want to optimize its structural parameters. The algorithm initially will calculate
the number of elements, numbers paths and the depth of each diagram. Then it
chooses, among the relations saved, from the benchmark, the relation to be
applied with respect to the characteristics of the schemas. Finally, ATA calculates
and displays the suggested values for each structural parameter involved in the
tool.
Besides EXSMAL, we applied a final test with COMA to check the results coming
from the tuning algorithm. The experiment consists in applying COMA to match
between two given schemas (never used in the test before). We calculated the
precision and Recall of the results obtained, figure 4 shows that the precision is
equal to 0,23 and that Recall is equal to 0,6 using standard value for structural
parameters. We have repeated the same matching after using the value of
parameters suggested by the tuning algorithm. The precision has improved to 0.33
and Recall improved to 0,8,matching shown in Figure 5.
The filtering Layer
In [HER06] we have extended EXSMAL by extracting the Path Similarity followed by
the internal Similarity. We have suggested the refinement of EXSMAL based on
these two similarities in order to obtain 1-1 cardinalities and eliminate irrelevant
similarities. The later are used for automating XML documents transformation.
© IBIS – Issue 2 (2), 2007
IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN: 1862-6378
Figure 4. The matching before the execution of
the tool
Figure5 The matching after the execution of the
tool
The mapping Layer
Based one the Mapping Expression Model defined in [RIF04], we have suggested in
[SEL06] the XME (XML Mapping Expression Model). XME model is expressed using
XML Schema and defines all the properties of the mapping model as well as the
used transformation operators. It helps identifying how to generate the value of a
target element/attribute from sources elments. In this respect, a target elment is
expressed as follows:
−
Ai =
∑ [ f ( a ,a
i
r
p ,...,ae
),li ,ci ] + τ i = α i ( A ) +τ i
i
Where :
Ai represents an element/field of the target representation,
αi(A) is a mapping expression applied on multi-sources to generate a part of the
target field A or to generate a sub-field of the field A, αi= < fi , li ,ci> where fi is a
mapping function,
li is a set of filters for sources rows, we could find a filter li(ar) for each source
field ar,
Ci is a condition on the mapped value of the field A. This enables us to ignore
mapped results that do not satisfy the condition,
τi represents a generic function, independent of the source. The generic function
covers all the cases that the mapping expression model doesn’t generate. This
function can be a constant value or, an incremental function, an automatic
function or a skolem function in DB.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Figure 6: Example of generated mappings between XML Schemas
The transformation and integration Layer
The mapping layer generates a set of mapping as shown in Figure 6. Once these
semantic mappings (source, target and operation) are defined, a script is
automatically created by the platform in in data transformation language such as
XSLT or XQuery. The result will be used for translating XML documents instance
valid with the source schema to an instance of the target schema. This part of the
project is still in progress.
We have implemented the front end of PLASMA prototype. We integrate the
EXSMAL algorithm with restrcturing the code source of EXSMAL.We have realized a
new Graphical user Interface with adding functionality of mapping editing between
matched elements. We have added to the initial GUI of EXSMAL two panels in the
interface and a canavas that accepts the drag and drop of objects representing
schemas, mapping functions, multiple schemas, etc. This development has been
realised using Java (J2EE JDK 1.5, and Eclipse). A snapshot of the PLASMA
prototype is presented in Figure 7.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN: 1862-6378
IBIS – Issue 2 (2), 2007
Figure 7: Snapshot of the matching generated by EXSMAL
Conclusion
The use of XML and XML schema is a major trump for interoperability. However,
the matching and mapping process is unavoidably required before performing any
interoperability in data integration and data exchange. We have presented through
the PLASMA framework a matching and mapping approach for documents exchange
and pre-integration. Our EXSMAL algorithm has been placed in the first layer core.
Its execution provides a matching set with (1-1, 1-n and n-1) cardinalities. We have
also developed an extension to the matching algorithm EX-SMAL that uses the
constraints existing in XML Schema. We have proposed and implemented an
algorithm for automatic tuning that uses the results of a matching benchmark. We
have carried an experiment to evaluate the precision of our algorithm. This
experiment showed that the optimal values suggested by the algorithm lead to a
better precision and better Recall. We have tried then to filter this set in order to
obtain 1-1 correspondences between the elements of XML schema and eliminate
irrelevant ones with two new calculations: Path similarity and internal similarity.
This filtering algorithm is not implemented yet in the current version of PLASMA.
The final mappings are then expressed in the data model XME (XML Mapping
Expression) using operations such as: Connect and Rename for the simple matching,
Merge and Split. Then, transformation layer can be applied to generate final results
from source instance. The implementation of the back end of the PLASMA
framework is still work in progress. We intend in the future to improve the
algorithm of constraints including all the constraints (intrinsic and process). We aim
to improve the prototype of PLASMA and to integrate the algorithm of constraints
and all the operators of transformation. Future works cover also extending
matching techniques to large scale matching between ontologies and taxonomies.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
References
[AUM05]
[BER04]
[BER99]
[BOU07]
[CHU05]
[HER06]
[DO02]
[MAD01]
[MEL02]
[MEL03]
[MIL01]
[MON02]
[RIF03]
[RIF04]
[RIF05]
[SEL06]
[WAN04]
Aumüller,D.; Do,H.H.; Massmann,S.; Rahm,E.: Schema and Ontology Matching with
COMA++. Proceedings SIGMOD 2005 (Software Demonstration), Baltimore, June
2005, pp 906-908.
Bernstein,P.; Melnik,S.; Petropoulos, M.; Quix,C.: Industrial-Strength Schema
Matching. In SIGMOD Record, Vol. 33, No. 4, December 2004, pp 38-43.
Bergamaschi,S.; Castano,S.; Vincini,M.: Semantic Integration of Semi structured
and Structured Data Sources. SIGMOD Record 28(1), 1999, pp 54-59.
Boukhebouze,M.; Rifaieh,R.; Benharkat,A.N; Amghar.Y.: Benchmarking XMLSchema Matching Algorithms for Improving Automated Tuning.
In ACS/IEEE
AICCSA-07, Amman, Jordan, May 13-16, 2007, pp 917-925.
Chukmol,U.; Rifaieh,R.; Benharkat,A.N.: EXSMAL: EDI/XML semi-automatic Schema
Matching Algorithm. In IEEE CEC 2005, July 19-22, 2005, Technische Universität
München, Germany, pp 422-425.
Herzi, K.; Benharkat,A.N.; Amghar,Y.: Refinement of Correspondences in EXSMAL
for XML Document Transformation. In DEXA Workshops 2006: 304-308
Do,H.; Rahm,E.: COMA-A system for flexible combination of Schema Matching
approaches. In Proceeding of the 28th VLDB Conference, August, 2002, Hong Kong,
China,pp 610-621
Madhavan,J.; Bernstein, P. A.; Rahm, E.: Generic Schema Matching with Cupid. In
Proceedings of the 27th VLDB Conference, 2001, Rome, Italy, pp.49-58.
Melnik,S.; Garcia-Molina, H.; Rahm,E.: Similarity Flooding: A versatile Graph
Matching approaches. In Proceeding (ICDE), 2002, San Jose, California, USA, pp
117-128.
Melnik,S.; Rahm,E.; Berstein,P.A.: Rondo: A programming Platform for Generic
Model Management. In Proceedings of SIGMOD Conference, June 9-12, 2003, San
Diego, California, USA, pp. 193-204.
Miller, R.J. ; Hernandez, M.A; Haas, L.M. ; Yan, L.L. ; Ho, C.T.H.; Fagin, R.;
Popa,L. : The Clio Project: Managing Heterogeneity. In ACM SIGMOD Record,
Volume 30, Issue 1 (March 2001), ACM Press, New York, NY, USA 200, pp.7883, ISSN: 0163-5808.
Mong-Li,L.; Liang Huai ,Y.; Wynne,H.; Xia,Y.: XClust: clustering XML schemas for
effective integration. In CIKM 2002, pp 292-299
Rifaieh,R.; Benharkat,A.N.: An Analysis of EDI’s Message Translation and
Integration Problems. In proceedings of the International Conference on Computer
Science, Software Engineering, Information Technology, e-Business, and
Applications (CSITeA’03), June 5-7, 2003, Rio de Janeiro, Brazil.
Rifaieh,R.: Using Contextual Ontologies for Semantic Sharing within Enterprise
Information Systems, PhD Thesis, National Institute of Applied Sciences of Lyon
(INSA-Lyon), Dec. 2004, http://docinsa.insa-lyon.fr/these/2004/rifaieh/these.pdf
Rifaieh,R.; Chukmol,U. ; Benharkat,A.N.,: A Matching Algorithm for Electronic Data
Interchange. In TES 2005, pp 34-47.
Sellami,S.; Benharkat,A.N.; Rifaieh,R.; Amghar,Y.: Schema Matching for Document
Exchange: A Constraint Based Approach. In (SITIS' 2006), Hammamet TUNISIE, pp.
299-309. springer Verlag, LNCS series . ISSN 2-9525435-1. 2006
Wang,G. and al.: Critical Points for Interactive Schema Matching. In Proceedings of
Sixth Asia Pacific Web Conference, Hangzhou, China, April 2004, Springer LNCS,
Springer Verlag, volume 3007, 2004, pp 654-664.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN: 1862-6378
IBIS – Issue 2 (2), 2007
About the authors
AÏCHA-NABILA BENHARKAT received her Ph.D. degree from the National Institute of
Applied Sciences of LYON in 1990. Since 1992 she is Associate Professor at the
Computer Science Department (INSA of LYON). Since 1993 she worked on the
Integration of heterogeneous databases using Description Logics. Currently, her
research interests include the interoperability domain, matching techniques in
small and large scale and ontologies.
RAMI RIFAIEH is a Research Fellow in San Diego Supercomputer Center; he received
his Ph.D. degree in Computer Science from National Institute of Applied Sciences of
Lyon in 2004. He is working on Data and Schema Integration, Matching/Mapping
Techniques, and resource integration in Life Sciences.
SANA SELLAMI is a PhD candidate at the INSA (National Institute of Applied Sciences
Lyon France). She received her Master degree in Computer Science from INSA-Lyon
in 2006 , and Engineering degree form INSAT, Tunisia in 2006. Her research interest
spans over the Large Scale Schema Matching domain.
MOHAMED BOUKHEBOUZE received his Master degree from Computer Science from
National Institute of Applied Sciences of Lyon in 2006 and currently is pursuing a
PhD in formal verification and business processes management.
YOUSSEF AMGHAR is Professor at the National Institute of Applied Sciences of LYON
since 2003. His research interests cover data engineering, interoperability and
formal models. His main field of teaching concerns information systems.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Using Ontologies to Model and Understand Product
Development
Axel Hahn, Kevin Hausmann, Stefan Häusler, Jan Strickmann
OFFIS – Institute for Information Technology
26121 Oldenburg, Germany
{hahn | hausmann | haeusler | strickmann} @ offis.de
Abstract: New product development is crucial for almost any company’s future. At the same time,
it is one of the most knowledge-intensive and complex business processes. Nowadays the high
complexity of most products requires the cooperation of engineers from various fields such as
mechanical engineering, electronics and software design to achieve outstanding results. Thus,
project management becomes a most important issue to guide a development project through
conflicting goals, misunderstandings or scarce resources.
To ease the interoperability between engineering domains, to understand the different phases of
development projects and translate technical properties into meaningful key figures for project
assessment, ontologies are a promising technology. The semantic approach they provide can help to
bridge the inconsistencies between the many domains involved and to deliver one integrated,
flexible model for engineering projects and processes.
This article shows how partial development information from domain-specific IT-systems is
converted into an integrated semantic model. Therefore, we discuss generic development processes
and ontologies representing individual phases and partial models. In this paper we focus on the task
of data integration when dealing with a set of partial product or project models: transformations
from source systems into an ontology, mapping of partial models and coherence of variable models.
The thereby arising integrated model grows with the product and serves as repository for various
questions concerning the development project such as: Is the project on time and in budget? Which
requirements are met? Are there conflicts between product elements? To answer these questions,
metrics and a method for their derivation are introduced and shown in an example from Software
Development.
The paper closes with the introduction of Permeter, a performance measurement framework
implementing the described methodology. The framework is under development at the University of
Oldenburg.
Introduction
Product development is a challenging research area due to its creativity,
complexity and its intricate management demands. At the same time, it is of vital
and growing importance to any company’s success in today’s fast paced and global
markets. There are several methods developed in science and industry which claim
to describe, manage and improve development processes. However, these solutions
are often very specific and of limited portability. As the business objects differ
from industry to industry, so do their design processes – developing a new laundry
detergent formula is naturally different from designing a packing machine or
coding a computer game. This is acknowledged by cross-industry standards such as
the ISO 9000 by different process groups. Our research goal is to understand the
mechanisms which influence the performance of product development. For
empirical analysis we propose new measurement method. In this paper we start
© IBIS – Issue 5(2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS Issue 2(2), 2007
with a characterization of development processes as a basis for our research
approach and introduce an ontology-based method integrating development results
in a flexible semantic environment for further analysis. It is designed to serve as a
knowledge base for a wide array of scientific and industrial analysis’s and problems
such as project controlling or the assessment of certain engineering methods by the
use of metrics. Special attention is given to explain the transformation and
integration from source data into a semantic model. The article closes with the
introduction of Permeter, a tool implementing the suggested methodology.
Development Processes
The notion of the ingenious inventor who single-handedly creates ground-breaking
new products is long gone – if it ever was true. Today new products and innovations
brought to the market are the result of many experts working together in a
somewhat structured, however complex, dynamic and sometimes chaotic course of
action: the development process. Development processes run from the first
product definition (requirements specification) to the start of (series) production
often going through a number of phases with specific goals. Each phase adds more
details and information to the product model ([Kah04], [Ehr03]). The content and
sequence of phases depends on the industry sector under consideration. Most
sectors have developed their own practice approach. Some of these will be
presented here to illustrate the complexity and entropy to be dealt with when
creating a development process model.
An example for an industry with a large supply of procedures to guide product
development is the software industry. Due to the intangibility and complexity of its
products, Software Engineering has been a dominating field of research ever since
the first computers were developed in the 1940s [Som04].
Classic procedures of Software Engineering are the waterfall-model of Royce and
the spiral model by Boehm [Boe88]. The waterfall-model is a sequential model
(with limited loopbacks) while the spiral model acknowledges the iterative nature
of product development. The most commonly known modern, object oriented
approach to software engineering is the Rational Unified Process (RUP) by
Jacobson, Booch, and Rumbaugh [JBR99]. It consists of the phases: Inception,
Elaboration, Construction and Transition (all with distinct milestones). The RUP is
divided into the work packages: Business Modelling, Requirements, Analysis &
Design, Implementation, Test and Deployment. All of these are supported by
Configuration & Change Management and Project management. Its success is
largely due to its sound foundation on the Unified Modelling Language (UML,
[JBR04]), which provides an ample supply of formal diagrams to model the
structure, behaviour and interaction of computer programs. Other well known
development procedures are the V-Model or eXtreme Programming (XP, [Bec03]).
The V-Model is a tailorable solution devised by the German government. XP is an
agile, incremental and iterative approach with short cycles and few formal
requirements. A thorough inspection of software development processes can be
found in [Bal00].
As soon as we leave exclusive software development and move into more
hardware-oriented industries such as mechanical engineering or electronics, the
level of interaction between different disciplines increases dramatically despite
available methodologies of construction and design. To cope with the explosion of
© IBIS Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
complexity and yet arrive timely at development results, Simultaneous Engineering
methods (SE) as described in [Bul98] and [Rib00] have been devised. SE is
characterized by a number of Simultaneous Engineering Teams working
concurrently on partial problems with integration teams taking care of
synchronizing the different groups, discussing issues and determining project
status. This creates a significant amount of communication overhead but currently
is the most common way to control large development projects as Sörensen points
out for the automotive industry in [Sör06]. Due to their innovative and unique
nature, product development activities are usually carried out in a project format,
often involving diverse organizations across the globe. The challenges faced by
global engineering networks and in integrated product developments are described
in detail in [GHK+06]. They are met by more structured approaches like design
flows in chip design and mechatronics or Gate Processes in the automotive
industry.
All development processes face the same challenges:
1. Dealing with technical complexity
2. Understanding the interactions and dependencies between phases and
partial models
3. Defining the status of the project in terms of time, cost and quality
(Evaluating performance)
The first point can be well mastered by CAx-Software, Integrated DeveIopment
Environments (IDEs) or Product Data Management Systems. However, engineers
often abandon the two latter points that are less easy to tackle.
IT support and scientific exploration of individual domains is available, but hardly
any research has been done on formally modelling and understanding development
processes based on the actual observation of development activities and their
(digital) results (= partial product models). Thus a sound, flexible and integrated
model enabling automatic transformation of technical qualities into economic key
figures is needed. A promising application domain for such a system is project
controlling - a crucial issue in development projects (see [STG94], [Har03]).
Another application is the contribution of firm metrics to development maturity
models such as CMMI [Car06] or SPICE [Aut05a].
Related Work
Prior to illustrating the use of metrics in the context of this paper, a short overview
of related work shall be given.
The foundation of metric use in software project management has been laid out by
the TAME project [BR88]. Basili and Rombach were the first authors to join the
project management perspective with actual product and process models in order
to record and reuse development experience and measure performance. However,
this early concept lacks inter-phase-relationships (for an integrated analysis of all
phases of the design process). Henninger presents an organizational learning
environment for software development, which captures project-related
information during the creation of individual software products to be later
disseminated to subsequent projects. While Henninger acknowledges the necessity
© IBIS – Issue 5(2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS Issue 2(2), 2007
of metrics, the prototype BORE (Building an Organizational Repository of
Experiences) only provides the infrastructure for capturing project events but
offers no evaluation functionality ([Hen97a], [Hen97b]).
The exploration of invisible software structures is tackled by LaSSIE, a semanticsbased approach to discover code relations, find related projects and promote reuse
from software libraries. LaSSIE models enhance the software development process,
but yet again lack metrics to analyze it. Examinations of metrics for conceptual
software models can be found in [PGP+05]. A more practical approach to
measurement in development processes is given in the “Capability Maturity Model
Integration” (CMMI [Car06]) and “Software Process Improvement and Capability
Determination” (SPICE, [HDH+06]), two approaches aiming to determine the
maturity level of defined processes. Even though both models rely heavily on the
use of metrics, they are based on informal assessments not formal models of actual
development results. Even though most of development performance research has
been done in the field of Software (due to the unique characteristics of SE) these
results now diffuse into classic fields of engineering such as the automotive
industry. A good example in this field is AutomotiveSPICE [Aut05b] which translates
the classic SPICE-procedures into the much wider field of automotive engineering.
Methods for the integration of metric results into a performance measurement
system and the derivation of KPIs are described in [And05]. What is missing is a
generic methodology to evaluate the process outcome of different phases of the
design
methodology
and
for
different
domains
of
engineering.
To derive the requirement for this generic methodology the authors stated with the
outcome of our measurement: aggregated and suitable metrics.
Metrics and Methods for the Integrated Model
As presented in the introduction a great number of metrics has been defined in
literature and was applied to actual projects. An overview of R&D metrics can be
found in [Hau96]. Still, despite all measuring efforts, a great number of projects
fails or exceeds its planned limits as shown in [HAS07] One reason is that the
metrics suggested are often indirect indicators for the project status, but do not
capture the “as is” status because they neglect actual development results. For
example the number of patents or the amount of R&D investments are somewhat
related to the engineering output, but allow no decision support in daily project
work. Another example is the number of research staff which does not indicate
whether they are working on the right projects. Other reasons why common R&D
metrics are not helpful are the complexity and interdependencies of the various
partial models found in engineering. Therefore, we need metrics that quantify the
interrelations and methods for finding and understanding them. The Goal Question
Metric approach from the TAME-project [BR88] is used to obtain these metrics. It
helps to find fitting metrics for an information request by asking appropriate
questions answered by a set of metrics as presented in figure 1 for fault reduction
metrics.
Metrics can be classified by a number of dimensions such as object of study,
research goal, quality characteristic, etc. (see [PGP+05]). The most common one is
the object of study. Usually, one differentiates between product and process
metrics. Combined models give insight into the overall status of the project.
© IBIS Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Goal
Fault Reduction
Question
How many faults
are in the product?
Metric
# faults
What are frequent
faults made?
Fault clasification
# faults made
Which overhead is
caused by faults?
Which method
causes few faults?
# duration of tests
# faults / method
# of iterations prior
to release
# faults per phase
Figure 1: The Goal Question Metric Approach for fault reduction.
A metric is defined to measure ONE product or process quality ([IEE98]). A single
metric gives little insight into the status of a project. Thus methods for
combination of metric values need to be acquired. A good example is the metric
editor described later. Other well known methods for combining metrics in order to
ease their interpretation are the earned value methods which contrasts work done
(=value achieved) with expenses occurred or trend diagrams for costs or milestones
(e. g. [Lit95]).
Product Metrics
As stated in the previous section, it is important to derive ones measurement
methods/metrics from the domain that it should be used in. This includes the
complete environment of a project. The information that is needed and the
source(s) that it comes from have to be clear before a metric is established. The
product that is developed and the process that it emerges from are the most
important factors in a project. There is no one size fits all metric set that can be
used for every type of product or every new project, because the information that
exists in a project and the specific properties of a product differ from one project
to another. However there are some points of interest and patterns that can be
applied to product measurement in general.
The reason why metrics are used is that there is an overflow of information in
general and a lack of specific information (which is needed for a better
understanding of project activities and their outcome) at the same time. In case of
the product it is important to secure its quality. Quality is a complex issue and its
meaning depends on the quality management used in the specific project. As a
simple example product quality can mean that the product is operating in a
reliable manner and at the same time the product has to fulfil its specifications,
which is not necessarily covered by the first point. Therefore, there is a need for
good documentation when developing and using metrics. To show some aspects of
product measurement it is suitable to use software metrics as examples, because
software engineering is the engineering discipline that makes most use of metrics.
As mentioned before it depends on the project and the product what should be
measured and how it can be measured. The object of measurement is called
attribute. Fenton [FP97] distinguishes between internal and external product
attributes. In case of the (software) product internal attributes mostly mean size
© IBIS – Issue 5(2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS Issue 2(2), 2007
and complexity. The Lines of Code (LOC) metric is the easiest example for
measuring size in software products. This very early metric evolves around doing a
simple thing as counting elements in the source code. The LOC metric is as
ambiguous as it is simple. In most cases where the metric is used it is not ensured if
the physical lines where counted, or the functional statements in the code (e.g.
the statements separated by a ";" in Java). It is also not possible to determine the
effort that went into a certain part of source code by just using LOC, as the value
of LOC depends largely on the program writing style of the person and the
programming language that is used. LOC shows the problems that are generated
while measuring even simple properties of software like its size.
In case of complexity you can name McCabe’s Cyclomatic Complexity [McC76],
which measures software by its control flow, counting the number of linearly
independent paths.
The external attributes of a product are often not easy to measure, because they
depend on the product itself as well as its environment (the project). Fenton
names reliability, usability, testability etc. One can also name quality as an
external product attribute, while it is for sure dependent on even more factors or
even includes the previously mentioned attributes.
The distinction between internal and external product attributes also has its
impact on the measurement process itself. As external attributes are often derived
from the products internal attributes in some way, one needs to distinguish
between direct and indirect measurement. The internal attribute size for example
can be measured directly as seen for the LOC metric. A simple example for indirect
measurement is the Module defect density ([FP97] p.40) which computes as
number of defects divided by the module size (in LOC for example).
While these examples are focused on software engineering, the basic patterns of
product metrics mentioned are transferable into other domains. Good
understanding of the data available in a project is an important factor for
successful measurement activities because of the high complexity of today’s
products. The usage of ontologies helps to organize and understand data and makes
metric definition easier.
Process metrics
Apart from understanding product properties, a firm understanding of the
development process itself is necessary. Common questions are: Which activities
were necessary to create a product? Which systems/methods were involved? How
much iteration was made. What did it all cost? Will we finish on time?
These questions may be answered by looking at the development process. In
[GRP05], Garcia et al. cite four steps before process evaluation can take place:
definition, evaluation, control and improvement of the process. An emphasis is put
on the existence of an integrated framework measurement beforehand – a set of
metrics suited for processes. The first task of process definition is the most
cumbersome. While for example a well tailored manufacturing process can be
statically modelled and expressed with flow charts or process languages such as the
Business Process Modelling Languages (BPML), it is hard to define a creative process
like product development beforehand. The same problem exists in regards to the
control of a process via a workflow management system. An alternative to explicit,
ex-ante modelling of processes well suited for our application context can be found
with process mining. Process models may also be derived through observing the
© IBIS Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
evolution of development results or the process protocol itself, if it is available e.
g. from a Product Data Management System (PDM) which logs all activities and
changes for the sake of configuration management and traceability in case of
liability claims. This has been presented for automotive development processes in
[Jen01].
Process Metrics
Resource per phase/Task
Indicates whether a project has too
much or too little manpower. If
categorized, it may help to identify key
personnel.
Cycle Time
Time for one revision cycle to be
completed
Concurrency Index
Number of parallel activities.
Fault metrics:
Faults are a good indicator of process
- Fault rate
quality. A “healthy” level of discovered
- Ratio
new/closed
problem defects shows that if faults are made,
reports
they are discovered. The ratio of new
vs. closed problem reports shows
whether faults are mended (< 1) or if
more problems arise.
Time metrics:
Indicators for quality of project planning
- % of tasks finished on time
and work achieved.
- % of passed checkpoints at first
review
Strategic Metrics
# of patents and copyrights
Show overall quality of design processes
and effectiveness.
% of accepted project proposals
Figure 2: Common Process Metrics.
For the evaluation, control and improvement tasks, a semantic model as presented
here is helpful in order to avoid slow and inaccurate manual reporting of process
properties since an integrated automatic feedback that is achieved by tapping into
the operative systems of product development is much more accurate. This leaves
us with the last important requirement: The measurement framework that defines
which properties need to be considered and how they are measured. This is
achieved by a number of process specific metrics. A selection of common process
metrics is presented in figure 2 based on [GRP05] and [Tha00].
A Methodology Towards the Integrated Product Model
As elaborated in the previous section, literature knows various methods to cope
with the technical and organizational complexity of product development. Despite
their individual maturity, they are hardly applicable for today’s requirements and
projects with numerous participants and artefacts, especially when it comes to
item 2 (understanding of the dependencies) and item 3 (status reporting) in the list
© IBIS – Issue 5(2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS Issue 2(2), 2007
of challenges stated in section 1 [Lit95]. The methodology presented in this
document is designed to meet these requirements.
Figure 3 sketches the proposed solution for a phase-connecting methodology of
product development. Based on conceptual ontologies of project phases and the
manual or (semi-) automatic interconnection between corresponding development
artefacts during the development process, it is possible to design metrics for a
project controlling dashboard and performance indicators.
KPI
Mapping on KPI‘s
Metrics
t-box
Analysis/ Assignment
Administration
a-box
Semantic Net
Linkage
Requirements
Linkage
Functions
Linkage
Linkage
Transformation
Transformation
Concept
Implementation
Test, QS
Project Management (PDM, CVS)
Figure 3: Overview of proposed methodology, an extension of [LR93].
A closer look at figure 3: At the bottom a generic example process runs with five
development phases along the timeline. The generic phases picked for the example
are:
Requirement
definition,
functional
design,
conceptual
design,
implementation, test and quality assurance.
For each phase, conceptual ontologies (t-boxes) are created. These ontologies, in
the following referred to as partial models, describe the artefacts created in the
corresponding step of the development process. The figure displays them in the
line labelled "t-box". In this specific project, the a-box is filled with corresponding
instances; e. g. a specific requirement for a product is an instance of the concept
"Requirement" established by the t-box. This procedure is shown in the line
labelled "a-box". In some phases, this modelling is done manually; in others an
automatic transformation from project data is possible. Requirements, for
example, are usually recorded manually. In this case, the a-box model may replace
the text file, allowing easy reuse in subsequent project phases, and is created and
maintained using an ontology editor. Other a-boxes can be created by a
transformation (indicated as arrows), for example an XSL transformation of UMLdiagrams. However, these are examples only: If the requirements are already
© IBIS Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
captured in a structured way, the manual maintenance can be replaced by a
computer-aided transformation with ease.
A similar procedure is used for horizontal links between items (artefacts) of
different partial models. To establish these links, the framework implementing the
methodology offers a specific editor as well as automatic linkage facilities – as
detailed in section ’Integrating partial product models’. The phase-spanning
semantic web established by these relations offers a representation of the
development process previously not available. Dependencies and interrelations
between individual phases and artefacts are made apparent.
Metrics and reporting tools find a rich playground to derive figures from this new
integrated information basis, as indicated in the upper part of figure 3. Hence,
surveillance, control and management of development processes can be improved.
Key figures are now derived from the project context and do not have to be
interpreted in the isolated frame of a certain phase. To implement the surveillance
and control functions, methods described in literature and research are applied as
conceptual and process oriented metrics, to be later transformed into key
performance indicators (KPI, see
[Sch02]). They form a basis for process
monitoring that allows both measurement and active control. Since the KPIs are
estimated from development processes and artefacts, any events and changes show
up directly, allowing early intervention by decision makers.
Aside from conceptual models of the development artefacts and their relations,
other relevant information for the project success is found in project management
and product data management systems. They offer a second pillar and additional
input for a profound evaluation system. Therefore, project management data, e.g.
task and resources involved in specific tasks are also integrated by the framework,
compare section ‘Permeter’.
This section provides an overview of the methodology proposed for the integration
of models in product development. The upcoming section focuses on how the
actual integration takes place in detail.
Integrating Partial Product models
After explaining how product data from different sources are loaded, the next step
in creating an integrated product model is the establishment of semantic relations
between the created instances of ontologies. As stated in the previous section,
both manual and (semi-) automatic establishing of relations is possible. After
exploring possibilities and (dis)advantages for these two approaches, this section
goes into details on both the manual method and the automatic linkage algorithms.
Manual vs. Automatic relation establishment
The manual and the automatic relation establishment approaches have both their
advantages. The following figure summarises the properties of both methods in
question:
© IBIS – Issue 5(2), 2007
http://www.ibis-journal.net ISSN:1862-6378
Manual linking
Suitable for small data sets only
High user interaction
High accuracy
Vulnerable to “metric gambling” and
abuse
IBIS Issue 2(2), 2007
(Semi-) Automatic linking
Suitable for small and large datasets
Low to medium user interaction
Medium to high accuracy
Figure 4: Manual vs. automatic linking
These properties given, the automatic linking approach is to be preferred.
Nevertheless, there are partial models where automatic algorithms can not be used
and a manual process is sufficient. For example, if artefacts of partial model
Requirements are to be connected with the artefacts of Functionality that fulfil
one or more requirements. In that case, the number of artefacts is rather small the
amount of time needed is acceptable. Furthermore it can be assumed that an
automatic algorithm would not lead to more accurate results, since only a
developer has the knowledge about what functionality fulfils which requirement.
On the other hand, if the amount of data grows, the manual mapping of partial
models is time-consuming and error-prone as it happens more often that a required
relation is overlooked. Automatic algorithms can ease the process of mapping
partial product models, especially for large data sets. These algorithms should be
combined with manual user interaction of a semi-automatic approach to verify the
results and make modifications, as it is the trend in actual ontology alignment and
mapping tools ([NM03], [ESS05]).
Regardless of the approach chosen, a high accuracy of correct relations between
artefacts can only be achieved if a domain expert is involved in the process.
Therefore, the usability of the linking tools needs to be increased (compared to the
state-of-the-art today) accordingly.
Manual linking
A graphical editor is needed to support the manual and semi-automatic
establishment of semantic relations between artefacts of different partial models.
The editor has to offer both the ability to explore large data sets of two ontologies
at the same time and also the functionality to establish defined relations between
two artefacts of these ontologies.
A suitable approach is the Matrix Browser [ZCB02] developed by IAO for the
visualization and navigation of networked information spaces like in ontologies.
The central idea of this approach is to map the underlying graph as an adjacency
matrix, an alternative representation from graph theory where the nodes are
shown along the horizontal and the vertical axis of a matrix. Each axis shows a
hierarchical structure of the ontology and can be interactively expanded and
collapsed.
© IBIS Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Figure 5: Matrix Browser
As shown in figure 5, we adopted this approach to use it for the manual
establishment of relations between artefacts of different partial models. The user
is enabled to select the partial model that should be set to one of the axes. For
example, it is possible to set partial model ‘Requirements’ to the vertical and
‘Function’ to the horizontal axis. By selecting two artefacts in the Matrix Browser,
they can be connected via one or more relations defined between their
corresponding ontology concepts in the t-box. This concept is based on the same
idea as the widespread Quality Function Deployment-method (QFD, [Aka04]).
Automatic linking
For our purpose, algorithms have to support the user in detecting which artefacts
of different partial models can be connected, e.g. which function realises which
requirement or, another example, which java class implements a class modelled in
UML.
As described above, it is unrealistic to perform this process fully automated,
further user interaction is needed.
In research, a lot of strategies were proposed to combine different ontologies along
with different terminology like ontology merging, aligning or mapping. Regarding
the definitions of Klein [Kl01], the strategies of aligning and mapping is more suited
for our purpose since we do not want to create a new ontology. Therefore, we
focus on using existing approaches concerning the alignment and mapping of
ontologies.
In recent years, different solutions for (semi-) automatic alignment of ontologies
were developed, using multiple methods to calculate the similarity between two
© IBIS – Issue 5(2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS Issue 2(2), 2007
ontology elements, e.g. terminological (string-based), structural, extensional
(instance-based) or semantic methods [EE03]. These solutions (e.g. PROMPT [NM03]
or APFEL [ESS05]) developed semi-automatic alignment processes around these
mapping methods.
The person applying the proposed framework in a development project is enabled
to use the research-proposed mapping and alignment strategies for the integration
on partial product models. This action is processed in the following manner:
1. Selection of two or more partial models to be linked
2. Selection of a semantic relation to be used for the correspondences
identified
3. Selection of a strategy for correspondence revelation (as proposed in
research) and configuration of its options and properties
4. Execution of the method and preview of the correspondences extracted
5. Manual addition or deletion
Before
After
Person
Person
com.test.Person
Freelancer
same as
com.test.Person
Freelancer
com.test.Employee
com.test.Employee
sameAs
same
as
Employee
com.test.Freelancer
com.test.Freelancer
Employee
Relation Establishment Process
1
Selecting
UML and Java
Partial Model
Selecting
„sameAs“
relation
2
Configure
string-based
alignment
3
4
Calculate
mappings
5
Adjust results
Figure 6: Semi-automatic relation establishment
Figure 6 illustrates an example instantiation of the described process. The partial
models selected for the process are UML and Java, each containing three instances.
These instances have to be connected by the “same as”-Relation for an UML class
(which the instance of the UML partial model represents), because it represents
the same artefact as the Java implementation of the class (on a different level of
abstraction). The strategy used for the example examines the identifier of the
instances in the models and proposes the creation of a semantic relation for the
artefacts corresponding.
Permeter
In this chapter we propose Permeter, an application implementing the proposed
methodology that is able to handle all described steps like loading partial model
specific project data, establishing relations between artefacts of different partial
models, running metrics on the integrated product model and displaying the
aggregated results in reports.
© IBIS Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Which partial models are applicable for specific projects depends on the domain.
E.g. the partial models for chip design projects will differ from the ones in
software engineering shown earlier. Even two projects of the same domain can
have different partial models. Creating an analogue component will differ from
creating a digital one. A project with the aim to design a mixed-signal chip will
need additional partial models compared to a project, where a digital chip without
analogue components is developed. In software development distinct ontologies
are needed addressing the particular programming languages used for the
implementation. Therefore, the Permeter architecture can be divided into two
parts, a domain and project independent core implementation and the domain
specific plug-ins that are used to configure the partial models, metrics, and
customized reports.
Permeter Rich Client
Analysis
View
ProductModel
View
Report
Engine (BIRT)
Metric
Engine
Metrics (XML)
ProjectModel
View
Ontology
Core
Reportdesigns (XML)
Report
Editor (BIRT)
Metric Editor
Modell
Repository
Domain-Specific
Plug-Ins
Extensions
Figure 7: Permeter Architecture
Figure 7 shows the architecture of the Permeter framework and its three main
views.
The central view and therefore the first main function of Permeter is the Product
Model View. This view offers the functionality to create an integrated product
model, representing the current status of the product under development.
For each partial model, a conceptual ontology (t-box) is defined that is located
online. In a specific project, the a-box is filled with corresponding instances, e. g.
directly from an ontology or via an implemented transformation process.
To establish links between artefacts of different partial models, Permeter offers a
specific editor as well as automatic linkage prototypes as described in ‘Integrating
Partial Product Models’ above. The partial model-spanning semantic web
established by these relations offers an overview of the process not previously
available. Dependencies and interrelations between artefacts are made apparent.
Thus, surveillance, control and management of development processes can be
greatly improved.
© IBIS – Issue 5(2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS Issue 2(2), 2007
The second important view supplied by Permeter is the Project Management View,
depicting the organizational context of the development project. Here, all tasks,
milestones and costs are collected and integrated with the product view. By
semantically bridging the technical and the organisational perspectives, the
derivation of KPIs into a “project cockpit” is made easier and founded on accurate
measures directly derived from development results.
The third main functionality of Permeter is therefore the analysis of the collected
data to achieve this improvement. Like partial models, reports and especially
metrics are only applicable to specific domains. It is even possible that some
metrics are important to one company and uninteresting to another one. For this
reason Permeter has a metric engine that can run metrics in a defined XML format
and a visual editor to create metrics.
Only base component metrics like math operators (addition, multiplication, etc.) or
instance/concept counters are directly implemented. These components can be
combined to metrics stored in a XML file. It is also possible to combine previously
defined metrics to new ones. The implemented base components, the derived XML
metrics and the relations between them are modelled in an ontology, which is used
by the editor to decide which combinations are valid and which are not. This
metric concept is a flexible approach to create domain specific metrics in a quick
and easy way.
In using the report system BIRT (Business Intelligence and Reporting Tools) it is
possible to aggregate and visualise results from executed metrics. The system
comes with a report designer that allows a free definition of reports and metrics
that serve as input for the defined reports.
Applying Permeter in Software Engineering
Besides Automotives and Chip Design, Software Engineering is the third domain
specific derivation available for Permeter. Here, one idea particularly stands to
reason: to examine the development of Permeter using the tool itself. We present
this evaluation as a case study here, mainly because it allows the illustration of the
methodology while reusing the concepts and the wording introduced in the above
sections.
In order to facilitate the case study, special care and emphasis has been laid on the
early phases of the development, i.e. the creation of requirement and functional
models. Due to the fact that the designers of Permeter are familiar with the
modelling of ontologies, requirements and functions have been represented
directly as OWL instances. Thus, their import to Permeter is straight forward. This
does not imply that Permeter cannot be used by other projects. Transformations
for requirement and functional models defined in other formalisms are available.
In the next step, three additional partial models were transformed and integrated:
the UML conceptual model, the Java source code and the JUnit test cases. Using
the matrix browser and the automatic mapping tooling provided by Permeter, the
five partial models were interconnected. This procedure was repeated weekly on
every Friday. Figure 9 shows relations inside the Permeter Java implementation
(here: the metric engine plug-in) by means of the Matrix Browser. The creation
date of this integration model capture is August 3rd 2007.
© IBIS Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Figure 9: Permeter Java implementation shown by the Matrix Browser
In the integrated product model, formally implicit relationships were made
explicit. Taking the requirements as root nodes, relation trees were spanned
reaching via functions, conceptual elements, source code entities to test cases.
Each test case in turn provides information on its status, e.g. whether it had the
expected result or whether it failed to run, which was used to infer the project’s
requirements fulfilment status. Obviously, this measure is profoundly improved
compared to its former definition that relayed on a simple, manual flag denoting a
requirement as (not) fulfilled. Similarly, other metrics concerning complexity of
design and implementation were improved respectively became applicable by the
integration of the partial models. As Permeter development is work in progress the
reports derived are not yet ready to be published. We will put more emphasis here
in future papers.
In a nutshell, the case study showed that Permeter is able to measure a project’s
progress and to increase the predictability of product development.
© IBIS – Issue 5(2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS Issue 2(2), 2007
Conclusion
The examination of development processes has shown that a great number of
development methods are suited to solve specific problems but fail to act as a
generic approach to understand and model the many interactions between domains
and partial models. This deficit is addressed by the proposed ontology-based
framework for integration and evaluation of development results. Its advantages lie
in the generic approach, the expandability, the use of semantics instead of data
models and the definition and use of metrics with the GQM-method. The
applicability of the framework is demonstrated with the Permeter tool which is a
work in progress at the University of Oldenburg.
Further research will consist of creating more appropriate metrics and applying
Permeter to real world-projects. The authors will cooperate with product
developing companies both in the chip design and in the automotive area.
Accompanying their activity, an assessment of the methodology and the prototype
implementation will be performed based on context-driven metrics and feedback
from the companies’ experts.
References
[Aka04]
Akao, Y.: QFD-Quality Function Deployment. 2004.
[And05]
Andy Neely, M. G. K. P.: Performance measurement system design: A literature
review and research agenda, pp. 1228-1263, 2005.
[Aut05a]
Automotive SIG: Automotive SPICE Process Reference Model. 2005.
[Aut05b]
Automotive SIG: Automotive SPICE Process Assessment Model. 2005.
[BR88]
Basili, V.R.; Rombach, H.D.: The TAME Project: Towards Improvement-Oriented
Software Environments. IEEE Transactions on Software Engineering, Vol. 14, No. 6.,
pp. 758-773. June 1988.
[Bal00]
Balzert, H.: Lehrbuch der Softwaretechnik. 2000.
[Bec03]
Beck, K.: Extreme Programming. 2003.
[Boe88]
Boehm, B. W.: A Spiral Model of Software Development and Enhancement. In
Computer, Vol.21, No.5 pp. 61-72, 1988.
[Bul98]
Bullinger, H.: Concurrent Simultaneous Engineering Systems. The Way to Successful
Product Development, Springer, 1998.
[Car06]
CarnegieMellon Software Engineering Institute: CMMI® for Development, Version
1.2. 2006.
[EE03]
Ehrig, M.; Euzenat, J.: State of the Art on Ontology Alignment. Knowledge Web
Deliverable 2.2.3, University of Karlsruhe, 2004.
[Ehr03]
Ehrlenspiel, K.: Integrierte Produktentwicklung. 2003.
[ESS05]
Ehrig, M.; Staab, S.; Sure, Y.: Bootstrapping ontology alignment methods with
APFEL. 4th International Semantic Web Conference (ISWC-2005), 2005.
© IBIS Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
[FP97]
N. E. Fenton, S. L. Pfleeger, Software Metrics - A Rigorous and Practical Approach
Second Edition, PWS Publishing Company, 1997.
[GHK+06]
Gausemeier, J., Hahn, A., Kespohl, H. D., Seifert, L.: Vernetzte
Produktentwicklung. 2006.
[GRP05]
Garcia, F., Ruiz, F., Piattini, M.: Metrics for Process Models. 2005.
[HAS07]
Hahn, A.; grosse Austing, S; Strickmann, J.: Metrics – The Business Intelligence Side
of PLM, in Product Lifecycle Management - Assessing the Industrial Relevance (Hrsg:
Garetti, M. et al.), Proccedings of the PLM 07, Mailand, Indersience, Genf, 2007 .
[Hau96]
Hauser, J.R.: R&D Metrics: An Annotated Bibliography, ICRMOT Working Paper,
M.I.T., Cambridge, MA 02142 1996
[Hen97a]
Henninger, S.: Case-Based Knowledge Management Tools for Software Development.
Journal of Automated Software Engineering, vol. 4, pp. 319-340. 1997.
[Hen97b]
Henninger, S.: Capturing and Formalizing Best Practices in a Software Development
Organization. The Ninth International Conference on Software Engineering and
Knowledge Engineering (SEKE '97). Madrid, Spain. 1997.
[HDH+06]
Hörmann, K., Dittmann, L., Hindel, B., Müller, M.: SPICE in der Praxis. 2006.
[Har03]
Harmuth, U.: Erfolgsfaktoren für Projekte. 2003.
[IEE98]
IEEE: IEEE Standard for a Software Quality Metrics Methodology Institute of
Electrical and Electronics Engineers. 1998.
[JBR04]
Jacobson, I., Booch, G., Rumbaugh, J.: The Unified Modeling Language Reference
Manual. 2004.
[JBR99]
Jacobson, I., Booch, G., Rumbaugh, J.: The Unified Software Development Process.
1999.
[Jen01]
Jenne, F.: PDM-basiertes Entwicklungsmonitoring. 2001.
[Kah04]
Kahn, K. B.: PDMA Handbook of New Product Development. 2004.
[Kl01]
M. Klein. Combining and relating ontologies: an analysis of problems and solutions.
In A. Gomez-Perez, M. Gruninger, H. Stuckenschmidt, and M. Uschold, editors,
Workshop on Ontologies and Information Sharing, IJCAI01, Seattle, USA, August 4-5
2001.
[LR93]
Lott, C. M., Rombach, H. D.: Measurement-based guidance of software projects
using explicit project plans, pp. 407-419, 1993.
[Lit95]
Litke, H.: Projektmanagement. Hanser Verlag, 1995.
[McC76]
McCabe, T.J. A complexity measure. IEEE Trans. Software Engineering SE-2, 4 , 1976.
[NM03]
Noy, N.F.; Musen, M.A.: The PROMPT suite: Interactive tools for ontoloy merging
and mapping. International Journal of Human-Computer Studies, 6(59), pp. 983 –
1024, 2003.
[PGP+05]
Piattini, M., Genero, M., Poels, G., Nelson, J.: Towards a Framework for Conceptual
Modelling Quality. 2005.
[Rib00]
Ribbens, J.A.: Simultaneous Engineering for New Product Development, Wiley &
Sons, 2000.
© IBIS – Issue 5(2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS Issue 2(2), 2007
[Sch02]
Schaefer, O. M.: Performance Measures in Value Management. 2002.
[Sör06]
Sörensen, D.: The Automotive Development Process, A Real Options Analysis, Stuttgart,
2006.
[Som04]
Sommerville, I.: Software Engineering. 2004.
[Tha00]
Thaller, G. E.: Software-Metriken einsetzen, bewerten, messen. 2000.
[STG94]
The Standish Group: The CHAOS Report. 1994.
[ZCB02]
Ziegler, J.; Kunz, C.; Botsch, V.: Matrix Browser – Visualisierung und Exploration
vernetzter Informationsräume. Mensch & Computer 2002: Vom interaktiven
Werkzeug zu kooperativen Arbeits- und Lernwelten. pp. 373-382, Stuttgart, 2002.
About the authors
Axel Hahn is professor for
business engineering at
the
University
of
Oldenburg.
R&D
management
and
interoperability are his
research topics.
Stefan Häusler is researcher at
Institut für Informatik OFFIS
Oldenburg. His research topics
requirement
engineering
performance measurement in
design.
the
in
are
and
chip
Jan Strickmann works at
Kevin
Hausmann
is
the Porsche Research and
researcher at the Institut
Development Center in
für Informatik OFFIS in
Weissach in the process
Oldenburg.
He
is
reengineering
responsible for the core
department. His research
development
of
interests
are
in
Permeter.
performance
assessment
of
simultaneous engineering processes and
virtual prototypes.
© IBIS Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
An Enterprise Application Integration Architecture
Supporting Ontology Based Approach for B2B
Collaboration
Razika Driouche, Zizette Boufaida
Lire Laboratory, Department of Computer Science
Mentouri University of Constantine
25000, Algeria
[email protected], [email protected]
Fabrice Kordon
Lip6 Laboratory, Pierre & Marie Curie University
4 Place Jussieu 75252, Paris
Cedex 05, France
[email protected]
Abstract: An important task in Enterprise Application Integration (EAI) is to resolve
syntactic and semantic heterogeneity. Especially difficult the latter is. Some existing
integration approaches have been criticized for their lack of semantic richness. They
focus mainly on the technical and syntactical integration. To overcome these
limitations, this paper introduces novel ontology based architecture to ensure both
sides of integration: inside Application to Application (A2A) and outside Business to
Business (B2B), in order to support larger scale of integrations.
Our proposal includes two important parts. First, we suggest an integration process
basing on mapping approach to bridge applications ontologies. Second, we introduce an
approach to build business process ontology using web services composition strategy
and an integration scenario. We also use BPEL4WS language as a specification language
for expressing business process control flow and constraints.
Introduction
Enterprise integration has gained considerable attention both from academia and
industry. Scholars as well as practitioners have emphasized that the so-called
Enterprise Application Integration (EAI) is one of the key challenges in information
technology [27]. In essence, EAI technologies provide tools to interconnect multiple
and heterogeneous Enterprise Application Systems (EAS) such as Customer
Relationship Management (CRM), Supply Chain Management (SCM), Enterprise
Resource Planning (ERP) and legacy systems. This interconnection is difficult to
achieve because integrated systems are developed independently [28].
Despite the whole range of available tools and widespread standards adoption, the
main goal of EAI, which is the semantically integration of EAS, is currently an
unsolved problem. Indeed, EAI still focuses mainly on technical and syntactical
solutions but does not address adequately the semantic problem, which constitutes
the real integration problem [28]. Most previous publications do not sufficiently
take into account the semantic integration.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
Semantic integration becomes very important in order to overcome heterogeneity
within EAI and which mainly concerns data, processes and services. Heterogeneity
may appear in the use of different programming languages, availability on different
hardware and operating system platforms. Moreover, the applications will most
probably rely on different data representation that increase difficulties when
exchanging them. They do not share the same semantics for the terminology of
their applications models. So, there is a disagreement about the meaning, i.e.
inconsistent interpretations.
B2B e-commerce is emerging as trading parties attempting to integrate
heterogeneous business processes. Collaboration of partners leads to the
interoperability issue [33], which represents a major obstacle in the business
sector. Heterogeneity arises from the fact that partners do not have the same
business process models. The partners cannot collaborate directly because their
processes are completely predefined. This means that a company, which models a
predefined order and selection of a group of activities cannot integrate directly
with enterprises having different orders and selections, even this company is
willing to conform to those orders. It also means that an enterprise sending and
receiving information, disposed in a set of messages and activities, cannot match
with another enterprise that exchanges the same information in a dissimilar
aggregate of messages and/or activities [29]. A solution to the problem is to model
enterprise processes in a flexible manner to support a larger scope of B2B
collaborations that companies my conform to, without necessity for model redesign
[25].
Previous approaches to explore EAI have focused on technical and syntactical
integration. Nevertheless, they must address the semantic level which is more
difficult but provide more added-value inside (A2A) and outside (B2B) the
enterprise [2]. On the one hand, integration helps to exchange data and to unify
software components. For this purpose, ontologies are often seen as a solution
providing exchange facilities of semantically enriched information. They
encapsulate heterogeneity of integrated applications and offer common
terminology. On the other hand, integration helps to streamline business processes.
The proposed solution in the literature is to model business processes in a flexible
manner to increase ability for process matching without requiring changes in their
design.
In this paper, we propose an approach that combines ontologies and web services
in order to address the semantic integration in the context of EAI, as well as both
sides of integration. In particular, we describe the integration architecture, which
has two levels: the applicative and the collaborative ones:
In the A2A integration level, a mapping approach is introduced to bridge
applications ontologies. The architecture supports several heterogeneous
applications accessing multiple local ontologies. A component, called wrapper,
generates a global ontology for managing heterogeneity based on the semantic
bridges concept [4]. The latter encapsulates all required information to translate
instances of the source ontology to instances of the target one.
In the B2B integration level, the focus is on partners collaboration basing on ebXML
extended scenario [35][44]. We also give details of our approach for building
business process ontology using Web services composition strategy. We define two
mechanisms to manage potential composition and monitor processes execution. We
specifically argue the idea that the BPEL4WS language can be used as a
specification language for expressing business process control flow and constraints.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
The next section outlines some important related work. Then, we describe the two
levels of the proposed architecture and present a description of the wrapper for
bridging the application ontology and its components. The following section focuses
on how to develop the application ontology and on the application integration
process. The main section of the paper gives details of business processes
collaboration. We explain the service ontology building and the business process
integration scenario based on a composition strategy of Web services. Finally, we
draw some conclusions from our work and sketch directions for some future
research.
Related work
EAI is a multifaceted field of research. It includes integration of heterogeneous
software applications, data and business processes. The integration is the process
of adapting a system to make distributed and heterogeneous applications work
together to carry out a common objective [27]. In companies, the essential
requirement for heterogeneous and distributed information systems is to be able to
exchange information and services in a semantically rich and sound way. Thus, the
semantics should be captured, verified and used to validate reliable information
exchange. This is commonly referred to as the problem of interoperability. The
integration is closely related – but not equal- to interoperability [33].
Ontologies, which seem to be a suitable technology of semantic web, are actively
used in order to deal with the semantic interoperability problem. According to
Gruber, an ontology is defined as an explicit specification of a conceptualization
[19]. Ontologies could be key elements in many applications such as the semantic
integration, web-services search and composition, and web site management for
organization and navigation [12]. Researchers do also believe that ontologies
provide a basis for enabling interoperability between software applications across
different organizations [22].
More recently, Web services have emerged with the advent evolution of the
Internet and provided a set of standards for EAI. Even if Web services are not fully
mature, they seem to become the linga franca of EAI, which makes integration
simpler and easier through using Web protocols and standards. Web services can be
used for integrating applications and processes via standards such as Business
Process Execution Language (BPEL) and Web Service Flow Language (WSFL) [24].
Furthermore, some new integration products are based on Web services such as
Cape Clear, XML Global, Colloxa, BPEL4WS Java Runtime (BPWS4J), etc. [31]. Even
if Web Services are promising to solve the integration problem, they do not
correctly address the semantic aspect supported by UDDI annuaire with the help of
some standards, like North American Industrial Classification Scheme (NAICS) [36]
and UNiversal Standard Products and Services Classification (UNSPSC) [3].
As a consequence, the combination of Web service technology with Semantic Web
technology will make EAI dynamically possible for enterprises of all types and sizes,
unlike the traditional technologies [35]. In addition, enterprise integration will
become more reliable and easier to achieve without the tedious low-level
implementation problems.
In the context of application and process integration, some important initiatives
that aim to bridge the gap between Web services and ontologies have been taken.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
OWL-S [38] provides an ontology mark up language to describe semantically
capabilities and properties of Web services. The ESSI WSMO working group, part of
the ESSI cluster aligns the research and development efforts in semantic Web
Services between the SEKT, DIP Knowledge Web and ASG research projects. This
group aims at developing a language called Web Service Modeling Language (WSML)
that formalizes the Web Service Modeling Ontology (WSMO) [39]. Both efforts
introduce for example the notions of process and Web Service composition to
describe behavior. Whereas, OWL-S has a process model. WSMO, in contrast,
distinguishes orchestration (problem solving through Web service invocation) from
choreography (communication pattern through message exchanges) [3]. More over,
there are differences: WSMO proposes the explicit representation of goals and
capabilities for dynamic discovery. In addition, data and process mediation are
explicit concepts applied by Semantic Web services to overcome the heterogeneity
problem.
As Web services become ubiquitous, there are few current efforts to define a
coherent model to compose them and build business processes. Some service
composition models have been proposed [1, 12] based on ontologies of services (for
sharing semantics of data/information), dynamics of component services (i.e.,
behavioral features, complex forms of dataflow, transactional attitudes,
adaptability to varying circumstances) and dynamic of the target service.
Papazoglou et al. [12] describe manual composition of services by specifying
composite service using the service scheduling language and the service
composition execution language. Bouguettaya uses services as atomic actions which
are semantically described in terms of their I/O interfaces and non-functional
properties such as their purpose, their category and their quality [1]. In Knoblock
work [8], services are modelled as views over data source. Each service is
represented in terms of I/O parameters, binding patterns and additional
constraints in the source. The composition of Web services may be developed using
a specification language such as BPML [24], BPEL [30] and BPEL4WS [29], WSCI
(Web Service Choreography Interface) [30], etc and executed by an engine such as
BPWS4J [14].
What has not yet been addressed by these work is the development of a
methodology for designing an actual integration scenario like travel planning or
goods purchasing.
The contribution of the proposed work lies in a comprehensive approach to flexible
modeling inside (A2A) and outside (B2B) integration. The work is based on
ontologies and Web services composition to ensure greater interoperability.
Ontologies offer the semantic aspects to exchange data and to unify terminology in
the models of applications. Web services can encapsulate heterogeneity in business
processes and their composition provides a flexible manner to support a larger
scope of B2B collaboration. Business processes are described by building service
ontologies and composition strategy. The latter focuses on a composition schema
managed by two mechanisms. The first one is the composition manager, which
establishes a classification of the potential compositions according to
organizational constraints and client requests. The second one is the supervisor
mechanism, which is the core of the composition schema. It monitors the process
execution engine in order to control constraints and handle possible exceptions.
The BPEL4WS language is used as a specification language for modeling business
process.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Application integration architecture
Within our ontology architecture, we identify several types of legacy, client/server
and Web applications, developed using different programming languages. They
work on different operating system platforms and use various format for the
exchange of data. Ontologies allow to heterogeneous and distributed systems to
interoperate seamlessly with each other. They are mainly used to provide
interaction between applications and also between applications and users. For
example, they enhance the invocation of application systems or the expression of
data and processing queries that can be defined by the user.
The integration system we propose aims at offering a support for integrating
heterogeneous and distributed applications, and accessing multiple ontologies
(Figure.1). It provides a communication framework as a central service. It permits
an appropriate exchange of information between applications ontologies and
generates the global one using mapping process for managing heterogeneity. The
integration process is based on the semantic bridges to indicate the semantic
equivalence of ontology entities for assembling them.
Figure 1: Integration system architecture.
Furthermore, we give an overview about the two-levels for application integration,
the application and the collaboration levels.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
The application level consists of heterogeneous, autonomous and distributed
application systems, where each application has its local ontology. The
application ontologies are related to each other with a global ontology. In the
architecture, we aim at overcoming the gap between application ontologies,
according to the semantic relations. A special component, named wrapper, is
invoked to perform its tasks for building the global ontology. The latter can be
seen as the enterprise ontology and permits the resolution of semantic conflicts
in both concepts and properties.
• The collaboration level takes place in the business process collaboration with
partners. Each company has a mobile agent that is responsible of requesting,
providing the services and the negotiation for selecting the best partner basing
on criteria (e.g., price limits, product configurations or delivery deadlines). It
uses an ebXML extended collaboration scenario for achieving business process.
The mobile agent permits to perform the integration tasks according to process
ontology and using optimized itinerary. The latter improves the quality of the
system and reduces the response time. EbXML extended scenario is based on the
integration of agent paradigm to guarantee a saving of search time, to
negotiate business parameters and to offer a great performance, especially in
the presence of the characteristic of mobility which, solves problems related to
the networks while decreasing consumption in resources. We have used jbuilder
(version X) language and Jade plateform to implement the extended
scenario[44].
Additionally, an overview about the wrapper component is given in the following.
The main task of the wrapper is to find semantic relations between concepts of
application ontologies and reuse mapping information for communication. It is
composed of four parts:
•
•
•
•
•
The identifier is used to find related concepts or attributes of ontologies and
the relations between them. This can be done automatically, semiautomatically or manually with the help of domain experts. For instance, it can
use lexical ontologies such as WordNet or domain specific thesaurus to define
synonym terms, antonym terms and encoding (e.g. miles/kilometers,
Fahrenheit/Celsius, hours/minutes) [4].
The adaptor is used to represent the identified relations of equivalence
between ontologies based on the semantic bridges. It combines many algorithms
to measure the similarity. It adopts a multi-strategy approach to compute the
concepts similarity at various levels, such as lexical, properties (roles and
attributes), hierarchical and instances similarities. Then, the adaptor checks
the domain and the cardinality inclusion.
The linker transforms instances from the source application ontology into
instances of the target application ontology by evaluating the equivalence
relations defined earlier by the adaptor. Two problems that may arise are that
the mappings are incomplete or the that the mapped entities differ in the
context. The missing mappings can be gained through inference mechanism.
The communicator is used to enhance the communication between global
ontology (enterprise ontology) and local ones (application ontologies). It is
responsible for information exchange to answer the user’s request using the
appropriate mapping. It translates the query sent by the user to an equivalent
one. This is done by using the mapping information in order to facilitate the
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
query interpretation, according to the concepts of the concerned application
ontology.
Application ontology integration process
The application integration process based on ontologies is a new approach that has
rapidly been developed with the emergent semantic web [36]. This is mainly due to
the fact that actually, the ontology provides a basis for enhancing interoperability
and integration of heterogeneous applications. The proposed iteration process for
integrating applications considers the addition of a new application and the
evolution of company requirements and market criteria. The process is divided into
four basic steps: capturing, building, integrating and exploiting.
The suggested process includes two sub-processes, essential for building and
integrating ontologies. The first one shows how to develop the application
ontologies. It includes many steps from meta-modeling to implementation. The
second one presents how to integrate application ontologies, starting from
collecting the applications needs and ending to a global ontology.
Capturing requirements
A company model is a computational representation of the structure including,
activities, processes, information, resources, people, behavior, goals and business
constraints. This step consists of taking information about the field of integration
while identifying the applications behavior as well as the relations which they
maintain between them. The goal is to capture the sets of the enterprise
applications, the activities that they perform, the required resources, the
manipulated data and the invoked messages. Then, we identify the information
flow, their structure and the technical infrastructure to support them for building
the application ontologies. This step is carried out by the industrial specialists and
domain experts.
Building the application ontology
A range of methods and techniques have been reported in the literature regarding
ontology building methodologies [15, 26]. Mike Uschold’s methodology [16],
Michael Grüninger and Mark Fox’s methodology [17] and Methontology [18] are the
most representative. Grüninger methodology is only limited to ontologies using
first-order-logic languages [17]. Uschold’s and Methontology have in common that
they start from the identification of the ontology purpose and the need for
knowledge acquisition. Whereas Uschold proposes a knowledge codification in a
formal language, in Methontology a set of independent intermediate
representations of the formal language to be used is expressed.
For our purpose, we use UML class diagram [41] to represent the meta-model
application ontology. In this purpose, the ontology building process contains all the
tasks to develop the application ontology based on the information identified in the
previous step. It consists in a classification of relevant applications characteristics
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
whereby each categorized class is linked with associated information. Then, the
formalization implies the representation of the ontology in a formal language. The
implementation builds computable model and checking consistency. Our building
process contains four steps: meta-modeling application, formalization,
implementation and adaptation [45].
Meta-modeling application
Let us notice that we retain the logical structure in layers where each application
is made up of data, activities and a user interface. The concepts of the application
ontology are represented as a set of UML class diagrams (Figure 2). Ontologies
include class/subclass hierarchies, relationships between classes, class attribute
definitions and axioms that specify constraints. This information is usually modeled
in a class diagram (the reason why UML is a promising notation for ontologies). Our
ontology makes it possible to offer an intelligent layer to manage semantics
heterogeneity in the integrated applications.
The meta-model in figure 2 illustrates the principle of the deployed ontology. It
specifies the activities called upon by messages (coming/outgoing), using resources
and manipulating data. The handling data are represented by schemes. Each
application has a behavior represented by links. The application properties (level,
link-type, data and activities) are made up of a certain number of concepts. Each
concept is described with properties, axioms, types and semantic relations
connecting the concepts between them. These relations can be relations of
aggregation, composition or synonymy. The axioms make it possible to capture
more complex rules on the concepts. These relations and axioms will enable us to
solve the conflicts related to semantic heterogeneity.
Figure 2:
Meta-model of application ontology.
The domain ontology gives a formal representation of concepts of the studied
domain as well as their relations. However, the domain ontology cannot describe
specific functionality neighter the differences between applications. It attempts to
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
find the set of concepts covering the domain. In our work, we consider business
enterprises which provide products and services for partners and potential
customers. We identify two overlapping domains, sale and purchase.
Let us give a description of the application ontology concepts, which is a 13-tuple:
Application Ontology = (APPLICATION, ACTIVITY, DATA, SCHEME, RESOURCE,
MESSAGE, LEVEL, LINK-TYPE, CONCEPT, RELATION, TYPE, AXIOM,
PROPERTY )
We give a semi-formal description of concepts in the UML class diagram. We create
the concept dictionary according to Methontology [18]. In this dictionary, we
define for each concept: synonyms, instances, attributes, roles. Table 1 represents
a concept dictionary for the application ontology.
The binary relations are showed as properties, which relate a concept to another.
For each role, we define its name, the name of the source concept, the name of
the target concept, the cardinality and the name of the inverse role. Table 2
represents a role table for the application ontology. Example: manipulates
(Activity, data) shows that the relation manipulates connects an activity to its
data.
Concept
name
APPLICATION
Synonyms
Instances
Attributes
PROGRAM
SOFTWARE
Sale
Marketing
Application-name
Application-profile
Application-domain
Application-behavior
ACTIVITY
FUNCTION
SERVICE
Request quote
Check stock
Invoke
purchase
Balance cash
Delivery article
DATA
PARAMETER
INPUT/OUTPUT
SCHEME
RESOURCE
DESCRIPTION
REQUIREMENT
NEED
Customer
Invoice
product
XML-shema
Memory
CPU
MESSAGE
QUERY
Activity-name
Activity-type
Precondition
parameters
Sub-activities
Time-begin
Time-end
Resource-assigned
postcondition
Data-name
Data-type
Access-right
Schema-name
Resource-name
Resource-type
Resource-state
Message-id
Message-content
communication-mode
Level-name
LEVEL
LINK-TYPE
..
MySQL Request
OWL API
-
How
relevant?
Compulsory
Important
Important
Important
Compulsory
Important
Important
Important
Important
NicetoHave
NicetoHave
Important
Compulsory
Important
Compulsory
compulsory
Compulsory
Important
Important
Compulsory
Important
Important
Compulsory
Applicative
Collaborative
CONNECTION
Data-exchange Link-name
Compulsory
Service-sharing
collaboration
..
..
..
..
Table 1: Concepts dictionary for the application ontology.
© IBIS – Issue 2 (2), 2007
Roles
from-a
to-a
has-type
is
composed
participates
..
in
out
is
manipulates
assigned to
composed
represented
manipulates
represented
assigned to
in
out
participates
has-type
..
http://www.ibis-journal.net ISSN:1862-6378
Role name
Source concept
composed
represented
is
manipulates
assigned to
from-a
from
has-type
participates
in
has-p
..
APPLICATION
DATA
APPLICATION
ACTIVITY
ACTIVITY
APPLICATION
CONCEPT
APPLICATION
APPLICATION
ACTIVITY
CONCEPT
..
Source
cardinality
1..N
0..N
1..N
0..N
0..N
1..1
0..N
0..N
1..1
0..1
1..1
..
Target concept
DATA
SCHEME
ACTIVITY
DATA
RESOURCE
APPLICATION
RELATION
LINK-TYPE
LEVEL
MESSAGE
PROPERTY
..
IBIS – Issue 2 (2), 2007
Target
cardinality
0..N
1..1
0..N
0..N
1..N
1..1
0..N
1..1
0..N
0..N
0..N
..
Inverse role
belongsto
describes
refers-to
manipulatedby
usedby
to-a
to
hasbehavior
indicates
out
concerns
..
Table 2: Roles description for the application ontology.
Let us remark that starting from the class diagram, the instantiation will vary
according to the application. In particular, each application has its activities and
its data. For example, a sale application contains activities as: receive order goods,
check availability, invoke purchase, make payment, delivery goods and
manipulates data as: customer, product, supplier, invoice.
Formalization
A standard framework for representing ontological knowledge is the theory of
Description Logics (DL). DL forms a language family of knowledge representation. It
allows to represent knowledge relating to a specific area using "descriptions" which
can be concepts, relations and instances. We consider the SH [9] family [13], which
includes the SHOIN(D) DL, extension of the influential SHOQ(D) DL. Ontology Web
Language (OWL) [38] is equivalent to SHOIN(D), i.e. SHOIN(D) is a syntactic variant
of OWL DL.
Formalization consists of two parts: terminological language TBOX in which
concepts and relations are defined; and an assertion language ABOX in which we
introduce the instances.
• TBOX construction: Here, we define here domain concepts and roles by means of
constructors provided by DL. For example, an activity must have one and only one
name. it refers to an application. It manipulates data to which a resource is
assigned. It reacts to incoming messages and produces an outgoing message to
send the response.
(∃ Activity-Name.String) ∩ (=1 Activity-Name.String) ∩ (≥1 Manipulates.Data )
∩ (≥1 Assignedto. Resource) ∩ (≥1 Refers-to.Application) ∩ (=1 In.Message) ∩
(=1 Out.Message).
Moreover, we specify subsumption relations which exist between various concepts;
for example to specify that the Activity class is subsumed by the Application class
we write: ACTIVITY ⊆ APPLICATION (cf. table 3).
• ABOX construction: The assertion language describes facts, by specifying the
instances (with their classes) and the relations between them in the following
way:
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
a: C
to indicate that a is an instance of class C.
Example: Sale-product: APPLICATION.
(a1, a2): r to indicate that the two instances a1 and a2 are connected by the role
r. Example: (Make-payment, Invoice): manipulates.
Concept
Definition
APPLICATION
(∃ Application-Name.String) ∩ (=1 ApplicationName.String) ∩ (≥1 is. ACTIVITY ) ∩ (≥1
Composed.DATA)∩ (=1 Participates.LEVEL)
∩ (≥1 Has- type.LINK-TYPE)∩ (=1 Froma.APPLICATION)∩(=1 From-a.APPLICATION)∩ (≥1 Is.
CONCEPT) .
(∃ Activity-Name.String) ∩ (∃ Activity-Type.Thing) ∩ (=1
Activity-Name.String) ∩ (=1 Activity-Type.Thing) ∩ (≥1
Manipulates.DATA ) ∩ (≥1 Assignedto. RESOURCE)
∩ (≥1 Refers-to.APPLICATION) ∩ (=1 In.MESSAGE)∩ (=1
Out.MESSAGE).
(∃ Data-Name.String) ∩ (∃ Data-Type.Thing) ∩ (=1 DataName.String) ∩ (=1 Data-Type.Thring) ∩ (≥1
BelongsTo.APPLICATION ) ∩ (≥1 ManipulatedBy.ACTIVITY)
∩ (=1 Represented.SCHEME)
ACTIVITY
DATA
Subsomption
relation
APPLICATION
THING
⊆
ACTIVITY
APPLICATION
⊆
DATA
APPLICATION
⊆
SCHEME
(∃ Schema-Name.String) ∩ (=1 Schema-Name.String) ∩ (=1 SCHEME ⊆ DATA
Describes.DATA)
RESOURCE
(∃ Resource-Name.String) ∩ (∃ Resource-Type.Thing) ∩
(=1 Resource-Name.String) ∩(=1 Resource-Type.Thing) ∩
(≥1 UsedBy.ACTIVITY )
RESOURCE⊆
ACTIVITY
MESSAGE
(∃ Message-id.String) ∩ (=1 Message-id.String) ∩ (=1 In.
ACTIVITY ) ∩ (=1 Out.ACTIVITY)
(∀ Level-Name.{ Collaborative-level, Applicative-level })
(∃ Link-Name.String) ∩ (=1 Link-Name.String) ∩ (=1 HasBehavior. APPLICATION )
..
MESSAGE
⊆
ACTIVITY
LEVEL ⊆ THING
LINK-TYPE
⊆
THING
LEVEL
LINK-TYPE
..
..
Table 3: Concepts definition for application ontology in TBOX.
Implementation and test
The implementation deals with building a computable model using the appropriate
language. The effort is concentrated on the suitability of OWL DL [9], which is
equivalent to SHOIN(D) [42]. For checking, we use the inference services provided
by many systems such as FACT [9], Renamed Abox and Concept Expression Reasoner
(RACER) [10] and DLP [9]. The RACER system can read OWL files, convert them in
the form of DL knowledge bases and provide inference services.
Protégé-OWL [21] is an open source ontology development environment with
functionality for editing classes, slots (properties) and instances. The current
version of Protégé (3.1.1) [21] is highly extensible and customizable. We have used
it for developing the application ontology. We also base on Protégé-visualisation
plug-in to browse the ontology and ensure its consistency. The visualisation is
particularly helpful when trying to understand hierarchical relations.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
Adaptation and evolution
Ontologies may change over time or may be adapted for application specific needs.
Therefore, mechanisms are required to discover changes and recognize different
versions (e.g. OnToKnowledge [7] or PROMPT [6]). Ontology versions may not be
compatible with each other. So, explicit information must be available to indicate
compatibility (backward, upward, full compatible, incompatible). OWL provides
some built-in constructs to mark version compatibilities.
For managing the evolution of ontology, a classification takes place once a
definition of concept is newly created. For this purpose, we propose the use of the
DL classification algorithm [11].
Integration of application ontologies
In order to achieve an ontology-based semantic integration, we propose a subprocess that includes several main tasks: comparison, bridging and inferencing.
•
The comparison allows to compare concept A in source ontology with concept B
in target ontology and to determine metric similarity. We adopt a multistrategy process [4] which computes similarities between ontology entities using
different algorithms. Using just one approach is unlikely to achieve as many
good mapping candidates as one; so, we combine several approaches. The first
one focuses on acquiring a lexical similarity. The next step computes the socalled property similarity that is responsible for acquiring the similarity
between concepts based on their properties, either attributes or relations.
Then, we propagate the similarity to other parts of the taxonomy, the super
and the sub-concepts. Finally, we check the domain and the cardinality
inclusion.
•
The bridging is responsible for establishing correspondence between entities
based on the similarities computed previously. Each instance represented
according to the source ontology is translated into the most similar instance
described according to the target ontology. It intends to associate a
transformation procedure to the translation.
•
The inference mechanism is used when the source concept has not a target
concept. It means that the source concept does not have a direct counterpart in
the target ontology. For example, the concept 4-stars room may differ between
several tourism information systems. Differences in the context can be solved
by making plausibilityn and consistency checks on the data. For example, the
concept 4-stars room differ between several Tourism Information System (TIS).
According to the given facilities of the room, one TIS may categorize the room
as 4-stars another TIS only as 3-stars because of the lack of certains services
[7].
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Exploitation
This step deals with the realization of the computational model of integration,
which makes it possible to link the enterprise applications. Once the system is
deployed, what remains is to maintain the technical infrastructure of integration.
Furthermore, this step covers the exploitation of the integration system which
function is not limited by time. It is necessary to address the evolutions of the
enterprise (e.g. integration of a new application) and to manage the ontology
versions.
In the next section, we outline a description of collaborative level for assessing B2B
integration. One objective of our business process model is to investigate and
assess the B2B collaboration level. As discussed in the introduction, close
integration may well reduce flexibility and agility of the enterprise to work with
other business partners. For this purpose, enterprises must react to the raised
innovation pressure and facilitate flexible collaboration on a global scale by
aligning their business processes. Thus, B2B collaboration is closely related to A2A
integration. Additionally, more details about business collaboration is given in the
following sections.
Business collaboration
Business enterprises are organizations that perform collective work. In order to
achieve necessary operational efficiencies, the work need to be governed by
processes that define the flow of work throughout the organization. According to
Coase [43], the exitence of the enterprise itself is dependent upon its ability to
achieve internal transactional. Unfortunately, enterprises cannot achieve every
need internally because, at some point, the overhead burden of these operations
exceeds their abilities. When this occurs, organizations form partnerships and
collaborative agreements with one another.
The field of investigation is referred to as collaborative business between partners.
It describes the internet based interlinked collaboration of all participants in an
added value network from the raw material supplier to the end consumer [32]. It
allows a comprehensive information exchange between employees and even
between organizations.
In the next section, we describe the business process modeling views. Then, we
focus on service ontology building and business process integration scenario.
Business process flexible model
Compared to traditional business processes, the complexity of inter-organizational
processes has risen considerably as a result of the strategic, structural and cultural
differences between the partners. The allocation of performances and resources of
the business partners, the determination of responsibilities for materiel and
financial exchange relationships, as well as the information and data exchange over
interfaces have to be planned and coordinated. Thus, the demands on Business
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
Process Management (BPM) increase [32]. Finding consensus between the different
partners is considerably facilitated by a general approach. It encompasses the
common model of a collaborative business process [5].
On purpose of process specification, the model takes into account the process
profile description. It specifies what the process does and how to invoke its
functionalities besides the process requester perspective, associated with a
particular partner and operating in a business context with special organizational
constraints, that specify which collaboration scenario is require. The model
considers business collaboration, functional, organizational and technical aspects.
•
•
•
•
The business collaboration view defines the choreography of the business
collaboration and the structure of exchanged information. A business
collaboration is performed by two (binary collaboration) or more (multi-party
collaboration) partners. It might be complex involving a lot of services between
business partners. However, the most basic collaboration is a binary one
realized by a requester from one side and a provider from the other side.
The functional view describes the business process profile. It shows the process
structure: the sets of services, I/O parameters, execution order and
participated partners.
The organizational view refers to collect information about partners and
organizational constraints in order to specify the business context. For this
purpose, the business process must be defined in the exact business context
that means the execution scenario used in its organization.
The technical view monitors the process execution based on partners goals. The
control flow is the most important aspect for describing a business process at
runtime. The collaborative partners have to compare the results of the
execution continuously with their goals and adjust deviations by handling
exceptions.
In our approach, functional aspects are used to build the business process ontology
and to permit process discovery on the basis of partner needs. Organizational
aspects are exploited to give the appropriate business context and further to
improve the process execution according to the business collaboration scenario and
control flow aspect in technical view.
Building the service ontology
A service is mainly described by a set of operations and I/O parameters providing
by the service interface description. Pre and post-conditions are stated on each
single operation and on the whole service. They are logical expressions on I/O
entities. They must be verified before the execution of an operation, or of the
service, and must be satisfied after the execution.
For example, a selling service may require as a precondition a valid credit card and
as input the credit card number and the expiration date. As output, it generates a
receipt and as post condition the card is charged. Further characterization of
service regards the order in which the operations are performed. Then, we define
what messages the service reacts and what messages the service produces.
Moreover, services maintain relations between them such as semantic ones
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
(synonymy, specialization, aggregation …) and conjunctural ones (localization,
composition, exchange, ...).
A service ontology is a 9-tuple consisting of :
Service
ontology
=
(SERVICE,
SERVICECLUSTER,
SERVICEPROFILE,
SERVICERESOURCE,
SERVICEMESSAGE,
SERVICEPARTNER,
SERVICELOCATION, SERVICEQUALITY, SERVICEEXCEPTION )
In our approach, the service ontology aims to define the semantic of services
composing the business process. In our work, three types of service are identified.
Atomic service: It is an elementary service directly invocable. It is the
description of the service proposed by the provider. It is described in terms of
interface (input, output, operations, error parameter and message types),
behavior, quality and implementation properties. In our ontology, we study the
description of atomic services to establish semantic relations between them and
to group them into sets (cluster) of similar services.
• Composite service: It represents a cluster of similar atomic services. This is
generally appropriate within large and complex enterprises with several
different business domains (production, marketing ...). Three kinds of relations
between composite services are considered. The is-similar relation holds when
a composite service has the same operations as another one. The is-equivalent
relation is maintained when the two considered services have the same
preconditions, inputs, operations, outputs and post-conditions. The is-part
relation is viewed when a composite service includes the other.
• Domain service: It is a composite service enriched with organizational
properties. They are augmented with information about provider/requester
partners, time availability, causality constraints, synchronization and the
manipulated objects: resource, inventory, orders and products. Three kinds of
relations between domain services are considered. The has-same relation holds
when a domain service has the same business domain as another one. The iscomplement that is obtained when the execution of a domain service implies
the execution of the other. The is-independent relation shows that two disjoint
services.
According to the presented service model, we develop a methodology to build the
service ontology, articulated in four phases: identification, description, mapping
and clustering.
•
Identification of concepts and roles
In this phase, the service characteristics are identified as concepts with properties
(roles and attributes), which are represented in table 4. Specific concepts of
services such as profile, cluster, resources, messages and partners are integrated in
the service ontology. How relevant means the importance of the concept in the
system (Important, Compulsory, NicetoHave). Table 5 shows the source and target
concept, source and target cardinality and the inverse role for each role.
© IBIS – Issue 2 (2), 2007
IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
Concept name
Synonyms
Instances
Attributes
SERVICE
FUNCTION
METHOD
Checkavailability
Invokepurchase
Makepayment
ServiceID
ServiceState
ServiceType
IsSimilarTo
How
relevant?
Compulsory
Important
Important
Important
SERVICECLUSTER
SERVICEGROUP
-
ClusterID
NumberOfServices
Compulsory
NicetoHave
SERVICEPROFILE
SERVICEDESCRIPTION
-
ServiceOperation
ServExeOrder
ServInput/Output
ServicePre/Post
Important
Important
Important
Important
SERVICERESOURCE
SER-REQUIREMENT
SERVICE-NEED
-
SERVICEMESSAGE
SERVICEINVOCATION
-
SERVICEPARTNER
COMPANY
SUPPLIER
-
ServiceResourceID
ServiceResourceType
ServiceResourceURL
MessageID
MessageContent
ProviderPartner
RequestorPartner
..
Compulsory
NicetoHave
Important
Compulsory
Important
NicetoHave
NicetoHave
..
..
..
..
Roles
issimilar
belongstocluster
containsservice
describes
hasserviceprofile
usedbyservice
reactstomessage
..
belongstocluster
containsservice
..
hasserviceprofile
describes
..
useresource
usedbyservice
..
producemessage
reactstomessage
provides
requests
..
Table 4: Concepts dictionary for the service ontology.
Role name
Source concept
belongstocluster
issimilar
describes
producesmessage
useresource
..
SERVICE
SERVICE
SERVICEPROFILE
SERVICEMESSAGE
SERVICE
..
Source
cardinality
0..1
0..N
1..1
1..N
1..1
..
Target concept
SERVICECLUSTER
SERVICE
SERVICE
SERVICE
SERVICERESOURCE
..
Target
cardinality
0..N
0..N
1..1
1..N
1..N
..
Inverse role
containsservice
issimilar
hasserviceprofile
reactstomessage
usedbyservice
..
Table 5: Roles description for the service ontology.
Service description and comparison
In this work, WSDL language [3] is used to represent the structure of Web services
in terms of interface, behavior and quality properties. In this phase, atomic
services are compared on the basis of their properties to evaluate their similarity.
Semantic relationships are established among them on the basis of the performed
similarity computation. Two atomic services are grouped in the same cluster when
the weight of semantic relationships connecting them is greater than a given
threshold.
Service mapping and inferencing
In order to perform a semantic analysis of services on the basis of their own
properties, we adopt a multi-strategy process to compute similarities between
entities using different algorithms. We use the same mapping process inspired from
semantic bridge concept employed in the application ontologies. We compute
similarity basing on interface, behavior and quality properties for each source and
target service:
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
•
Interface similarity: The service interface includes the I/O entities, operations
and message types. To perform a semantic analysis of service interface, three
types of the entities similarity are distinguished. For each type, the domain
experts affect a value in the interval [0..1].
- I/O similarity refers to the evaluation of the input/output information
entities names of the two services Si and Sj, denoted by I/OSim (Ei, Ej).
Lexical ontologies such as WordNet or domain specific thesaurus are used
to define terminological relationships (e.g. synonym and antonym) for
acquiring similarity [34]. Names ni and nj of I/O entities are compared to
evaluate their degree of affinity A().
1 : ∃ at least one path of terminological relationships
in the thesaurus between ni and nj.
A (ni, , nj) =
0
: Otherwise.
We define Ei as a vector of I/O entities of service Si and Ej as a vector of
I/O entities of service Sj.
Ei = (ei1, ei2, …, eik) ⇒ | Ei |= k
Ej = (ej1, ej2, …, ejl) ⇒ | Ej |= l
I/OSim (Ei, Ej) = ∑ i,j (A (ni, , nj))/ | Ei |+|Ej|
- Operation similarity: It refers to the evaluation of the operations of the
two services Si and Sj denoted OSim(Oi, Oj) is computed by comparing their
names, their input/output information entities based on WordNet or
domain specific thesaurus. Two operations are similar if their names and
their I/O entities have affinity in the thesaurus. The similarity value of two
operations is obtained by summing up the affinity values of their
corresponding elements. OSim (Oi, Oj) value is the average of similar pairs
of operations, one from the first service and the other from the second.
We define Oi as an operations vector of service Si and Oj as a vector of
operations of service Sj.
Oi = (oi1, oi2, …, oim) ⇒ | Oi |= m
Oj = (oj1, oj2, …, ojn) ⇒ | Oj |= n
OSim (Oi, Oj) = ∑ i,j (A (ni, , nj))/ | Oi |+|Oj|
- Message similarity: It focuses on the message name and type to measure
similarity MSim(Si, Sj). Two services having the same type of received
(resp. sent) messages and their names having affinity in the thesaurus, are
similar.
The global interface similarity ISim(Si, Sj) for each pair of services is evaluated by
taking a weighted sum of I/O, operation and message similarities. Services where
interface similarity is greater than a threshold are the only ones taken into
account.
ISim(Si, Sj)= (I/OSim (Ei, Ej)+ OSim (Oi, Oj)+ MSim(Si, Sj))/3
•
Behavior similarity: It concerns the evaluation of sets of operations for
services. A service is defined as a vector S with N dimension:
S= (O1, O2, …, ON)
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
∀ i ∈ {1, 2, …, N}/ Oi is an operation of service S.
The behavior similarity relies on the idea of semantic bridge concept [4], which
means that for each source operation in the first vector, there is a target operation
in the second one, giving two services Si and Sj where their vectors have the same
dimension; operation oik of service Si is bridged to operation ojh of service Sj if:
1- oik and ojh have the same name;
2- oik and ojh have the same I/O parameters;
3- oik and ojh have preconditions (resp. postcondition) logically equivalent.
This means that the two operations are equivalent.
We can now extend the definition of semantic bridge to two services. Two services
are bridged if there is a bijective relation between their associated vectors. Each
operation in the one service must be compatible with a corresponding operation in
the other service.
When the source operation does not have a target operation, this means that the
source operation does not have a direct counterpart in the target vector. For
example, if an enterprise sends a purchase order and receives the confirmation as
one single operation in the service, it cannot match with another enterprise that
sends order and receives confirmation as two separate operations in another
service. For this purpose, the source operation can be bridged with the composition
of the two target operations.
Quality similarity: Each service is characterized by a set of quality parameters
provided by available classification standard (for example, ISO 9000 STANDARD).
Some services may be very good, reliable and quick to respond; others may be
unreliable, sluggish, or even malevolent. Specific application dependent quality
parameters are considered (for example, the number of credit cards accepted
by an on-line ticket reservation service). Each quality parameter is described by
means of a type, a metric and a predicate. Quality type means quantitative and
qualitative feature of parameter. The metric indicates the measure unit (e.g.
second …). The predicate is a logical expression (e.g. response time < 1s). The
quality similarity of the two services Si and Sj denoted QSim(Si, Sj) is evaluated
by comparing their properties. Si and Sj have similar quality iff:
1- Si and Sj have the same type;
2- Si and Sj have the same metric;
3- Si and Sj have predicates logically equivalent.
•
Finally, global similarity GSim for each pair of services Si and Sj is evaluated by
taking a weighted sum of ISim(Si, Sj), BSim(Si, Sj) and QSim(Si, Sj). Only services
where similarity is greater than a threshold are considered.
The inference mechanism is used when the service does not have a similar service.
That means the service does not belong to any cluster.
Service clustering and classifying
The result of similarity-based description is a set of clusters of similar services with
defined similarity relationships. A hierarchical clustering algorithm [40] is used to
determine clusters based on global similarity established among services. Some
problems appear, when the service exists in many clusters. If the intersection of
the clusters is not the empty set so, clusters should be amalgamated.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
After the application of the four phases, the effort is concentrated on the
suitability of the OWL DL, which is equivalent to the SHOIN(D) to formally describe
the obtained ontology [9]. A concept definition is a statement of the form A:=D,
where A is a concept name and D is a concept description. In the following, the
formalization of SERVICE, SERVICEPROFILE and SERVICECLUSTER concepts is
presented in table 6.
Concept
SERVICE
Definition
SERVICE:=(∃ ServiceID.String) ∩ (∃ Has
ServiceProfile. SERVICEPROFILE) ∩ (≥1 Has
ServiceCluster. SERVICECLUSTER) ∩ (∃
ReactsToMessage. SERVICEMESSAGE) ∩ (≥0
IsSimilarTo. Service) ∩ (≥1UseResource.
SERVICERESOURCE).
Subsomption relation
SERVICE ⊆ THING.
SERVICEPROFILE
SERVICEPROFILE := (∃ ServiceOperation.String)∩
( ∃ ServExeOrder.Integer)∩(∃ Descibes.SERVICE)∩
(∃ ServiceInput/Output.String)∩
(∃ ServicePre/Post.Boolean).
SERVICEPROFILE
SERVICE.
⊆
SERVICECLUSTER
SERVICECLUSTER: = ( ∃ ClusterID.String) ∩ ( ∃
NumberOfService.Integer) ∩ (≥1 ContainsService
SERVICE).
..
SERVICECLUSTER
SERVICE.
⊆
..
..
Table 6: Concepts definition for service ontology in TBOX.
For checking, we need to use the inference services provided by the DL formalism
for supporting the building process and improving our ontology quality. The use of
the RACER system[10] can make it possible to read OWL file and to convert it in the
form of a DL knowledge bases. It can also provide inference services. To
manipulate the application ontology, Protégé-OWL [21], version 3.1.1 offers a
convivial graphical user interface. It permits to visualize and to edit the ontology
hierarchical relations. For example, the transitive property of the IsSimilar role is
described in OWL DL as follows:
<owl: ObjectProperty rdf: ID="IsSimilar">
<rdf: type rdf:resource ="http://www.w3.org/2002/07/owl#TransitiveProperty"/>
<rdfs: range rdf: resource ="#Service"/>
<rdfs: domain rdf: resource ="#Service"/>
<owl:inverseOf rdf:resource ="#IsSimilar"/></owl:ObjectProperty>
Composition strategy
Collaborative processes are defined as the abstraction of complex business
processes involving different partners. They specify the interaction between Web
services and their coordination. In this work, business processes are defined as a
composition of domain services. Only at runtime, for a given domain service, a
composite service is retrieved and invoked. If the needed service is not available,
then another similar service in the same cluster is invoked. For this purpose, a
composition strategy is defined as follows:
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
A CS (Composition Strategy) of Web services is formally defined as a set of triplets:
CS= (S, R, C)
S: a set (S1, …, Sn) of domain services.
R: a client service request.
C: a composition schema that fulfils R by suitably orchestrating (S1, …, Sn).
The composition schema is defined as a set of triplets:
C= (Pc, Pr, Co)
Pc: potential composition of domain services.
Pr: Partners implied in the composition.
Co: Constraints used for the selection of the current composition (service
availability, quality of service: response time, cost …).
The composition schema is monitored by two mechanisms:
Composition manager
This mechanism uses the information provided by the UDDI registry to establish a
classification of the potential compositions according to organizational constraints
and client request. It tries to find a conformity between organizational and
customer constraints. The latter analyzes the constraints required by the customer
and checks the availability of services to deduce the best combination from the set
of composition. After making a decision, it sends it to the supervisor mechanism.
The behavior of the manager mechanism is described in the Figure. 3
Figure. 3. Composition manager mechanism.
Composition supervisor
The supervisor mechanism is the core of the composition schema. It monitors the
process execution engine in order to control constraints and to handle possible
exceptions. It invokes the appropriate services after receiving the best
composition. This mechanism uses the constraints of selection to monitor the
execution of services. It is based on mechanisms provided by BPEL4WS language. In
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
case of errors caused by failure in service discovery, the compensation mechanism
(scope) is used to cancel the execution of the current composition and to send an
alarm signal to the manager mechanism in order to select a new combination.
Moreover, the exceptions mechanism (throw, catch) refers to manage the problems
at runtime (Figure. 4.).
Figure. 4. Composition supervisor mechanism.
Business process integration scenario
In this work, the integration scenario includes the process views: business,
functional, organizational and technical ones. The composition schema shows the
management of the business process (invoking services and scheduling the different
steps at runtime). Each service is augmented with constraints on treatment and
data based on the service provider. Then, we define the constraint causality to
give execution order monitoring by partners goals. The obtained dynamic schema is
consistent with partners’ business contexts. For this purpose, the schema
composition is the program steps to be executed for managing the business
process.
In the following, we present the process integration scenario.
Let’s now give the process integration scenario. As shown in Figure. 5, the
integration scenario begins once the Web services have been described using the
WSDL language (1). Then, they are published both in the service ontologies
registries and in the UDDI registry (2, 3). Next, the user requirement is used to
enhance the requested services (4). When the services are identified, they can be
identified by service discovery. If a given service is not available at the execution
time, then another similar service in the same cluster is selected and invoked in
the appropriate service ontology (5, 6). Once the needed services are discovered,
they are based on a composition strategy to form business process using
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
composition schema. Once the set of desired services have been discovered, they
are ordered according to implied partners and constraints to select the best
composition (7). In this approach, the trace of each composition schema is kept in
the repository to minimize response time. The current composition is executed
under the control of the supervisor mechanism in the orchestration engine (8). The
latter monitors the service execution and handles possible exceptions (9). Finally,
the result is returned to the user (10).
Figure. 5. Business process integration scenario.
BPEL4WS and business processes
Current process based languages such as BPEL4WS [14] support some flexibility,
with the common control flow which constructs such sequence, iteration, and/or
splits and joins. BPEL4WS represents the union of WSFL [24] and XLANG [3]. This
merger has created the market consolidation necessary to make BPEL4WS the de
facto standard for expressing process consisting of Web services. Structurally, a
BPEL4WS file describes the order by stating whom the participants are, what
services they must implement in order to belong to the composition, and the
control flow of the process.
For example, in a tourism agency, the reservation process needs several services
orchestrated in this order: service for booking a flight, service for hotel reservation
and service for payment. The tourism agency collaborates with airlines companies,
hotels and banks to satisfy the customer request. For each service, the provider
partner gives its organizational constraints to execute it. Each constraint is
described by name (constraint name), type (operation-constraint, data-constraint),
partner (partner name), predicate (logic expression), order (the service before, the
service after) and frequency (the number of times to check constraint). We use the
BPEL4WS language to describe the business process:
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
- Definition of partners
<partners>
< partner name=”custmor” partnerRole= ”clienttourismagency”>
< partner name=”webserviceairlinecompany” partnerRole =”reservationticket”>
< partner name=”webservicehotel” partnerRole =”reservationroom”>
< partner name=”webservicebank” partnerRole =”confirmationpayment”>
</partners>
- Definition of organizational constraint
< name= “c1- card-credit” constraint type=”data-constraint” partner=” webserviceairline”
predicate=”card credit = valid” order =”before: reservationticket” frequency=”1” constraint/>
< name=”c2-card-credit” constraint type=”operation-constraint” partner=”webservicehotel”
predicate=” card credit = valid” order=”after: reservationticket” frequency=”1” constraint/>
< name=”c3-confirmationpayment” constraint type=”message-constraint” partner=”customer”
predicate=”card credit = valid” order =” after: reservationroom” frequency=”3” constraint/>
- Definition of data types and messages
<variables>
<variables name= “request” messagetype =”typerequest”>
< variables name= “recepisse” messagetype =”typerecepisse”>
< variables name= “reservationticket” messagetype =”typereservationticket”>
< variables name= “reservationroom” messagetype =”typereservationroom”>
< variables name= “confirmationpayment” messagetype =”typeconfirmationpayment”>
<variables/>
- Definition of business process
<sequence name=”seqsejourreservation” name= “c1- card-credit” >
<receive partner=”customer” operation “sejourreservation” variable= “request”>
<flow>
<invoke partner=”webserviceairline” operation =”reservationticket”>
<invoke partner=”webservicehotel” name=”c2-card-credit” operation =”reservationroom” >
</flow>
<invoke partner =”webservicebank” operation = “confirmationpayment”>
<reply
partner
=”customer”
name=”c3-confirmationpayment”operation=”sejourreservation”
variable=”recepisse”>
</sequence>
Conclusion and perspectives
Semantic integration is a major problem that can affect data, applications and
processes. A solution is to use an ontology-based approach associated to Web
services. The ontologies provide the semantic basis to unify terminology in
applications models. The Web services can encapsulate heterogeneity in business
processes and their composition provides a support to orchestrate the execution.
In this paper, we have outlined an architecture for applications integration with
two levels: applicative and collaborative ones. It is comprehensive in the sense
that it (a) comprises different integration levels and (b) includes ontologies for
managing heterogeneity and ensuring flexibility in business processes. It is generic
because it is not tailored to specific technologies or business applications.
We believe that the results presented here give many interesting insights for
researchers who wish to explore B2B integration more thoroughly. Our architecture
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
will also be useful for practitioners who wish to assess the flexibility of B2B
integration.
We have organized our architecture along two dimensions, namely, A2A and B2B
integration. In the A2A integration, we have focused on a mapping approach to
bridge applications ontologies. It supports several heterogeneous applications
accessing multiple local ontologies. A component, called wrapper, generates a
global ontology for managing the enterprise information heterogeneity based on
the semantic bridges concept. In the B2B integration, we have built service
ontologies based on domain services, enriched with information about
organizational aspects to keep business context for each partner. These ontologies
are developed for supporting the design and execution of business processes using
Web services composition strategy.
There are numerous of interesting directions for future research. One area is to
focus on the dynamic reconfiguration of the Web service composition strategy to
improve service quality. Another area is to further detail and tailor the
architecture for specific technologies. For this purpose, we will consider the
heterogeneity in the standards of exchange like RosettaNet and BizTalk in order to
propose a portal to support partners collaboration [24]. A third area is to detail and
to empirically test similarity in the mapping process using context properties. A
last direction might address some perspectives for service ontologies: (i)
experimentation to validate the service ontology; (ii) using the framework MAFRA
[7] to support the integration of service ontologies; and (iii) improving services
similarities based on OWL aspects.
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
Medjahed, B.; Bouguettaya, A.; Elmagarmid, A. K.: Composing Web Services on the
Semantic Web. Very Large Data Base Journal, 12 (4), 2003.
Linthicum, D. S.: Next Generation Application Integration, Addison-Wesley edition,
2004.
Chauvet, J.: Services Web avec SOAP, WSDL, UDDI, ebXML. Eyrolles edition, 2002.
Maedche, A.; Motik, B.; Silva, N.; Volz, R.: MAFRA A MApping FRAmework for
Distributed Ontologies in the Semantic Web. In ECAI’02 Workshop Knowledge
Transformation, Lyon, France, 2002.
Gronau, N.; Uslar, M.: Skill Management Catalogues Built via KMDL- Integrating
knowledge and Business Process Modeling. ECAI’04 Workshop Knowledge Management
and Organizational memory, Valencia, Spain, 2004.
Noy, N. F.; Musen, M.A.: PROMPT-Algorithm and Tool for Automated Ontology Merging
and Alignment. Seventeenth National Conference on Artificial Intelligence, AAAI,
Austin, TX, 2000.
Gahleitner, E.; Wob, W.: Enabling Distribution and Reuse of Ontology Mapping
Information for Semantically Enriched Communication Services. The 15th International
Workshop on Database and Expert Systems Applications (DEXA), Zaragoza, Spain,
2004.
Michalowski, M.; Ambite, J.L.; Thakkar, S.; Tuchinda, R.; Knoblock, C.A.; Minton S.:
Retrieving and Semantically Integrating Heterogeneous Data from the Web. IEEE
Intelligent Systems, 19(3), 2004.
Horrocks, I.; Patel-Schneider, P. F.; Harmelen, F. V.: From SHIQ and RDF to OWL: The
making of a web ontology language. In Web Semantics Journal, 1 (1), 2003.
Haarslev, V.; Möller, R.: RACER System Description. In Proc. of the Int. Joint Conf. on
Automated Reasoning (IJCAR 2001), volume 2083 of Lecture Notes in Artificial
Intelligence, 2001, pp 701–705.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
[33]
[34]
[35]
[36]
Baader, F.; Calvanese, D.; McGuinness, D.; Nardi, D. The Description Logic Handbook
Theory, Implementation and Applications. Cambridge University Press, 2003.
Yang, J.; Papazoglou, M.P.: Service Components for Managing the Life cycle of Service
Compositions. In Information Systems, 29 (2), (2004)
Knublauch, H.; Mugen, M.; Rector, A.: Editing Description Logic Ontologies with
Protégé OWL Plugin, In DL’04, Workshop on Description Logic, Canada. 2004.
BPEL4WS: http://xml.coverpages.org/bpel4wx.html.
Lopez, M. F.; Perez, A. G.: Overview and Analysis of Methodologies for Building
Ontologies, In Knowledge Engineering Review. 17(2), 2002.
Uschold, M.; Grüninger, M.: Ontologies Principles Methods and Applications, In
Knowledge Engineering Review. 11(2), 1996.
Grüninger, M.; Fox, M. S.: Methodology for the Design and Evaluation of Ontologies, In
IJCAI’95 Workshop on Basic Ontological Issues in Knowledge Sharing, Montreal, 1995.
Fernandez, M.; Gomez-Perez, A.; Juristo, N.: Methontology from Ontological Art
Toward Ontological Engineering, In AAAI’97 Spring Symposium Series on Ontological
Engineering, USA, 1997.
Gruber, R. T.: A Translation Approach to Portable Ontology Specification. In
Knowledge Acquisition, 5:199-220, 1993.
Borgida, A.; Serafini, L.: Distributed Description Logics Assimilating Information from
Peer Sources, In Journal of Data Semantics, 2003.
Protégé OWL, version 3.1.1, 2005,
http://protege.stanford.edu.
Albani, A.; Dietz, J.; Zaha, J.: Identifying Business Components on the Basis of
Enterprise Ontology. Konstantas, D., Bourrieres, J.-P., Léonard, M., Boudjlida, N.
(ed), Interoperability of Enterprise Software and Applications, Geneva, Switzerland,
2005.
Abels, S.; Haak, L.; Hahn, A.: Identification of Common Methods Used for Ontology
integration tasks. IHIS’05, November 4, Bremen, Germany, 2005.
Holfreiter, B.; Huemer, C.: Modeling Business Collaborations in Context. In the
Proceedings of On The Move to Meaningful Internet Systems OTM Workshops,
Springer-Verlag, 2003.
Bauer, B.; Müller, J. P.; Roser S.: A Decentralized Broker Architecture for
Collaborative Business Process Modelling and Enactment. Second International
Conference I-ESA’06, Interoperability for Enterprise Software and Application,
Bordeaux, 2006.
Kalinichenko, L.; Missikoff, M.; Schiappelli, F.; Skvortsov, N.: Ontological Modelling.
Fifth Russian Conference on Digital Libraries, RCD’03, St-Petersburg, Russia, 2003.
Stelzer, D.; Fischer, D.; Nirsberger, I.: A Famework for Assessing Inter-organizational
Integration of Business Information Systems. International Journal of Interoperability
in Business Information Systems, IBIS, 2(2), 2006, pp. 9-20.
Izza, S.; Vincent, L.; Burlat, P.: Unified Framework for Application Integration, in
seventh International Conference on Enterprise Information Systems, USA, 2005.
Martin, H.; Zdravkovic, J.; Johannesson, P.: Service-Based Processes Design for
Business and Technology. ICSOC’04, November 15-19, New York, USA, 2004.
Alonso, G.; Casati, F.; Harumi, K.; Machiraju, V.: Web Services-Concepts,
Architectures and Applications. Springer-Verlag edition, Berlin, 2004.
Acharya, R.: EAI- A Business Perspective. eAI Journal, www.bijonline.com/, 2003.
Zang, S.; Hofer, A.; Adam, O.: Cross Enterprise Business Process Management
Architecture–Methods and Tools for Flexible Collaboration, In OTM Workshop, LNCS,
vol. 3292, Springer-Verlag, Heidelberg R. Meersman et al. edition, Berlin, 2004.
Panetto, H.; Scannapieco, M.; Zelm, M.: INTEROP NoE- Interoperability Research for
Networked Enterprise Application and Software. OTM Workshop, Vol 3292,
Heidelberg, R.; Meersman et al edition, Berlin, 2004, pp. 866-882.
Bianchini, D.; De Antonellis, V.; Melchiori, M.: Domain Ontologies for knowledge
Sharing and Service Composition in Virtual Districts. In the Proceedings of the 14th
International Workshop on Database and Expert Systems Applications (DEXA), 2003.
Mertz, D.: Understanding ebXML, Understanding the Business Web of the Future
Phenomenological Unifier. Gnosis Software, Inc, 2001.
Dogac, A.: Semantic Web Services. In the Proceedings of International Conference
on Web Engineering, Munich, 2004.
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
[37]
[38]
[39]
[40]
[41]
[42]
[43]
[44]
[45]
IBIS – Issue 2 (2), 2007
Dogac, A.; Laleci, G.; Kirbas, S.; Kabak, Y.; Sinir, S.; Yildiz, A.; Gurcan, Y.: ArtemisDeploying Semantically Enriched Web Services in the Healthcare Domain ,
http://www.ing.uinbs.it/~deantone/interdata_tema3/Artemis/artemis.html , 2004.
Martin, D.; Burstein, M.; Lassila, O.; Paolucci, M.; Payne, T.; Mcllraith, S.: Describing
Web Services Using OWL-S and WSDL. http:// www.daml.org/services/owl-s/1.0/,
2003.
Roman, D.; Lausen, H.; Keller, U.; Oren, E.; Bussler, C.; Kifer, M.; Fensel, D.: Web
Service
Modeling
Ontology,
WSMO
working
draft.
http://www.wsmo.org/2004/d2/v1.0/20040920/, 2004.
Jain, A. K.; Dubes, R. C.: Algorithms for Clustering Data. Prentice-Halle, 1988.
OMG:
OMG
Unified
Modeling
Language,
Version
2.0.
2003.
http://www.omg.org/technology/documents/modeling spec catalog.htm#UML.
Horrocks, I.; Sattler, U.: Ontology Reasoning in the SHOQ(D) Description Logic. The
17th Int. Joint Conf. on Artificial Intelligence (IJCAI 2001), Morgan Kaufmann, 2001,
pp 199–204.
Williamson, O. E.; Winter, S.G.; Coase, R. H.: The Nature of the Firm: Origins,
Evolution and Development. Oxford University Press, New York, 1991.
Driouche, R.; Boufaida, Z.; Kordon, F.: Towards Integrating Collaborative Business
Process based on Process Ontology and EbXML Collaboration Scenario. In
6th
International Workshop on Web Based Collaboration’06, IEEE Computer Society,
Krakow, Poland, 2006.
Driouche, R.; Boufaida, Z.; Kordon, F.: An ontology based architecture for integrating
enterprise applications. In fourth Forth International Workshop on Modelling,
Simulation, Verification and Validation of Enterprise Information Systems, Paphos
Cyprus, May 2006.
Biography
Razika Driouche is a Ph. D. candidate in LIRE laboratory at Mentouri
University of Constantine, Algeria. She is currently working in the
area of application integration for the last three years. Her research
interests include building and integration of ontologies, Web
services composition and business processes modeling.
Zizette Boufaida is a full professor since 2003 in the Computer
Science department of the university of Constantine, Algeria. She is
actually, vice-dean of Faculty of Engineer, in charge of research and
external relationships. She is also in the program committee of
DEXA. Her current research activities are conducted in SIBC team of
LIRE Laboratory and include advanced information systems and
semantic web.
Fabrice Kordon is a full Professor at the University P. & M. Curie and
the head of the team "Modeling and Verification" in the "Network
and Distributed Systems" department at LIP6 lab. His research
interests mainly concern the development of high-level specification
languages and their relationship with formal modeling and analysis
in the context of model driven development for concurrent
software.
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Fourth International Conference
I-ESA 2008
Interoperability for Enterprise Software and Applications
Science meets Industry
Berlin, Germany
Pre-conference: March 25th, 2008
Conference: March 26th – 28th, 2008
Call for Papers
Scope
Interoperability in the context of enterprise applications is the ability of a system
or a product to work with other systems or products without special effort from the
customer or user. The capability to interact and exchange information both
internally and with external organisations (partners, suppliers, customers) is a key
issue in the economic sector. It is fundamental in order to produce goods and
services quickly and at lower cost, while ensuring higher levels of quality and
customisation.
Today, companies maximise flexibility and speed of response to changing market
conditions, by focusing on core innovation and information centred activities,
outsourcing capital and labour intensive activities to less advanced economies.
They develop knowledge and links with other companies with which they can
collaborate to provide the products and services that the market demands – the
networked enterprise, or virtual organisation. Successful enterprises in this
knowledge economy may be small, but are highly productive, focussed on
innovation and capable of flexibly and rapidly sharing information and knowledge
to participate in highly competitive networked enterprises. The issue of
interoperability within the enterprise is therefore no longer limited to the
interoperability between silos of systems within single companies, but has become
one of interoperability throughout a value chain.
Furthermore, it is now recognised that interoperability of systems and thus sharing
of information is not sufficient to ensure common understanding between
enterprises. Knowledge of information meaning and understanding of how is to be
used must also be shared if decision makers distributed between those enterprises
in the network want to act consistently and efficiently. Enterprise interoperability
must therefore now embrace sharing of collaboration knowledge and core
competencies as well as the administration of such shared knowledge – knowledge
oriented collaboration.
Initiated in 2005 by two major European research projects of the 6th Framework
R&D Programme of the European Commission, the ATHENA IP (Advanced
Technologies for Interoperability of Heterogeneous Enterprise Networks and their
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (1), 2007
Applications, Integrated Project) and the INTEROP NoE, (Interoperability Research
for Networked Enterprise Applications and Software, Network of Excellence), the IESA conferences have been recognized as a tool to lead and generate an extensive
research and industrial impact in the field of interoperability for enterprise
software and applications.
I-ESA brings together the world’s leading researchers and practitioners in the area
of enterprise interoperability, and it is a unique forum for the exchange of visions,
ideas, research results and industrial experiences, dealing with a wealth of
interoperability research subjects for business, enterprise applications and
software.
The industry will find at I-ESA’08 an outstanding opportunity to exchange
experience and problems on business interoperability in their daily operations. IESA series of conferences have become a forum where industry and research worlds
meet and share experience, ideas and problems on different interoperability
aspects. In the I-ESA 08 arena the industry will point out the difficulties
encountered in their business life and show the way taken to solve those problems
while interrelate with other organizations sharing experience. The industry will be
able to meet interoperability experts and learn from them about emerging new
solutions and even establish relationships for further research and development
and problem solving. Special sessions will address industrial interoperability issues
for large scale industry like Automotive and Aeronautic and medium as well as
small sized ones like Furniture. Additionally, I-ESA’08 will feature a high-quality
pre-conference programme consisting of workshops, tutorials, the International
School on Interoperability and a doctoral symposium. During the main conference,
there will be an attractive exhibits programme and an additional brokerage event,
including a hands-on "industrial use case challenge".
The Doctoral Symposium is an opportunity for students involved in the preparation
of their PhD in any area of Interoperability for Enterprise Software and Applications
to interactively discuss their research issues and ideas with senior researchers, and
receive constructive feedback from members of the research community and other
Doctoral Symposium participants. Besides scientific matters, students will also have
the opportunity to seek advice on various aspects of completing their PhD and
performing research as a young professional in groupware.
Topics
Full and original scientific papers of high quality are invited to I-ESA’08. The
conference encourages submissions in (but not restricted to) the following areas:
Business Aspects of Interoperability

Micro-/Macro Economic impact of interoperability

Business interoperability requirements

Business Process Management
Interoperability Scenarios

Success stories and lessons learnt

Initiatives and communities

Problem statements

Case studies in interoperability solutions and their impact on day2day
business

How to reduce application gaps of new interoperability results

New Interoperability Scenarios and derived requirements – e.g.
Interoperability of Business Services
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Business Interoperability

Computer-supported negotiation, agreement, and contracting for
interoperability

Managing differences between networked enterprises in terms of culture,
languages and work processes

Business Interoperability frameworks and their application

Interoperability issues in electronic markets

Measuring impact of new interoperable solutions

Knowledge management in networked enterprises

creation, transfer, distribution, management and application

Legal issues for interoperability

Security issues in interoperability

Business models for solutions and services

The human factor in interoperability
Systems engineering in the context of interoperability

Integration and federation of Product, Systems and Project Management

How to act as supplier for different OEMs and their systems

Catalogues and libraries

Managing complexity and variety

Corporate Standards

Control systems engineering and integration/federation

Reducing gaps between Sectors – e.g. Production, Logistics, and Finance

Reconciling requirements between different kind of partners
Specification and Implementation

Enterprise Application Integration in the context of networked enterprises

Self-organization and emergent behavior in the context of interoperability

Requirements engineering for interoperable enterprises

Design Methodologies

Validating, and verifying solutions and methodologies

Service Engineering and Methodology
Architectures and Frameworks

Model Driven Architectures

Service oriented (Enterprise) Architectures

Agent based approaches to interoperability

Ubiquity, mobility, open architectures and large scale interoperability
Enterprise Modeling for Enterprise Interoperability

Distributed Modeling

Synchronization of models

Meta-data and meta-models

Methods, tools and frameworks for (networked) enterprises

Semantic mediation and enrichment of enterprise models

Service Modeling for Business
SW - Platforms

Open platforms supporting collaborative businesses

Non-functional aspects of interoperable solutions

Intelligent infrastructure and automated methods for business system
integration
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (1), 2007
Interoperability of operational Systems

Aspects of Service Management
Semantics for Enterprise Interoperability

Semantic web based approaches

Enterprise applications analysis and semantic elicitation

Reasoning methods and tools for model transformation
reconciliation

and
data
Workshops and Tutorials
The 4th I-ESA conference invites proposals for workshops to be held in conjunction
with the main conference. Following the successful organization of the conference
and workshops in the first three I-ESA conferences, this year the workshop proposal
topics are open to other areas not covered by the main conference topics.
Workshop proposals should explore new interoperability issues, problems and
solutions in areas coming from not only enterprise software and applications, but
also from eHealth, mobile services, and end-user software and applications to
social and financial issues related to the introduction of interoperability.
I-ESA 2008 is the fourth in a conference series that provides one of the most
prestigious forums for researchers and practitioners in all aspects of
interoperability of enterprise software and applications. The conference has a
record of attracting innovative research of high quality related to all aspects of
interoperability (theory, frameworks, methods, techniques, software architectures,
empirical findings).
The conference and workshops will be complemented by tutorials and an
exhibition.
Workshops are meant to facilitate the exchange of ideas and experiences between
active researchers, and stimulate discussions on new and emerging issues in topics
related to interoperability. The workshops should be organized in a way that will
leave sufficient time for brainstorming and discussions between attendees, so that
at the end new research directions and topics can be identified. Workshops may
concentrate in-depth on research topics, or may also be devoted to application
and/or standardization issues.
Committees
GENERAL CONFERENCE CHAIR
Chair: Prof. Kai Mertins, Fraunhofer IPK Berlin, DE
Co-Chair: Dr. Rainer Ruggaber, SAP Research, DE
INTERNATIONAL PROGRAMME COMMITTEE
Chair: Prof. Keith Popplewell, Loughborough University, GB
Co-Chair: Prof. Xiaofei Xu, Harbin Institute of Technology, CN
STEERING COMMITTEE
Prof. Guy Doumeingts, Emeritus Professor of University Bordeaux 1, GFI Group, FR
Ms. María José Núñez. AIDIMA, ES
Prof. Kai Mertins, Fraunhofer IPK Berlin, DE
Ms. Petra Frenzel, SAP Research, DE
Dr. Rainer Ruggaber, SAP Research, DE
ORGANIZING COMMITTEE
Chair: Thomas Knothe, Fraunhofer IPK Berlin, DE
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Co-Chair: Amparo Roca de Togores. AIDIMA, ES
Co-Chair: Cathy Lieu, University Bordeaux 1, FR
Co-Sponsorship
I-ESA 2008 is co-sponsored by:
IFIP- International Federation for Information Processing
Forum on the Technology Platform on the future manufacturing technologies:
MANUFUTURE, Germany
EIC - Enterprise Interoperability Centre, Belgium , Belgium
IFAC - INTERNATIONAL FEDERATION OF AUTOMATIC CONTROL
Deutsches Forum für Interoperabilität e.V., Germany
Important Dates
Conference
Full paper submission deadline: September 30th, 2007
Author notification due: October 31st, 2007
Camera-ready copies due: December 15th, 2007
Workshops & Tutorials
Workshops & Tutorials submission: November 9th, 2007
Acceptance notification due: December 15th, 2007
Information for authors
We encourage authors to submit original papers via the electronic I-ESA conference
management system. The paper submission deadline is September 30th, 2007. We
ask authors to register the electronic metadata of their publications (title, coauthors, keywords, short abstract) by September 20th, 2007. Only full paper
submissions will be accepted.
Previous I-ESA Conference proceedings have been published by Springer, this year
the proceedings will be again published by a recognised international publisher
with ISBN registration. Accepted papers will be included in the Conference
Proceedings book, only if they are presented by one of the authors at the I-ESA
2008 conference.
The maximum paper length is 12 pages according to Springer LNCS style (see the IESA homepage for details or http://www.springer.de/comp/lncs/authors.html).
International Scientific Journals have been supporting and observing the I-ESA
conferences, inviting authors of selected papers to publish there.
Submit NOW your Expression of Interest by www.i-esa.org, or contacting the IESA’08 Conference Chair ([email protected]).
Website
http://www.i-esa.org
© IBIS – Issue 2 (2), 2007
http://www.ibis-journal.net ISSN:1862-6378
IBIS – Issue 2 (2), 2007
© IBIS – Issue 2 (2), 2007
IBIS – Interoperability in Business Information Systems
Call for Articles
New issues of the IBIS journal are released three times a year. Authors are
encouraged to submit articles for inclusion in the next issue of the International
Journal of Interoperability in Business Information systems.
We are interested in both theoretical and practical research. We are also
interested in case studies and results of recently finished research projects.
Although our focus is on research results, we also accept a small amount of
practical articles. Contributions should be original and unpublished and need to
have a high quality.
Topics of the journal include, but are not limited to:
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Integration of business information systems
Enterprise modeling for Interoperability
Interoperability architectures and frameworks
Interoperability of business standards
Intelligent mediators
Coupling of information systems
Interoperability of classification systems
(Semi-)Automatic transformation of standards
Interoperability of (meta) models
Semantic integration concepts
Interoperability between domain specific standards
Semantic analysis of structured, semi-structured, and unstructured data
Interoperability of catalog based information systems
Cooperation in heterogeneous information systems
Ontology matching, merging and aligning
Semantic combination of heterogeneous standards
Ontology- and model management
Interoperability of sector specific systems
Authors are encouraged to submit high quality articles online. Please carefully look
at the submission guidelines to prepare your submission. All submissions should be
made online using our online submission system. In case of any problems, please
contact us via email.
For additional information and deadlines, please visit our website at
http://www.ibis-journal.net
© IBIS – Issue 2 (2), 2007