Preliminary results of using an Intelligent Agent and Case Based Reasoning for Quadratic Portfolio Optimization Edward Falconer, Abel Usoro, Mark Stansfield, Brian Lees There have been a number of intelligent agent systems that have assisted portfolio management decisions. The majority of these systems have focused on analysis methods, such as fundamental analysis, technical analysis or analysis of trader behavior. According to Clarke et al, because of the efficient markets hypothesis, portfolio managers should limit the time that they spend on analysis of this type and instead concentrate on other aspects of portfolio management, such as diversification, asset allocation or cost management. This study investigates the application of intelligent agents and case based reasoning to strategic asset allocation. Traditionally, strategic asset allocation is calculated using tools for quadratic optimization and determining an investor’s utility towards risk. Associated with strategic asset allocation are many complexities, such as the distributed nature of problems that make a solution difficult to implement. This paper outlines results of an evaluation of whether intelligent agent and case based reasoning could overcome these complexities. From a review of literature there does not appear to have been a study that has considered this problem in the way proposed in this paper. Keywords: Intelligent agents, quadratic optimization, portfolio management, asset allocation, case based reasoning. 1. Introduction The efficient markets hypothesis has led academics and practitioners to divide investment management approaches into two primary types: i) active investment management, where an investment manager uses available information and forecasting techniques to seek better performance than a diversified portfolio; and ii) passive investment management, where an investment manager has minimal expectations about the returns that can be achieved and relies instead on diversification to match the performance of a market index (Clarke et al., 2001; Fama, 1991). Although each of these approaches has benefits, there has been strong evidence to support the claim that the profitability of active investing is minimal, especially after transaction costs, and therefore passive investing is more profitable (Clarke et al., 2001; Bernstein, 2000; Bernstein, 2002). One of the implications of choosing a passive investment approach is that portfolio managers must concentrate on other strategies for improving the return from their portfolio (Clarke et al., 2001). Examples of these strategies include cost management, diversification and asset allocation strategies. This research concentrates on asset allocation. In general, asset allocation is a process whereby an overall sum of cash is separated into different portions. Each portion is then allocated to a different category of asset class. According to Sharpe (1992, 2001) each asset class should represent a capitalisation-weighted portfolio of securities in order to mimic the return variation created by different weights of asset class in the returns of the portfolio under evaluation. There exist many different approaches to asset allocation. These variations include strategic, tactical and dynamic asset allocation, amongst others. This research concentrates on strategic asset allocation. The optimal strategic asset allocation can be determined in one of two ways: by maximizing return for a given risk level or alternatively, minimizing risk for a particular return objective (Nemati and Iyer, 1999; Kaplan, 1998; Todd and Markowitz, 2000). A strategic asset allocation decision is made up of a quadratic component and a utility component (Nemati and Iyer, 1999). The quadratic component is concerned with constructing the portfolio under quadratic constraints whilst the utility component is concerned with determining the investor’s attitude towards risk. Moreno et al., (2005) described Markowitz’s mean variance approach as the classical model for risk forecasting and the preferred approach for making strategic asset allocation decisions. An earlier version of the work described in this paper was presented at the International conference on intelligent agents, web technologies and internet commerce – IAWTIC – in Vienna Austria (20061). This research considers the application of a combined intelligent agent and case based reasoning (CBR) approach to the strategic asset allocation problem. Broadly speaking literature on intelligent agent and case based reasoning in finance has focused on three areas: i) portfolio monitoring, ii) stock selection, and iii) behavioural finance. Portfolio monitoring is defined as the ongoing, continuous, daily provision of an up-to-date financial picture of an existing portfolio (Decker and Sycara, 1997). Stock selection is defined as an analysis technique which uses valuation and forecasting techniques to identify mis-priced assets (Davis and Liu, 2002). Behavioural finance is defined as the computational study of economies modelled as evolving systems of autonomous interacting agents (Tetsfatsion, 2001). All of the artificial intelligent studies that have been found have focused on analysis methods (such as fundamental analysis, technical analysis or analysis of 1 The conference proceedings are currently in print therefore page numbers cannot be provided. Expected publication date is January 2006. 1 trader behaviour), which are aimed at identifying investment opportunities with above average returns. The determination of an investor’s utility or support of quadratic programming does not appear to have been considered. This research therefore assesses the effectiveness of intelligent agents and CBR in dealing with the complexities associated with this problem. This paper concentrates on the quadratic program calculations under mean variance efficiency. Utility theory will be considered in a subsequent paper. The paper reports preliminary results on the effectiveness of a prototype CBR and agent based system towards the quadratic programming problem. The rest of this paper provides: (i) an introduction to strategic asset allocation; (ii) a background to the research problem and an overview of the data preparation steps; (iii) an overview of agent and CBR characteristics; (iii) an outline of the agent architecture; (iv) an application of case based reasoning to the problem and preliminary results; (v) conclusions and future directions. 2. Strategic Asset Allocation Strategic asset allocation refers to a situation where a portfolio is constructed using mean-variance optimisers and is considered a long-term investment strategy which is modified at regular intervals. Tactical asset allocation is the process of diverging from the strategic asset allocation when an investor’s short-term forecasts deviate from the long-term forecasts used to formulate the strategic allocation. Dynamic asset allocation refers to strategies that continually adjust a portfolio’s allocation in response to changing market conditions. The most compelling evidence for the use of risk-based strategic asset allocation is provided by Brinson, Hood and Beehower (referred to as BHB). According to Bitters (1997: page 56) “the contribution they made was revolutionary”. BHB questioned the importance of asset allocation and compared the returns that are achieved by an investment manager from factors such as: market timing, stock selection and asset allocation policy. BHB questioned an investment manager’s abilities to choose stocks, and time deals, against their ability to allocate assets through an asset allocation policy. BHB compared those adopting a strategic asset allocation2 approach with those adopting a dynamic or tactical asset allocation approach by measuring the portfolio return achieved by introducing active analysis after making a risk-based strategic portfolio choice. According to Bitters (1997), BHB observed that: (a) Investment policy return explained on average 93.6 percent of the total variance in actual plan return. (b) Returns due to policy and market timing resulted in average variance explained by 95.3 percent. (c) Returns achieved by policy and security selection returned in an average variance explained by 97.8 percent of the overall return. BHB showed that on its own, risk-based strategic asset allocation produced an average return of 93.6% and when BHB actually used the term “asset allocation policy” which was in vogue until Brennan et al (1997) coined the term “strategic asset allocation”. 2 combined with non-strategic approach, the increase in return was insignificant (4.2% at the most, ie the third case above). BHB also discovered that investment managers opting for an active approach, that is those using market timing or stock selection techniques and tactical asset allocation decisions, resulted in a 1.10 percent reduction in the annual overall plan return (Bitters, 1997). In addition to this, they discovered that involving active management actually increased the risk associated with a portfolio and produced less returns than portfolios constructed purely through strategic asset allocation (Bitters, 1997). The works of BHB highlight one of the major questions facing the investment community, namely whether or not the abilities of an investment manager can add any additional profit to a portfolio. BHB concluded that investors are better-off if they adopt a strategic asset allocation and ignoring the alternatives. Bitters even went as far as to say that investors are actually disadvantaging themselves by opting for anything else. BHB provided compelling evidence for the claim that investors should therefore choose the passive approach and use strategic asset allocation. Unfortunately according to Rekenthaler (2000) the BHB article was widely misquoted and led to the belief that asset allocation was the answer the investment community had been looking for. However, a preoccupation with asset allocation led to a pre-occupation with asset allocation methods, often to the detriment of the investor. According to Rekenthaler (2000), the debate raged on until Ibbotson and Kaplan (2000) published a paper that suggested the returns could be 40, 90 or 100 percent of performance depending on how the data was interpreted. Current opinion is that strategic asset allocation remains a central element in the process Rekenthaler (2000). However, attempts at determining how useful are thought to be fruitless (Rekenthaler, 2000). The view of many is that 100% of the return achieved on a portfolio is through asset allocation (Jankhe, 1999). The evidence therefore remains stacked in favour of the strategic approach. 3. Mean Variance Efficiency Portfolio selection (also known as Mean Variance Efficiency) considers the problem of efficiency and the identification of what Markowitz (1992) defines efficient set analysis. Portfolio selection and portfolio efficiency are about variance, return and risk. A portfolio is considered efficient if the mix of assets between bonds and stock investments, and the variance in return from those investments, meet the risk profile of an individual investor. Thereby a portfolio is considered inefficient if another portfolio can be constructed with a higher expected return and the same variance, or with lower variance and the same expected rate of return (Markowitz, 1992). In 1990 Markowitz was awarded a share of the Nobel Prize in economics for his contribution. According to Markowitz the investor could be seeking more risk or they could be seeking less of it. This depended on their expected rate of return. The problem therefore became a quadratic equation with the amount invested in each security determined by the risk preferences of the 2 investor and the risk and return associated with the securities. Investing in securities with different rates of risk and return balances out the risks involved if the investor chose only one security for their portfolio. The belief is that the return on a portfolio would be protected from losses if an investor chose securities that have different variances and covariances. Markowitz proved that this was the case under quadratic constraints. The aim of quadratic programming in the portfolio selection problem is to calculate the amount that should be invested in different securities based on the covariance and expected returns related to those securities. The quadratic program uses a value representing the investor’s risk aversion as an input variable and adjusts the amount that is invested in each security based on either the expected returns or the covariance of return, within a portfolio. Markowitz (1952, 1959, 1992) defined the following objective function to determine how to maximise returns or minimise risk subject to an individual’s risk preferences. Min(max)imise Σ µi xi – p Σ Σ σij xi xj Subject to: Σ xi = 1 & xi >= 0 for all i. In the function, p represents the investor’s aversion to risk, µi represents the return on the ith security, xi represents the amount invested in the i’th security and σij refers to the covariance of the ith and jth security. The simplest way of implementing a solution under quadratic and linear constraints is to substitute the investor’s risk aversion for a value representing the minimum portfolio return (Fylstra, 2005). This minimum return value then replaces the letter p in the equation and allows the portfolio to be determined based on return and risk. By taking this approach the quadratic programming problem is about which asset classes have been chosen and the minimum portfolio return. 4. Data Preparation The expected return and covariances used within the quadratic calculation are determined using historical index pricing information. In contrast to the standard approach to portfolio selection, the securities, in this case, are in the form of global asset prices. The global approach differs slightly to the standard asset allocation approach as the focus is placed upon prices that reflect all equities within a specified global market. In contrast, asset allocation approaches usually focus on much broader asset allocation classes, such as real estate, stocks, bonds and cash. Despite the differences, there are many advantages to adopting a global approach. These advantages include global asset coverage and the availability of daily price updates. This decision is also supported by the globalisation of markets discussed in literature and the importance of global diversification (French, 1991; Lummer et al., 1994a; Lummer and Riepe, 1994b; Kaplan, 1998; Dempster et al., 2003). Gerber (1994) and Black and Litterman (1992) provide compelling evidence for the use of global asset classes. In particular, they discuss the benefit that is gained from broad diversification across capital markets and the high degree of correlation in return that appears at times within these markets. Lummer and Riepe (1994a; 1994b) and Kaplan (1998), also discuss how global asset allocation provides the opportunity to apply Mean Variance Efficiency more effectively because of the general stability of global markets. The historical global asset prices were obtained from Morgan Stanley Corporate International (MSCI©). According to the MSCI website, ‘close to 2000 organisations worldwide currently use the MSCI international benchmark indices…MSCI estimate that over USD 3 trillion are currently benchmarked to these indices on a worldwide basis’. The following global equity asset classes were used for the calculation: UK, Japan, Australia, Thailand, Colombia, Brazil, USA, Canada and Germany. This selection of asset classes provides broad coverage across developing and established markets. Although the assets are limited to those specified, the model could easily be extended to include additional asset classes, cash assets or bond assets, with minimal effort. Under portfolio selection theory, return and risk is determined by obtaining the following values for each of the chosen asset classes (Markowitz, 1992): Ci – Pi + Ii Ci The letter C represents the closing price from the current year, the letter P represents the closing price from the previous year and the letter I represents the income from the index. The lower case i represents the year. Each mean return is then represented in terms of a percentage returned from a specific year. In this study, the mean variance calculation was carried out using ten years worth of asset prices. Annual returns and mean returns are used to determine the extent to which an individual asset class deviates from the average asset class price. In addition to this, the variances from the average return are also used in the calculation. The following equation is used to determine the deviation of the annual return: Deviation from the Average Return = (ri) – (r’i) The letter r is the return variable (annual return), r’ is the average return and lower case i represents the year. The standard deviation of return is then determined from the overall portfolio variance. Table 1 provides an example of the annual return, average return and standard deviation of return, of the UK market over a ten year period (priced in US dollars). Name: UK Price: USD 3 Date Annual Price Dec 1993 Dec 1994 Dec 1995 Dec 1996 Dec 1997 Dec 1998 Dec 1999 Dec 2000 Dec 2001 Dec 2002 Dec 2003 Dec 2004 641.457 611.077 716.422 883.481 1,052.09 1,208.25 1,325.94 1,146.25 962.025 791.076 1,006.08 1,162.44 Average (Mean) Standard deviation = Annual Return Deviations Squared deviations -0.049715502 0.147043223 0.189091786 0.160264197 0.129239303 0.088759186 -0.156755832 -0.19150126 -0.21609681 0.213705457 0.134511542 0.040776844 0.154781582 0.090492347 -0.106266379 -0.148314941 -0.119487353 -0.088462458 -0.047982342 0.197532677 0.232278105 0.256873654 -0.172928613 -0.093734697 0 0.008188865 0.011292543 0.021997322 0.014277227 0.007825607 0.002302305 0.039019158 0.053953118 0.065984074 0.029904305 0.008786193 0.023957338 Table 1:Computation of Standard Deviations, Mean Returns and Variances, using broad asset classes (Calculations based on Markowitz's theory of meanvariance efficiency) Covariance and correlation coefficient were then calculated across each of the global market asset classes (these correlation coefficient matrix and expected returns have been detailed in Appendix 1). Groups of assets were then chosen and the coefficients from the countries were loaded into a quadratic program (QP) which performed optimisation calculations. The QP was then configured to maximise return or minimise risk using a variety of pre-determined constraints. The quadratic program, as previously explained, uses the variables that it is given to calculate how much must be invested in each asset class to achieve a certain level of return for a certain level of risk. A quadratic program named MOSEK© was used to perform the calculations. In order to choose assets a front end application was developed. The user could also choose a specified minimum portfolio return. The data was then fed into the quadratic program for the calculations to be made. The covariance and mean return values for each of the chosen assets are collected from a variety of flat CSV (comma separated variables) files. Prior to passing the information to the quadratic program the data was formatted in a way that could be interpreted by the quadratic program. The outputs of the quadratic program are suggested investment weightings which would maximise return subject to the constraints applied. The prototype software agent and CBR are written in VB.NET and use a class library (via a VB.NET API link to MOSEK©) to link to the quadratic program. 5. Agent & CBR Characteristics Strategic asset allocation is characterised by a variety of complexities that have to be considered in the design of a decision support system (Nemati and Iyer, 1999). These complexities are related to the problem of matching investors with portfolios and the overall application of the strategic asset allocation theory. The complexities are listed below. (i) The strategic asset allocation task is distributed over various parts; this means that different factors contribute to the matching of an investor with a portfolio. (ii) Some tasks require input from other tasks to arrive at a conclusion; thereby necessitating the need to share data and control the flow of information between tasks. (iii) There is a chance that the same set of data inputs could be used multiple times; this means there is a potential for task duplication. (iv) There is not one solution that can be applied to every set of circumstances; this means that each portfolio or investor could be either unique or at best similar but often not exactly the same as those encountered previously. (v) In order to match up an investor with a portfolio it is necessary to have input from a user; and this user could require a different level of assistance depending on their preferences. In the light of these observations a review was conducted of available technologies and intelligent agent and case based reasoning were identified as displaying the types of attribute that could deal with the complicated nature of the strategic asset allocation decision. Agent and CRB technologies have been used to successfully deal with the complexities of portfolio management (Davis and Luo, 2001; Davis et al., 2000; Decker and Sycara, 1997). 5.1 Intelligent Agents 4 Intelligent agents are software artefacts that have a number of characteristics that distinguish them from other technologies. In particular, an agent can be given a goal or task which it can complete autonomously (Turban and Aronson, 2001). Agents may also carry out tasks or goals in parallel with the existing computer operations and ask for advice from humans if they become “stuck” in a problem (Wooldridge, 2002). Agents can act proactively and communicate with human users, providing a personalised service to some users and communicating with other agents in a multi-agent environment (Turban and Aronson, 2001). incorporate this autonomy by taking the collection of knowledge and forming the suitably structuring the data. The overall aim is usually for the intelligent agent system to support the problem solving activities of the human expert. Turban and Aronson (2001) also discuss how over time it may be necessary to maintain the knowledge base, by identifying and correcting any inconsistencies, and by filling in any perceived gaps through the acquisition and inclusion of additional knowledge. The ability to develop knowledge and interact with a user is also required within the solution. Wooldridge (2002) discuss how intelligent agents can interact with, and react to, a user’s preferences. This means that software can be designed to recognise a user’s actions and modify the user/agent interaction to ensure the level of support is appropriate to that user. Decker and Sycara (1997) also discuss how intelligent agents can incorporate a separate control component to regulate the flow of information between distributed tasks. This control component can organise the flow of information that passes between the various parts of the problem and ensures that a conclusion is reached. This ability to control distributed tasks is a key attribute of the anticipated solution. Intelligent agents therefore demonstrate various abilities that suggest they may be appropriate for resolving this problem. In particular the ability of intelligent agents to interact with human experts, organise data flow, construct knowledge bases and learn from experience are perceived to be important for overcoming the complexities associated with matching portfolios with investors. In order to build the proposed solution it is necessary to acquire knowledge from a variety of knowledge sources: human expertise and recorded knowledge of previous successful problem solutions. This means the system must have an element of autonomy. Intelligent agents can be 5.2 Case Based Reasoning Case-based reasoning (CBR) is a methodology that can allow software to learn from experience. Case based reasoning subdivides a problem into a series of tasks and then combines the results of each task to form a case. Each case is then compared with previous cases to determine whether new experiences can be learned. Fig 1 illustrates a general CBR cycle (Althoff, 2001) the figure illustrates the four tasks that exist as a part of all CBR programs. New Problem Situation Retrieve Similar Cases Reuse case information Case Base Retain experience Revise proposed solution Proposed Solution Fig 1: General CBR Program Cycle Fig 1 identifies four tasks: Retrieve task – to retrieve similar cases; reuse task – to reuse information and knowledge from previous cases to solve a problem; revise task – where the problem is revised in line with previous experiences; retain task – where experience is retained that could be used for future cases. Using this general cycle a software program can learn from its experiences. CBR was found to offer the types of capabilities that may make it appropriate for dealing with the complexities of the strategic asset allocation problem. For instance Althoff (2001) discusses how CBR systems gather case data from distributed tasks and join the outputs of these tasks together to formulate a combined state. This combined state then contributes to any future development. In this way a 5 distributed problem can be resolved, with each distributed solution contributing to all future solutions. Leake (1996) also discusses how CBR systems can be used to subdivide each solution into different parts and then pass requests for analysis onto other components. By initiating this task division CBR software can perform distributed tasks at the same time as other tasks are performed and then combine the parts to arrive at a more complete solution. Althoff (2001) identifies that CBR systems evaluate each set of results to determine whether the current case is new, or whether it matches a previous case. The CBR system then adds new cases and modifications to existing cases to the case base for the future benefit of the software. Each case is made up of a set of responses that are unique to each problem and solution. By comparing current information and approaches with new sets of input CBR programs are able to identify similarities and differences with each new set of inputs. In so doing the CBR system therefore learns from its experiences. CBR was therefore found to offer additional capabilities that would strengthen the intelligent agent approach. It is the ability of CBR to organise distributed tasks, subdivide solutions into categories, historically evaluate solutions and learn from experience that mean it may be applicable for dealing with the complexities associated with strategic asset allocation. 6. Intelligent Agent Architecture The purpose of the agent architecture is to define how the tasks are separated and organised. The agent architecture is designed to carry out distributed tasks separately then combine the results to make a strategic asset allocation decision for the user. The agent architecture is shown in Fig 2. There are five parts to the agent: (i) investor profiler; (ii) asset selector; (iii) control component; (iv) case based reasoning component; and (v) maths calculator. Each component is outlined below. User Question Database Investor Profiler Asset Selector Asset Pricing Files Control Component QP (Mosek) Maths Calculator CBR Program Case Base Fig 2: Prototype Architecture for Asset Allocation Investor Profiler: The investor profiler retrieves questions from a questionnaire class linked to a database. The database stores a series of questions which it asks the user. Each question also has a set of responses. The investor profiler asks the user the questions and presents some possible answers. The investor then chooses one of the answers from those presented. The investor profiler then writes these results to the database and sends the investor’s risk profile to the control component. Asset Selector: The asset selector component is tasked with choosing assets that match the profile of the investor. The choice of asset is evaluated based on the information passed by the control component which initially comes from the investor profiler. The asset selector retrieves input in the form of asset prices from asset pricing files. It passes the relevant information to the control component. The asset selector component also communicates the results to the user. Control Component: The control component is tasked with controlling the flow of information between the agent’s various parts. The control component is used primarily to control the flow of information within the agent environment. The control component is tasked with passing information for processing and passing processed information around the agent environment. The Maths Calculator: The maths calculator component interacts with the quadratic program and modifies the 6 constraints, number of variables and other inputs, to allow the quadratic program to make its calculations. The maths calculator receives requests from the control component about possible inputs to the quadratic program. The maths calculator determines the inputs and reads the solution. It plays the part of a driver program for information being passed to the quadratic program (QP) program. Mosek was used to perform the quadratic calculations. The Case Based Reasoning component: The CBR component interacts with the control component and tracks information received from the different parts of the agent architecture. It uses inferences around the inputs that have been received previously to determine whether information can be used again, or whether parts of a previous problem can be used again. The CBR sends its results to the control component. A unique reference method is used to ensure the appropriate results are sent to the appropriate component. The agent program is designed to support the activities associated with the asset allocation task. The remainder of this paper outlines preliminary results for the maths calculator and provides an overview of the work undertaken into the application of CBR to the problem. The additional components will be covered in a separate paper. 7. Preliminary Results There are two sets of results to report, findings related to: (i) the math calculator component and its interface with the quadratic program; (ii) the application of CBR to quadratic programming problem. Although the results are related, they are considered separately as they were found to each pose a different set of challenges, and were complex to explain when combined together. 7.1 The Quadratic Calculator Program and Math The QP used a quadratic function to determine the optimum investment amount to place in each asset class. The following objective function was used to maximise return or minimise risk (from an article by Fylstra (2005). maxmise cTx - xTQox subject to Ax >=b Σx = 1 xi >= 0 for all i. o Q is a matrix containing covariances for the assets held within the portfolio. cT is a vector containing returns on each of the assets held within the portfolio. A is also a vector containing the returns on each of the assets held within the portfolio. b is a value that represents the minimum portfolio return. x is the amount to be invested in the ith asset class within the portfolio. Within the function, xTQox represents the risk associated with the portfolio and cTx is the return on the portfolio. The values associated with the portfolio are then adjusted based on the minimum portfolio return and the asset selector’s choice of asset class. Within the equation the constraints (Ax >=b; Σx = 1; xi >= 0 for all i) remain the same, with the values associated with the Qo, cT and A, adjusting in line with the input of the problem variables. The maths calculator has a front end application which allows a user to choose asset classes, choose a minimum portfolio return, view covariances and mean returns, and view the optimised result. The program is illustrated in Figure 3. The math calculator retrieves the values for the matrices Qo, cT and A from a set of flat files. A separate flat file exists for each of the mean returns (cT and A) and the covariances (Qo). The data tables are detailed in appendix I. In order to make the calculations, the user first chooses two or more asset classes from those available and selects a minimum portfolio return. Once the user has chosen the asset classes and minimum portfolio return, they press the optimise button and initiate the quadratic calculations. On selecting the optimise button the math calculator first retrieves from the asset selector which obtains the data from the flat files. The maths calculator then formats the data for the quadratic program. This formatting stage involves locating and retrieving the mean returns and covariances associated with each of the chosen assets and leaving the irrelevant data. The math calculator then interfaces with the quadratic program and requests the quadratic results. The calculations that are then returned are displayed in the investment amount textbox. As well as this the math calculator also displays the mean returns and covariances in the respective textboxes. 7 Covariances (Qo) Mean Returns (cT, A) Minimum Return (b) Optimal Amount (xi) Figure 3: Maths Calculator with Example Quadratic Variables and Results When interfacing with the quadratic program the math calculator uses the values shown in Table 2 to initiate and format the data. These values are determined from either the data held within the matrices or the chosen asset classes. Each value serves a dual purpose: (i) to enable the maths calculator to input the appropriate data to allow the MOSEK tool to calculate the results; and (ii) to allow the CBR program to determine inferences about the data. The two constraints - namely Ax >=b and Σx = 1 - remain the same at all times with b changing in line with the investor’s minimum portfolio return. The size of the three matrices and vectors (Qo, cT and A) increase in line with the number of variables included in the portfolio. Value Number of variables Number of variables in the matrix A Number of variables in the matrix Qo Values within the matrix A Values for the matrix cT Values within the matrix b Values within the matrix Qo Description The number of asset classes chosen by the asset selector The same as the number of asset classes chosen by the asset selector Calculated using the following equation (where N is the number of asset classes chosen by the user): N(N + 1) 2 The expected return for each asset class The same as the expected return for each asset class The minimum portfolio return The covariance for each value Ni, Nj Table 2: Variables used by the MOSEK tool and determined by the maths calculator Despite the initial simplicity of the quadratic program in calculating the optimum results, from close inspection, it was apparent that at times it was not possible to produce results that were useful. An extensive evaluation was therefore conducted of the quadratic program and the outputs that were produced under various constraints. This section introduces results of an extensive evaluation that was conducted of the quadratic program and the outputs that were produced under various constraints. This evaluation involved evaluating 675 variations of portfolio choice. The range of choices was created by adjusting the minimum portfolio return (which ranged from 0.01 to 0.15) for the full range of 45 different asset allocation variations (with 9 asset classes). This evaluation was useful as it outlined the types of limitations that could be encountered by using a quadratic program. In particular from this evaluation it was revealed that at times it was not possible to produce results that were useful. This problem is discussed by Fylstra (2005) who indicates that analysts take great care to ensure that their portfolio models remain calculable under quadratic or linear constraints. These problems are encountered when an invalid choice of asset class and minimum portfolio return combination are chosen. The following four categories identify the types of result that were returned. 1. In the first category the optimal result is split based on the variables and constraints as they have been defined. The program therefore supplies optimal amounts that provide beneficial advantage and are sensible from an asset allocation perspective. Table 3 provides an example of a set of quadratic results that meet this condition. Country USA Canada Germany UK Total Variable X X0 X1 X2 X3 Optimum investment amount 0.289805523 0.306632754 0.189994446 0.213567277 1.000000000 Table 3: Optimal results for the quadratic code and sample inputs 8 2. In the second category the results are infeasible. This arises because the variables make the equation nonlinear and therefore not solvable by a quadratic equation. The example in table 3 shows the results returned for infeasible solution. Specifically, the total shows as 4 and thus fails the first test of feasibility, ie the constraint that summation of x’s should be equal to 1 (i.e. Σx=1). Country Japan Australia Thailand Columbia Total Variable X X0 X1 X2 X3 Optimum investment amount 1 1 1 1 4 Table 4: Results obtained for a Infeasible quadratic equation 3. In the third category although the results have passed the first test of feasibility (Σx=1), they are so skewed towards one or two asset classes that the costs of investing in the remaining asset classes outweigh the benefits. An example of this is shown in table 5. In this example the optimal result would involve investing 0.99999998504% in Germany and UK, but investing only 0.00000000576% in the USA and Canada In this instance In . This although mathematically accurate (Σx=1) is not beneficial from an investing perspective because the cost would outweigh the benefit. Country USA Canada Germany UK Total Variable X X0 X1 X2 X3 Optimum investment amount 0.00000000472 0.00000000104 0.06442543178 0.93557455326 1.00000000000 Table 5: Results obtained from a combination of asset classes 4. In the fourth scenario the results break the constraints that the sum of x must equal 1. Table 6 shows an example of this scenario. This was found to occur whenever a combination of asset classes and minimum portfolio returns caused the equation to become nonlinear in some places but linear in others. Country USA Canada Germany UK Total Variable X X0 X1 X2 X3 Optimum investment amount 0.1284890814 0.1415283229 0.0682738176 0.0803402387 0.4186314686 Table 6: Results obtained where the portfolio is not fully invested It was found from the evaluation that each quadratic result fell into one of these four categories. If category two occurred the user was asked to change their choice. If category three occurred the amount was proportionally adjusted so that the amount was equally split. If category four occurred the user was asked to change their choice. The CBR program is also being incorporated to resolve these problems by allowing the program to evaluate the user’s choice and identify suitable alternatives. These alternatives are aimed at improving the solution. 7.2. Case Based Reasoning Search and Retrieve Method This section outlines the development of the case-basedreasoning solution and the identification of the case features. The primary purpose of the CBR component is to assist with the problem of matching investor’s with portfolios and ensuring that the agent system makes efficient use of resources and time. The CBR program operates by first having the quadratic program and investor profiler carry out their tasks then providing the problem and solution variables to the CBR component. The CBR program then evaluates and monitors cases as new problems and solutions are introduced. The intension will be that as time passes the quadratic program, investor profiler and stock selector program will be used less and will be replaced by the solution that is provided by the CBR program. In general the CBR tool carries out the following steps: 1. Searches from its memory to find portfolios that have been previously been chosen for investors with specific responses and makes an assessment of similarity. 2. Infers an answer from the most similar matches. 3. Adjusts the portfolio to investor matching solution for changes in circumstances by inferring relationships between multiple sets of the historical data. 4. Modifies solutions from step three and records them for future use. The search and retrieve stages are completed in the first and second steps and will use the nearest neighbour retrieval method. The nearest neighbour retrieval method identifies possible matches by defining values that can be used to index each case. The indexes are then weighted in terms of importance. Each weight is used to calculate how far away a previous case is from the current case. Using the nearest neighbour retrieval method there are three states that the CBR program can find itself in when evaluating alternatives: (i) an exact match that is obtained by identifying that the same set of input values within the problem have been used before; (ii) no match that is obtained if the current case does not match any of the previous cases; or (iii) a partial match that is obtained if the current case is similar to previous cases, but not exactly the same. The distance between previous cases and a current case is proportional to the similarity between the features of each case. Previous cases given a partial match can be far away or close to the current case. However, the closer a previous case is to the current case the greater the chance that it will be considered as a possible match. Rule Description CountryName NumberOfCountries GlobalInfluence AverageVolume Exact Match 10 10 5 5 Partial Match 0 5 2 2 No Match 0 0 0 0 9 LengthEstablished EmergingRating Covariances MinimumPortfolioReturn MeanReturn Mosek Result RiskAttitude RiskPerception KnowledgeAndExperience OptimalRiskLevel Complexity InvestmentGoal InvestmentObjective InvestmentTimeHorizon InvestmentAmount EstimatedYearReturn InvestorExpectedReturn Total 5 5 10 10 10 10 10 10 5 5 10 10 10 10 10 5 5 180 2 2 5 5 5 5 5 5 2 2 7 5 5 7 7 2 2 87 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 Table 7: Case Features as they apply to the quadratic problem and investor profiler The purpose of the CBR program will be to categorise the portfolios chosen for specific investors so that previous cases can be compared as possible alternatives. Each case has a set of features that are used to measure how similar a previous case is to a current case. The features that relate to this particular problem have been identified and are shown in table 7. Each feature has been allocated a weight based on the importance of that feature to the problem. Therefore for instance the country name and number of countries are considered more important than global influence, average volume etc. This is apparent because country name and number of countries have a score of 10 whilst global influence and average volume have a score of 5. Each feature either relates to an aspect about either the portfolio or the investor. The features are used to determine how close a previous case is to a current case. The purpose is to categorise the portfolios based on the case information that is attained. The portfolios will be categorised with the score relating to each portfolio and investor combination. This score is then used to determine how close this case may be to previous cases. The portfolios will be categorised by the level of risk they contain. The values will be interpreted by considering the potential risks associated with the individual portfolio as it compares with other portfolios. An example of the matching strategy being adopted is noted below. Step 1: User chooses a portfolio chosen. During this step the user is asked to choose asset classes from the checkbox and selects a minimum portfolio return and selects the optimise button. In this example the user has chosen a portfolio containing the USA, UK, Germany and Canada with a minimum portfolio return of 0.15. The results of this calculation have already been shown in table 6. The user also provides input to the program that allows the investor to be characterised in terms of investment objectives, time horizon and knowledge and experience of investing. Step 2: Previous cases retrieved. During this stage all previous cases are compared with the portfolios chosen in step 1 for the type of investor identified by the user. This retrieval is based on the values identified in table 7. In total three portfolios have been found that have been chosen for investors with similar characteristic. These cases are each considered as a potential match for the current case. The first portfolio contains USA, UK, Thailand and Canada and has the same portfolio return as the current portfolio. In investor has some similarities to the current investor however the time horizon and risk profile are different. The first portfolio/investor combination achieved a 58% similarity and received the following scores: CountryName NumberOfCountries GlobalInfluence AverageVolume LengthEstablished EmergingRating Covariances MinimumPortfolioReturn MeanReturn MosekResult RiskAttitude RiskPerception KnowledgeAndExperience OptimalRiskLevel Complexity InvestmentGoal InvestmentObjective InvestmentTimeHorizon InvestmentAmount EstimatedYearReturn InvestorExpectedReturn Score 5 10 2 2 2 2 5 10 5 10 5 5 2 2 10 5 10 0 10 2 2 103 The second portfolio contains USA, UK, Canada and has a different portfolio return, however, the investor has different goals and investment amount. The second portfolio/investor combination achieved a 60% similarity and received the following scores: CountryName NumberOfCountries GlobalInfluence AverageVolume LengthEstablished EmergingRating Covariances MinimumPortfolioReturn MeanReturn MosekResult RiskAttitude RiskPerception KnowledgeAndExperience OptimalRiskLevel Complexity InvestmentGoal InvestmentObjective 10 0 5 5 5 5 5 10 5 10 5 5 5 5 10 0 5 10 InvestmentTimeHorizon InvestmentAmount EstimatedYearReturn InvestorExpectedReturn Score 7 0 5 2 109 The third portfolio contains USA, UK, Canada, Germany and has a different portfolio return. The investor however is very similar to the current set of investor responses provided by the investor. The third portfolio/investor combination achieved an 87% similarity and received the following scores: CountryName NumberOfCountries GlobalInfluence AverageVolume LengthEstablished EmergingRating Covariances MinimumPortfolioReturn MeanReturn MosekResult RiskAttitude RiskPerception KnowledgeAndExperience OptimalRiskLevel Complexity InvestmentGoal InvestmentObjective InvestmentTimeHorizon InvestmentAmount EstimatedYearReturn InvestorExpectedReturn Score 10 10 5 5 5 5 5 5 10 10 10 10 5 5 10 10 10 7 10 5 5 157 Step 3: Reuse case information. As described in diagram 1 this stage is completed based on the results of the previous step. Of the three portfolio/investor combinations that have been identified as possible matches the third portfolio most accurately meets the set of criteria. In order for the current problem to be identified as a possible alternative the minimum-portfolio-return and the features relating to the investor are all required to be very close to that of the current case. All three portfolios are displayed as possible matches to the user with the score displayed for each portfolio. The final stages of general CBR cycles (identified in diagram 1 as revise proposed solution and update case base and retain experience) have not been considered in this paper and will be considered at a later stage. One of the advantages of adopting CBR is that it is possible to allow the system to evolve and grow with time. This makes the quadratic solution particularly interesting for determining alternative solutions to a given problem. The distributed nature of the problem means that there are an exponential number of different combinations that could occur. Although the laws of averages would suggest that the same combination of results will be chosen over and over again, this cannot be confirmed. By incorporating CBR it is possible to cater for both the scenario where the situation has or has not occurred before and allow the system to learn from its encounters. This is especially beneficial due to the number of calculations that must be made to produce a complete data set. For instance, even with the numbers of asset classes reduced to 9 (as is the case here) the volume of calculations required for any portfolio optimiser to complete a proper evaluation of these cases reaches into its thousands. With the introduction of additional asset classes the process could therefore become even more problematic. 8. Conclusions and Future Direction This paper outlines a decision-support system to assist with strategic asset allocation decisions that is based on an intelligent agent. The system uses case-based reasoning to make portfolio optimisation calculations under quadratic constraints. These calculations are performed as part of an optimisation of the expected return and the minimisation of risk within a portfolio of broad asset classes. The results attained from an evaluation of the maths calculator present sufficient evidence to support the decision to use intelligent agents for the problem of matching investors to portfolios. In addition to this the ease with which the variables associated with the maths calculator fit with the general approach to CBR also support the decision to use CBR within the approach. Combining the capabilities of intelligent agent and CBR based systems therefore present the opportunity to support strategic asset allocation decisions. By combining these technologies the system could learn from its interactions with the user and also its experiences when evaluating problems. The next piece of work to be undertaken will involve the development of the additional agent components. Specifically this refers to the investor profiler which will consider utility theory in its determination of a final solution. Work will also be undertaken on extending the maths calculator so that it carries out conic optimisation calculations as well as quadratic calculations. References Althoff, K (2001), “Case-Based Reasoning”, ICCBR-01 Workshop On Case-Based Reasoning Authoring Support Tools, 31st July. Bernstein, W (2000) The Intelligent Asset Allocator: How to Build Your Portfolio to Maximize Returns and Minimize Risk, McGraw-Hill Education. Bernstein, W (2002) The Four Pillars of Investing: Lessons for Building a Winning Portfolio, McGraw-Hill Education. Bitters, W.E (ed.) (1997) The New Science of Asset Allocation, Glenlake Business Monographs. Black, F; Litterman, R (1992) “Global Portfolio Optimization”, Financial Analysts Journal, Sept, 28-43. Brinson, G.P., Hood, L.R., Beebower, G.L., 1986, “Determinants of portfolio performance.” Financial Analysts Journal. July/August, pages 39-44. 11 Campbell, JY; Viceira, LM (2002) Strategic Asset Allocation: Portfolio Choice for Long-Term Investors, Oxford University Press. Clarke, J; Mendelker, G; Jandik, T (2001) Expert Financial Planning : Investment Strategies from Industry Leaders, (Edited by Robert C Arffa), 2001, chapter 10, Wiley and Sons. Davis, D; Luo, Y (2001) “Using KADS to Design a Multi-Agent Framework for Stock Trading”, Multi-Conference Event Agent for E-Business on the Internet, Las Vegas, June 2001. Davis, D; Luo, Y; Liu, K (2000) “A Multi-Agent Framework for Stock Trading”, World Computing Conference, Beijing, August. Decker, KS; Sycara, K (1997) “Intelligent Adaptive Information Agents”, Journal of intelligent Information Systems, Kluwer Academic Publishers, volume 9, pp 239 – 260. Dempster, Germano, Villaverde, (2003) “Global Asset Liability Management”, British Actuarial Journal, Volume 9, Number 1, pp. 137-195(59). Fabozzi, FJ; Markowitz, HM (2002) The Theory and Practice of Investment Management (Frank J. Fabozzi Series), John Wiley & Sons. French, K (1991) “Investor Diversification and International Equity Markets”, American Economic Review, MIT Press. Fylstra, D (2005) “Introducing Convex and Conic Optimization for the Quantitative Finance Professional”, Wilmott Magazine, pp. 18-22. Gerber, G (1994) Equity Style Allocations: Timing Between Growth and Value, in Global Asset Allocation: Techniques for Optimising Portfolio Management, New York: John Wiley & Sons. Kaplan, P (1998) “Asset Allocation Models Using the Markowitz Approach”, http://www.misp.it/doc/materiali_doc/Curti1EN.pdf, accessed January 23rd January 2006. Lummer, S; Riepe, MW (1994a) "Taming Your Optimisation: A Guide through the pitfalls of Mean Variance Optimisation", Global Asset: Techniques for Optimising Portfolio Management, Edited by J.S. Lummer, M.W. Riepe, , John Wiley & Sons. Lummer, S; Riepe, MW (1994b) "The Role of Asset Allocation in Portfolio Management", Global Asset Allocation: Techniques for Optimising Portfolio Management, Edited by J.S. Lummer, M.W. Riepe, John Wiley & Sons. Luo, Y; Davis, D; K, Liu (2001) “Information and Knowledge Exchange in a Multi-Agent System for Stock Trading”, Appendix 1 Covariances and Mean Returns for 9 Classes (MSCI©) Correlation Coefficients USA Canada Germany UK Japan Australia Thailand Colombia Brazil USA 0.03502 0.03547 0.04672 0.02897 0.04817 0.04817 0.17519 0.07985 0.07527 IEEE/IEC Enterprise Networking Applications and Services Conference, (EntNet2001), Atlanta, USA, July. Luo, Y; Davis, D; Liu, K (2002) “A multi-agent system framework for decision making support in stock trading”, The IEEE Network Magazine Special Issue on Enterprise Networking Services, Volume 16, No 1, Jan/Feb. Moreno, D; Marco, P; Olmeda, I (2005) “Risk forecasting models and optimal portfolio selection”, Applied Economics, Volume 37, Number 11, 20, June, pp. 1267-1281(15), Routledge, part of the Taylor & Francis Group. Nemati, H; Iyer, LS (1999) “Intelligent decision support system for asset allocation”, Americas Conference on information Systems. Paolucci, M; Niu, Z; Sycara, K; Domashnev, C; Owens, SR (2002) Van Velsen. M., “Matchmaking to Support Intelligent Agents for Portfolio Management”, In Proceedings of AAAI (Demo Session), 2002. Rekenthaler, J (2000) “A Different Twist on Asset Allocation”, Journal of Financial Planning, January Issue. Seo, Y. W; Giampapa, J. A; Sycara, K (2002) "Text classification for intelligent agent portfolio management," First International Conference on Autonomous Agents and MultiAgent System. Sharpe, W (1992) “Asset Allocation: Management Style and Performance Measurement”, The Journal of Portfolio Management, Winter 1992. Sharpe, W (2001) Budgeting and Monitoring the Risk of Defined Benefit Pension Funds, http://www.stanford.edu/~wfsharpe/art/q2001/q2001.htm, 2001. [accessed 10 October 2006] Sharpe, W (2004) http://www.stanford.edu/~wfsharpe/art/princeton/prince2.pdf [accessed 11 January 2006]. Todd, GP; Markowitz, HM; (2000) Mean-Variance Analysis in Portfolio Choice and Capital Markets; Frank J. Fabozzi Associates. Tseng, C; Gymytrasiewicz, P.J (2000) “Real Time Decision Support System for Portfolio Management”, 35th Annual Hawaii International Conference on System Sciences (HICSS’02), Volume 3, January 07 – 10. Turban, E; Aronson, AE (2001) Decision Support Systems and intelligent systems, Sixth Edition, Prentice-Hall Incorporated. Wooldridge. M (2002) An Introduction to Multiagent Systems, John Wiley and Sons Ltd. Country Expected Return USA 0.06634338 Global Equity Asset Canada 0.070273699 Germany 0.02871881 UK 0.040776844 Japan -0.043841495 Canada Germany UK Japan Australia Thailand Colombia Australia 0.053515971 0.03547 0.04672 0.02897 0.04817 0.04817 0.17519 0.07985 Thailand -0.34793296 0.03593 0.04732 0.02934 0.04879 0.04879 0.17745 0.08088 Colombia -0.026992001 0.04732 0.06232 0.03864 0.06426 0.06426 0.23370 0.10652 Brazil -0.003288647 0.02934 0.03864 0.02396 0.03984 0.03984 0.14490 0.06604 0.04879 0.06426 0.03984 0.06626 0.06626 0.24097 0.10983 0.04879 0.06426Edward 0.03984 0.06626Dr Abel 0.24097 Falconer is0.06626 a research student; Usoro, Dr0.10983 Mark and Dr Brian Lees are0.24097 his academic 0.87637 supervisory team. 0.17745 0.23370Stansfield 0.14490 0.24097 0.39945 0.08088 0.10652 0.06604 0.10983 0.10983 0.39945 0.18207 0.07624 0.10041 0.06226 0.10353 0.10353 0.37653 0.17162 12 Br 0.0 0.0 0.1 0.0 0.1 0.1 0.3 0.1 0.1 Semantic Databases: An Information Flow (IF) and Formal Concept Analysis (FCA) Reinforced Information Bearing Capability (IBC) Model Yang Wang and Junkang Feng Database Research Group Semantic Database (SDB) seems hitherto somehow overlooked in the literature compared with its ‘big brother’, Semantic Web. What are the hindrances to the development of SDB, which hence have to be taken into account as we observe, include information representation, knowledge management, meaning elicitation, constraints/regularity identification and formulation, and also partiality preservation. We propose an architecture, which is a result of reinforcing the notion of the Information Bearing Capability (IBC) that we put forward elsewhere before by applying the theory of Information Flow (IF) and that of Formal Concept Analysis (FCA). We believe that this architecture should enable SDB to cover a number of these aspects, which build upon and go beyond the relational database (RDB). 1. INTRODUCTION Semantic Web (SW) is the supreme elegance of topics, which covers numerous fields, such as knowledge organization and management, network technology and even data modeling. Comparing to this prosperous triumph, the seemingly evident lack of attention to Semantic Database (SDB) would appear rather peculiar. Whereas it is well known that SDB aims at capturing, modeling and yielding meanings rather than raw data, we observe that the short of robust theoretical modeling foundation and guidance lies as a gulf before the ‘fortune’. In our opinion, if we want to achieve a satisfactory SDB, not only primary pre-requisites such as capturing more semantics and constraints, but also profound concepts of information, representations and partiality, need to be addressed. To get across this gulf, the foundation of this research is a series of theories (we refer to them as ‘SIT’, short for Semantic Information Theories) concerning semantic information and information flow including Dreske (1981), Devlin (1991), and in particular, Barwise and Seligman’s (1991) information channel theory (IF for short). We believe that an Information Flow (hereafter IF for short) and Formal Concept Analysis (FCA) reinforced Information Bearing Capability (IBC) model (We will say more about it shortly) provides a new prospective to SDB, which both assures traditional requirements of design and brings up some philosophical and mathematical insights. This would, therefore, promote SDB to be compatible with Knowledge base (KB) and hence to be a strong support for SW. 1.1 A Short Review of Semantic Databases (SDB) A database system is a representation system, which should be able to reflect real objects in the circumstance being modeled. The content of a database rests with what actually exists in the modeled domain while any change operates on this content should correspond with what happens to those real world objects. Sustaining this tie is not easy as at the first glance. Designing a data model that captures as much as meaning as the modeled domain is the solution of many researchers (Hammer and McLeod 1981, Jagannathan et.al, 1988, Tsur and Zamolo 1984). To this end, concepts around SDB came into the scene. Bearing the goal of representing, describing and structuring more semantics and meanings than contemporary database (viz. Relational Database) in mind, SDB needs to be closely related to the modeled domain. Hammer addresses a number of criteria that should be enforced during SDM design (Hammer and McLeod 1981): The constructs of the database model should provide for the explicit specification of a large portion of the meaning of a database. So called semantic expressiveness is not sufficiently achieved by many current data modeling techniques, such as hierarchical, network, and relational models 13 A database model must support a relativist view of the meaning of a database, and allow the structure of a database to support alternative ways of looking at the same information. Being capable of capturing more meaning requires never rigid definitions and distinctions between ‘entities’, ‘attributes’ and ‘association’. A database model must support the definition of schemata that are based on abstract entities. This point, in fact, addresses that a database should have the mechanism to support possible semantic constraints. In the related literature, there are mainly two most interesting streams identified by the authors in SDB modeling. The first one is that some of the researchers are developing their SDM structure on the root of available modeling techniques. Most related to this research, some systems are inheriting the basic modeling constructs of RDM’s apparatus, for example, Iris Data Model (Lyngback and Vianu 1987), Generic SDM (Chen and McLeod 1989) and SDB management System SIM (Boyed 2003). Meanwhile, Rishe and his group build up a Semantic Wrapper over RDB which produces set of SDB tools including Knowledge database tool, Knowledge base and Query Translator (http://n1.cs.fiu.edu/SemanticWrapper.ppt). The second is that some research shows that SDB is more likely linked to Ontology and Knowledge base (http://www.fmridc.org/f/fmridc/dmt/sdm.html). This would seem to orientate SDB to flourishing the development of SW. Besides this, currently, research around SDB encounters numerous obstacles. The bottleneck, as we have identified, resides in lack of certain infrastructure to retrieve semantics and formulate semantic constraints, not from traditional database point of view but follow vigorous guidance of Semantic Information Theory (SIT in short). We believe that by philosophically separating truly information from raw data, dually grasping semantic constraints and partially representing semantic information relation, an advance model of SDB can be achieved. 1.2 IF and FCA Based IBC Prospect of SDB In 1998, we identified a research problem, namely the ‘information content’ of a formalized information system (Feng 1998). In that paper numerous works were cited and it was shown that the main cause of this problem seemed that information had been treated as ‘mystical liquid’. We then argued that the lack of clearly expressed and defined ‘information content’ of a conceptual data schema was responsible for many difficulties in data modeling and analysis as a process of inquiry, which is a basis for the design of an information system. Then in 1999 we formulated a notion called ‘information bearing capability’ (IBC for short) by drawing on interdisciplinary views of information creation and transmission (Feng 1999). A four-facet principle currently elaborates this notion, which is concerned with a set of sufficient and necessary conditions for the IBC of an information system. The conditions are: information content containment, distinguishability, accessibility and derivability (Feng 2005). The principle about IBC and their associated concepts that have been put forward in a series of research papers (such as Xu and Feng 2002, Feng and Hu 2002, Xu 2005, and Wang and Feng 2005a) may be seen as forming an innovative perspective for looking at information systems. Now, IBC as a cornerstone is applied to a number of research problems that are being looked at by our group such as schema mapping, data exchanging and modeling. The ideas around IBC however should be further developed and tested in real world applications. To this end, it seems that the most appropriate tool to reason about and verify IBC would be IF combined with FCA. We envisage that endeavor along this line will uplift the articulation of what might be called ‘the microscopic infrastructure’ of the IBC principle to an adaptable, adoptable and applicable level in SDB modeling. This paper proceeds as follows. In the next section, we highlight some aspects of SDB modeling that seem to have been overlooked in the light of SIT rooted IBC model. Our approach of combined use of IF and FCA in the IBC model, which would, we believe, advance the state of the art of SDB, is introduced in section 3. Following this, a conceptual picture of IF and FCA reinforced IBC model for SDB described and elucidated in section 4. 2. What Should a SDB Represent and Provide? Model, 14 As aforementioned, SDB is proposed in the literature to address those problems encountered in other forms of data modeling. As summarized by Boyed (2003), there are several essential goals, which need to be sustained, during SDB development. The SDB is a high-level semanticsbased database description and structural formalism for databases (Hammer, 1981). Although attempting to capture all the semantics of the modeled domain is unattainable, SDB should endeavor to incorporate most of the semantics. SDB advances RDB and other database models in terms of its real-world perception of the problems, different perspectives of queries, and most importantly its inheritancebased hierarchical modeling structure. In addition to these known characteristics, following the insight of IBC based on SIT, we would propose more significant features for SDB. Only when these features are delivered can we say that SDB is satisfiably achieved. 2.1 Data, Information and Semantics Database is the vehicle for storing and providing information. Without the guidance of interdisciplinary philosophical semantic information theory, it is not surprising that contemporary database modeling dose not separate data and truly information. Notwithstanding modeling methods like RDB being many and varied, as far as SDB is concerned, it should broaden its edge to tackle the truth of data, information, meaning and semantics in order to capture semantics and to solve some difficult issues, for example, query answering, lossless transformation, etc. In a typical contemporary database, ‘what you see is what you get’ is the prevailing feature. Relation between data and information remains scrupulously bypassed. For a long decade, data with its meaning is treated as information in the context of database (Checkland, 1981). A famous schema transformation approach, i.e., ‘information capacity’ (IC) (Miller 1993), straightly takes data instances of schemata as information. Fusing Organizational Semiotics (OS) into database, ‘meaning is created from the information carried by signs’ (Mingers 1995). A veritably practical SDB should take the challenges that lie in several aspects around definitions of information, information content and meaning. Some of my colleagues have provided an analysis about this (Wang and Feng 2005). Firstly, instances are not always faithful to their semantic types. Traditionally, the schema of a database is thought to represent the type level of information while database instances fill into these type level classes whereby receive their semantics or meaning from the classes. However, this view overlooks the facts that instances may not loyal to their respective semantic infrastructures. These instances do not represent any information that originated the types (Dretske 1991). Secondly, the meaning of data in the database is not necessarily to be part of their information content. SDB should be able to use alternative ways to represent the same information. Therefore, a data construct represents a piece of information only when the information content of the data construct includes that piece of information. It is not convincing to use meaning as the criteria for the information content of a piece of data. Finally, it is not adequate to take the ability of accommodating instances into the schema as the information capacity of data constructs in the database (Wang and Feng 2005). The fewer constraints being modeled, the less specific the instances are. Hence, less information there is. SDB modeling should take this point into consideration and facilitate it. 2.2 Constraints and Representations No matter what form it is in; a database is after all need to represent objects and relations in the represented domain. The modes of representation (Shimojima 1996) obey structural constraints that mirror the regularities that govern things going on in the represented domain. Any representation involves certain kind of information flow. Information flow results from the regularities in a distributed system (Barwise and Seligman 1997, P.8). Contemporary database like RDB limit themselves into a particular structure of constraints such as relational objects and associated relations. SDB should go beyond these limits in the way of finding the best fit between the representing system and the represented domain. Apart from this aspect, SDB should also ensure that its reasoning be consistent with the represented domain. In other words, reasoning over constraints needs great care. Wobcke (2000) identifies the differences between schema-based 15 and information flow based reasoning. The former is partly subjective and defeatable contrasting to the objectiveness and nondefeatability holding by the latter. If given a fixed context by discarding all alternative situations, schema-based reasoning and information flow based reasoning are transferable. Shimojima (1996) uses basic mathematical instruments to model constraints in order to perform a rigorous investigation on a wide range representation issues. His research provides a sound theoretical foundation for developing our IBC model for SDB in virtue of inferential reasoning intimate to what happens in the domain to be modeled. 2.3 Partiality Talking about semantics, it is evident to many researchers, especially those who are familiar with logics and linguistics, that there are ‘holes in reality’ (Duzi 2003). These holes reside in our abstract way of modeling particular dependency relations among real world objects. Many attempts have been made to philosophically address such issues as Possible World Semantics and Situation Semantics. As aforementioned, in database, there exit instances that do not inherit semantics from its corresponding class types. Following Duzi (2003), if we take these instances as the logical construction C (not unlike the notion of ‘concept’ of Dretske 1981) for the ‘mode of representation’, which is discussed in previous section, it should link the expression E and its denotation D. Problems arise when we use empty concepts, the construction C will fail to achieve anything, not even any meaning. As a result, the denotation D will fail to give any truth-value in an argument. Macroscopically, it is necessary for SDB to be equipped with partial order to handle overall informational relationships. Based on Dretske’s information flow (1991) and Barwise and Perry’s situation semantics (1993), Wobcke (2000) argues that using conditionals as basic appliance, people could evaluate the subjectiveness and intentionality of a collection of schemata. The idea is to treat those conditionals as expressing constraints which are actually informational relations between facts and events of the kind that can be modeled using structures of situations (Wobcke 2000). The order of situations for the collection of constraints is in the form of partial order supporting subjective reasoning. In certain circumstances, i.e., providing certain fixed context (situation), reasoning on this order is identical to the reasoning of information flow. Also, in Duzi’s thesis (2001), she points out that information content inclusion relations (in relation to attributes) are of partial order. Most specifically, she formalizes informational capability in a complete lattice based on the power set of the attributes in question. Furthermore, it is interesting that this lattice is proved to be isomorphic to its substituting partial ordered set of equivalence classes. Therefore, for the sake of manipulating informational scenarios, the need of supporting partial order of the IBC model both philosophically and mathematically should not be ignored. Moreover, we believe that such a work would be aligned with issues in knowledge representation in the AI field. 3. Architecture based upon Information Flow (IF) and Formal Concept Analysis (FCA) The central idea of IBC is called the IBC principle. This principle, is made up of conditions of information content containment, distinguishability, accessibility and derivability and it is put forward by Feng (2005) and his colleagues through a period of arduous work in the sense of drawing interdisciplinary views of information creation and transmission (Feng 1999, Xu and Feng 2002, Feng and Hu 2002, Xu 2005, and Wang and Feng 2005a). IF is first introduced into IBC for reasoning about and for verifying the principle (Wang and Feng 2005). As being successively compatible and content with the IBC, IF has become a headstone for further development and application of the IBC model. For the purpose of elevating implementation, FCA is probed and found that it is adaptable, applicable and adoptable both theoretically and practically with IF. 3.1 Channel-Theoretical Information Flow The Channel-theoretical Information Flow theory (IF) is a mathematical model of semantic information flow. Information flow is possible due to the regularities among normally disparate components of a distributed system. It is known that such a theory succeeds in capturing partial 16 order of classifications (Kalfoglou and Schorlemmer 2005) that underlies the flow of information. Sophisticated notions (we do not go into details here) stemming from IF now have been formulated for explorations on semantic information and knowledge mapping and exchanging. Kent (2002a, 2002b) exploits semantic integration of ontologies by extending a first order logic based approach (Kent 2000) which is also based on IF. An information flow framework (IFF) has been advocated as a metalevel framework for organising the information that appears in digital libraries, distributed databases and ontologies (Kent 2001). From Kent’s work, Kalfoglou and Schorlemmer (2003a) develop an automated ontology mapping method in the field of knowledge sharing and cooperation. IF and its surrounding concepts are also relevant to solving problems of semantic interoperability (Kalfoglou and Schorlemmer 2003b). Apart from this main stream of applications, IF supports various research efforts from defensible reasoning (Cavedon 1998); endoperspective formal model (Gunji et al 2004) to semiconcept and protoconcept graphs (Malik 2004). Besides the effective effort of using IF to represent, capture and model constraints for a given modelled domain, it is also observed that IF ‘was not developed as a tool to be used in real world reasoning’ (Devlin 1999) and we observe that it is on its own insufficient for describing domain information or knowledge. To fill these gaps, Formal Concept Analysis (FCA) was proposed as a silver bullet. 3.2 Formal Concept Analysis (FCA) FCA was developed by Rudolf Wille (Wille 1982) as a method for data analysis, information management, and knowledge representation (Priss 2005a). Presumably due to its applicable nature, it does not take long for FCA to become a common interest in many research communities, for example, social net work analysis (Freean and White 1993), linguistics (Priss 2005b), and software engineering (Fischer 1998, Eisenbarth et al. 2001). As aforementioned, FCA provides solid foundations for not only information and knowledge retrieval by its underlying mathematical theory (Godin et al. 1989, Kalfoglou et al. 2004) but also for respective representations by concept lattice (Wille 1982, 1992, 1997b) along with concept graphs (Prediger and Wille 1999). We maintain that the use of FCA will supplement with IF in SDB modeling. By using IF along, it would appear that the construction of an ‘information channel’ in many cases is difficult when applying IF to real information system problems. To alleviate it, we envisage that ‘Conceptual Scaling’ techniques (Ganter and Wille 1989, Prediger and Stumme 1999)’, which are affinity with FCA, will be useful. Furthermore, reasoning and inference over difference levels of a channel can be characterized by ‘Concept Graph’ (Prediger and Wille 1999) in the light of FCA-based ‘Concept Lattice’ (Wille 1982, Wille 1992, Wille 1997b). In other words, FCA provides the investigation with a basis for extraction, representation and demonstration of informational aspect of semantics, and at the same time IF-based techniques/methods can be charged with the task of information flow based reasoning. As a result, the combined use of IF and FCA can shed some light on solving problems around the IBC within the context of SDB, which is also harmonious with knowledge discovery and representation. 3.3 Prospect of Combined Use of IF and FCA The essential element of our IBC mode for SDB is the combined use of IF and FCA. They provide vital insights for our SDB model. The compatibility between them is crucial for any combined use. We give reasons below for using IF theory and the theory of FCA in combination. Firstly, both IF and FCA share the same origin, i.e., category theory with the means of Chu space (Gupta 1994, Barr 1996 and Pratt 1995). As Wolff (2000) observes, ‘it is really astonishing that these tools (IF and FCA) are not mutually taken into account in each other’s theory’. Priss (2005a) treats the ‘classifications’ in IF as a general sense of ‘concept lattices’ in FCA. Following this line of thinking, secondly, nearly all fundamental concepts invented by both of IF and FCA can find counterparts in each other. For example, the notions of ‘classifications’ in IF matches that of ‘formal context’ in FCA; ‘information channels’ in IF matches ‘scaled many-valued contexts’ in Conceptual Scaling (Ganter and Wille 1989, Ganter and Wille 1999) associated with FCA. Other basic notions presented in IF, such as ‘state space’, ‘refinement of channels’, and ways of handling ‘vagueness’ 17 are also delivered in FCA mathematically (Wolff 2000). Finally, IF bears epistemological resemblance to FCA. To be explicit, starting from the same algebraic category, IF together with FCA aim at formulating and justifying ‘partial order’ that relies on agreed understanding of the existence of ‘duality’ between separated situations, which is exactly why information flow commences. Combined use of IF and FCA is beneficial to constructing the IBC model of SDB. SDB highly needs to capture more semantics. In IF and FCA reinforced IBC model, FCA would serve as the linkage between IF reasoning and the modelled domain. Due to the ‘non-directly-applicable’ nature of IF (Devlin 1999), applying it directly to modeling informational semantics proves to be problematic. In contrast, a number of works stemming from FCA around knowledge discovery and information retrieval have been put forward. For example, Stumme and his colleagues have encouraged the use of FCA in exploration and representation of implied information and facilitating the conversion of information into knowledge (Hereth et al. 2000, Stumme et al. 1998). We would use the ‘Conceptual Scaling’ techniques (Prediger and Stumme 1999, Prediger and Wille 1999) to combine FCA with IF reasoning because of FCA’s logical equivalence with ‘Information Channel’. The results of reasoning would be presented in Concept Graphs, which has advantages in representing semantics in partial order. Also, a combined use of IF and FCA can satisfactorily model more semantic constraints identified by Hammer (1987). To tackle information flow, IF insists on analyzing relations between tokens and types. According to the second principle of information flow, i.e., ‘information flow crucially involves both types and their particulars’ (Barwise and Seligman 1997, P.27). Originally and largely following Dretske (1981), we thought that semantics are presented on the type level which further provides the meanings to the tokens involved in information flow. However, from the paper of Kalfoglou and Schorlemmer on IF-map (2003a), we find the important role of tokens, e.g., the same set of rivers and streams, played in determining semantics or constraints of the whole system in terms of semantic correspondences between the types. We observe that in fact, Kalfoglou and Schorlemmer has employed primary thinking of FCA in exploring ‘intension’ and ‘extension’ of formal concepts within a given formal context. That is from either set, i.e., intensions or extensions; we can define its counterpart in the context, and thus the formal concepts. Therefore, using relations in tokens (extensions), we would gain relation of concepts and hence arrive at a set of constrains, which reflect a type of regularities of the whole system in the given context. This is exactly how tokens take part in defining the semantics of a system, and in achieving semantic interoperability. Further to this point, we envisage that duality held by both IF and FCA enables us to support alternative ways for the user to view even the same information in SDB. Start with the relations that reside in types and we would end up with relation of tokens and vice versa. Therefore, depending on what aim we want to achieve, we could selectively take either tokens or types as our starting point in different analysis. Explicitly, if we want to solve the semantic interoperability problem, as Kalfoglou and Schorlemmer did, we shall investigate tokens-determined relations in order to achieve the relations on types. On the other hand, if we want to find out why and how data constructs represents (or conveys) the information about a given semantic relation (i.e., a relation between some real world objects), in most cases, we will take the semantics on types of this structure as a foundation. 4. Outline of IF and FCA Reinforced IBC Model for SDB Based on previous sections, we can now start describing the IF and FCA reinforced IBC model designed for SDB. We will begin with data schemata as we believe that original databases and schemata is too valuable to be retained (Figure 1). The original database schema together with a serial of dependencies held by the schema would be analyzed by using IF and FCA. This analysis needs to be assisted by obtained initiative business constraints e.g. stake holder views, presented in the format of scales, so that subjectiveness is preserved at this early stage. The construction of ‘information channel’ of IF will benefit from the technique of ‘conceptual scaling’ of FCA. The output of investigation is a 18 conceptual space which contains all the constraints (semantics) captured by every information channel. This space is called by us as the ‘kernel of IBC’. When the user puts a query for a piece of information to this kernel, if there is no direct answer, an inference will be carried out by means of a set of ‘information content inference rules’ (Feng and Hu 2002). Then, final results are added into a separate conceptual space following the decision of the user. Connected with knowledge representation and management, the consequent results could be transformed using XML-extended Information Flow Framework (IFF) (http://www.ontologos.org/IFF/The%20IFF%2 0Language.html) language. Figure 1. Overall Picture of IF and FCA Reinforced IBC model There are two most important parts of this model to construct corresponding IF channels. The which show in two boxes in Figure 1. To clarify many-valued context will then become singlewhat actually happens inside of them, we will use valued context as a result. Following this, using two more diagrams. dependencies that are determined by business In Figure 2, there is a detailed process for arriving rules as the other scales, another scaling, i.e., the at the kernel of IBC. Both primary database ‘relational scaling’, will be accomplished by a schemata and instances are translated into manyfinal lattice layout also with a crowd of valued context by FCA. Then, two scaling information channels. The ultimate results are sets processes are performed. The first one called of ‘IF’ theories derived from all of the channels. ‘conceptual scaling’. It is based on the idea that This is what we want to model as the system embedded structural constraints are used as scales regularities. 19 Figure 2. How to Achieve Kernel of IBC In addition, another significant part in our model (Wang 2005 and Xu 2005), we found that these is inference on information content (Figure 3). inference rules can be justified by theory of IF. The information content based inference rules In the future, we will generalize these are put forward by Feng and Hu (Feng and Hu verifications by not only IF but also FCA. 2002). Furthermore, through two MSc projects 20 Figure 3. Information Inference Rules (IIR) 21 Computing and Information Systems © University of Paisley 2006 5. CONLUSIONS This paper represents our first step towards satisfactorily modeling a SDB by means of an IF and FCA reinforced IBC. Three more criteria, i.e., extracting information, modeling semantic constraints and also partially representing information flow, have been proposed in addition to traditional SDB requirements. The overall idea of the IBC model for SDB is shown with diagrams that heavily draw on concepts from both IF and FCA. This attempt seems worthwhile for the development of SDB, and it is also compatible with most modern knowledge management systems, and therefore relevant to the area of semantic web. References Barr, M. (1996). The Chu construction. Theory and Applications of Categories, 2(2):17–35. Barwise, J. and Seligman, J. (1997) Information Flow: the Logic of Distributed Systems, Cambridge University Press, Cambridge. Barwise, J. and Perry J. (1983). Situations and Attitudes, Cambridge, Mass.: Bradford-MIT. Boyed, S. (2003). A Semantic Database Management System: SIM. The University of Texas at Austin, Department of Computer Sciences. Technical Report CS-TR-03-43. Cavedon, L. (1998). Default Reasoning as Situated Monotonic Inference, Minds and Machines 8. Checkland, P. (1981). Systems Thinking, Systems Practice. Chichester, UK: Wiley. Chen, I. A. and McLeod, D. (1989). Derived Data Update in Semantic Databases, Proceedings of the fifteenth international conference on Very large data bases, p.225-235, July, Amsterdam, The Netherlands. Devlin, K. (1991). Logic and Information, Cambridge. Devlin, K. (2001). Introduction to Channel Theory, ESSLLI 2001, Helsinki, Finland. Dretske, F. (1981). Knowledge and the Flow of Information, Basil Blackwell, Oxford. Duží, M. (2001). Logical Foundations of Conceptual Modelling. In VŠB-TU Ostrava. Duží, M. (2003). Do we have to deal with partiality?. In Miscellania Logica, vol. Tom V, 45-76. Eisenbarth, T., Koschke, R., & Simon, D. (2001). Feature-driven Program Understanding using Concept Analysis of Execution Trace. In Proceedings of the Ninth International Workshop on Program Comprehension. International Conference on Software Maintenance. Feng, J. (1998). The "Information Content’ Problem of a Conceptual Data Schema, SYSTEMIST, Vol.20, No.4, pages 221-233, November 1998. ISSN: 0961-8309 Feng, J. (1999). An Information and Meaning Oriented Approach to the Construction of a Conceptual Data Schema, PhD Thesis, University of Paisley, UK. Feng, J. and Hu, W. (2002). Some considerations for a semantic analysis of conceptual data schemata, in Systems Theory and Practice in the Knowledge Age. (G. Ragsdell, D. West. J. Wilby, eds.), Kluwer Academic/Plenum Publishers, New York. Feng, J. (2005). Conditions for Information Bearing Capability, Computing and Information Systems Technical Reports No 28, University of Paisley, ISSN 1461-6122. Fischer, B. (1998). Specification-Based Browsing of Software Component Libraries. Proc. Automated Software Engineering, Hawaii, 246-254. fMRIDC, The Semantic Database Model, http://www.fmridc.org/f/fmridc/dmt/sdm.html. Ganter, B. and Wille, R. (1999). Formal Concept Analysis: mathematical foundations. Springer. ISBN: 3-540-62771-5. Godin, R., Gecsei, J., & Pichet, C. (1989). Design of Browsing Interface for Information Retrieval. In N. J. Belkin, & C. J. van Rijsbergen (Eds.), Proc. SIGIR ’89, 32-39. Gupta, V. (1994). Chu Spaces: A Model of Concurrency. PhD thesis, Stanford University, 1994. Hammer, M., McLeod, D. (1981). Database Description with SDM: A Semantic Database Model, ACM Trans. Database Syst. 6 (3): 351-386. Hereth, J., Stumme, G., Wille, R. and Wille, U. (2000). Conceptual Knowledge Discovery in Data Analysis. In B. Ganter, & G. Mineau (Eds.), Conceptual Structures: Logical, Linguistic and Computational Issues. LNAI 1867. Berlin: Springer, 421-437. Information Flow Framework (IFF) Language, http://www.ontologos.org/IFF/The%20IFF%20Lan guage.html. Jagannathan, D., Guck, R. L., Fritchman, B. L., Thompson, J. P., Tolbert, D. M. (1988). SIM A Database System Based on the Semantic Data Model. SIGMOD Conference: 46-55. Kalfoglou, Y. and Schorlemmer, M. (2003a). IFMap: an Ontology Mapping Method based onIinformation Flow Theory. Journal on Data Semantics, 1(1):98–127. Kalfoglou, Y. and Schorlemmer, M. (2003b) Using Information Flow Theory to Enable Semantic Interoperability, In Proceedings of the 6th Catalan Conference on Artificial Intelligence (CCIA '03), Palma de Mallorca, Spain, October 2003. Kalfoglou, Y., Dasmahapatra, S., & Chen-Burger, Y. (2004). FCA in Knowledge Technologies: Experiences and Opportunities. In P. Eklund (Ed.), Concept Lattices: Second International Conference on Formal Concept Analysis, LNCS 2961. Berlin: Springer, 252-260. Kalfoglou, Y., Schorlemmer, M. (2005). Using Formal Concept Analysis and Information Flow for Modeling and Sharing Common Semantics: lessons learnt and emergent issues, In Proceedings of the 13th International Conference on Conceptual Structures (ICCS2005), Kassel, Germany, July 2005 Kent, R. E. (2000). The Information Flow Foundation for Conceptual Knowledge Organization. In: Dynamism and Stability in Knowledge Organization. Proceedings of the Sixth International ISKO Conference. Advances in Knowledge Organization 7 111–117. Ergon Verlag, Würzburg. Kent, R. E. (2001). The Information Flow Framework. Starter document for IEEE P1600.1, the IEEE Standard Upper Ontology working Group, http://suo.ieee.org/IFF/. Kent, R. E. (2002a.) The IFF Approach to Semantic Integration. Presentation at the Boeing MiniWorkshop on Semantic Integration, 7 November 2002. Kent, R. E. (2002b). Distributed Conceptual Structures. In: Proceedings of the 6th International Workshop on Relational Methods in Computer Science (RelMiCS 6). Lecture Notes in Computer Science 2561. Springer, Berlin. Kollewe, W., Skorsky, M., Vogt, F., and Wille, R. (1994). TOSCANA – ein Werkzeug zur begrifflichen Analyse und Erkundung von Daten. In R. Wille, &19 M. Zickwolff (Eds.), Begriffliche Wissensverarbeitung - Grundfragen und Aufgaben.Mannheim: B.I.Wissenschaftsverlag, 267-288. Lyngbaek, P. and Vianu, V. (1987). Mapping a Semantic Database Model to the Relational Model, Proceedings of the 1987 ACM SIGMOD international conference on Management of data, p.132-142, May 27-29, San Francisco, California, United States. Malik, G. (2004.) An Extension of the Theory of Information Flow to Semiconcept and Protoconcept Graphs. ICCS 2004: 213-226. Miller, R. J., Ioannidis, Y. E. and Ramakrishnan, R. (1993). The Use of Information Capacity in Schema Integration and Translation, in Proceedings of the 19th International Conference on Very Large Data Base, Morgan Kaufmann, San Francisco. Mingers, J (1995). Information and Meaning: Foundations for an Intersubjective Account. Journal of Information Systems 5 285-306. Pratt, V. (1995). The Stone gamut: A coordinatization of mathematics. Logic in Computer Science, pages 444–454. Prediger, S. and Stumme, G. (1999). Theory-driven Logical Scaling. Conceptual Information Systems meet Description Logics. In P. Lambrix, A. Borgida, M. Lenzerini, R. Muller, & P. PatelSchneider (Eds.), Proceedings DL’99. CEUR Workshop Proc. 22. Prediger, S. and Wille, R. (1999). The Lattice of Concept Graphs of a Relationally Scaled Context. In W. Tepfenhart, & W. Cyre (Eds.), Conceptual Structures: Standards and Practices. Proceedings of the 7th International Conference. Priss, U. (2005a). Formal Concept Analysis in Information Science. Annual Review of Information Science and Technology. Vol 40. Priss, U. (2005b). Linguistic Applications of Formal Concept Analysis. In: Ganter; Stumme; Wille (eds.), Formal Concept Analysis, Foundations and Applications. Springer Verlag. LNAI 3626, p. 149160. Rishe, N., Semantic Wrapper over Relational Database. http://n1.cs.fiu.edu/SemanticWrapper.ppt. Shimojima, A. (1996). On the Efficacy of Representation, Ph.D. Thesis. The Department of Philosophy, Indiana University. Stumme, G., Wille, R., and Wille, U. (1998). Conceptual Knowledge Discovery in Databases using Formal Concept Analysis Methods. In J. M. Zytkow, & M. Quafofou (Eds.), Principles of Data Mining and Knowledge Discovery. LNAI1510. Berlin: Springer, 450-458. Tsur, S. and Zaniolo, C. (1984) An implementation of GEM Supporting a Semantic Data Model on a Relational Back-end, Proceedings of the 1984 ACM SIGMOD international conference on Management of data, June 18-21, Boston, Massachusetts. Wang, X. and Feng, J. (2005b.) The Separation of Data and Information in Database System under an Organizational Semiotics Framework. The 8th International Workshop on Organizational Semiotics, Toulouse, France. Wang, Y. and Feng, J. (2005a). Verifying Information Content Containment of Conceptual Data Schemata by Using Channel Theory. The 14th International Conference on Information Systems Development, Karlstad, Sweden. Springer-Verlag. 23 Wille, R. (1982). Restructuring lattice theory: an Approach based on Hierarchies of Concepts. In I. Rival (Ed.), Ordered sets. Reidel, DordrechtBoston, 445-470. Wille, R. (1992). Concept Lattices and Conceptual Knowledge Systems. Computers & Mathematics with Applications, 23, 493-515. Wille, R. (1997a). Conceptual Graphs and Formal Concept Analysis. In D.Lukose, H. Delugach, M. Keeler, L. Searle, & J. F. Sowa (Eds.), Conceptual Structures: Fulfilling Peirce’s Dream. Proc. ICCS’97. LNAI 1257. Berlin:Springer, 290-303. Wille, R. (1997b). Introduction to Formal Concept Analysis. In G. Negrini (Ed.), Modelli e modellizzazione. Models and modelling. Consiglio Nazionale delle Ricerche, Instituto di Studi sulli Ricerca e Documentazione Scientifica, Roma, 39-51. Wobcke, W. (2000). An Information-Based Theory of Conditionals. Notre Dame Journal of Formal Logic 41(2): 95-141. Wolff, K. E. (2000). Information Channels and Conceptual Scaling. In Working with Conceptual Structures. Contributions to ICCS 2000, Shaker Verlag. Xu, H. and Feng, J. (2002). ‘The "How" Aspect of Information Bearing Capability of a Conceptual Schema at the Path Level’. The 7th Annual Conference of the UK Academy for Information Systems, UKAIS'2002. Leeds . ISBN 1-898883149, pp.209-215 Xu, Z. (2005). Verifying Information Inference Rules by using Channel Theory, MSc dissertation, University of Paisley. Wang.Y is a Researcher and Dr. Feng J. a Senior Lecturer at the University of Paisley 24 Computing and Information Systems © University of Paisley 2006 An Investigation into Trust as an Antecedent to Knowledge Sharing in Virtual Communities of Practice Abel Usoro, Mark W Sharratt and Eric Tsui Information Systems Group This study focuses on the role of trust in knowledge sharing within the context of virtual communities of practice. Trust is widely accepted as an important enabler of knowledge management (KM) processes. We conceptualise trust across three dimensions, namely: competence, integrity, and benevolence; we test hypotheses as to the effect of these facets of trust on knowledge sharing by surveying an intra-organisational global virtual community of practitioners. The results indicate that all three dimensions of trust are positively related to knowledge sharing behaviour. Trust based on the perceived integrity of the community was found to be the strongest predictor of knowledge sharing behaviour. Our findings suggest that the dimensions of trust buttress each other; although they are theoretically distinct, they appear to be empirically inseparable. We propose that in order for knowledge sharing to be enabled, trust must concurrently exist in all three dimensions. Implication to organisations in their recruitment policy is to include competence, integrity and benevolence in their sought-for attributes of new employees. KM practitioners also have to encourage these attributes in existing employees, who are potential members of on-line communities of practice. 1 INTRODUCTION The management of the knowledge base of organisations is becoming an area of strategic focus for many knowledgeintensive organisations (Ruggles 1998; Beaumont and Hunter 2002). Indeed, knowledge is unique as an organisational resource in that most other resources tend to diminish with use; the potential for growth in knowledge resources increases with use, as “ideas breed new ideas, and shared knowledge stays with the giver while it enriches the receiver” (Davenport & Prusak 1998: p. 16-17). Hence, understanding key knowledge management (KM) processes is a growing area of organizational research efforts (Scarbrough et al. 1999). Within the KM field, there is widespread acceptance of the role of communities of practice as a key KM enabler (Wenger et al. 2002; Scholl et al. 2004). CoPs have been described as groups of people “informally bound together by shared expertise and passion for a joint enterprise” (Wenger and Snyder 2000: 139), “playing in a field defined by the domain of skills and techniques over which the members of the group interact” (Lesser and Storck 2001: 831). Such communities provide a rich locus for the creation and sharing of knowledge both within and between organisations (Brown and Duguid 1991; Lesser and Everest 2001). Information and communication technologies are now extending the boundaries of traditional face-to-face communities by creating virtual communities that enable global asynchronous and real-time collaboration (Hildreth and Kimble 2000). These communities may exist primarily, or solely, online and become largely dependent on computer-mediated communication to connect members. However, the availability of information systems does not automatically induce a willingness to share information and develop new knowledge (Nahapiet and Ghoshal 1998). Indeed, research has found that social interaction that is mediated by online technologies can be less effective than in face to face meetings (Preece 2000). Despite such limitations of technology, research has shown that emotional attachments can develop online, despite a lack of face-to-face contact (Rheingold 1993; Sharp 1997; Preece 2000). The building of trust is an important social process that is widely accepted as a prerequisite to cooperation (for example: Gambetta 1988; Ring and Van de Ven 1994; Mayer et al. 1995; Nahapiet and Ghoshal 1998; Wang and Ahmed 2003). Research has shown that levels of trust influence the exchange of resources between intra-organisational business units (Tsai and Ghoshal 1998) and research investigating knowledge sharing has found trust to be important in the receipt of useful knowledge (Levin et al. 2002). It logically follows that virtual communities that fail to develop trusting relations will restrict the development of knowledgesharing activities. While Levin et al.’s research focused on the receipt of knowledge, there appears to be little current understanding as to the importance of trust in the provision of knowledge in general, and specifically within the context of virtual communities of practitioners. Ardichvili et al. (2003) conducted an exploratory study into the factors that affect knowledge sharing within intra-organisational virtual communities of practice. Their findings identify a lack of understanding of the role of trust in the provision of knowledge. Our paper takes up this challenge by investigating the role of trust, in its many guises, in the provision of knowledge within a virtual community of practitioners. The following section sets out our operational definitions of knowledge sharing and trust. This leads to the development of a number of hypotheses as to the relationship between three facets of trust and knowledge sharing. We discuss the design of the empirical research, and test the hypothesised relationships. Findings are then presented and reflected upon. We conclude the paper with future research directions and discuss the limitations of this study. 2 OPERATIONAL DEFINITIONS Knowledge is an intangible resource and is thus difficult to measure. Indeed, a review of the literature has established that knowledge sharing in not well defined for the purposes of empirical research. In order to understand knowledge sharing, it is necessary to define what we mean by knowledge, and how this relates to information and data. 2.1 Knowledge, Information and Data Knowledge can only exist in the mind of the individual (Van Beveren 2002). It is through knowledge that we perceive the world around us; knowledge largely determines how we interpret and respond to external stimuli. Hence, knowledge often determines action. Knowledge is acquired through a process of action and reflection (Argyris & Schon 1978). Within this process, the interaction and communication of two or more individuals can lead to the exchange and sharing of knowledge. Here, information facilitates this process and acts as a communicatory representation of knowledge. Data is the raw component of information. Intrinsically, data contains no meaning, data becomes information when framed within a meaningful context. On their own, the numbers 56 and 1648 are just items of data. Framed within a context, such data may provide information, for example, the number 56 bus is due at 16:48 hours. Hence data is transformed into information. It is knowledge which provides the context that creates information from data and it is through the interpretation of such information that new knowledge may be acquired. Hence, information and knowledge intrinsically facilitate the knowledge-sharing process. 2.2 Knowledge Sharing As we have described, knowledge per se cannot exist outside the mind of the individual. Knowledge sharing involves a process of communication whereby two or more parties are involved in the transfer of knowledge. This is a process that involves the provision of knowledge by a source, followed by the interpretation of the communication by one or more recipients. The output of the process is the creation of new knowledge. Hence, knowledge sharing is defined as a process of communication between two or more participants involving the provision and acquisition of knowledge. Indeed, the communication can take many forms, using both verbal and non-verbal mechanisms, with or without the use of technology. 2.3 Trust Trust is a concept about which there is much debate (Kramer & Tyler, 1996). Amongst the scientific community, the definition of the concept lacks consensus (Fisman and Khanna 1999; Adler 2001), although there appears to be agreement that the construct of trust is both complex and multifaceted (Ford, 2003; Simons 2002). Fukuyama (1995) views trust as “the expectation that arises within a community of regular, 26 honest, and cooperative behaviour, based on commonly shared norms, on the part of the members of the community" (p. x). This view of trust based on the expectation of honest and cooperative actions is shared by many (for example, Gambetta 1988; Mishra 1996; Bhattacherjee 2002). Mayer et al. (1995) define trust as “the willingness of a party to be vulnerable to the actions of another party based on the expectation that the other will perform a particular action important to the trustor, irrespective of the ability to monitor or control that other party” (p. 712). Both definitions above relate to a confidence in others’ future actions, with Mayer et al. extending their definition by arguing that in order to be vulnerable, one must be willing to take a risk based on the trusting relationship. An important distinction is that Fukuyama views trust as based on shared norms within a group; Mayer et al. approach their definition from a dyadic level, analysing the existence of trust between two individuals. Simons (2002) identifies Mayer et al.’s (1995) definition of trust as often cited in the literature. Mayer et al. identify three Benevolence attributes of another party in which perceptions of trust can be based, namely, benevolence, integrity and ability. This can be seen to extend Fukuyama’s definition of trust by incorporating ability as a specific antecedent to trust. Mishra (1996) goes further by defining four dimensions of trust: concern, reliability, competence and openness. While concern, reliability and competence can be seen to mirror benevolence, integrity and ability, respectively, Mayer et al. (1995) highlight the overlap of the openness dimension with their benevolence and integrity dimensions, arguing that “Mishra's openness is measured through questions about both the trustee's general openness with others and openness with the trustor, which could be expected to be related to either integrity or benevolence, respectively" (p. 722). This overlap is supported by McKnight et al. (2002) who presented confirmatory factor analysis. Figure 1 presents the overlap (mapping) of the two definitions of trust. Given the discussion in this section, and in the interests of parsimony, we adopt Mayer’s conceptualisation of the three characteristics of another party in which trust may be held. Concern Reliability Integrity Competence Ability Mayer et al Openness Mishra Figure 1: Mapping between Mayer et al. (1995) and Mishra’s (1996) definitional components of Trust 3 CONCEPTUAL MODEL: TRUST AS AN ANTECEDENT TO KNOWLEDGE SHARING In virtual communities, trusting relations can emerge without any direct social interaction due, in part, to the transparency of online communications. A newcomer typically has access to an electronic record of previous discussions and access to knowledge-based assets held in the community’s common repositories. Perceptions of trustworthiness based on 27 competence, honesty, benevolence and behavioural reliability provide confidence in future actions, and can be fostered by the high degree of openness and visibility surrounding online communications within virtual communities. This in turn fosters greater levels of cooperation, and discourages opportunistic behaviour (Fishman & Khanna, 1999). Ability- or competence-based trust exists when an individual believes that another party has knowledge and expertise in relation to a specific domain (Jarvenpaa et al. 1998). This facet of trust can be related to the fear of losing face that Ardichvili et al. (2003) identified as one of the main barriers to knowledge sharing in online communities of practice. For example, if a member’s perception of her own competence is significantly lower than the level of competence that she associates with the virtual community, then the motivation to publicly share her knowledge may be affected due to the fear of criticism or ridicule. This suggests a causal link between one’s perceptions of the community’s ability and engagement in knowledge sharing, whereby high levels of competence-based trust could restrict the knowledge shared with a community. However, whereas this argument appears to make logical sense, the converse does not logically hold. Where a member perceives a community to be of low competence, such a perception is unlikely to encourage knowledge sharing. In fact, such perceptions are more likely to discourage any form of voluntary participation in the community. 28 Computing and Information Systems © University of Paisley 2006 Another dynamic emerges when we consider the perspective of competencebased trust in motivating community participation. According to Lave and Wenger’s (1991) theory of situated learning, a newcomer to a community of practice becomes involved in a transition, over time, from peripheral participation in the practice towards becoming a masterful practitioner. This process involves a member learning by becoming situated within the field of the community’s practice. Within virtual CoPs, the process of situating in the community’s practice involves members creating and sharing knowledge by engaging in intellectual exchange through their participation in the community’s computer-mediated communications. By sharing and developing ideas, by testing and validating assumptions, by discussing, problem solving and generally striving to become more competent practitioners, the community members are able to engage in the mutual development of both their own knowledge and the community’s pool of expertise. With this ongoing process, members engaged in the development of cooperative and trusting relations whilst simultaneously developing knowledge of what it means to be a competent and masterful practitioner (Nahapiet and Ghoshal 1998). Within this process, we define a member’s passion for the practice, as the desire to become a more competent practitioner and to engage in the community’s practice. Where a member’s passion for the practice is high, the member will be more likely to seek engagement and interaction with a community of competent practitioners. From this perspective, the role of competence-based trust can be seen both as a potential enabler and barrier to community participation. When motivated by a passion for the practice, sharing knowledge with a community one perceives as highly competent becomes an intrinsically rewarding experience. The rewards derived from the consensual validation received from a community held in high regard can act as a motivator to members to share their ideas, thoughts, and insights. MacMillan and Chavis (1986) discuss the process of consensual validation, describing how “people will perform a variety of psychological gymnastics to obtain feedback and reassurance that ... what they see is real and that it is seen in the same way by others” (p. 11). In communities of practice, consensual validation may go further, acting as a mechanism that represents a member’s transition from peripherality to central participation; in effect validating a member’s standing within the community. Thus, the community’s consensual validation may act as a form of recognition, establishing and confirming one’s status as a knowledgeable practitioner. Conversely, sharing knowledge with a community one perceives to be of low competence will be an inherently less rewarding experience. Not only is the value derived from the community’s validation reduced, member’s perception of their self worth is also diminished by identifying and participating within a community perceived to be of limited competence. From the above discussion, we have identified a different dynamic that the role of competence-based trust may play in terms of enabling and disabling community participation in general, and knowledge sharing specifically. It is possible that the motivating effect driven by the opportunity to become a competent practitioner could potentially overcome the fear of losing face identified by Ardichvili et al. (2003). Likewise, it is equally possible that fear of losing face will be the dominant force. In other words, we can consider fear of losing face as a moderating variable such that if it exists at intense levels could change from positive to negative the relationship between competence and knowledge sharing. Based on this uncertainty, we propose the following hypothesis. H1: One’s degree of trust in the competence of a community is positively related to one’s engagement in knowledge sharing with the community. A shared interest in the community’s practice can foster the development of a sense of community amongst members through a process of identification between members. The process of identifying with a community enhances the individual’s concern with collective processes and group outcomes (Kramer, 1996, in Nahapiet & Ghoshal, 1998), and has been found to relate to expectations of benevolent behaviour and community participation (Chavis & Wandersman, 1990). Where the sense of community is strong and benevolence is high, community members are more likely to perceive knowledge as a public good, owned and maintained by the community. Wasko and Faraj (2000) note: “With a public good, the economically rational action is to free-ride. [However,] the motivation to exchange knowledge as a public good goes beyond the maximisation of self-interest and personal gain. People do not act only out of self-interest, but forego the tendency to free-ride out of a sense of fairness, public duty, and concern for their community … People often behave altruistically and pro-socially, contributing to the welfare of others without apparent compensation” (Wasko & Faraj, 2000: 161-2). From the public good perspective, knowledge sharing can be viewed as self-motivated through a sense of moral obligation and a general desire to be part of something larger. Such pro-social behaviours lead to the emergence of trust based on the perceived benevolence of the community , whereby members expect that help will be reciprocated should it be requested. Conversely, if one’s sense of a community’s benevolence is low, expectations of future reciprocity may likewise be low, and knowledge sharing is unlikely to be fostered. Furthermore, if low perceptions of a community’s benevolence are combined with high perceptions of the community’s competence, this may exacerbate the fear of losing face barrier discussed above . Benevolence-based trust will contribute to overcoming the fear of losing face by creating the confidence that one will not be criticised or made to look foolish when engaging publicly in sharing one’s knowledge. Further to this, knowledge sharing can be viewed from the perspective of BarTal’s (1976) theory of Generalised Reciprocity. From this perspective, the beneficiaries of knowledge-contributions are likely to seek to reciprocate benevolent actions with the collective, where direct reciprocation is not possible. Hence, we put forward the following proposition: H2: One’s degree of trust in the benevolence of a community is positively related to one’s engagement in knowledge sharing with the community. Integrity is a much debated concept within the trust literature. Sitkin and Roth (1993) discuss how perceptions of integrity-based trust are engendered within organisations by the perception of congruence between an individual’s values and the core cultural values of the organisation; the authors’ premise being that perceptions of value incongruence will foster feelings of distrust. This perspective sits close to Mayer’s et al.’s definition of integritybased trust based on “perceptions that the trustee adheres to a set of principles that the trustor finds acceptable” (p. 719). Mayer et al. elaborate on this understanding by defining a number of factors that influence the creation of integrity-based trust, such as: the independent verification of the trustee’s integrity from reputable third parties; perceptions that the trustee holds an acceptable level of moral standards; and demonstration of consistent behaviour including congruence between a trustee’s actions and words. The focus on the alignment between an actor’s actions and words is what Simon’s (2002) has defined as behavioral integrity; he describes this as the extent an individual is perceived to “walk her talk”, adding that, conversely, it 30 reflects the extent to which she is seen to be “talking her walk” (p. 19). Hence, trust in the integrity of a virtual CoP might be thought of as based in part on the compatability of the community’s cultural values with those of the trusting member, the credibility of the community’s reputation, and the consistency of community members’ past behaviour such as the extent to which actions are congruent with words. What can be derived from this understanding of integrity is that such perceptions are rooted in past behaviour. Consistent and reliable past behaviour creates confidence in future actions. If a member expects that other members’ future behaviour may lack integrity, for example, by acting dishonestly, unreliably or in a manner that is otherwise incongruent with her personal values, she is not likely to readily engage in sharing knowledge with the community. Conversely, she is likely to be more willing to engage in cooperative interaction where perceptions of honesty and expectations of behavioural reliability are high. Hence, H3: One’s degree of trust in the integrity of a community is positively related to one’s engagement in knowledge sharing with the community. Integrity-based trust H1 Competence-based trust H2 Online Knowledge Sharing H3 Benevolence-based trust Figure 2: The Antecedence of Trust to Knowledge Sharing 4 4.1 METHODOLOGY Data Collection For the fieldwork, CSC, a Fortune 500 global IT services organisation provided access to a suitable virtual community of practitioners. CSC employs over 79,000 people worldwide specializing in business process and information systems outsourcing, systems integration and consultancy services. The company have been focussed on knowledge management, and have been operating multiple online ‘knowledge communities’, for a number of years. Interviews with a number of community leaders were held to develop an understanding of the role of the communities and the mechanisms used to share knowledge. Access was granted to survey the organisation’s Systems Thinking Community, a global online competence-based group of over 400 members that had been in existence for over 4 years. The community’s main purpose is to improve the organisation’s business performance by applying the tools of systems thinking. The community develops decision simulation models by running online systems thinking courses via the Portal, with online workshops using Lotus SameTime. Membership and participation is entirely voluntary, and when the survey was conducted, 120 of the members were actively engaged in the 31 current course. Members received a link to the survey sent out by email. 4.2 Measurement Development 4.2.1 Knowledge sharing As knowledge sharing involves two or more participants engaged in the provision and acquisition of knowledge, it can therefore be measured from the perspective of both the source and the recipient of the exchange. However, this research addresses the role of trust in the provision of knowledge. Given that in the context of online communities there may be multiple recipients of shared knowledge, some of whom the provider may be unaware of, a logical approach to measuring knowledge sharing would be to understand how to tap into the construct from the source’s perspective. Hence a number of metrics were devised to measure the provision of knowledge from the perspective of the knowledge source3. When measuring specific behaviour, frequency of engagement in that behaviour is often used as an indicant (for example, Yoo et al. 2002). However, we argue that in relation to knowledge sharing, such an approach in itself is deficient. Knowledge is intangible, therefore it cannot be easily quantified. The frequency of engagement in knowledge sharing behaviour does not indicate the quality, usefulness or value of the knowledge provided or acquired. For example, a single contribution could have more value than ten contributions combined. Hence, measures that tap into both the quality (A5-A8) and quantity (A1A2) of an individual’s provision of knowledge were developed. Finally, the degree to which an individual feels that they engage in knowledge sharing will provide an indication of the individual’s knowledge sharing orientation. Hence, knowledge sharing focus was measured as a third dimension of knowledge sharing behaviour; two measures were developed to tap into this dimension (A3-A4). 3 See Appendix – Section A 4.2.2 Trust McKnight et al. (2002) developed the Trusting Beliefs Scale, which measures the degree an individual believes another party to be trustworthy. The original scales were designed for examining levels of trust held in online vendors and have been adapted to fit the context of this study4. C1-C3 measure the degree to which one perceives a community to be highly benevolent, C4C7 measure the degree to which one perceives a community to behave with high integrity, and C8-C11 measures the trusting belief in the competence of the community. 5 DATA ANALYSIS After partially completed and spoiled questionnaires were removed, this exercise yielded 75 usable responses, representing a response rate of 18%. This response was appreciated given that only 27% of the community members were actively engaged in the current course. Furthermore, Saunders et al. (2000) have pointed out how time poor modern organizations can be with regards to research, and this is emphasized with multi-national organizations such as this study was based. The responses were received from members primarily based in the US (45%), UK (34%) and Australia (11%). Members based in Switzerland, Spain, Denmark and India made up the remaining 10%. The respondents were predominantly male (81%), with the average length of tenure being 5 years, 10 months. 5.1 Validity and Reliability of Measures By ensuring that the measurement scales are both valid and reliable, we can increase the credibility of the conclusions that can be drawn by minimising the risk of measurement error confounding the results. The validity of a measurement scale is “the degree to which it is free from any systematic or non-random error” (Hair et al., 1998: 90). In other words, the 4 See Appendix – Section C 32 extent to which the measures accurately represent the construct that they are intended to measure. Netemeyer et al. (2003) elaborate on this understanding, describing construct validity as “the degree to which inferences legitimately can be made from measures ... to the theoretical constructs on which those operationalizations are based” (p. 71). As the measures used for trust were adapted from previously validated scales, it was particularly important to assure the validity and reliability of the newly developed knowledge sharing scale. Content validity provides a good foundation upon which to establish the validity of construct measurements (Litwin, 1995); this was tested by asking two of the community leaders and two senior academics (the co-authors) for feedback on the validity of the scales. Following this feedback, a number of the measures were reworded to aid clarity. A negatively loaded question was incorporated into the survey as a reliability check. be evaluated via the loadings of each indicator (Wasko & Faraj, 2005). Each item loaded higher on the expected construct than on other constructs. The cross-dimension loadings were high (.66 - .80) across the four dimensions of our knowledge-sharing construct, which is conceptually somewhat unsurprising. However, the crossdimension loadings were also high (.48 .70) across the three facets of trust, which demonstrated that the three trust subconstructs are only mildly discriminant from each other. In order to better understand the relationship between the theoretical constructs, a principal components analysis was conducted. As we hypothesised a relationship between the trust and knowledge sharing constructs, an oblique rotation was used (Pallant, 2005). Following theory and specifying 1 knowledge management and 3 trust factors we have the results in Table 1. Considering Table 1, the 3 items generally cluster at their expected factor grouping except the integrity item 6 T_INT which rather loads on the benevolence factor. The statement for that item is “The community would keep its commitments”. A follow up interview is a possible means of discovering why the respondents understood that question to refer to community benevolence rather than integrity. Meanwhile, the discussion section has provided some explanation why this could be the case. 5.1.1 Convergent and Discriminant Validity Convergent validity is a measure of the degree to which two measures that are purported to measure the same construct are correlated; discriminant validity measures the degree to which two conceptually similar constructs are distinct (Hair et al., 1998). Convergent and discriminant validity can K-SHR7 K-SHR9 K-SHR6 K-SHR1 K-SHR2 K-SHR3 K-SHR4 K-SHR5 K-SHR8 11 T_CMP 8 T_CMP 1 0.92 0.91 0.90 0.88 0.82 0.82 0.81 0.79 0.79 2 3 4 0.90 0.89 33 10 T_CMP 0.86 9 T_CMP 0.82 1 T_BEN 0.85 2 T_BEN 0.73 6 T_INT 0.70 3 T_BEN 0.67 5 T_INT 0.38 4 T_INT 0.33 0.37 7 T_INT 0.40 Rotation Method: Oblimin with Kaiser Normalization. Rotation converged in 11 a iterations. 0.45 0.57 0.50 0.47 Table 1: 4 Factor Extraction - Oblimin 6 DISCUSSION AND FUTURE RESEARCH DIRECTIONS All the three trust factors positively relate to knowledge sharing in on-line communities of practice thus upholding all the 3 hypotheses of this study. Each will now be discussed. 6.1 Competent-based Trust The positive hypothesis tests true that competentbased trust positively influences knowledge sharing in virtual communities. This finding is in line with recent organisational research that regards integrity (and competence) as paramount to trust (Adler, 2001). As pointed out by Ardichvili et al (2003, p 73) there was need to verify that competence (and integrity) are major -Fear of losing face Competence Trust components of trust in virtual communities of practice. This verification is what this finding of the study has done. Of the three trust factors, competence is the one that has all its original items clearly cluster together. The more a virtual community is competent (knowledgeable, capable and effective in sharing knowledge), the more its members will be inclined to share knowledge. At least from the sample used, it can be said that the passion for knowledge had a greater effect than fear of losing face. It is however suggested that a future study takes a closer look at the moderating influences of these two variables. Figure 3 outlines how their influences could be conceptualised. Knowledge sharing +Passion for knowledge Figure 3: Moderating influences of fear of losing face and passion for knowledge 34 perceived as only integrity but also as competence and benevolence; hence, the existence of some cross-loading on those items. Overall, we can conclude that integrity is a component of trust which affects knowledge sharing in on-line communities of practice. This conclusion agrees with earlier research that recognise integrity as a trust component (cf Tschannen-Moran, 2000, pp 314, 318) Fear of losing face should dampen the positive influence of competence trust on knowledge sharing whereas passion for knowledge should do the opposite (cf Ardichvili et al. 2003; Nahapiet and Ghoshal 1998). For arguments supporting the existence of these moderating variables, please refer to section 3. 6.2 Benevolence-based Trust Items C1 (community would act in my best interest) and C3 (community is interested in my well-being, not just its own) clearly cluster together and support prove hypothesis 2 that benevolent trust positively affects the level of knowledge sharing. When it comes to C2 (if required, the community would do its best to help me), the highest loading (0.73) is on the benevolence factor but other loading (0.45) is on the integrity factor. This suggests that respondents perceived the community’s willingness to do the best to help its members as not only benevolence but also an integrity attribute. On the whole, benevolence stands out as a component of trust that affects knowledge sharing in on-line communities of practice. This finding is in consonant with earlier research that include benevolence as a component of trust (cf Tschannen-Moran, 2000, pp 314, 318) 6.3 Integrity-based trust Though they mostly cluster on the expected factor, the integrity items are the most problematic. In the first place, the fact that the community would keep its commitments (C6) is perceived by respondents as benevolence instead of integrity. This could be because, even though they also acted as givers of knowledge to each other, the respondents regarded themselves more as students of the community; taking knowledge from the community to apply to their work situation (see section 4.1). Moreover, honesty and sincerity of the community (C4, C5 and C7) are not 7 CONCLUSIONS The research has upheld the three trust factors as positively affecting knowledge sharing in online communities. As discussed in 5.1.1 the cross loading of the factors suggest that though they are theoretically distinct, practically, they buttress each other to influence knowledge sharing. 7.1 Implications for Knowledge Management Professionals Current KM research and practice has recognised that the informal and trusting nature of communities of practice is the bedrock of knowledge sharing. The implication is that KM practitioners need not be pro-active in forming and running communities of practice. Nonetheless, an understanding of the trust components would guide practitioners on how to create and support knowledge sharing environment. This study has confirmed that trust is composed of the three components of competence, benevolence and integrity which are positively related to knowledge sharing in on-line communities of practice. The implication is that practitioners should support the three values if they are to encourage trust and knowledge sharing. This research did not investigate how to support the three values but a priori, organisations should look for these three qualities in their would-be recruits. Also, KM practitioners can lead by example: exhibit these qualities in their dealing with other members of the organisation. They 35 can also take advantage of every opportunity to encourage these attributes in existing employees, who are potential members of on-line communities of practice. The clearer-cut and most significant component appears to be competence but the findings show that the factors enhance each other and therefore, KM practitioners should be interested in all the three components of trust. ACKNOWLEDGMENTS We would like to express our thanks to Computer Sciences Corporation for their collaboration during this project, in particular, for the company’s participation in the empirical research. C3 C4 C5 C6 C7 C8 C9 C10 APPENDIX. QUESTIONNAIRE C11 SECTION A: KNOWLEDGE SHARING A1 I frequently share my knowledge with others in the community. A2 I am one of the more active contributors within the community. A3 I make a conscious effort to spend time engaged in activities that contribute knowledge to the community. A4 I try to share my knowledge with the community. A5 Other community members find my knowledge-sharing contributions to be useful. A6 My contributions to the community enable others to develop new knowledge. A7 I am a knowledgeable contributor to the virtual community. A8 The knowledge I share with the community has a positive impact on the business. A9 Overall, I feel the frequency and quality of my knowledge-sharing efforts are of great value to the community. SECTION C: TRUST C1 I believe that the competency group’s virtual community would act in my best interest. C2 If I required help, the community would C12 do its best to help me. The community is interested in my wellbeing, not just its own. The community is truthful in its dealings with me. I would characterise the community as honest. The community would keep its commitments. The community is genuine and sincere. The community is a competent and effective source of expertise. The community performs its role of sharing knowledge very well. Overall, the community is a capable and proficient source of expertise and knowledge. In general, the community is very knowledgeable. I trust the community when I ask them not to forward or share any CSC or client sensitive material. References Adler, P. S. (2001). "Market, Hierarchy, and Trust: The Knowledge Economy and the Future of Capitalism." Organization Science, 12(2), 215-234. Ardichvili, A., Page, V., and Wentling, T. (2002). "Motivation and Barriers to Participation in Virtual Knowledge-sharing Communities of Practice." Reprint: <12 Reprint Status> Ardichvili, A., Page, V., and Wentling, T. (2003). "Motivation and barriers to participation in virtual knowledge-sharing communities of practice." Journal of Knowledge Management, 7(1), 64-77. Argyris, C. and Schon, D. A. (1978). Organizational Learning: A Theory of Action Perspective, Addison-Wesley Publishing, Reading, MA. Beaumont, P. and Hunter, L. C. (2002). "Managing Knowledge Workers." CIPD, London. Bhattacherjee, A. (2002). "Individual Trust in Online Firms: Scale development and initial test." Journal of Management Information, 19(1), 211-242. Brown, J. S. and Duguid, P. (1991). "Organisational Learning and Communities of Practice: Towards a unified view of working, learning and innovation." Organization Science, 2(1), 41-57. 36 Davenport, T. H. and Prusak, L. (1998). Working Knowledge, Harvard Business School Press, Boston, Massachusetts. de Vaus, D. A. (1996). Surveys in Social Research, Unwin Hyman, London. Fisman, R. and Khanna, T. (1999). "Is trust a historical residue? Information flows and trust levels." Journal of Economic Behavior & Organization, 38, 79-92. Fukuyama, F. (1995). Trust: The social virtues & the creation of prosperity, The Free Press, New York. Gambetta, D. (1988). "Can we trust trust?" Trust: Making and breaking cooperative relations, D. Gambetta, ed., Basil Blackwell, New York, 213238. Hair, J. S., Anderson, R. E., Tatham, R. L., and Black, W. C. (1995). Multivariate Data Analysis, Prentice-Hall, Engelwood Cliffs, NJ. Herzberg, F. (1968). "One More Time: How Do You Motivate Employees?" Harvard Business Review,(Special Issue - January 2003), 87-96. Hildreth, P. and Kimble, C. (2000). "Communities of Practice in the distributed international environment." Journal of Knowledge Management, 4(1), 27-38. Jarvenpaa, S. L., Knoll, K., and Leidner, D. E. (1998). "Is Anybody Out There? Antecedents of Trust in Global Virtual Teams." Journal of Management Information Systems, 14(4), 29-64. Jarvenpaa, S. L. and Staples, D. S. (2000). "The Use of Collaborative Electronic Media for Information Sharing: an exploratory study of determinants." Journal of Strategic Information Systems, 9, 129154. Kramer, R. M. and Tyler, T. R. (1996). Trust in Organisations: Frontiers of theory and research, Sage, Thousand Oaks, CA. Lave, J. and Wenger, E. (1991). Situated Learning: Legitimate Peripheral Participation, Cambridge University Press, Cambridge, MA. Lesser, E. and Everest, K. (2001). "Communities of Practice: Making the most of intellectual capital." Ivey Business Journal, 65(4). Lesser, E. L. and Storck, J. (2001). "Communities of Practice and Organisational Performance." IBM Systems Journal, 40(4), 831-841. Levin, D. Z., Cross, R., and Abrams, L. C. (2002a). "The Strength of Weak Ties You Can Trust: The mediating role of trust in effective knowledge transfer." Rutgers University. Marques de Sa, J. P. (2003). Applied Statistics using SPSS, STATISTICA, and MATLAB, SpringerVerlag, Berlin. Mayer, R. C., Davis, J. H., and Schoorman, F. D. (1995). "An Integrative Model of Organisational Trust." Academy of Management Review, 20(3), 709-734. McKnight, D. H., Choudhury, V., and Kacmar, C. (2002). "Developing and Validating Trust Measures for e-Commerce: An Integrative Typology." Information Systems Research, 13(3), 334-359. McKnight, D. H., Cummings, L. L., and Chervany, N. L. (1998). "Initial Trust Formation in New Organizational Relationships." Academy of Management Review, 23(3), 473(18). McMillan, D. W. and Chavis, D. M. (1986). "Sense of Community: A Definition and Theory." Journal of Community Psychology, 14(1), 6-23. Mishra, A. K. (1996). "Organisational Responses to Crisis: The Centrality of Trust." R. M. K. &. T. R. Tyler, ed., Sage, Thousand Oaks, CA, 261-287. Nahapiet, J. and Ghoshal, S. (1998). "Social Capital, Intellectual Capital and the Organizational Advantage." Academy of Management Review, 23(2), 242-266. Preece, J. (2000). Online Communities: Designing Usability, Supporting Sociability, John Wiley & Sons, Chichester, England. Rheingold, H. (1993). The Virtual Community: Homesteading on the Electronic Frontier, AddisonWesley, Reading, MA. Ring, P. S. and Van de Ven, A. H. (1994). "Developmental Processes of Cooperative Interorganizational relationships." Academy of Management Review, 19, 90-118. Ruggles, R. L. (1998). "The State of the Notion: knowledge management in practice." California Management Review, 40(3), 80-89. Saunders, M., Lewis, P., and Thornhill, A. (2000). Research Methods for Business Students, Pearson Education, Harlow. Scarbrough, H. (1999). Knowledge Management: A literature review, Institute of Personnel and Development. Scholl, W., Konig, C., Meyer, B., and Heisig, P. (2004). "The Future of Knowledge Management: an international delphi study." Journal of Knowledge Management, 8(2), 19-35. Simons, T. (2002). "Behavioural Integrity: The Perceived Alignment Between Managers' Words 37 and Deeds as a Research Focus." Organization Science, 13(1), 18-35. Tsai, W. and Ghoshal, S. (1998). "Social Capital and Value Creation: The Role of Intrafirm Networks." Academy of Management Journal, 41(4), 464-476. Tschannen-Moran (2001) “Collaboration and the need for trust” Journal of Educational Administration 39(4), 308-31. Van Beveren, J. (2002). "A Model of Knowledge Acquisition that Refocuses Knowledge Management." Journal of Knowledge Management, 6(1), 18-22. Wang, C. L. and Ahmed, P. K. (2003). "Structure and structural dimensions for knowledge-based organizations." Measuring Business Excellence, 7(1), 51-62. Wasko, M. M. and Faraj, S. (2000). ""It is what one does": why people participate and help others in electronic communities of practice." Journal of Strategic Information Systems, 9(2-3), 155-173. Wenger, E. (1998). Communities of Practice: Learning, meaning and identity, Cambridge University Press, Cambridge. Wenger, E., McDermott, R., and Snyder, W. (2002). Cultivating Communities of Practice: A guide to managing knowledge, Harvard Business School Press. Wenger, E. and Snyder, W. M. (2000). "Communities of Practice: The organisational frontier." Harvard Business Review,(Jan-Feb 2000), 139-145. Yoo, W.-S., Suh, K.-S., and Lee, M.-B. (2002). "Exploring Factors Enhancing Member Participation in Virtual Communities." Journal of Global Information Management, 10(3), 55-71. Dr Abel Usoro is a Lecturer in the School of Computing, Mark W Sharratt was his MSc student, and Eric Tsui is a Professor at the HK Polytechnic University in Hong Kong. Prof Tsui facilitated access to the sampling frame and also collaborates in the research. 38 Revise New Reuse Proposed Retrieve Retain Problem proposed case Case Base Similar information experience Situation Solution solution Cases 39
© Copyright 2026 Paperzz