2015 48th Hawaii International Conference on System Sciences Toward a Theory of Knowledge Economics: an information systems approach Martin Hilbert University of California, Davis; [email protected] engineering and computer science. This article applies concepts from theories developed for the engineering of technologies systems to social systems and economic dynamics. The idea is to use our understanding of how information systems work with knowledge to describe aspects of how the knowledge economy works. If we are able to formulate both technological systems and economic processes within the same analytical language, we will also be able to create a theory that treats socio-economic and technologic information processes as the mutually dependent whole that they reassemble in the knowledge economy. The approach taken in this article shows that the reformulation of general economic evolution in terms of information theory sheds a new light on several aspects of economic dynamics. It follows a long tradition in taking an evolutionary and multilevel approach to knowledge [32,33]. More than using these concepts as analogies, it formalizes them in concrete mathematical representations. For example, economic natural selection (such as through market dynamics) can be modelled as uncertainty reduction through negative entropy. Concepts like ‘economic fitness’ obtain a new meaning as the literal information ‘fit’ (the quantifiable amount of mutual information) between the evolving socio-economic system and its environment. Such fit occurs over multiple levels. As a natural result of this approach, it shows that absolute knowledge of economic dynamics can maximize economic payoff. The presented derivations work with formal notions of information and knowledge (such as entropy, mutual information, and Kolmogorov complexity), which can be quantified (for example in bits and bytes). This effectively converts the amount of knowledge into a quantifiable ingredient of economic growth. Intuitively the argument of this article is as follows. If you have information about a specific event, you have the potential to act on it and to extract economic benefit from it; if you don’t, you don’t. How you obtain this information (be it through empirical analysis, a humanly communicated cue, or any other hint) is another question. Fact is that the amount of information is quantifiable (for example in bits), and that the equations of economic evolution can show a Abstract A comprehensive theory of knowledge societies and knowledge economics is still missing. This article shows that the analytical tools from computer science, information systems and information theory provide an adequate language to work toward such theory. The presented formalization follows an evolutionary and multilevel approach, and embraces both the concepts of information and knowledge. It is shown that economic evolution can be modeled as a process that selects superior strategies from a set of possibilities through natural (market driven) selection. This reduces uncertainty by the production of negative entropy (‘negentropy’). Furthermore, economic ‘fit-ness’ can be modelled in terms of the informational ‘fit’ of the economic system with its environment, which takes the form of Shannon’s mutual information. In the case of a deterministic dynamic, uncertainty reduction takes the form of a predictable algorithm, which is quantified through Kolmogorov complexity. The length of such algorithm represents the knowledge about the unfolding dynamic. This shows that we might be able to eventually describe both, economics and technological information systems within the same analytical framework. 1. Introduction “Many of the most general and powerful discoveries have arisen, not through the study of phenomena as they occur in nature, but, rather, through the study of phenomena in man-made devices, in products of technology… Our knowledge of aerodynamics and hydrodynamics exists chiefly because we created airplanes and ships, not because of the existence of birds and fishes. Our knowledge of electricity came mainly not from the study of lightning, but from the study of man’s artefacts” [1]. In this sense, our understanding of what information and knowledge is benefits significantly from the study of artificial devices that replicate intelligent processes. This suggests that knowledge economics can best be understood not by studying it as it occurs in the complex business setting of companies, but by looking at it from the perspective of theoretical communication 1530-1605/15 $31.00 © 2015 IEEE DOI 10.1109/HICSS.2015.461 3841 On the other side, adopting a formal definition from information theory, information is related to the reduction of uncertainty and uncertainty implies probabilities. One bit of code in a telecommunication message provides the reduction of uncertainty of a probability space by half [4-5]. The crux of the presented argument here consists in the fact that both quantities (i.e. both bits) are equivalent asymptotically. This asymptotic equivalence of knowledge and information is due to the asymptotic equivalence of deterministic and probabilistic processes. The relevant proof is quite subtle and goes back to Kolmogorov in 1968 [2], and was fleshed out by Zvonkin and Levin two years later [6]. It has been popularized in the information theory community by Leung-Yan-Cheong and Cover [7] (see also the standard textbook on information theory ([5], Ch. 14.3) and among computer scientists and physicists by a series of papers that eventually exorcised Maxwell’s 120 year old demon by the hands of Zurek [8-10] (for an introduction see [11]). While we will not go into the relevant math of this proof at this point, we review the general intuition behind the argument. Theoretically, the right measure to quantify the notion of a deterministically unfolding process (an algorithm) has been established by Andrey Kolmogorov [2,12], and was independently developed by Solomonoff [13] and Chaitin [14], and is known as “Kolmogorov complexity” [2]. Kolmogorov defined the amount of information in a deterministic algorithm to be equal to the minimum number of symbols that we need to efficiently describe an object to a specific level of detail. Knowledge about the unfolding of a dynamical system with two environmental states requires a binary code (e.g. 1 and 0), resulting in an algorithm that describes the occurrences of either the one, or the other environmental state, like: 100100100100… . If this pattern is stable, we can compress (e.g. by noticing that a 1 is always followed by two 0s, resulting in a compressed algorithm that looks something like: [repeat 100]). This algorithm defines a stable endless series in a deterministic manner. The length in bits of this optimally compressed algorithm is its Kolmogorov complexity. If the sequence would be totally random, we would need to know bit by bit, and the algorithm would be just as long as the sequence. The final length of the algorithm quantifies the amount of knowledge the agent needs to unequivocally describe such dynamical environmental system. Shannon [4], the founding father of information theory, basically looked at the same issue the other direct relation between this amount of information and the economic potential to grow. If you know about a series of unfolding events in the future, you can even extract more economic benefit. This converts (probabilistic) information, into (deterministic) knowledge). Here we work with socalled algorithmic information theory (in the sense of Kolmogorov complexity), not anymore with probabilistic information theory (in the sense of Shannon’s entropy). The more knowledge about a process, the more potential to grow. The amount of knowledge can also be quantified in bits (the number of bits of the shortest program that describes the algorithm). It turns out that the amount of information about a possible event (such as measured in bits) is the same as the amount of information needed to describe the deterministic unfolding of a specific event (such as measured in bits). In the first case, uncertainty is reduced through communication (provision of probabilistic informational bits). In the second case it is provided by knowledge about the process (provision of a deterministic algorithm encoded in bits). 2. How to formulize information and knowledge The natural branches of science to tackle the task of formalizing knowledge and information are computer science and information theory. It turns out that from the point of view of these solid theoretical frameworks, both knowledge and information are two sides of the same coin. Just like two sides of a coin, they are not identical, but they are two faces that show the same quantity. On the one side, adopting a strict computer science definition, knowledge implies a step-by-step recipe of doing something, a deterministic algorithm. If there is knowledge, there is no doubt, there is a deterministic procedure [2-3]. One bit of code in an informatics program provides a deterministic instruction on a binary choice on how to proceed in the algorithm. The number of bits required to represent the shortest expression of an algorithm is called Kolmogorov complexity [12-14]. As such, knowledge (represented in form of an algorithm) can be quantified in bits.1 1 It might be that the explicit form of the algorithm is not revealed [20]. This does not change this logic, as some kind of underlying process is surely executed (even so it is not disclosed). The literature of knowledge economics is full of examples where such tacit knowledge algorithms are made explicit (such as in the case of the famous home-baking machine in the “knowledge creating company” of Nonaka and Takeuchi [34]). 3842 way around. He asked: what is the likelihood that we will pick one specific environmental state if we draw randomly from the universe of all possible states? He concluded that this likelihood is equivalent to the amount of uncertainty that gets reduced when we are ‘informed’ about the correct state. Shannon’s information is defined as the reduction of uncertainty and uncertainty quantified with probabilities. He defines that one bit reduces uncertainty exactly by half. On the one hand, Shannon’s measure identifies the object of interest based on the likelihood that this object will appear amidst all possible objects (probabilistic). On the other hand, Kolmogorov’s measure involves a step-by-step description, an executable procedure, or unambiguous recipe, in short, an algorithm. When executed, such algorithms unfold in time, because time is required to execute them. Economic life is full of (more or less well defined and approximate) algorithms. Economic routines [15] and behavioral patterns [16] are often mentioned examples, as are cultural norms [17] or institutionalized mechanisms, procedures and habits [18-19]. The agent might not be aware or might not be able to formulate the details of the procedure (as with tacit knowledge, [20]1), but it still has to be sorted and executed from somewhere. Some procedures might also only be approximate heuristics [16], which means that they are not the optimal solution. However, in all of those cases, the process is still deterministic (not probabilistic). Kolmogorov’s measure quantifies the amount of information in an algorithm and is therefore often called “algorithmic randomness” [8] and “algorithmic information” [9]. In short, Kolmogorov describes something by contrasting it against all other possible things, and Shannon evaluates the probability that something appears among all other possible things. Since the likelihood and the description are linked to the number of alternative possible choices, both are approximated on average by the information theoretic equivalent of the law of large numbers (called the ‘asymptotic equipartition property, see [5]). One consists of ‘knowledge’ (the ability to describe/recreate something by following a deterministic procedural structure in time) and the other one of ‘information’ (the reduction of probabilistic uncertainty about a structure). In the words of Cover and Thomas, who wrote the standard text book on information theory: “It is an amazing fact that the expected length of the shortest binary computer description of a random variable is approximately equal to its entropy” [5], and in the words of Li and Vitanyi, who wrote the standard textbook on Kolmogorov complexity: “It is a beautiful fact that these two notions turn out to be much the same” [3]. 3. How economic evolution processes information by reducing uncertainty We will now apply both concepts to the general notions of economic evolution (see also [21]). On the one hand, if we have complete knowledge about any economic dynamic, we can predict the deterministic unfolding of the environment step-by-step. This means that we have knowledge about the dynamic. If we know the future, it is straightforward to maximize economic growth within a given reality by reallocating 100% of the resources to the best available strategy. On the other hand, if we have uncertainty about the unfolding of the environment, economic evolution will identify the superior strategy (or company) by means of natural selection. The fittest strategy of all applied strategies will survive. We can model this in terms of conventional information theory as follows (see also [22]). We start with the traditional interpretation of fitness as the factor of reproduction or economic growth, and define the number of most fine-grained units at time as , with: , whereas the units of the evolving population are freely selectable, indivisible, positive integers (number of US$ of gamble bets and payoff, number of US$ of GDP, or number of economic agents, etc.). Without loss of generality, we can represent growth factors on a logarithmic scale. This is common practice in economics and in our case normalizes fitness at 0 for . For an unchanging population, in example, a logarithm of base 2 represents fitness terms of the number of population doublings at each time step. This simple set up is sufficient to express fitness in terms of Shannon entropies (in the sense of [4-5]): 3843 Equation (1.1) shows that if we assume that each individual of the evolving population has its own, uniquely distinguishable type, fitness measures the uncertainty inherent in the future population, minus the uncertainty of the present population (both consisting with different ‘alphabets’, or ‘number of types’ in this case, which can be due to minuscular mutations, etc.). In the case of many mutations that result in many more , uncertainty increases, unique types at step (see equation (1.1)). In the and case that evolution leads to the survival of one single member at the final equilibrium stage at time (‘survival of the fittest’), there is no uncertainty anymore about who of all population members is the . In this chosen one and case: fitness on a higher level more than the sum of its parts on a lower level [22]. The population fitness during an environmental state is its expected average fitness (over all types) . The long given the environmental state : term logarithmic population fitness over periods can then be partitioned in time into the product of the growth factors in the distinct environmental states , each appearing in its proportion . For example, the overall growth rate of an industry over the entire year is the product of the share of growth rates on days with much growth, and days with little growth. The negativity of entropy (sometimes also referred to as “negentropy” [23-24]) shows that economic evolution --just like any process of communication -reduces uncertainty. While a communication processes reveals a desired symbol of an uncertain alphabet, evolution can be understood as a process that reveals the fittest member of a population. The amount of uncertainty reduced by evolution is measured by the . amount of negative entropy This logic can be generalized to evolutionary processes that are out of equilibrium and still ongoing (no determination of the ‘fittest’ (yet)) [22]. Relative entropy (the so-called Kullback-Leibler distance [5]) turns out to be the correct information theoretic metric in this case. Being a relative metric, it allows to quantify the resolved uncertainty relative to the present or relative to the future. From the perspective of the future, uncertainty has already been resolved, resulting in the subtraction of relative uncertainty, while from the perspective of the present, the descriptive complexity of the involved uncertainty (such as expressed in informational bits) is a positive ingredient of fitness, as it is still to be exploited. In this notation the single-bar emphasizes the (arithmetic) ‘space-average’ over the different population members, while the double-bar emphasizes the additional (geometric) ‘time-average’ over the observed period. We can expand and reformulate this in space and average population fitness time: In this notation the single-bar emphasizes the (arithmetic) ‘space-average’ over the different population members, while the double-bar emphasizes the additional (geometric) ‘time-average’ over the observed period. We can expand and reformulate this in space and average population fitness time: 4. How fitness is about the informational fit We can expand this logic to include the dynamic between the evolving system and the environment. By doing his, ‘fit-ness’ attains an intuitive, but formal interpretation as the amount of ‘fit’ between the evolving population and its environment. ‘Fit-ness’ is related to the amount of informational ‘fit’ on a certain level. This amount of mutual information represents the emergent quantity that makes the total population If the environment and the future state of the population are independent, Shannon’s much 3844 celebrated mutual information [4] between the environment and the present population emerges naturally from the equation [22]. This is always the case at the final end state of an evolutionary process in which only the fittest has survived: population fitness (with and therefore (for related work see also [26-28]. As such, equation (2.1) is a generalization of Kelly’s special case result, since it also holds for the less restrictive case with variety of growth factors among types (for more recent results on Kelly’s criteria see. The basic idea behind Kelly’s bet-hedging (also called Kelly gambling, or the Kelly criterion) is to not maximize short-term utility, but to consider the long term return over time. To optimize the growth rate, we have to adjust our shares of types (which refer to a variety in “space”) to the shares of the environment (which refer to a variety in “time”). Kelly’s result in equation (2.3) shows that the best we can do is to maintain the shares of types proportional to the occurrence of their respectively favorable environments. In other words, if we have two possible environments that occur with a distribution of 70 % and 30 % of the time, we should maintain a stable share of 70% of the type that grows faster in the first environmental state, and 30 % of the type that grows faster in the other environmental state (this can be shown with a straightforward Lagrange multiplier maximization for the asymptotic case, see [5], or simply consider the fact that in this case the term in equation (2.3) becomes 0). This is called a proportional betting strategy. It is a clear-cut one-to-one relationship between what is likely to be expected (“in time”) and the allocation of current preferences right now (“in space”). The technical reason behind the superiority of proportional betting is that a geometric expansion always overtakes arithmetic expansion [5], which is a well-known result in economic literature. The superiority of geometric expansion has become second nature to most economists. We can also reduce uncertainty with additional side information about the future state of the environment. The two sources for possible side information can be the observation of the past (“memory”) and observations of third events from the present that correlate with the future (“cues”). Let suppose that we have a certain amount of memory that allows us to extract information about the future. The importance of data analysis to predict future patterns becomes increasingly important in a big data world [29]. We will ask how much such information from memory can increase our maximal growth rate. Aiming at maximal growth rates we stick to proportional betting, and the DKL term of equation (2.3) falls out. According to equation (2.3), the remaining focus is therefore set on the reduction of the uncertainty of the future (its entropy H(T)) due to memory. The memory about the past represents a new conditioning variable . Equation (2.2) shows that mutual information between the current population and its environment adds to current population fitness on a certain level. Fitness obtains an intuitive explanation as the amount of mutual information ‘fit’ between the evolving system and its environment [22]. The amount of mutual information also quantifies the amount of levelspecific emergence involved in an evolutionary is exactly process: the total more than the (weighted) sum of its parts. We could now assume an extreme case of relative fitness in which each environmental state has only one surviving type with superior fitness, , while in this environmental state all other types die out with . Only one type fits the particular environment. In this case the fittest type represents the entire population after updating, , and equation (8.2) simplifies to: Equation (2.3) is a well-known result derived by Kelly in 1956 [25]. Since both and are always positive, it says that the attainable average population fitness is reduced by the level of uncertainty inherent in , as well as by the distance the environment, of the distribution of the evolving population from its . Among other things, this leads to environment, the celebrated result in portfolio theory that a proportional bet hedging strategy maximizes 3845 on basis of which we need to condition the likelihood of the occurrence of a certain environmental state and the ensuing distribution among our types. For example, the past could have been a prolonged bear or bull market. Let us suppose that they occurred with a likelihood q(t-1) and (1-q)(t-1), respectively. Given a certain past (bull or bear) the probability of future bull or bear will be different (conditioning). We therefore end up with two separate distributions, which are distinguished by the conditional variable q(t│t-1), the likelihood of environmental state t, conditioned on the past (t-1). Therefore, the long term optimal growth rate, conditioned on this memory is: This measure is again Shannon’s celebrated mutual information [4]. In our case it is the mutual information between the information extracted from the past, and the current environment. It quantifies how much the memory of the past can tell us about the future (this relation is of course noisy and therefore probabilistic). In words, it says that information about the past can increase the optimal growth rate exactly by as much as the past can tell us about the future. If the past tells us a lot about the future, the mutual information between past and future will be high, and fitness can be increased decisively. If the future is independent from the past, the mutual information is 0 [4], and no increase is possible. This one-to-one relationship between the amount of information (as quantified by Shannon) and the increase in growth rate provides information with a value, a fitness value. Much in the same sense, we can now extent the analysis and ask about how much an outside cue from a third, but related event might increase fitness [21]. For example, we might observe the behavior of other economic agents, or obtain a tip from an expert. In the simplest case, the logic basically stays the same and we can replace the conditioning variable of the past (T-1) with any other conditioning variable of a cue, and obtain the mutual information between the cue and the . environment: 5. Empirical application of the knowledge economy The resulting term is the conditional entropy H(T│T-1), that quantifies the amount of uncertainty that is left after the conditional variable is known [5]. For the difference between the optimal long term growth rate with, and without memory we get: This section takes the ideas of the previous section and applies it to the empirical case of the knowledge economy. In specific we calculate the amount of bits processed by the export trade of the U.S. economy over a period of twenty years. Fig. 1 is based on the data of international export of 512 product groups of the USA between 1982 and 1987, measured in nominal value of thousand US$. Fig. 1 aggregates the 9 groups (code 0 to 8) of the official Standard International Trade Classification (SITC), revision 2 [35-37] into two higher level groups: on the highest level we distinguish between two groups (non-manufacturing and manufacturing). The first consists of Edibles (code 0: Food and live animals chiefly for food; code 1: Beverages and tobacco) and Substances (code 2: Crude materials, inedible, except fuels; code 3: Mineral fuels, lubricants and related materials; code 4: Animal and vegetable oils, fats and waxes); and the second consists of manufactured goods (code 6: Manufactured goods classified chiefly by materials), and Machinery (code 7: Machinery and transport equipment; code 8: Miscellaneous manufactured articles) (see Fig. 1). 3846 In other words, following our binary population structure on the first level of fine-graining, during the period between 1979 and 2000 the U.S.’s export economy reduced 0.0007 bits of relative uncertainty. This illustrates that equation (2.1) (unusual as it may seem at first) is not mere theory, but can readily be applied to practical cases. We can quantify how many bits and bytes are processed by the knowledge economy. Fig. 1. USA’s export from 1979-2000 in nominal value of thousand US$ (based on the 512 most completely recorded export items of the first digit classification of SITC revision 2). 6. Conclusion The previous sections have presented a formal way to shows that information from past experience and from present cues can allow us to predict the future better and therefore allows us to increase the growth rate of the population if we allocate resources accordingly. This makes intuitively sense: if we have more information about the environment, we can adjust the evolving population to environmental conditions (therefore optimize the informational ‘fit’ between the environment and the evolving system) and optimize population growth/fitness. Each bit of information allows us to increase fitness accordingly. The fact that Shannon’s probabilistic information metric arises in these equation can be intuitively understood as follows [30]: imagine complete uncertainty among 8 equally likely choices (e.g. investment options). Having US$ 8, the bet hedging investor would distribute US$1 to each all of those 8 choices and surely walk out with one winning dollar. Let us assume that we receive 1 bit of information, which is defined by Shannon [4] as the reduction of uncertainty by half). As a result, the investor would know which half of the investment options fail, and which of the remaining 4 options are still equally likely to succeed. He would now equally distribute US$2 of his US$8 to each of the remaining 4 choices and walk out with two winning dollars. One bit of information reduced uncertainty by half and multiplied the obtained gain by two. The equation shows that the logic of dividing uncertainty by 2 will increase gains by 2 is not a numerical coincidence, but a solid relation between uncertainty reduction and achievable growth. Using a less stylized example, imagine that several bakeries discovered hidden relationships between the weather and the demand for specific baked goods: the demand for cake grows with rain, the demand for salty goods with sunshine. This insight can be obtained through a big data driven information system, or simply through human observation. Conditioned on a cue for rain or sunshine, a bakery can calculate likelihoods of how many salty and sweet products it will sell and there optimize its economic growth. In practice, productivity increments of up to 20 % have The overall compound annual growth factor over the involved 21 periods is 2.4284. The respective evolutionary fitness matrix for the first level (nonmanufacturing vs. manufacturing) is presented in Table 1. It shows that one third of the time the environment is characterized by superior fitness of the nonmanufacturing sector (7 out of 21 periods) and two thirds vice versa. The respective compound annual growth rates during these kind of periods are presented in the matrix. Table 1. USA’s export fitness and environmental distributions Level 1. 1.055 0.989 1.022 1.085 Calculating the required values following the logic outlined above in equation (2.1), gives: 3847 environment [8-11]. This article has shown an analogy for economic evolution: the amount of information and knowledge the social system has about the environment is equal to its potential to grow. This is intuitively pleasing: the more knowledge and information we have, the more our economic system can grow. This leads to the final question of how much true and solid information we can obtain about the shortand long term patterns. In essence, this question is at the heart of what economic analysis does: identifying compressible structure and patterns in market forces and human behavior in order to extract laws and rules of thumb that allow us describe and predict the dynamics of the system and act accordingly. In this sense, economics itself becomes an input ingredient of economic growth. The more we know about the economy and its dynamics, the more we can allocate resources to enable growth. This shows that concepts and analytical tools from computer science, information systems and information theory provide an adequate language to describe, model and quantify the general role of knowledge in economic dynamics. The result is promising, since it proposes the possibility that eventually we will be able to describe both, socioeconomic-, and technological information systems with the same analytical language. This will not only provide theoretical coherence to describe sociotechnological systems and their interplay, but will surely also deepen our understanding of the role of information systems in the knowledge economy. been reported for individual bakeries that fine-tune their strategy with the help of the binary conditioning random variable rain/sunshine [31]. Fitness depends on what is known about the uncertain future environment. That is the basic argument of this article, and equations (1.1) – (2.4) provide a formal way of how to present this logic in terms of information theory. It is straightforward to close the argument by reconnecting this logic back to Kolmogorov’s notion of algorithmic information: if we have a deterministic algorithm about the series of sunny and rainy days, we can optimize long-term growth by simply allocating economic resources to match the algorithmic sequence of the identified environmental pattern. Deterministic knowledge is the extreme form of probabilistic information. In possession of deterministic knowledge, there is no uncertainty. Going back to equation (1.2), this implies that all uncertainty can be resolved immediately. Instead of waiting for evolution to produce negentropy through natural selection, planned intervention can shift all required resources accordingly and optimize fitness through ‘knowledgeable intervention’. In the presence of knowledge, there is no uncertainty. In other words, knowledge and information are a quantifiable ingredient of economic growth. To achieve this superior growth rates, we can use deterministic knowledge, or at least probabilistic information about the environment. By obtaining additional information about the future state of the environment (from past experience or from cues that correlate with the future), the achievable growth rate can even be increased on average. Shannon’s probabilistic measure of information emerged naturally from our equations, which provides a very concrete quantification of the contributions of information to economic growth. The amount of information about the future is exactly equivalent to the achievable increase in growth rate. Since Shannon’s measure of information converges with Kolmogorov’s algorithmic measure of information, this means that the algorithmic/ deterministic “law” that describes the unfolding of the environment. In this case we “know” something about the dynamical system that governs the environment and can predict it into the far future. This algorithmic knowledge about the unfolding future reduces uncertainty and allows for optimizing the allocation of resources to extract work from the environment and grow. Ergo, knowledge increases fitness. The same logic is at the heart of the argument that finally exorcised Maxwell’s notorious demon by the hands of Zurek [8]. Either informational observations or deterministic knowledge about the dynamical system allows the demon to extract work from the 6. References [1] 1.Pierce, J. R. An Introduction to Information Theory. (Dover Publications, 1980). [2] Kolmogorov, A. N. Three approaches to the quantitative definition of information. International Journal of Computer Mathematics 2, 157–168 (1968). [3] Li, M. & Vitanyi, P. An Introduction to Kolmogorov Complexity and Its Applications. (Springer, 1997). [4] Shannon, C. A Mathematical Theory of Communication. Bell System Technical Journal 27, 379–423, 623–656 (1948). [5] Cover, T. M. & Thomas, J. A. Elements of Information Theory. (Wiley-Interscience, 2006). [6] Zvonkin, A. K. & Levin, L. A. The complexity of finite objects and the development of the concepts of information and randomness by means of the theory of algorithms. Russian Mathematics Surveys (Uspekhi Mat. Nauk) 25, 83– 124 (1970). 3848 [7] Leung-Yan-Cheong, S. & Cover, T. Some equivalences between Shannon entropy and Kolmogorov complexity. Information Theory, IEEE Transactions on 24, 331–338 (1978). [22] Hilbert, M. Evolutionary Fitness as Emergent Multilevel Complexity. under review (2014). [23] Schrödinger, E. What is life? Mind and Matter. (Cambridge University Press, 1944). [8] Zurek, W. H. Algorithmic randomness and physical entropy. Phys. Rev. A 40, 4731 (1989). [24] Brillouin, L. The Negentropy Principle of Information. Journal of Applied Physics 24, 1152–1163 (1953). [9] Caves, C. in Complexity, Entropy and the Physics of Information (Zurek, W. H.) 91–115 (Westview Press, 1990). [25] Kelly, J. . A new interpretation of information rate. Bell System Technical Journal 35, 917–926 (1956). [10] Zurek, W. H. Complexity, Entropy and the Physics of Information. (Westview Press, 1990). [26] Donaldson-Matasci, M. C., Bergstrom, C. T. & Lachmann, M. The fitness value of information. Oikos 119, 219–230 (2010). [11] Baeyer, H. C. V. Warmth Disperses and Time Passes: The History of Heat. (Modern Library, 1999). [27] Rivoire, O. & Leibler, S. The Value of Information for Populations in Varying Environments. Journal of Statistical Physics 142, 1124–1166 (2011). [12] Kolmogorov, A. Interpolation und Extrapolation von stationären zufälligen Folgen. Bulletin ofthe Academy of Sciences of the USSR Series on Mathematics, 3–14 (1941). [28] Permuter, H. H., Kim, Y.-H. & Weissman, T. Interpretations of Directed Information in Portfolio Theory, Data Compression, and Hypothesis Testing. IEEE Transactions on Information Theory 57, 3248 –3259 (2011). [13] Solomonoff, R. J. A formal theory of inductive inference. Part I,. Information and Control 7, 1–22 (1964). [14] Chaitin, G. J. On the Length of Programs for Computing Finite Binary Sequences. J. ACM 13, 547–569 (1966). [29] Mayer-Schönberger, V. & Cukier, K. Big data: a revolution that will transform how we live, work and think. (John Murray, 2013). [15] Nelson, R. R. & Winter, S. G. An Evolutionary Theory of Economic Change. (Belknap Press of Harvard University Press, 1985). [30] Bergstrom, C. T. & Lachmann, M. The fitness value of information. arXiv:q-bio/0510007 (2005). at <http://arxiv.org/abs/q-bio/0510007> [16] Tversky, A. & Kahneman, D. Judgment under Uncertainty: Heuristics and Biases. Science 185, 1124 –1131 (1974). [31] Christensen, B. Smarter Analytics: Der Bäcker und das Wetter [the baker and the weather]. (2012). at <https://www.youtube.com/watch?v=dj5iWD2TVcM> [17] Boyd, R. & Richerson, P. J. The Origin and Evolution of Cultures. (Oxford University Press, USA, 2005). [32] Popper, K. R. Objective Knowledge: An Evolutionary Approach. (Oxford University Press, 1972). [18] North, D. C. Structure and Change in Economic History. (W. W. Norton & Company, 1982). [33] Cooke, P. Knowledge Economics: Clusters, Learning and Cooperative Advantage. (Routledge, 2002). [19] Powell, W. W. & DiMaggio, P. J. The New Institutionalism in Organizational Analysis. (University Of Chicago Press, 1991). [34] Nonaka, I. & Takeuchi, H. The Knowledge-creating Company: How Japanese Companies Create the Dynamics of Innovation. (Oxford University Press, 1995). [20] Polanyi, M. The Tacit Dimension. (University of Chicago Press, 1966). [35] UNSD (United Nations Statistics Division). UN COMTRADE database. (United Nations Statistics Division, 2012). at <http://comtrade.un.org/> [21] Hilbert, M. Linking Information, Knowledge and Evolutionary Growth: A multilevel interplay between natural selection and informed intervention. presented at ‘Advances in Dynamic Economics and Development’, Nov. 2013. at <http://www.economia.ufpr.br/Eventos/Complexity2013/Hil bert%20%20Linking%20information%20,knowlege%20and%20evol utionary%20Growth.pdf> [36] Feenstra, R. C., Lipsey, R. E., Deng, H., Ma, A. C. & Mo, H. World Trade Flows: 1962-2000. (NBER, 2005). at <http://www.nber.org/papers/w11040> [37] Hidalgo, C. A. & Hausmann, R. The building blocks of economic complexity. Proceedings of the National Academy of Sciences 106, 10570 –10575 (2009). 3849
© Copyright 2025 Paperzz