A Bayesian Model for Sales Forecasting at Sun Microsystems Phillip M. Yelland, Shinji Kim, Renée Stratulate1 Sun Microsystems Laboratories, MS UMPK16-158, 16 Network Circle, Menlo Park, U.S.A. Abstract An accurate short-term forecast of product sales is vital for the smooth operation of modern supply chains, especially where the manufacture of complex products is outsourced internationally. As a vendor of enterprise computing products whose business model has long emphasized extensive outsourcing, Sun Microsystems Inc. has a keen interest in the accuracy of its product sales forecasts. Historically, the company has relied on a judgment-based forecasting process, involving its direct sales force, marketing management, channel partners, and so on. Management recognized, however, the need to address the many heuristic and organizational distortions to which judgment-based forecasting procedures are prey. Simply replacing the judgmental forecasts by statistical methods with no judgmental input was unrealistic; short product life cycles and volatile demand demonstrably confounded purely statistical approaches. This article documents a forecasting system developed in Sun’s research laboratory and currently deployed by the company that uses Bayesian methods to combine both judgmental and statistical information. We discuss the development and architecture of the system, including steps that were taken to ease its incorporation into the company’s existing forecasting and planning processes. We also present an evaluation of the system’s forecasting performance, as well as possible directions for future development. Key words: Sales forecast, Bayesian statistics, prior elicitation, Markov chain Monte Carlo Email address: [email protected], [email protected], [email protected]. 1 The authors gratefully acknowledge the invaluable contributions of Dr. Eunice Lee (Sun Worldwide Operations) to the work described here. 1 A Bayesian Model for Sales Forecasting at Sun Microsystems Introduction Sun Microsystems Founded in 1982, and currently headquartered in Santa Clara, California, Sun Microsystems Inc. is a premier supplier of enterprise computing products, employing some 33,500 people worldwide. In the fiscal year ending in June 2008, its revenues amounted to approximately $14 billion. Sun’s sales offerings range from microprocessors to IT services, but the bulk of its income is derived from sales of computer servers and storage systems, which range in price from below $1,000 to more than $10 million each. 2 Throughout Sun’s history, the company’s business model has been distinguished by a very high degree of outsourcing, with a heavy reliance on contract manufacturers and resellers. Baldwin and Clark (1997) relate that on the whole, this strategy has proven very advantageous to Sun, and indeed the past few years have witnessed the adoption of similar approaches by the company’s competitors (Davis 2007, Dean 2007). It does, however, present significant challenges for supply chain management—challenges that are compounded by the character of computer products themselves, which have lifecycles measured in months or even weeks, and parts that can depreciate catastrophically in inventory. Accordingly, Sun has devoted considerable effort to the engineering of its supply chain, as exemplified by its recently-implemented One Touch program, which involves extensive use of drop shipment and cross-docking (Whiting 2006). Nonetheless, as articles such as (de Kok, Janssen, van Doremalen, van Wachem, Clerkx, and Peters 2005) recount, the effectiveness of even the most agile and well-coordinated supply chain rests on accurate forecasts of demand—in fact in a survey documented by Wisner and Stanley (1994), managers of lean supply chains like Sun’s placed a greater reliance on demand forecasts than did those operating more traditional supply chains. Forecasting at Sun Sun’s supply chain forecasting process, in common with many of those in industry, revolves around predictions for product sales over a quarterly time horizons—generally the current and next two fiscal quarters, in Sun’s case. These numbers serve as input to planning and procurement processes. The quarterly sales of a small selection of Sun’s server and storage products are displayed in figure 1. 3 2 All information current as of February 2009—c.f. (SMCI). In the interests of confidentiality, sales data appearing in the figure are mildly disguised. A label such as on the horizontal axis of a graph in the figure denotes the third financial quarter of the second year of the product’s life. 3 2Q3 2 8000 30 6000 20 4000 10 2000 0 0 Figure 1. Sample product demands (units) 3 6Q1 40 Q4 10000 Q3 12000 50 Q2 60 5Q1 0 Q4 150 Q3 200 Q2 300 Q4 Q3 6Q1 Q3 5Q1 Q3 4Q1 Q3 3Q1 0 4Q1 500 2Q1 2000 Q3 1Q1 Q3 Q4 2Q1 Q2 Q3 Q4 3Q1 Q2 Q3 Q4 4Q1 Q2 Q3 Q4 5Q1 Q2 Q3 Q4 6Q1 2500 Q3 Q4 2Q1 Q2 Q3 Q4 3Q1 Q2 Q3 Q4 4Q1 Q2 Q3 Q4 5Q1 Q2 Q3 Q4 6Q1 Q2 Q3 Q4 2Q1 Q2 Q3 Q4 3Q1 Q2 Q3 Q4 4Q1 Q2 Q3 Q4 5Q1 Q2 Q3 Q4 6Q1 3000 Q2 3Q1 Q3 Q4 2Q1 Q2 Q3 Q4 3Q1 Q2 Q3 Q4 4Q1 Q2 Q3 Q4 5Q1 Q2 Q3 Q4 6Q1 A Bayesian Model for Sales Forecasting at Sun Microsystems 2500 2000 1500 1500 1000 1000 500 0 250 8000 6000 100 4000 50 2000 0 A Bayesian Model for Sales Forecasting at Sun Microsystems Before the rollout of the program described in this paper, Sun’s sales forecasting rested heavily on the production of so-called judgmental forecasts (Lawrence, O’Connor, and Edmundson 2000), assembled by polling the direct sales force and channel representatives and combining the results in committee with input from marketing, sales and senior executives. Sun is hardly alone in such a practice; judgment has long formed the basis of commercial sales forecasting. It figures prominently, for example, in an instructional article written in the 1950s by Boulden (1958), and surveys published in recent decades— including (Dalrymple 1987, Wisner and Stanley 1994, Sanders and Manrodt 1994, Klassen and Flores 2001, Sanders and Manrodt 2003, Fildes, Goodwin, Lawrence, and Nikolopoulos 2009)—attest to the continuing prominence of judgment in the forecasting processes of the vast majority of companies. According to Sanders and Manrodt (2003), for example, of 240 US corporations surveyed, almost 90% based their sales forecasts exclusively on judgment, and of those who used quantitative techniques, some 60% routinely made judgment-based adjustments to the forecasts produced. Certainly, some of the reasons for the continued widespread use of judgment in industry have little to do with improved forecast accuracy: Sanders and Manrodt (1994) cite a lack of acquaintanceship with mathematical techniques and a desire to conflate forecasting with organizational goal-setting. But Lawrence, Goodwin, O’Connor, and Önkal (2006) note that of late, the research community has come to recognize that judgment has an important and technically legitimate role to play in forecasting. This is especially true— as Sanders and Ritzman (2001) observe—in highly dynamic environments such as Sun’s, where structural changes may undermine a statistical forecasting model. Moon, Mentzer, and Smith (2003) insist that both judgment-based and statistical elements are indispensable to the sales forecasting process in particular—a point illustrated vividly in the episode related by Worthen (2003), when Nike Corporation was compelled to take massive writeoffs against inventory, thanks to an exclusive reliance on statistical forecasts. Our experience at Sun also led us to conclude that judgmental input was vital, in that exclusively statistical methods that dispensed with judgmental inputs were found to be largely ineffective at forecasting company product sales. There is, however, a wealth of literature documenting the biases and errors associated with judgment-based forecasting—c.f. (McGlothlin 1956, Tversky and Kahneman 1974, Wright and Ayton 1986) or (Bolger and Harvey 1998), for example. Mentzer and Bienstock (1998) and Tyebjee (1987) point out that in addition to these problems, judgmental sales forecasts may be distorted by other factors, such as organizational pressures. That commercial sales forecasts are indeed subject to distortion is confirmed by recent survey data from Fildes et al. (2009). Mindful of this, supply chain managers in Sun’s “Worldwide Operations” business unit looked for statistical techniques that could be used to enhance—but 4 A Bayesian Model for Sales Forecasting at Sun Microsystems not supplant—the company’s judgmental forecasts, and engaged the research unit of Sun (“Sun Labs”) to this end. The ensuing joint project culminated in the Sun Labs Forecasting System, the suite of software described in this paper. Related Work Many researchers and practitioners have sought to combine judgmental and statistical forecasting techniques. Broadly speaking, such efforts may be categorized under the headings: (1) adjustment, (2) combination, (3) correction, (4) judgmental bootstrapping and (5) Bayesian methods. Work in each of these categories is reviewed briefly below. Adjustment Articles such as (Sanders and Manrodt 1994, Fildes and Goodwin 2007, Franses and Legerstee 2009) and (Fildes et al. 2009) evidence the widespread manual adjustment of statistical forecasts ex post facto—frequently, by “eyeball” analysis of forecast graphs, as described in (Bunn and Wright 1991) and (Webby and O’Connor 1996). Without doubt, therefore, this approach has intuitive appeal. However, Armstrong and Collopy (1998) point out that if adjustments are carried out in an unstructured or undisciplined fashion, this approach risks simply reintroducing the distortions of judgmental forecasting. In fact the accounts of Fildes and Goodwin, Franses and Legerstee, and Fildes et al. (ibid.) suggest that more often than not in practice, adjustments reduce the accuracy of a statistical forecast. Bunn and Wright (1991) also aver that without sufficient ancillary detail, it can be difficult to see how an apparently arbitrary adjustment was justified, leading to possible contention within the organization employing the forecast. Combination An alternative approach—discussed by Blattberg and Hoch (1990), Webby and O’Connor (1996), Lawrence et al. (2006) and Franses (2008)—is to originate statistical and judgmental forecasts independently and use a mechanical procedure to combine them. This builds upon work on forecast combination which began with (Bates and Granger 1969), and which is surveyed by Clemen (1989) and Timmermann (2006). Combination is generally effected by taking a weighted sum of the values of the constituent forecasts in each period. A number of methods have been proposed to estimate the combination weights: Granger and Ramanathan (1984) suggest least squares regression, for example, whereas Bates and Granger (1969) compute weights based on the forecast errors of the constituent forecasts, 5 A Bayesian Model for Sales Forecasting at Sun Microsystems and Diebold and Pauly (1990) use Bayesian shrinkage regression. In practice, a simple unweighted average of the component forecasts—as expounded by Blattberg and Hoch (1990) and Armstrong (2001b) amongst others—has been found to perform consistently well. Timmermann (2006) lists a number of factors that recommend forecast combination: It synthesizes the information sets used to produce the component forecasts, it dilutes bias in individual forecasts and it increases robustness with respect to model misspecification and structural breaks. Set against this, Timmermann also notes that estimated combination weights can be very unstable in practice—which helps explain the remarkably good relative performance of the simple average. In principle, avers Diebold (1989), where the information sets underlying the component forecasts are available, 4 it is always preferable to construct a single encompassing forecast model, rather than simply to combine the forecasts themselves. Correction Rather than combining a statistical and a judgmental forecast, some authors—including Theil (1971), Ahlburg (1984), Moriarty (1985), Elgers, May, and Murray (1995) and Goodwin (1996, 2000)—have explored statistical methods for correcting judgmental forecasts in the light of observed outcomes. Generally, such methods are based on Theil’s (ibid.) “optimal linear correction,” which involves regressing observed outcomes on forecasts, using the estimated regression coefficients to produce a revised prediction from new forecasts; Goodwin (1997) accommodates time-varying coefficients using a weighted regression. A technique related to statistical correction originates with Lindley (1983), and is applied explicitly to time series forecasting by West and Harrison (1997, sec. 16.3.2). Lindley’s methodology is an example the so-called supra-Bayesian approach to the reconciliation of expert opinion developed by Pankoff and Roberts (1968) and Morris (1974, 1983). Here, the value of a judgmental forecast is construed as a linear function of the actual value, and Bayesian updating is used to produce a revised forecast. Judgmental bootstrapping Researchers—particularly in psychology—have long sought to capture judgmental reasoning in a tractable mathematical form. Efforts centered on linear models date back at 4 Such a set might comprise a judgmental bootstrap model of the forecaster in the case of a judgmentbased forecast—see later. 6 A Bayesian Model for Sales Forecasting at Sun Microsystems least to (Hughes 1917); important contributions were made by Meehl (1957), Hammond, Hursch, and Todd (1964), Hursch, Hammond, and Hursch (1964) and Tucker (1964). Surveying this work, Dawes (1971) coined the term bootstrapping to describe the process by which an expert’s judgment is modeled by a linear expression involving the environmental factors (usually referred to as cues) that enter into the expert’s consideration. Recent authors exploring the application of such a process to judgment-based forecasting, including as Armstrong (2001a), O’Connor, Remus, and Lim (2005) and Batchelor and Kwan (2007) use the qualified term judgmental bootstrapping to avoid confusion with the (quite distinct) statistical bootstrap technique developed in the late 1970s by Efron (1979). 5 Armstrong (ibid.) also applies the term to models that go beyond simple linear combination of cues. Evidence for the efficacy of judgmental bootstrapping in forecasting is mixed: Ashton, Ashton, and Davis (1994) find a bootstrap model out-performed by a statistical forecasting model, and Åstebro and Elhedhli (2006) and Batchelor and Kwan (2007) cannot conclude that a bootstrap model forecasts more accurately than the experts it represents. Lawrence and O’Connor (1996) and Fildes et al. (2009) assert that bootstrapping is less effective in the context of time series extrapolation, where cue information tends to be autocorrelated; the “error bootstrapping” technique developed in response by Fildes (1991) seeks to model the errors in a judgmental forecast, much like the forecast correction approach described above. Bayesian methods The Bayesian paradigm of statistical inference (see e.g. (Gelman, Carlin, Stern, and Rubin 2003) for an overview), with its incorporation of subjective information in the form of prior distributions, seems a natural means of combining judgmental and statistical elements in forecasting. Indeed, a substantial number of Bayesian models have been devised for product demand, both at the inventory level (Silver 1965, Hill 1997, Dolgui and Pashkevich 2008a,b), and in the aggregate (Lenk and Rao 1990, Montgomery 1997, Moe and Fader 2002, Neelamegham and Chintagunta 1999, van Heerde, Mela, and Manchanda 2004, Neelamegham and Chintagunta 2004, Lee, Lee, and Kim 2008)—amongst many others. Sales applications also feature prominently in the seminal work of Pole, West, and Harrison (1994) and West and Harrison (1997) on Bayesian forecasting. Bunn and Wright (1991) note that despite the apparent attractions of Bayesian modeling, and a research literature that dates back several decades, there is a dearth of Bayesian 5 Alternatives terms, such as “the statistical approach,” “actuarial modeling” (both of which appear in the clinical literature) or the (mildly grandiloquent) “paramorphic representation” of Hoffman (1960) would doubtless invite even greater confusion. 7 A Bayesian Model for Sales Forecasting at Sun Microsystems models with judgmental priors used routinely in forecast applications. Bunn and Wright suggest that the chief impediment is the expense (in time and effort) of repeatedly eliciting subjective priors of sufficient quality—a point reinforced by researchers such as Wright and Ayton (1986), who highlight the difficulties involved in obtaining reliable judgmental priors. In fact many of the models cited in the previous paragraph circumvent the need for informative priors by relying on a hierarchical (Gelman and Hill 2006) structure to pool information from analogous historical situations in order to produce forecasts. System Overview As the title of the paper suggests, we began the Sun Labs Forecasting System with a Bayesian model of product sales. Mindful of the need to incorporate judgmental information, we built a model (described in detail in the next section) intended to formalize the judgmental framework used by the company’s forecasters. The model included parameters specifying the level of sales achieved by a product in its mature phase, the time required to achieve maturity after launch, and so on. Our intention was to elicit priors for these parameters from the forecasters, using Bayesian updating to reconcile these priors with actual sales and to extrapolate future sales. 6 This system relied on the frequent elicitation of priors from the forecasting staff—ideally, each time they came into possession of new information. We tried to minimize the work required of the forecasters to produce the priors, constructing the model so that its parameters were easy to interpret, and implementing a number of graphical tools to assist with estimation. Despite this, we fell afoul of a “Catch-22” (Heller 1961) that affirmed the conjecture of Bunn and Wright highlighted above: Our colleagues were unwilling to invest significant time and effort until the efficacy of the Forecasting System had been proven, but without good priors, the System’s performance was bound to be subpar. Faced with this problem, we attempted first to deploy a hierarchical Bayesian model that drew priors from historical records, 7 combining the statistical forecasts produced by the model with the judgmental forecasts prepared by company forecasters. The heterogeneity of Sun’s product line and rapid changes in its product markets made the performance of the hierarchical model disappointing, however, and combinations of the model’s forecasts with judgmental ones performed inconsistently. The solution eventually adopted for use in the Forecasting System resembles the converse of the judgmental bootstrapping process described in the previous section. For consider 6 7 An early version of the Forecasting System based on this idea is described in (Yelland 2004). Again, see Yelland (ibid.) for details. 8 A Bayesian Model for Sales Forecasting at Sun Microsystems that the company’s forecasters routinely produce judgmental sales forecasts, and that the model used in the Forecasting System embodies the way in which those forecasts are assembled from beliefs about sales at maturity, time to reach maturity and so on. Therefore fitting the model to the judgment-based sales forecasts produces estimates of the forecasters’ that underlie them—“cues” in the parlance of judgmental bootstrapping, or parameter priors, to a Bayesian. In the Forecasting System, these prior estimates are then corrected using a device similar to the linear correction used by Theil and his successors: Using historical records of forecasts and actual sales, parameter estimates derived from actual sales are regressed on estimates for the same parameters derived from forecasts of those sales. The coefficients of the regression are used to produce adjusted prior estimates from the new judgmental forecast. Finally, the adjusted prior estimates are revised by Bayesian updating in the light of actual sales to produce the System’s forecast. A functional schematic of the Sun Labs Forecasting System is given in figure 2. Details of the forecasting model and a full description of prior estimation and correction are provided in the next section. System Details Model Dynamic Linear Model From its inception, the Sun Labs Forecasting System centered on the use of dynamic linear models (DLMs), as described by West and Harrison (1997) (hereinafter abbreviated “W&H”). Bayesian formulations of the structural time series models also discussed from a classical statistical viewpoint by Harvey (1989), 8 DLMs have been employed for sales forecasting by a number of authors, including W&H themselves, van Heerde et al. (2004) and Neelamegham and Chintagunta (2004). For reference, the general formulation of a DLM representing a scalar time series (W&H, p. 102) is presented below: Observation equation: Evolution equation: Initial information: yt = Ft> θt + νt , θt = Gt θt−1 + ωt , θ0 ∼ N(m0 , C0 ). 8 νt ∼ N(0, Vt ), (1a) ωt ∼ N(0, Wt ), (1b) (1c) The more recent account of Durbin and Koopman (2001) also concentrates on a classical treatment of structural models, with some attention is given to the Bayesian perspective. 9 A Bayesian Model for Sales Forecasting at Sun Microsystems Processing step Prior deduction Data Sales force forecasts Initial parameter estimates Forecasting model Parameter correction Historical records Parameter priors Bayesian updating Actual sales Revised forecast Figure 2. System schematic Here, observation equation (1a) takes the form of a regression of the time series value, yt (which in the case of the Forecasting System is quarterly product sales), on (the transpose of) a design vector, Ft , of known covariates. The coefficients in this regression, θt , may vary over time; they comprise the state vector of the model. Changes in the state vector itself are described by the evolution equation (1b): In each time period, the state is subject to a linear transformation and randomly perturbed by the addition of the noise component ωt . 9 The specification is completed by the provision of a multivariate normal prior for the state vector in the period preceding that of the first observation. 9 Noise terms νt and ωt are assumed to be uncorrelated. 10 A Bayesian Model for Sales Forecasting at Sun Microsystems The DLM framework is extraordinarily flexible: Harvey, West and Harrison, Durbin and Koopman (ibid.) demonstrate the encoding of trend, seasonal and dynamic regression components—singly or in combination—as DLMs, as well as providing DLM versions of the familiar ARIMA models (Box, Jenkins, and Reinsel 1994). As Harvey points out, structural time series models are generally more intuitively accessible than alternatives such as the ARIMA models—an important consideration in view of our original intention to parallel the reasoning of Company forecasters. Unfortunately, we struggled to represent the important characteristics of Sun’s product demands directly in DLM form; in particular, we found DLMs ill-suited to the distinctive life cycle curve illustrated in figure 1, and the multiplicative seasonality evinced by many products (the DLM formulation provides only for additive seasonality). Obvious devices, such as log transformations, produced forecasts that were far from satisfactory, and elaborations of the basic DLM such as “mixture” or linearized models (W&H, chp. 12, 13, resp.) quickly became unwieldy. 10 The model currently used in the Sun Labs Forecasting System is actually derived from the most basic form of DLM, called the first-order polynomial model by West and Harrison (p. 32), and the random walk plus noise model by Harvey (ibid. p. 19). 11 The model is defined as follows: yt = S(t)L(t)µt + νt , νt ∼ N(0, V ), (2a) µ t = µ t −1 + ω t , ωt ∼ N(0, W ), (2b) µ0 ∼ N(m0 , C0 ). (2c) Here the state vector consists of a single scalar quantity, µt —usually referred to as the level of the process—which evolves according to the random walk set out in equation (2b). In a further simplification, the variances of the noise terms in equations (2a) and (2b) are constant over time. The only departure from the DLM framework is in the (scalar) quantity corresponding to the design vector Ft in equation (1a); instead of the known (if possibly time-varying) values assumed in the usual DLM, the design vector in this model is determined by the product of two stochastic quantities, S(t) and L(t), which represent respectively the effects of seasonality and life cycle stage in the period t. These are described in further detail below. 10 See (Yelland and Lee 2003) for a description of some of our early investigations. Incidentally, West and Harrison (loc. cit.) indicate that the first-order polynomial model is widely used for forecasting in inventory and operations management applications like that of the Forecasting System. 11 11 A Bayesian Model for Sales Forecasting at Sun Microsystems Seasonality Seasonality is described straightforwardly using dummy variables (see Ghysels, Osborn, and Rodrigues (2006), for example). Parameters Q1t , Q2t and Q3t give the size of the multiplicative changes in the respective quarters of the financial year, relative to sales in the fourth quarter of the year: S(t) = 1 + Q1t ϕ1 + Q2t ϕ2 + Q3t ϕ3 , where Q jt = 1 if t is the jth quarter of a year, 0 otherwise. Life cycle Mirroring the conceptual framework generally used by Company forecasters, the model divides a product’s life cycle into three phases: (1) a ramp up phase, in which product sales increase consistently period over period, (2) a mature phase, in which product sales are roughly constant and (3) a ramp down phase, in which product sales dwindle again. In the model, lengths of these phases are delimited by the three positive random variables τ1 , τ2 and τ3 , respectively. The value of the life cycle coefficient varies between 0 and 1, the latter corresponding to sales in maturity. The ramp up and down phases are characterized by geometric growth and decay, defined by the positive constants α and β. 12 For the sake of identification, β is set (arbitrarily) to 0.05, so that sales dwindle to 5% of mature sales during the ramp down. In symbols, the life cycle coefficient is defined as follows: (τ1 −t)/τ1 α L(t) = 1 (t−τ1 −τ2 )/τ3 β if 0 ≤ t < τ1 , if τ1 ≤ t < τ1 + τ2 , if τ1 + τ2 < t. Figure 3 illustrates the general shape of the life cycle curve in the model, and Table 1 summarizes its interpretation. Formal modeling of product life cycles has a long history, dating back (at least) to the work of Rogers (2003, first published in 1962) and Bass (1969); see (Mahajan, Muller, and Wind 2000a) for a compendium of work in the area. The vast majority of models of this 12 It is generally the case that α and β are less than 1. Values of α and β in excess of 1 are permissible, however, though in such instances the value of L(t) may exceed 1 in some periods, and the epithets “ramp up” and “ramp down” are probably inappropriate. 12 A Bayesian Model for Sales Forecasting at Sun Microsystems L(t) 1 α Ramp up Ramp down Maturity β 0 τ1 0 τ1 + τ2 τ1 + τ2 + τ3 t Figure 3. Life cycle coefficient over time Life cycle phase Span Behavior Ramp up 0 ≤ t < τ1 Geometric growth from a value α when t = 0 to 1 as t → τ1 . Mature τ1 ≤ t < τ1 + τ2 Constant value 1. Ramp down τ1 + τ2 < t Geometric decay from 1 when t = τ1 + τ2 to β as t → τ1 + τ2 + τ3 . Table 1. Interpretation of the life cycle coefficient sort are structural in nature; product demand emerges endogenously as a result of consumer responses to the experience of other consumers and to the marketing actions of the producing firm. By contrast, the life cycle representation used in the Sun Labs Forecasting System is a purely empirical account of what Mahajan, Muller, and Wind (2000b) refer to as strategy-driven diffusion—the life cycles of the products modeled by the Forecasting System do not result from consumer actions, but from strategic and operational 13 A Bayesian Model for Sales Forecasting at Sun Microsystems decisions taken by Sun, its suppliers and its distributors. In particular, the ramp up phase in the model reflects product quality assurance processes and supply chain inertia, and the ramp down phase is initiated not by market saturation, as in the Bass model and its successors, but by a decision on the part of the Company to withdraw the product in the face of technological development. We should note that of late, mainstream work in the area of life cycle modeling has employed piecewise representations like that used in the Forecasting System—see (Niu 2006, Huang and Tzeng 2008) and (Sood, James, and Tellis 2009). Prior Estimation Regarding prior estimation, assume that a judgment-based forecast has been produced at the end of quarter t a product’s life cycle for sales of the product in quarters t + 1, . . . , t + h. Let ϑ ∈ {α, β, τ1 , τ2 , τ3 , ϕ1 , ϕ2 , ϕ3 , µ0 } 13 denote a model parameter for which a prior is required. The prior estimation procedure is as follows: (1) Prepend actual sales (y1 , . . . , yt ) to the forecast, producing a combined series ŷ of length t + h. (2) Using a non-linear least-squares regression, fit to the series ŷ the model (listing explicitly all the parameters that enter into S(t) and L(t) ): yt = S(t; ϕ1 , ϕ2 , ϕ3 )L(t; α, τ1 , τ2 , τ3 )µ0 + νt . (3) Let ϑˆ + be the estimated value of ϑ in the above regression. The model in equation (3) is a restriction of the full model in (2), without the level evolution specified in equation (2b). In practice, we found the restriction necessary to ensure reliable fitting of judgment-based forecasts—especially early in a product’s life, where few actual demand points are available to supplement the forecast demands. The quality of the resulting priors was not substantially compromised. (3) Collect together a set ŷ1 , . . . , ŷ N of combined forecasts and actuals for a set of recentlyobsolete products similar to the current one, 14 where the forecasts were made in the corresponding quarter t of the products’ lives. Since these products are obsolete, their entire sales histories y1 , . . . , yN are also available. (4) Fit the model in equation (3) (using non-linear least squares again) to the series collections ŷ1 , . . . , ŷ N and y1 , . . . , yN . For i ∈ 1, . . . , N, let ϑˆ i be the estimate of the parameter of interest that results from fitting to ŷi , with ϑi the estimate from the corresponding 13 The prior for µ0 actually constitutes the values m0 and C0 in (2c). Sun’s products are generally categorized by market, such as “enterprise storage”, “low-end server”, etc., and this categorization is used to identify like products. 14 14 A Bayesian Model for Sales Forecasting at Sun Microsystems series yi . Thus ϑi represents the parameter value that characterizes the actual trajectory of the ith product’s sales, whereas ϑˆ i is the value suggested by the forecast in the tth quarter of its life cycle. (5) Fit a linear regression ϑi = a + bϑˆ i + υi to the estimates from step 4, 15 yielding coefficient estimates â and b̂ and an estimate of the residual standard deviation ς̂. (6) The prior for ϑ derived from the value ϑˆ + estimated for the new forecast in step 2 is given by standard results for classical normal regression (e.g. Gelman and Hill 2006, p. 48)—i.e., a normal distribution with mean a + bϑˆ i and standard deviation: # 12 1 (ϑˆ + − mean ϑˆ i )2 ς̂ 1 + . + N ∑iN=1 (ϑˆ i − mean ϑˆ i )2 (7) The set of priors required for the model is completed by priors for the observation and evolution noise variances in equation (2a) and equation (2b), respectively. Following a rule of thumb we have found to work well in practice, these are set to relatively diffuse scaled inverse χ2 distributions (see Gelman et al. 2003, p. 480) Inv–χ2 (4, 0.2 × m0 ) and Inv–χ2 (4, 0.1 × m0 ), resp., where m0 is the prior mean for µ0 . " Bayesian Updating The Bayesian updating step in figure 2 (which revises the forecast priors in light of observed sales to date) is carried out using a Gibbs sampling routine, schematic descriptions of which now abound in the literature—see (Gilks, Richardson, and Spiegelhalter 1996, chp. 1), for example. The individual steps of this particular sampler are described in the Appendix. Many of the steps rely on standard results concerning conjugate updating in Bayesian analysis, which may be found in reference texts such as (Gelman et al. 2003) or (Bernardo and Smith 1994). Where such closed-form updates are not available, we resort to Metropolis-Hastings sampling (also discussed by Gilks et al.); proposals are generated using Geweke and Tanizaki’s (2003)’s Taylored chain procedure, details of which are also provided in the Appendix. The Gibbs sampler is run for 10,000 iterations, with the first 3,000 samples discarded; convergence is verified using Geweke’s (1992) single chain diagnostic (in additional to occasional sample-path plots). In fact execution of the sampler for 10,000 iterations is quite conservative—we are almost invariably able to establish convergence by about 4,000 iterations. 15 For parameters restricted to the positive half-line, the linear regression is fit on a log scale to ensure coherent predicted parameter values. The resulting estimates are actually used to derive truncated (not log-) normal distributions on the original scale (by matching first and second moments), as we have found that truncated normal priors make for far better forecast performance than log-normal ones. 15 A Bayesian Model for Sales Forecasting at Sun Microsystems Forecast Performance To demonstrate the efficacy of the Sun Labs Forecasting System, we compare its forecast performance with that of alternative forecasting methods, using the demand histories of a sample of Sun’s products. Setup The test is conducted using a collection of 32 products, representing a cross-section of Sun’s recent product lines. Of these products, 27 are randomly selected for calibration, and holdout forecasting is performed on the remaining 5. 16 Point forecasts at horizons of 1, 2 and 3 quarters are prepared for each quarter of the holdout products’ demand histories, yielding some 77 forecast values in total for each method at each horizon. Four forecast methods are compared in the test: Sys This forecasting method relies on posterior predictive distributions calculated by the Forecasting System in the Bayesian updating step of figure 2. 17 With calibration data (i.e. demand histories and judgmental forecasts) available at time t collectively denoted Dt , the posterior predictive distribution for the forecast value yt+h at horizon h is given by p(yt+h | Dt ), which is itself calculated by marginalizing over the posterior density for the model parameters (here represented by ϑ): p ( y t + h | Dt ) = ∝ Z Z p(yt+h |ϑ ) p(ϑ | Dt )dϑ (4) p(yt+h |ϑ ) p( Dt |ϑ ) p(ϑ )dϑ. (5) The integral expression in equation (5) is readily approximated using the Gibbs sampler specified in the Appendix, as described by Albert (2008), for example. For the point forecast ŷt+h|t (the forecast for yt+h made in period t), we use the mean of the posterior predictive distribution of yt+h . DLM 16 This method is a straightforward implementation of a univariate dynamic linear model, as set out in (W&H, chp. 5–8). Following West and Harrison’s prescription, the model comprises a second-order polynomial- (“local level with trend”) The relatively large number of products in the calibration collection is necessary to ensure reliable operation of forecast methods , and , which require a representative set of products from each of the company’s product groups in order to correct priors and forecasts. 17 This is the standard approach to forecasting with Bayesian models—see (Neelamegham and Chintagunta 1999, 2004), for instance. Sys DLM CJudg 16 A Bayesian Model for Sales Forecasting at Sun Microsystems and four-period seasonal component. As Gardner and McKenzie (1989) suggest, to improve forecasts at over longer horizons, the trend in the polynomial component is damped, so that it decays over time; the design and evolution matrices necessary to accomplish this are adapted from analogous structures described for the closely-related single source of error (SSOE) structural time series models in (Hyndman, Koehler, Ord, and Snyder 2008, p. 48). Formulae for updating and forecasting with this DLM are standard: We use the “unknown, constant variance” results summarized on (W&H, p. 111), which incorporate the estimation of a time-invariant observation noise variance in equation (1a). Multiple discount factors are used to specify the evolution noise component in equation (1b)—c.f. op. cit. pp. 196–198 for details. Discount and damping factors are derived using a grid search based on forecast accuracy for a sample of the calibration set, and initial conditions are set from the same corrected priors used for the Sys method, using a procedure similar to that described by Hyndman et al. (ibid., sec. 2.6.1). All components in the state are uncorrelated in the prior, so that the matrix C0 in (1c) is diagonal. Judg CJudg This method simply reproduces the company’s judgmental forecast—the forecasts from which the priors for methods Sys and DLM are derived. For this method, the company’s judgmental forecast is “corrected” using Theil’s (1971) optimal linear correction. Recall from the above that this involves regressing actual sales on predicted sales for products in the calibration set, using the estimated coefficients to compute revised prediction from forecasts in the holdout set—see Theil (1971, p. 34), or more recently, Goodwin (2000), for a detailed discussion. Separate regressions are calculated for each forecast horizon and each of the product categories identified for prior correction. Test Results Table 2 summarizes the performance of the candidate methods in the forecasting test. We have used the mean absolute scaled error (MASE) of Hyndman and Koehler (2006) as a performance metric; the Appendix defines the MASE, and sets out the considerations that lead to its adoption. MASEs are given for each method at each forecast horizon—a smaller entry indicates better performance. The rank (smallest to largest) of each method at each horizon is recorded in parentheses after the corresponding entry. As the table shows, the Forecast System consistently exhibits superior overall performance, with the purely statistical DLM method generally turning in the worst performance, and the corrected judgmental forecast also consistently outperforming its uncorrected counterpart. The MASE is 17 A Bayesian Model for Sales Forecasting at Sun Microsystems Horizon Sys DLM Judg CJudg 1 2 3 0.53 (1) 0.84 (1) 0.80 (1) 0.73 (3) 1.13 (4) 0.97 (4) 0.73 (4) 1.10 (3) 0.96 (3) 0.66 (2) 1.00 (2) 0.94 (2) Table 2. Mean absolute scaled error by forecast horizon defined such that a value less than 1.00 indicates a performance better (on average) than that of the “naïve” random-walk forecast, which simply uses the most recent observation to produce a forecast. Thus according to the table, only the Sys method consistently improves on the benchmark. For a more detailed perspective, Figure 4 is a box-and-whisker plot (Tukey 1977) of the distribution of absolute scaled errors of the forecast methods at each forecast horizon. Though the distributions of the ASEs are clearly skewed, the plot broadly confirms the superiority of the Sys method established by the MASEs. Implementation A depiction of the current implementation of the Forecast System is given in figure 5. Remarkably, we found it possible to implement the System almost entirely using open source and/or freely-available software. The core of the System is written in the statistical programming language R (Venables and Smith 2002). It resides on a dedicated server and carries out the following operations: • Calculating priors from judgmental forecasts, • Updating the priors given actual product sales, and • Forming sales predictions from the updated priors. This portion of the System is run once a week to incorporate the latest sales figures. Input to the R code comes from the company’s data warehouses by way of Java interface modules, and output is stored in a dedicated MySQL database (Williams and Lane 2004) on the same server. 18 19 0 1 2 3 Sys ● DLM ● ● Judg ● ● Horizon: 1 0 1 2 Sys ● ● DLM Judg CJudg ● 0 1 2 3 ● ● ● Sys Figure 4. Distribution of ASEs by model and forecast horizon CJudg ● ● 3 Horizon: 2 DLM ● ● ● Judg ● ● ● Horizon: 3 CJudg ● ● ● ● A Bayesian Model for Sales Forecasting at Sun Microsystems A Bayesian Model for Sales Forecasting at Sun Microsystems Manufacturing Actual sales Sun Labs Forecasting System Data warehouse Sales, Forecasts Web interface Revised forecasts Sales, Forecasts Revised forecasts Forecast meeting Supply chain management Forecasts Plans Figure 5. How the Forecast System supports Sun’s supply chain management process Access to the forecast data is Web-based, relying on the Adobe Flash/JavaScript/PHP/Apache combination common in modern Web applications (Wiedermann 2006, Schafer 2005). Figure 6 provides a (mildly disguised) snapshot of the forecast display for a typical product: On the left of the page, the set of forecast products is arrayed by category, while at the top of the page is a selection of the “base week” in which the forecast was produced, allowing the user to review historical records as desired. The animated chart (provided by amCharts: http://www.amcharts.com/) provides an interactive pictorial representation of forecasts and actual sales, which are also presented in tabular form at the bottom of the page. The system also provides forecasts in the form of PDF reports, produced on the server using LATEX (Lamport 1994). 20 A Bayesian Model for Sales Forecasting at Sun Microsystems Figure 6. Snapshot of Forecast System user interface Concluding Remarks The Sun Labs Forecasting System has been in use at Sun for over a year at the time of writing. The System provides an effective combination of judgmental and statistical forecasting information that consistently improves upon the forecast accuracy of both its constituents. The system operates almost entirely unattended, extracting the requisite information from the Company’s data warehouses and delivering forecasts in a regular and expeditious manner through a simple, intuitive user interface. Furthermore, with Bayesian priors derived automatically from the output of the existing sales forecasting process, the Forecasting System imposes no additional procedural burdens on its users. Taken to- 21 A Bayesian Model for Sales Forecasting at Sun Microsystems gether, these factors no doubt account for the warm reception the System has received from the Company’s supply chain managers. Development of the Forecasting System continues. Most immediately, we are investigating means by which forecasters can inform the System of upcoming events such as very large orders or promotions that are likely significantly to perturb the sales of one or more products. The technical mechanisms by which such interventions may be achieved are fairly straightforward to implement—see (West and Harrison 1997, chp. 11) for a discussion. More delicate are the organizational issues involved, for articles such as Sanders and Ritzman (1991), Fildes et al. (2009) and Sanders (2009) testify that although management intervention generally increases forecast accuracy when managers are possessed of specific and accurate information about special events, forecasts are usually degraded if managers are allowed to intervene willy-nilly. Our current intentions are to provide some form of decision support system—along the lines of that described by Lee, Goodwin, Fildes, Nikolopoulos, and Lawrence (2007)—to help forecasters make only considered interventions. The Forecasting System makes no attempt to provide normative information—such as inventory levels or purchase quantities (Zipkin 2000)—based on the forecasts it produces. This is quite deliberate, since the complexity of Sun’s supply chain and the sourcing arrangements upon which it rests makes the derivation of policy recommendations from demand forecasts formidably complex; inventory-, depreciation- and expediting costs, costs of lost sales, contractual supply arrangements, sales of complementary products and many other factors vary from product to product, between different configurations of the same product, and across quarters. The Forecasting System system therefore passes off responsibility for operating decisions to the users of its forecasts. A somewhat vexatious upshot of this is that it is very difficult to place a firm dollar amount on the benefits that accrue from superior accuracy of the System’s forecasts. Certainly, the suggestions of Mentzer (1999) and Kahn (2003) provide means of reaching plausible rough estimates, but a truly definitive figure would require a thorough-going appraisal of the company’s forecasting and supply arrangements along the lines proposed by Moon et al. (2003). Policy prescriptions are easier to make at a less aggregate level of demand: Purchasing and inventory management recommendations are much more tractably computed from a demand forecast for a single manufacturing part. In regard of this, we have begun to apply many of the techniques used in the Forecasting System (such as Bayesian updating and a semi-parametric life cycle representation) to the forecasting of part demands. At this level of disaggregation, priors derived from judgmental forecasts of product demand are frequently unreliable, because a multitude product configurations and frequent substitu- 22 A Bayesian Model for Sales Forecasting at Sun Microsystems tion of manufacturing parts due to changes in technology or sourcing arrangements makes the correspondence between product and part demands too unstable. However, we have found hierarchical priors (Gelman and Hill 2006)—which pool information across sets of parts—to be effective replacements. On a more pragmatic note, we should note that in developing a parts forecasting solution, our dealings with Sun’s operating units have been greatly enhanced by the credibility garnered by the Sun Labs Forecasting System. 23 A Bayesian Model for Sales Forecasting at Sun Microsystems Appendix Gibbs Sampler The following details the Gibbs sampler used for Bayesian updating in the Forecasting System. Each step in the sampler is introduced by the full conditional distribution from which a sample is to be drawn. Variables of which the sampled quantity is conditionally independent are omitted from the conditioning set. µ0 , . . . , µ T yt , λ, ϕ, V, W, m0 , C0 Conditionally on the other parameters of the model, µ0 , . . . , µ T constitutes a sample path of the state vector of a first-order polynomial dynamic linear model. Procedures for sampling the state of a DLM (generally under the moniker forward filtering/backwards sampling algorithms) are described by Frühwirth-Schnatter (1994), Carter and Kohn (1994), West and Harrison (1997) and Durbin and Koopman (2001), amongst others. W µ0 , . . . , µ T Again, a standard step in DLM sampling—c.f. (W&H, p. 568), for example. α τ1 , τ2 , τ3 , yt , µt , ϕ, V The full conditional is proportional to the expression: " T ∏ N(yt |L(t)S(t)µt ,V ) # × N[0,∞) (α|µα , σα2 ). t =1 This is sampled using a Taylored chain proposal in a Metropolis-Hastings step—the calculation of the Taylored chain proposal is described in detail in the following section. ϕ yt , µt , λ, V Draw from the posterior distribution of the coefficients ϕ1 , ϕ2 , ϕ3 in the linear regression: yt − L(t)µt = [L(t)µt Q1t ] ϕ1 + [L(t)µt Q2t ] ϕ2 + [L(t)µt Q3t ] ϕ3 + νt , given independent priors—see e.g. (Gelman et al. 2003, chp. 8). 24 νt ∼ N(0, V ), A Bayesian Model for Sales Forecasting at Sun Microsystems V yt , λ, ϕ, µt The conditional likelihood of V is ∏tT=1 N(yt |S(t)L(t)µt ,V ); with prior V ∼ Inv–χ2 (ν, σ), the posterior is given by conjugacy—see e.g. (Gelman et al. 2003, chp. 2). The Geweke -Tanizaki (2003) 18 Taylored Chain With x the current state of the sampler, to produce a proposal x + for a target kernel p(z), let q(z) = log p(z), with q0 (z) and q00 (z) the first and second derivatives thereof. Proceed by cases: Case 1: q00 ( x ) < −ε, where ε is a suitable small constant, such as 0.1 . 19 Rewrite the Taylor expansion of q(z) around x: 1 q(z) ≈ q( x ) + q0 ( x )(z − x ) + q00 ( x )(z − x )2 2 2 1 q0 ( x ) = q( x ) − (−q00 ( x )) z − x − 00 2 q (x) | {z } ‡ Since q00 ( x ) < 0, the component (‡) of the latter expression constitutes the exponential part of a normal distribution, which implies that the target kernel in the vicinity of x may be approximated by a normal distribution with mean x − q0 ( x )/q00 ( x ) and p standard deviation 1/ −q00 ( x ); sample x + accordingly. Case 2: q0 ( x ) ≥ −ε and q0 ( x ) < 0 Approximate q(z) by a line passing through x and x1∗ , the largest mode of q(z) smaller than x: q( x1∗ ) − q( x ) q(z) ≈ q( x1∗ ) + (z − x1∗ ) x1∗ − x | {z } ‡ 18 Sampling techniques similar to the Taylored chain are also discussed by Qi and Minka (2002). By ensuring that | q00 ( x ) | > 0, using ε rather than 0 reduces the occurrence of proposed values that depart too markedly from the current state. 19 25 A Bayesian Model for Sales Forecasting at Sun Microsystems In this case, the component (‡) indicates an exponential distribution, and the proposal is: 20 x + = x̂1 + w, where w ∼ Exp(λ1 ), q( x1∗ ) − q( x ) , λ1 = x1∗ − x x̂1 = x1∗ − 1/λ1 Case 3: q0 ( x ) ≥ −ε and q0 ( x ) > 0 Approximate q(z) by a line passing through x and x2∗ , the smallest mode of q(z) larger than x. The proposal is developed in a manner parallel to that in Case 2: x + = x̂2 − w, where w ∼ Exp(λ2 ), q( x2∗ ) − q( x ) , λ2 = x2∗ − x x̂2 = x2∗ − 1/λ2 Case 4: q0 ( x ) ≥ −ε and q0 ( x ) = 0 In this instance, x + is sampled from a uniform distribution over a range [ x1 , x2 ], such that x1 < x < x2 . End points x1 and x2 are set to suitable modes of q(·), if they can be found, and to user-supplied values otherwise. Measures of Forecast Accuracy As operationalizations of the quadratic loss criterion that pervades statistical inquiry, the mean square error and its related metric, the root mean square error—we refer to both metrics jointly as the (R)MSE—have long been a staple of academic research in forecasting—see (Granger and Newbold 1973), for example. Unfortunately, though the (root) mean square error is analytically attractive, researchers such as Armstrong and Collopy (1992) point to a number of practical problems with its use: (1) Performance with respect to the (R)MSE may be significantly affected by outliers. Such problems may be ameliorated by eliminating outliers, though this might draw the objectivity of the procedure into question. (2) More seriously, the (R)MSE is inherently scale dependent, in that its magnitude depends not only on forecast accuracy, but also on the level of the underlying series; ceteris paribus, a forecast 10% in excess of an actual value of 1,000,000 will result in a substantially greater (R)MSE than one 10% above an actual of 1,000. This largely invalidates summary The origin of the proposal, x̂1 , is offset from the mode x1∗ in order to guarantee irreducibility of the resulting Markov chain; see (Geweke and Tanizaki 2003) for details. 20 26 A Bayesian Model for Sales Forecasting at Sun Microsystems measures based on the (R)MSE of performance across a heterogeneous collection of series. 21 Since the series in this test described in the paper vary in maximum between 200 and 4000, the (R)MSE is unsuited to this application. The mean absolute percentage error (MAPE) favored by practitioners, which expresses error as a fraction of the associated actual value, avoids the scale-dependency of the (R)MSE. The MAPE has disadvantages of its own, however: (1) In contradistinction to the (R)MSE, summary MAPE measures may be skewed by small actuals—indeed, the MAPE is infinite for an actual value of 0. Some researchers, such as Coleman and Swanson (2007) have suggested taking logarithms as a way of mitigating the problem, though this makes the resulting metric more difficult to interpret. (2) The MAPE exhibits a rather counter-intuitive asymmetry; a forecast of 5 units on an actual of 10 produces an absolute percentage error of 50%, whereas a forecast of 10 units on an actual of 5 gives an APE of 100%. Attempts to amend the MAPE in order to overcome this problem (Makridakis 1993) have met with limited success (Koehler 2001). In light of these problems, some authors—including Fildes (1992) and Armstrong and Collopy (1992)—have proposed the use of metrics based on relative absolute errors (RAEs). These are absolute forecast errors divided by the corresponding error from a benchmark method (normally the “naïve” or “random walk” method, which simply repeats the last observation). Hyndman and Koehler (2006) point out, however, that RAE-based metrics suffer from some of the same problems from which the (R)MSE and MAPE suffer: They are sensitive to outliers, and may be skewed by small benchmark forecast errors (again, the RAE is infinite if the forecast error is 0); remedies involving outlier elimination and log-transformation are subject to the same criticisms, too. Hyndman and Koehler (ibid.) propose the mean absolute scaled error (MASE) as a robust, scale-independent metric that largely avoids the problems set out above. Formally, for series y1 , . . . , y T , and denoting by ŷt+h|t the forecast for yt+h made in period t, the absolute scaled error ASEth is defined: ASEth = |ŷt+h|t − yt+h | 1 T −h −h ∑tT0 = 1 | y t0 + h − y t0 | Then the mean absolute scaled error, MASEh at horizon h for the entire series is simply the mean T −1 ∑tT=1 ASEth , and a summary metric for a collection of series may be calculated by taking the mean of the ASEs across all the series. In recommending the MASE, Hyndman 21 Chatfield (1988), for instance, points out that Zellner’s (1986) analysis was unduly affected by the outsized contribution to summary MSE metrics of only 5 out of 1,001 test series. 27 A Bayesian Model for Sales Forecasting at Sun Microsystems and Koehler note that normalization with respect to the benchmark forecast error confers scale-independence, while use of the in-sample aggregate in the denominator of scaled errors makes the MASE metric more stable than those based on relative errors. References D. A. Ahlburg. Forecast evaluation and improvement using Theil’s decomposition. Journal of Forecasting, 3:345–351, 1984. J. Albert. Bayesian Computation with R. Springer, 2008. J. Armstrong and F. Collopy. Error measures for generalizing about forecasting methods— empirical comparisons. International Journal of Forecasting, 8:69–80, 1992. J. M. Armstrong. Judgmental bootstrapping: Inferring experts’ rules for forecasting. In J. M. Armstrong, editor, Principles of Forecasting: A Handbook for Researchers and Practitioners, pages 171–192. Kluwer, 2001a. J. M. Armstrong. Combining forecasts. In J. M. Armstrong, editor, Principles of Forecasting: A Handbook for Researchers and Practitioners, pages 417–439. Kluwer, 2001b. J. M. Armstrong and F. Collopy. Integration of statistical methods and judgment for time series forecasting: Principles from empirical research. In G. Wright and P. Goodwin, editors, Forecasting with Judgment, chapter 10, pages 269–293. John Wiley & Sons, New York, 1998. A. H. Ashton, R. H. Ashton, and M. N. Davis. White-collar robotics: Levering managerial decision making. California Management Review, 37:83–109, 1994. Carliss Y. Baldwin and Kim B. Clark. Sun wars: Competition within a modular cluster, 1985-1990. In David B. Yoffie, editor, Competing in the Age of Digital Convergence, pages 123–158. Harvard Business School Press, 1997. F. Bass. A new product growth model for consumer durables. Management Science, 15: 215–227, 1969. R. Batchelor and T. Y. Kwan. Judgemental bootstrapping of technical traders in the bond market. International Journal of Forecasting, 23:427–445, 2007. J. M. Bates and C. W. J. Granger. The combination of forecasts. Operational Research Quarterly, 20:451–468, 1969. J. M. Bernardo and A. F. M. Smith. Bayesian Theory. Wiley, 1994. R. C. Blattberg and S. J. Hoch. Database models and managerial intuition: 50% model + 50% manager. Management Science, 36(8):887–899, August 1990. F. Bolger and N. Harvey. Heuristics and biases in judgmental forecasting. In G. Wright and P. Goodwin, editors, Forecasting with Judgment, chapter 5, pages 113–137. John Wiley & Sons, New York, 1998. J. Boulden. Fitting the sales forecast to your firm. Business Horizons, 1:65–72, 1958. 28 A Bayesian Model for Sales Forecasting at Sun Microsystems G. E. P. Box, G. M. Jenkins, and G. C. Reinsel. Time Series Analysis. Prentice-Hall, 3rd edition, 1994. D. Bunn and G. Wright. Interaction of judgemental and statistical forecasting methods: Issues & analysis. Management Science, 37(5):501–518, May 1991. C. Carter and R. Kohn. On Gibbs sampling for state space models. Biometrika, 81:541–543, 1994. C. Chatfield. Apples, oranges and mean square error. Journal of Forecasting, 4:515–518, 1988. R. T. Clemen. Combining forecasts: A review and annotated bibliography. International Journal of Forecasting, 5:559–583, 1989. C. D. Coleman and D. A. Swanson. On MAPE-R as a measure of cross-sectional esimation and forecast accuracy. Journal of Economic and Social Measurement, 32:219–233, 2007. D. Dalrymple. Sales forecasting practices: Results from a United States survey. International Journal of Forecasting, 3:379–391, 1987. Jessica Davis. Dell takes wraps off channel partner program, 2007. eWeek.com, December 5, http://www.channelinsider.com/article/Dell+Takes+Wraps+O+Channel+Partner+ Program/220950\_1.aspx. R. M. Dawes. A case study of graduate admissions: Application of three principles of human decision making. American Psychologist, 26:180–188, 1971. Ton de Kok, Fred Janssen, Jan van Doremalen, Erirk van Wachem, Methieu Clerkx, and Winfried Peters. Philips electronics synchronizes its supply chain to end the bullwhip effect. Interfaces, 35(1):37–48, January–February 2005. Jason Dean. The forbidden city of Terry Gou, 2007. The Wall Street Journal, August 11, http://online.wsj.com/public/article/SB118677584137994489.html. F. X. Diebold. Forecast combination and encompassing: Reconciling two divergent literatures. International Journal of Forecasting, 5:589–592, 1989. F. X. Diebold and P. Pauly. The use of prior information in forecast combination. International Journal of Forecasting, 6:503–508, 1990. A. Dolgui and M. Pashkevich. On the performance of binomial and beta-binomial models of demand forecasting for multiple slow-moving inventory items. Computers and Operations Research, 35:893–905, 2008a. A. Dolgui and M. Pashkevich. Demand forecasting for multiple slow-moving items with short requests history and unequal demand variance. International Journal of Production Economics, 112:885–894, 2008b. J. Durbin and S. J. Koopman. Time Series Analysis by State Space Methods. Oxford University Press, Oxford, 2001. B. Efron. Bootstrap methods: Another look at the jackknife. The Annals of Statistics, 7:1–26, 1979. 29 A Bayesian Model for Sales Forecasting at Sun Microsystems P. T. Elgers, H. L. May, and D. Murray. Note on adjustments ot analysts’ earning forecasts based upon systematic cross-sectional components of prior-period errors. Management Science, 41:1392–1396, 1995. R. Fildes. Efficient use of information in the formation of subjective industry forecasts. Journal of Forecasting, 10:597–617, 1991. R. Fildes. The evaluation of extrapolative forecasting methods. International Journal of Forecasting, 8:81–98, 1992. R. Fildes and P. Goodwin. Good and bad judgement: Lessons from four companies. Foresight: The International Journal of Applied Forecasting, 8:5–10, 2007. R. Fildes, P. Goodwin, M. Lawrence, and K. Nikolopoulos. Effective forecasting and judgmental adjustments: An empirical evaluation and strategies for improvement in supplychain planning. International Journal of Forecasting, 25:3–23, 2009. P. Franses. Merging models and experts. International Journal of Forecasting, 24:31–33, 2008. P. H. Franses and R. Legerstee. Properties of expert adjustments on model-based SKUlevel forecasts. International Journal of Forecasting, 25:35–47, 2009. S. Frühwirth-Schnatter. Data augmentation and dynamic linear models. Journal of Time Series Analysis, 15:183–102, 1994. E. S. Gardner and E. McKenzie. Seasonal exponential smoothing with damped trends. Management Science, 35(3):372–376, March 1989. A. Gelman and J. Hill. Data Analysis Using Regression and Multilevel/Hierarchical Models. Cambridge University Press, 2006. A. Gelman, J. Carlin, H. Stern, and D. Rubin. Bayesian Data Analysis. Chapman & Hall/CRC Press, Boca Raton, 2nd edition, 2003. J. Geweke. Evaluating the accuracy of sampling-based approaches to calculating posterior moments. In J. M. Bernardo, J. O. Berger, A. P. Dawid, and A. F. M. Smith, editors, Bayesian Statistics, volume 4, pages 169–193. Clarendon Press, 1992. J. Geweke and H. Tanizaki. Note on the sampling distribution for the Metropolis-Hastings algorithm. Communications in Statistics – Theory and Methods, 32(4):775–789, 2003. E. Ghysels, D. R. Osborn, and P. M. M. Rodrigues. Forecasting seasonal time series. In G. Elliott, C. Granger, and A. Timmermann, editors, Handbook of Economic Forecasting, volume 1, pages 660–706. North-Holland, 2006. W. R. Gilks, S. Richardson, and D. J. Spiegelhalter. Markov Chain Monte Carlo in Practice. Chapman & Hall, London, 1996. P. Goodwin. Statistical correction of judgmental point forecasts and decisions. Omega: International Journal of Management Science, 24:551–559, 1996. P. Goodwin. Adjusting judgemental extrapolations using Theil’s method and discounted weighted regression. Journal of Forecasting, 16:37–46, 1997. P. Goodwin. Correct or combine? Mechanically integrating judgmental forecasts with statistical methods. International Journal of Forecasting, 16:261–275, 2000. 30 A Bayesian Model for Sales Forecasting at Sun Microsystems C. W. J. Granger and P. Newbold. Some comments on the evaluation of economic forecasts. Applied Economics, 5:35–47, 1973. C. W. J. Granger and R. Ramanathan. Improved methods of combining forecasts. Journal of Forecasting, 3:197–204, 1984. K. R. Hammond, C. J. Hursch, and F. J. Todd. Analyzing the components of clinical inference. Psychological Review, 71:255–262, 1964. A. C. Harvey. Forecasting, Structural Time Series Models and the Kalman Filter. Cambridge Univ. Press, Cambridge, 1989. Joseph Heller. Catch-22. Simon & Schuster, New York, 1961. R. Hill. Applying Bayesian methodology with a uniform prior to the single period inventory model. European Journal of Operational Research, 98:555–562, 1997. P. J. Hoffman. The paramorphic representation of clinical judgment. Psychological Bulletin, 57:116–131, 1960. C. Huang and G. Tzeng. Multiple generation product life cycle predictions using a novel two-stage fuzzy piecewise regression analysis method. Tehnological Forecasting and Social Change, 75:12–31, 2008. H. D. Hughes. An interesting seed corn experiment. The Iowa Agriculturist, 17:424–425, 428, 1917. C. J. Hursch, K. R. Hammond, and J. L. Hursch. Some methodological considerations in mutliple-probability studies. Psychological Review, 71:42–60, 1964. R. J. Hyndman and A. B. Koehler. Another look at measures of forecast accuracy. International Journal of Forecasting, 22:679–688, 2006. R. J. Hyndman, A. B. Koehler, J. K. Ord, and R. D. Snyder. Forecasting with Exponential Smoothing: The State Space Approach. Springer, 2008. K. Kahn. How to measure the impact of a forecast error on an enterprise? Journal of Business Forecasting, pages 21–25, 2003. Spring. R. Klassen and B. Flores. Forecasting practices of Canadian firms: Survey results and comparisons. International Journal of Production Economics, 70:163–174, 2001. A. B. Koehler. The asymmetry of the sAPE and other comments on the M3-competition. International Journal of Forecasting, 17:570–574, 2001. L. Lamport. LATEX: A Document Preparation System. Addison-Wesley Professional, 2nd edition, 1994. M. Lawrence and M. O’Connor. Judgement or models: The importance of task differences. Omega: International Journal of Management Science, 24:245–254, 1996. M. Lawrence, P. Goodwin, M. O’Connor, and D. Önkal. Judgmental forecasting: A review of progress over the last 25 years. International Journal of Forecasting, 22:493–518, 2006. M. J. Lawrence, M. O’Connor, and R. H. Edmundson. A field study of sales forecasting accuracy and processes. European Journal of Operational Research, 122(1):151–160, April 2000. 31 A Bayesian Model for Sales Forecasting at Sun Microsystems C.-Y. Lee, J.-D. Lee, and Y. Kim. Demand forecasting for new technology with a short history in a competitive market: The case of the home networking market in South Korea. Technological Forecasting and Social Change, 75:91–106, 2008. W. Y. Lee, P. Goodwin, R. Fildes, K. Nikolopoulos, and M. Lawrence. Providing support for the use of analogies in demand forecasting tasks. International Journal of Forecasting, 23:377–390, 2007. P. Lenk and A. Rao. New models from old: Forecasting product adoption by hierarchical Bayes procedures. Marketing Science, 9:42–57, 1990. D. V. Lindley. Reconciliation of probability distributions. Operations Research, 31:866–880, 1983. V. Mahajan, E. Muller, and Y. Wind, editors. New-Product Diffusion Models. Kluwer, 2000a. V. Mahajan, E. Muller, and Y. Wind. New-product diffusion models: From theory to practice. In New-Product Diffusion Models Mahajan et al. (2000a), chapter 1. S. Makridakis. Accuracy measures: Theoretical and practical concerns. International Journal of Forecasting, 9:527–529, 1993. W. H. McGlothlin. Stability of choices among uncertain alternatives. American Journal of Psychology, 69:604–615, 1956. P. E. Meehl. When shall we use our heads instead of the formula? Jounal of Counseling Psychology, 4:268–273, 1957. J. T. Mentzer. The impact of forecasting improvement on return on shareholder value. Journal of Business Forecasting, pages 8–12, 1999. Fall. J. T. Mentzer and C. C. Bienstock. Sales Forecasting Management: Understanding the Techniques, Systems and Management of the Sales Forecasting Process. Sage Publications, Thousand Oaks, CA, 1998. W. W. Moe and P. S. Fader. Using advance purchase orders to forecast new product sales. Marketing Science, 21(3):347–364, Summer 2002. A. Montgomery. Creating micro-marketing pricing strategies using supermarket scanner data. Marketing Science, 16:315–337, 1997. M. A. Moon, J. T. Mentzer, and C. D. Smith. Conducting a sales forecasting audit. International Journal of Forecasting, 19(1):5–25, 2003. January–March. M. Moriarty. Design features of forecasting systems involving management judgments. Journal of Marketing Research, 22:353–364, 1985. P. A. Morris. Decision analysis expert use. Management Science, 20:1233–1241, 1974. P. A. Morris. An axiomatic approach to expert resolution. Management Science, 29:24–32, 1983. R. Neelamegham and P. Chintagunta. A Bayesian model to forecast new product performance in domestic and international markets. Marketing Science, 18(2):115–136, 1999. R. Neelamegham and P. Chintagunta. Modeling and forecasting the sales of technology products. Quantitative Marketing and Economics, 2:195–232, 2004. 32 A Bayesian Model for Sales Forecasting at Sun Microsystems S.-C. Niu. A piecewise-diffusion model of new-product demands. Operations Research, 54: 678–695, 2006. M. O’Connor, W. Remus, and K. Lim. Improving judgmental forecasts with judgmental bootstrapping and task feedback support. Journal of Behavioral Decision Making, 18:247– 260, 2005. L. D. Pankoff and H. V. Roberts. Bayesian synthesis of clinical and statistical prediction. Psychological Bulletin, 80:762–773, 1968. A. Pole, M. West, and J. Harrison. Applied Bayesian Forecasting and Time Series Analysis. Chapman Hall, 1994. Y. Qi and T. P. Minka. Hessian-based Markov chain Monte Carlo algorithms. In First Cape Cod Workshop on Monte Carlo Methods, 2002. T. Åstebro and S. Elhedhli. The effectiveness of simple decision heuristics: Forecasting commercial success for early-stage ventures. Management Science, 52:396–409, 2006. E. M. Rogers. Diffusion of Innovations. Free Press, 5th edition, 2003. N. R. Sanders. Comments on “Effective forecasting and judgmental adjustments: An empirical evaluation and strategies for improvement in supply-chain planning”. International Journal of Forecasting, 25:24–26, 2009. N. R. Sanders and K. B. Manrodt. Forecasting practices in US corporations: Survey results. Interfaces, 24:92–100, 1994. N. R. Sanders and K. B. Manrodt. Forecasting software in practice: Use, satisfaction, and performance. Interfaces, 33:90–93, 2003. N. R. Sanders and L. P. Ritzman. On knowing when to switch from quantitative to judgmental forecasts. International Journal of Operations and Production Management, 11:28–37, 1991. N. R. Sanders and L. P. Ritzman. Judgmental adjustment of statistical forecasts. In J. M. Armstrong, editor, Principles of Forecasting: A Handbook for Researchers and Practitioners, pages 405–416. Kluwer, 2001. Steven M. Schafer. Web Standards Programmer’s Reference: HTML, CSS, JavaScript, Perl, Python, and PHP. Wrox, Chicago, IL, 2005. E. Silver. Bayesian determination of the reorder point of a slow moving item. Operations Research, 13:989–997, 1965. SMCI. Sun microsystems company info, 2009. http://www.sun.com/aboutsun/company/ index.jsp. A. Sood, G. James, and G. Tellis. Functional regression: A new model for predicting market penetration of new products. Marketing Sciences, 28:36–51, 2009. H. Theil. Applied Economic Forecasting. North-Holland, 1971. A. Timmermann. Forecast combinations. In G. Elliott, C. Granger, and A. Timmermann, editors, Handbook of Economic Forecasting, chapter 4, pages 135–194. North-Holland, 2006. 33 A Bayesian Model for Sales Forecasting at Sun Microsystems L. R. Tucker. A suggested alternative formulation of the developments by Hursch, Hammond and Hursch and by Hammond, Hursch and Todd. Psychological Review, 71:528– 530, 1964. J.W. Tukey. Exploratory data analysis. Addison-Wesley, 1977. A. Tversky and D. Kahneman. Judgment under uncertainty: Heuristics and biases. Science, 185:1124–1131, 1974. T. T. Tyebjee. Behavioral biases in new product forecasting. International Journal of Forecasting, 3:393–404, 1987. H. J. van Heerde, C. F. Mela, and P. Manchanda. The dynamic effect of innovation on market structure. Journal of Marketing Research, 41:166–183, 2004. W. N. Venables and D. M. Smith. An Introduction to R. Network Theory, 2002. R. Webby and M. O’Connor. Judgmental and statistical time series forecasting: A review of the literature. International Journal of Forecasting, 12:91–118, 1996. M. West and P. J. Harrison. Bayesian Forecasting and Dynamic Models. Springer-Verlag, New York, 2nd edition, 1997. Rick Whiting. Sun takes build-to-order to the next level, 2006. InformationWeek, September 12, http://www.informationweek.com/story/showArticle.jhtml?articleID=192700331. J. Wiedermann. Web Design Flash Sites. Tachen, 2006. H. E. Williams and D. Lane. Web Database Applications with PHP and MySQL. O’Reilly Media, 2nd edition, 2004. J. Wisner and L. Stanley. Forecasting practices of JIT and non-JIT purchasers. European Journal of Purchasing and Supply Management, 1:219–225, 1994. B. Worthen. Future results not guaranteed. CIO, 16:1–4, 2003. URL http://www.cio.com. au/article/168757/future_results_guaranteed. G. Wright and P. Ayton. The psychology of forecasting. Futures, 18:420–439, 1986. P. M. Yelland. A model of the product lifecycle for sales forecasting. Technical Report 127, Sun Microsystems Laboratories, 2004. URL http://research.sun.com/techrep/2004/smli_ tr-2004-127.pdf . P. M. Yelland and E. Lee. Forecasting product sales with dynamic linear mixture models. Technical Report TR-2003-122, Sun Microsystems Laboratories, March 2003. A. Zellner. A tale of forecasting 1001 series: The Bayesian knight strikes again. International Journal of Forecasting, 2:491–494, 1986. Paul H. Zipkin. Foundations of Inventory Management. McGraw-Hill/Irwin, New York, 2000. 34
© Copyright 2026 Paperzz