Profit Persistence An Application of a Dynamic Panel Method Term paper for the course “Econometric Methods for Panel Data” Vienna, June 30, 2005 Gürkan Birer Matr.nr.: 0254010 Michael Weichselbaumer Matr.nr.: 9640369 2UK 300632 Ökonometrische Methoden für Paneldaten Lehrveranstaltungsleiter: O. Univ.-Prof. Dr. Robert Kunst Contents 1 Introduction 1 2 Method 2.1 Variables . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 2 3 3 Models 3.1 Time Dummies . . . . . . . . . . . . . . . . . . . . . 3.2 Dynamic Panel . . . . . . . . . . . . . . . . . . . . . 3 4 4 4 Results 4.1 Time Dummies . . . . . . . . . . . . . . . . . . . . . 4.2 Dynamic Panel . . . . . . . . . . . . . . . . . . . . . 7 7 8 5 Conclusion 10 1 Introduction In a competitive economy firm profits above the average should diminish over time. Hence, the empirical question of interest is if this can be observed and thus confirmed for a given set of data or if profits tend to persist. The problem is normally analysed as a time-series problem, since in the structural model profit persistence is dominated by the influence of past profits (Goddard & Wilson 1999). This means that for every individual firm a particular ARparameter is estimated from the corresponding individual time series of profits. In our work we want to investigate this question using panel data methods. Our approach differs from the time-series approach in two important ways: first, we hypothesise that persistent profits do not exist and hence exclude the intercept; second, we model a single coefficient of disappearance of profit deviations from the average, meaning that the speed of convergence is the same for all firms – and thus over all industries. In terms of model specification this implies the formulation with a single parameter for the influence of lagged profits. The next section gives a bit more of the underlying method to provide the reader with some understanding of the time-series approach and the arising differences from our approach. Additionally, it defines the variables included and describes the data. Section 3 describes the estimated models: a time-dummy model and a dynamic panel. This splitting is retained in section 4, where the results are presented. We conclude with section 5. 2 Method First we describe the dependent variable in a bit more detail. Then, a summary for the explanatory variables is given for those that are usually associated with the possibility of profits above the average and we were able to construct from our database. Afterwards, two different models are presented for estimation of persistence of profits above average over time: a model with time dummies hoping that one can observe the effect of competition in decreasing time dummies; a dynamic model with the alternative and dominantly used modelling strategy of persistence shown in a positive autoregressive coefficient. Both models are used in a random effects version, since we assume that the firms that are in our database and remain in the sample after the construction process of our explanatory vari1 ables is like a sample obtained from random draws of all firms in the economy. In both estimations we apply the same set of explanatory variables. 2.1 Variables First, the dependent variable: The profit of a firm that we use is defined as the ratio of profit after taxes and before interest to total assets. Taxes are excluded since effective tax-rates may differ and firms should then base their decision on entry or exit on the after-tax profit. Interest paid is included since total assets in the denominator do also encompass those assets financed by debt (argumentation follows Maruyama & Odagiri 2002). Profit of firm i at time t is denoted πit . The variable of interest, denoted ρit , is the deviation of firm i’s profit at time t from the average profit of all firms in year t. It is calculated as follows: ρit = where π̄t = πit − π̄t π̄t N 1 X πit N n=1 Next, a description of the calculation of the explanatory variables that we used and the basic intuition. Industry Dummy Our data is classified into four-digit codes following the Standard Industrial Classification (SIC) of the United States Government. It classifies every firm for every year according to their main field of production and is thus called ”Primary SIC-Code”1 . We use the industries classified in the large groups 2 and 3, the classes of manufacturing and construction, and reduce them to the three-digit-level, since only this level is specified accurately for all usable observations. Market Share This is calculated as share in sales of aggregate sales in the three-digit-level industry where a firm is classified. Though all firms with availability of sales are used to compute aggregate sales and not only those firms which are used in the very final data set for computation (i.e. where all variables are available), it remains an approximation. There 1 Though this classification system was replaced by the North American Industry Classification System (NAICS), we use the old SIC-Codes, since our data is originally classified in SIC and the last year we use is 1996. 2 are different causes that support the inclusion of market share as an explanatory variable, for example higher market power, higher efficiency due to economies of scale. We expect it to be positively related to profitability. Risk As a proxy for risk we calculate the standard deviation of industry profits, again at the three-digit-level. This variable is also expected to have a positive relationship. Growth of Industry It is calculated as the percentage growth in aggregate sales per three-digit-industry-level. We adopt the usual expectation that the relationship is positive, since more rapid growth relieves competitive pressure due to supposedly higher demand. Growth in the Number of Competitors Calculated as the percentage change in the absolute number of competitors in the industry. If it is high, competition should also increase and higher profits are expected to vanish. Entry/Exit Profits, in turn, depend on both actual and threatened entry. Since one has at most data about actual entry, entry’s impact on profits must always be understated. Thus, in timeseries models of profit persistence entry is dropped as an explanatory variable and is treated as a latent variable (Goddard & Wilson 1999). 2.2 Data The data stems from the Standard & Poors database Compustat. After adaptations necessary due to missing data for some variables and after the construction process that is as usual subject to the trade-off between number of years and number of observable individuals, we obtain a balanced sample of 900 U.S.-American firms for the time period of 1985 to 1996. The operations of the firms are spread over 98 three-digit-classes of SIC. 3 Models Now we will describe the two different models that we want to use for measuring the effect of competition over time: a simple random effects model with time dummies and a dynamic panel with random effects. 3 3.1 Time Dummies In a first trial, we model the persistence with time dummies and hope that the increase in competition will be reflected in diminishing time effects. As described above, we chose some other explanatory variables to control for important influences on profits. The estimation equation is: πit = Constant + MSit β1 + RiskIndustry,t β2 + ∆CompetitorsIndustry,t β3 +∆OutputIndustry,t β4 + Zλ,it λt + uit with 3.2 uit = µi + νit µi ∼ i.i.d.(0, σµ2 ) νit ∼ i.i.d.(0, σν2 ) µi , νi independent Dynamic Panel As mentioned above, from the economic point of view the inclusion of a lag of the dependent variable as an explanatory variable is legitimate. The model then takes the form (Baltagi 1995, pp.125): 0 yit = γyi(t−1) + xit β + uit , i = 1, . . . , N t = 1, . . . , T where uit = µi + νit – a one way error component model. For the estimation of the dynamic panel, it is well documented in the literature that the OLS estimation will lead to biased and inconsistent parameter estimates, even if the νit are not serially correlated, because yit and yi(t−1) are both a function of µi and hence yi(t−1) is correlated with the error term. For the fixed effects estimator the within transformation – the multiplication with the mean-adjusting matrix – wipes out the µi , but still the estimators will be biased, since νi(t−1) is contained in ν̄i , which is thus correlated with yi(t−1) . The same holds for random effects models. Only if T tends to infinity, the fixed effects estimator of γ and β will be consistent for the dynamic error component model. Since our data has the shape of a cross section panel with N large and T small, we do not rely on this property. A possibility to avoid the bias is the first-difference transformation of the data, because it wipes out the individual effects but does not create the problem of correlation. Among the alternative methods that use first-differencing, we choose for our dynamic panel regressions the method developed by Arellano & Bond (1991). First 4 we describe the method and than we discuss some of the advantages of using this method than the other methods like the well known method of Anderson & Hsiao (1981). In their paper, Arellano & Bond start with the simplest autoregressive specification with random effects in the form: yit = γyi(t−1) + ηi + νit , |γ| < 1 They assume lack of serial correlation in the error terms but not necessarily independence over time, i.e. they assume that E(νit ) = E(νit νis ) = 0. The model works as follows. First they take the first differences to get rid of the individual effects that leads to an equation of the form yit − yi(t−1) = γ(yi(t−1) − yi(t−2) ) + vit − vi(t−1) So for the first period (t=3) we have yi3 − yi2 = γ(yi2 − yi1 ) + vi3 − vi2 and for the second period (t=4) we have yi4 − yi3 = γ(yi3 − yi2 ) + vi4 − vi3 and if we continue in this iterative way we get for the last period (t=T): yiT − yi(T −1) = γ(yi(T −1) − yi(T −2) ) + viT − vi(T −1) They use the GMM (Generalized Method of Moments) for obtaining an optimal estimator of γ. The advantage of using GMM is that it will still work even if we have no information about the initial conditions and the distributions of vit and µi – hence it is preferred to other estimation methods like Maximum likelihood or OLS. The GMM estimator γ̂ is based on sample moments, defined as N −1 N X 0 0 Zi ν̄i = N −1 Z ν̄ i=1 0 0 0 0 where ν̄ = (ν̄1 , . . . , ν̄N ) – the column vector of all ν̄i , which, in turn is the column vector of the first differences of the νit for all t. Thus, 0 ν̄i has length T − 2 (since it is reduced by one observation due to first differencing and another one by formulation of a model with lagorder one) and ν̄ has length T − 2 times the number of individuals, 5 0 0 N . Z is a N (T − 2) × m matrix consisting of (Z1 , . . . , ZN ), which are used as instruments. The estimator γ̂ is given by: 0 0 γ̂ = argminγ (ν̄ Z)AN (Z ν̄) where AN is the so-called weighting matrix. Any such matrix would lead to unbiased estimates of γ, provided that the instruments are orthogonal to the errors. The choice of the instruments is an important factor because they should be highly correlated with the dependent variables and uncorrelated with the errors. Under these assumptions for the simplest AR(1) specification Arellano & Bond provide a set of instruments for the GMM estimation. The instrument matrix will be of the form: [yi1 ] 0 [yi1 , yi2 ] Zi = ... 0 [yi1 , . . . , yi,T −2 ] These instruments are different from the ones that suggested by Anderson & Hsiao (1981). From their Monte-Carlo study, Arellano & Bond provide evidence that the AH-estimates are either biased due to reasons of singularity or, in the best case, are more inefficient. That is why we have chosen to use the Arrelano-Bond method explained so far. What now remains among the central issues that have to be considered in the chosen estimation method are two tests that we want to describe shortly: first, a test procedure that can be used for testing if the errors vit are serially uncorrelated or not. This test is very important because the GMM estimators are only consistent if these errors are serially uncorrelated; second, since for T > 3 it will be an over-identified system – in other words, there will be more equations than unknowns – we need an over-identifying restrictions test. They suggest the test that has been developed by Sargan. Now, a few more words about the extended model with exogenous variables, which has the following form: Model with exogenous variables 0 yit = γyi(t−1) + β xit + ηi + νit , |γ| < 1 where νit are not serially correlated. The GMM estimator of the 6 coefficient vector will be 0 0 0 0 0 0 (γ̂, β̂) = (X̄ ZAN Z X̄ )−1 X̄ ZAN Z ȳ where the matrix of instruments Z will include the same instruments ˆ This as before for γ̂ plus an additional set of instruments for beta. additional set can be formed in two ways: if xit are strictly exogenous, in other words, E(xit , νis ) = 0 for all t, s, then all the x are valid instruments for all the equations; if the xit are predetermined in the sense that E(xit , νis ) 6= 0 for s < t and zero otherwise, then only xi1 , . . . , xi(s−1) are valid instruments. 4 Results Now we want to present the results of the two models introduced above: 4.1 Time Dummies For the first panel regression where we had four explanatory variables and time dummies the results are quite fuzzy (see table 1). Out of the four explanatory variables only two of them – market share and risk – turn out to be significant at the five percent level with risk unexpectedly having a negative coefficient. The only significant explanatory variable with the correct sign turns out to be the market share (2.04 with z=5.48). The two variables that were not significant at any meaningful level – ∆competitors and ∆sales – had the correct sign. Only four out of 12 time dummies were significant at the five percent level and they had different signs. The industry dummies were not significant. Thus the estimations below did not include them. Theoretically, the time dummies should all be insignificant, because we subtracted yearly averages from our dependent variable. Hence, the time-dummies can not explain differences between the years by construction. But since we calculated the deviations from the yearly means before dropping observations where not all necessary data was available, though the data to calculate profits was available, positive or negative average yearly deviations can still remain. What we can conclude now is that the data that remains after dropping some observations due to data limitations is representative with respect to the sample from which we calculated the yearly profit averages for most years at least. Besides that, the overall model can explain only up to five percent of the change in the dependent variable. The first explanation for 7 Table 1: Results from the time-dummy-estimation on profits above the mean (1985-1996) Explanat. variable Market Share Risk ∆Competitors ∆Sales Year86 Year87 Year88 Year89 Year90 Year91 Year92 Year93 Year94 Year95 Year96 constant συ σ² ρ Observations Coefficient 2.040 -.211 -.814 -.0528 0.097 0.601 -0.753 -0.371 -0.001 -0.448 -0.094 -0.020 0.071 -0.306 -0.289 -0.549 (Std. Err.) (0.373) (0.072) (1.458) (0.123) (0.204) (0.201) (0.213) (0.205) (0.200) (0.198) (0.199) (0.208) (0.200) (0.201) (0.201) (0.153) z 5.48 -2.93 -0.56 -0.43 0.48 2.99 -3.53 -1.81 0.00 -2.26 -0.47 -0.10 0.35 -1.52 -1.44 -3.58 P > |z| 0.000 0.003 0.577 0.668 0.633 0.003 0.000 0.070 0.997 0.024 0.637 0.922 0.723 0.128 0.150 0.000 0.996 4.002 0.058 (fraction of variance due to ui ) 900 groups 12 obs./group the results is that the model was not correctly specified. When we check the correlation structure of the explanatory variables we can say that they are not highly correlated. Another reason can be that the data suffers heavily from outliers. As a result we find that this simple method could not capture the effect of competition over time and continue with the results of the dynamic panels. 4.2 Dynamic Panel The performance of the dynamic panels with or without the set of explanatory variables from above – we assume that all are strictly exogenous – is slightly better than the models from above. At least we see (in tables 2 and 3) a positive (rather small) and highly significant autoregressive parameter γ. In the dynamic setting risk has the correct sign but now it turns out to be insignificant and unexpectedly the market share now turns out to have a big negative and highly significant coefficient which is totally contradicting economic theory. As we mentioned before Stata gives two test results which 8 may be interpreted as follows: the Sargan test which is used to test the validity of the over-identifying restrictions (or the validity of the instruments) has been rejected. The results of the serial autocorrelation tests are quite good. In both panels the null hypothesis that there is no second order serial autocorrelation in the error term could not be rejected. Table 2: Results from the dynamic-panel-estimation (1985-1996) Explanat. variable ρt−1 Coefficient 0.168 (Std. Err.) (0.013) z 13.06 P > |z| 0.000 χ2(54) =321.76 Prob≥ χ2 =0.000 average autocovariance in residuals of order 1 is 0: z = -50.54 Prob>z = 0.000 average autocovariance in residuals of order 2 is 0: z = -0.89 Prob>z = 0.375 900 groups 10 obs./group Sargan test of over-identifying restrictions: Arellano-Bond test that H0 : no autocorrelation Arellano-Bond test that H0 : no autocorrelation Observations Table 3: Results from the dynamic-panel-estimation with exogenous variables (1985-1996) Explanat. variable ρt−1 Market Share Risk ∆Competitors ∆Sales Coefficient 0.168 -5.374 0.011 -1.861 0.140 (Std. Err.) (0.013) (1.878) (0.096) (2.043) (0.164) z 13.05 -2.86 0.12 -0.91 0.86 P > |z| 0.000 0.004 0.904 0.363 0.392 χ2(54) =317.94 Prob≥ χ2 =0.000 average autocovariance in residuals of order 1 is 0: z = -50.57 Prob>z = 0.000 average autocovariance in residuals of order 2 is 0: z = -0.89 Prob>z = 0.375 900 groups 10 obs./group Sargan test of over-identifying restrictions: Arellano-Bond test that H0 : no autocorrelation Arellano-Bond test that H0 : no autocorrelation Observations The coefficients of the lagged dependent variables are the same in both estimations. The cases of the wrong sign of market share and non-significant parameters for all of the exogenous variables made us suspicious for the reliability of our database. The rela- Discussion 9 tionships seem to be reasonable and a lot of empirical evidence could be found for them. Additionally, a persistence of profit-deviations from the mean by only 16.8 percent from one year to the next seems too low and is also too low in light of estimates from the dominantly used method of estimating the autoregressive parameters from every time-series individually (as done, e.g. in Goddard & Wilson 1999, Maruyama & Odagiri 2002; see also the references there.). 5 Conclusion We conclude that our method is not useful to model the persistence of profits. Probably one could try to model a different data set, hoping that the expected relationships with the exogenous variables will hold then and be significant. Additionally, the estimation of individual coefficients of persistence, γi , could be necessary and thus yield better results. This is suggested by the argument that the internal structure and functioning of firms is likely to differ as well as that different firms face different entry and mobility barriers – though there is no strong evidence that the convergence process differs between above- and below-mean profit firms (Mueller 1990, p.194-196). Still, the doubt that our method should be preferred over the time-series approach mentioned remains strong. 10 References Anderson & Hsiao (1981), ‘Estimation of dynamic models with error components’, Journal of the American Statistical Association 76, 598–606. Arellano, M. & Bond, S. (1991), ‘Some tests of specification for panel data: Monte carlo evidence and an application to employment equations’, Review of Economic Studies 58, 277–297. Baltagi, B. H. (1995), Econometric analysis of panel data, John Wiley and Sons, New York. Goddard, J. & Wilson, J. (1999), ‘The persistence of profit: a new empirical interpretation’, International Journal of Industrial Organization 17, 663–687. Maruyama, N. & Odagiri, H. (2002), ‘Does the ’persistence of profits’ persist?: a study of company profits in japan, 1964-97’, International Journal of Industrial Organization 20, 1513–1533. Mueller, D. C. (1990), The dynamics of company profits. An international comparison, Campridge University Press. 11
© Copyright 2026 Paperzz