Friday, December 27, 2013

Steve Keen: Economists Are "Insufficiently Numerate"

"Curiously, though economists like to intimidate other social scientists with the mathematical rigor of their discipline, most economists do not have this level of mathematical education...

...One example of this is the way economists have reacted to 'chaos theory'... Most economists think that chaos theory has had little or no impact - which is generally true in economics, but not at all true in most other sciences. This is partially because, to understand chaos theory, you have to understand an area of mathematics known as 'ordinary differential equations.' Yet this topic is taught in very few courses on mathematical economics - and where it is taught, it is not covered in sufficient depth. Students may learn some of the basic techniques for handling what are known as 'second-order linear differential equations,' but chaos and complexity begin to manifest themselves only in 'third order nonlinear differential equations.' Steve Keen (2011). Debunking Economics: The Naked Emperor Dethroned?, Revised and expanded ed. Zed Books, p. 31

The above quotes are also in the first edition. Before commenting on this passage, I want to re-iterate my previously expressed belief that some economists, including some mainstream economists, understand differential and difference equations.

I misremembered this comment as being overstated for polemical purposes. But, in context, I think it is clear to those who know the mathematics.

I took an introductory course in differential equations decades ago. Our textbook was by Boyce and DiPrima. As I recall, we were taught fairly cookbook techniques to solve linear differential equations. These could be first order or second order and homogeneous or non-homogeneous. They could also be systems of linear differential equation. I recall some statement of an existence theorem for Initial Value Problems (IVPs), although I think I saw a more thorough proof of some such theorem in an introductory1 real analysis course. We might have also seen some results about the stability of limit points for dynamical systems. Keen is not claiming that economists do not learn this stuff; this kind of course is only a foundation for what he is talking about.

I also took a later applied mathematics course, building on this work. In this course, we were taught how to linearize differential equations. We definitely were taught stability conditions here. If I recall correctly, the most straightforward approach looked only at sufficient, not necessary conditions. We also learned perturbation theory, which can be used to develop higher order approximations to nonlinear equations around the solutions to the linearized equations. One conclusion that I recall is that the period of an unforced pendulum depends on the initial angle, despite what is taught in introductory physics classes2. I do not recall much about boundary layer separations, but maybe that was taught only in the context of Partial Differential Equations (PDEs), not Ordinary Differential Equations (ODEs). This is still not the mathematics that Keen is claiming that economists mostly do not learn, although it is getting there.

You might also see ordinary differential equations in a numerical analysis course. Here you could learn about, say, the Runge-Kutta method. And the methods here can apply to IVPs for systems of non-linear equations3. I believe in the course that I took, we had a project that began to get at the rudiments of complex systems. I think we had to calculate the period of a non-linear predator-prey system. I believe we might have been tasked with constructing a Poincaré return map.

According to Keen, a sufficiently numerate economist should know the theory behind complex dynamical systems, chaos, bifurcation analysis, and catastrophe theory4. I think such theory requires an analysis able to examine global properties, not just local stability results. And one should be interested in the topological properties of a flow, not just the solution to a (small number of) IVPs. Although this mathematics has been known, for decades, to have applications in economics, most economics do not learn it. Or, at least, this is Keen's claim.

Economists should know something beyond mathematics. For example, they should have some knowledge of the sort of history developed by, say, Fernand Braudel or Eric Hobsbawm. And they should have some understanding of contemporary institutions. How can they learn all of this necessary background and the needed mathematics5, as well? I do not have an answer, although I can think of three suggestions. First, much of what economists currently teach might be drastically streamlined. Second, one might not expect all economists to learn everything; a pluralist approach might recognize the need for a division of labor within economics. Third, perhaps the culture of economics should be such that economists are not expected to do great work until later in their lifetimes. I vaguely understand history is like this, while mathematics is stereotypically the opposite.

Footnotes
  1. As a student, I was somewhat puzzled by why my textbooks were always only Introductions to X or Elements of X. It took me quite some time to learn the prerequisites. How could this only be an introduction? Only later work makes this plain.
  2. Good physics textbook are clear about linear approximations to the sine function for small angles. Although our textbook motivated perturbation theory in the context of models of solar systems, I have never seen perturbation theory applied here in a formal course. Doubtless, astrophysicists are taught this.
  3. Stiff differential equations is a jargon term that I recall being used. I do not think I ever understood what it meant, but I am clear that the techniques I think I have mostly forgotten were not universally applicable without some care.
  4. Those who have been reading my blog for a while might have noticed I usually present results for the analysis of non-linear (discrete-time) difference equations, not (continuous-time) differential equations.
  5. There are popular sciences books about complex systems.

Monday, December 23, 2013

Alan Greenspan, Fool or Knave?

Robert Solow quotes Greenspan's new book:

"'In a free competitive market,' [Greenspan] writes, 'incomes earned by all participants in the joint effort of production reflect their marginal contributions to the output of the net national product. Market competition ensures that their incomes equal their "marginal product" share of total output, and are justly theirs.'" -- Robert M. Solow

I am not going to waste my time reading banal balderdash from Greenspan. Solow feels the same way about Ayn Rand:

"I got through maybe half of one of those fat paperbacks when I was young, the one about the architect. Since then I have found it impossible to take Ayn Rand seriously as a novelist or a thinker." -- Robert M. Solow

Anyway, as I have explained repeatedly, marginal productivity is not a theory of distribution, let alone justice. Firstly, in a long-run model, endowments of capital goods are not givens and marginal productivity conditions fail to pin down the functional distribution of income. A degree of freedom remains, which one might as well take to be the interest rate.

Second, ownership of capital goods does not contribute to the product, even though decisions must be made how to allocate capital, both as finance and as physically existing commodities. The New Republic is written for a popular audience, and Solow is plainly trying to avoid technicalities. Is his comment about property in the following an echo of my - actually, Joan Robinson's - point, albeit mixed in with other stuff, including comments about initial positions:

"Students of economics are taught that ... the actual outcome, including the relative incomes of participants, depends on 'initial endowments,' the resources that participants bring when they enter the market. Some were born to well-off parents in relatively rich parts of the country and grew up well-fed, well-educated, well-cared-for, and well-placed, endowed with property. Others were born to poor parents in relatively poor or benighted parts of the country, and grew up on bad diets, in bad schools, in bad situations, and without social advantages or property. Others grew up somewhere in between. These differences in starting points will be reflected in their marginal products and thus in their market-determined incomes. There is nothing just about it." -- Robert M. Solow

As far as I am concerned, Greenspan's job, for decades, has been ensuring, on behalf of the rulers of the United States, that workers do not get too big for their britches.

Wednesday, December 18, 2013

Period Doubling As A Route To Chaos

Figure 1: An Example of Temporal Dynamics for the Logistic Equation
1.0 Introduction

This post illustrates some common properties of two dynamical systems, chosen out of a large class of such systems. Two quite different functions are drawn, and I demonstrate that, qualitatively, certain behavior arising out of these functions looks quite alike. Furthermore, I point out a mathematical argument that a certain quantitative constant arises for both functions.

I do not claim that the iterative processes here characterize any specific economic model (but see here and here) or physical process. Feigenbaum (1980) mentions "the complex weather patterns of the atmosphere, the myriad whorls of turmoil in a turbulent fluid, [and] the erratic noise in an electronic signal." Such processes have an emergent macro behavior consistent with a wide variety of micro mechanisms. The mathematical metaphors presented in this post suggest that if economic phenomena were described by complex dynamic processes, economists should then reject microfoundations, reductionism, and strong methodological individualism.

2.0 The Logistic Equation

This post is about the analysis of a sequence x0, x1, x2, ... in discrete time. Successive points in this sequence are defined by the repeated iteration of a function fi. (The index i allows one to specify a specific function.) The first few terms of the time series are defined as follows, for a given index i and a given initial value x0:

x1 = fi(x0)
x2 = fi(x1) = fi(fi(x0))
x3 = fi(x2) = fi(fi(fi(x0)))

The logistic function is defined as follows:

f1(x) = a x(1 - x), 0 < a < 4.

Note the parameter a. For a given value of a in the indicated region, the long term behavior of the iterative process defined above is independent of the initial value. This long term behavior varies dramatically, however, with a. In other words, the long term behavior exhibits a kind of dynamic stability, but is structurally unstable.

The behavior of such a sequence can be nicely illustrated by certain diagrams. Figure 1, above, displays the temporal dynamics for one sequence for one value of the parameter a and one initial value. The abscissa and the ordinate in this diagram both range from zero to unity. The 45-degree line then slopes upward to the right from the point (0, 0) to the point (1, 1). Any distance measured upward on the axis for the ordinate can be reflected through the 45 degree line to project the same distance horizontally on the axis for the abscissa. That is, draw a line horizontally from the Y-axis rightward to the 45 degree line. Then draw a vertical line downward from that intersection with the 45 line to the X axis. You will have measured the same distance along both the abscissa and ordinate.

Values for the time series are shown in the diagram by vertical lines. When projected downward to the axis for the abscissa, one will have a plot of x0, x1, x2, etc. In the case shown in Figure 1, the initial value, x0, is 1/2. The logistic function is shown as the parabola opening downward. A line is drawn upward from the axis for the abscissa to intercept the logistic function. The value of the ordinate for this point is x1. To find this value, as measured on the abscissa, a line is drawn leftward from the point of interception with the logistic function to the 45 degree line. Next, draw a line downward from this point on the 45 degree line to the logistic function. The value of the ordinate for this new point on the logistic function is then x2. The step function in Figure 1 going down to the labeled point is a visual representation of the entire time series. Can you see that, in the figure, all times series for the given value of the parameter a, no matter the initial value, will converge to the labeled point? In the jargon, the times series for the logistic function for this value of a is said to have a single stable limit point.

As a matter of fact, the long term behavior of every time series for the logistic function is generically independent of the initial value. It makes sense then, not to plot the first, say, 20,000 points of the time series and only plot the next say 5,000 points. This would lead to a boring graph for Figure 1; the only point in the non-transient part of the time series would be at the stable limit point. Figure 2 shows a more interesting case, for a larger value of the parameter a. Notice the upside-down parabola now rises to a higher value. Because of the form of the logistic function, the plotted function remains symmetrical around x = 1/2.) For the parameter value used for Figure 2, no stable limit points exist for the time series. Rather, the time series converges to a limit cycle of period 3. That is, the cycle illustrated with the structure with the black lines has three vertical lines and repeats endlessly.

Figure 2: A Cycle with Period 3 for the Logistic Equation

Figures 1 and 2 demonstrate that the limiting behavior of an iterative process for the logistic equation varies with the parameter a. Figure 3 displays this variation from a value of a somewhere under 3.0 to 4.0. In Figure 3, the value of a is plotted along the abscissa. For each value of a, non-transient values of a time series are plotted along the ordinate. To the left of the figure, the time series converges to a single stable limit point. Somewhere to the right, this limit point becomes unstable, and the limiting behavior consists of a cycle of period 2. Moving further to the right - that is, increasing a, limit cycles of period 4, 8, 16, etc. appear. The limit cycle of period 3 shown in Figure 2 corresponds to a parameter value of a somewhere to the center left of the region shown in the blown-up inset.

Figure 3: Structural Dynamics for the Logistic Equation

In some sense, this is recreational mathematics. Computers these days make it fairly easy to draw a more complete representation of Figure 4 in May (1976). The blow-up in Figure 3 demonstrates that the structural dynamics for the logistic function is fractal in nature. We see the same shape repeated on increasingly smaller and smaller scales. Chaos arises for parameter values of a between the period doubling cascade and the period-3 cycle. (Chaotic behavior is shown in Figure 3 by the dark shaded regions.)

3.0 A Exponential-Logistic Equation

I repeated the above analysis for what I am calling an exponential-logistic function:

f2(x) = (x/c) ea(1 - x), 0 < a

where:

c = 1, if a - ln a ≤ 1
c = -1/(a ea - 1), if 1 < a - ln a

This exponential-logistic function was suggested to me by a function in May (1976). I introduced the scaling provided by c such that the maximum value of this function never exceeds unity. This function, like the logistic function, is parametrized by a single parameter, which I am also calling a. Figure 4 shows the non-transient behavior for a specific value of the parameter a for the exponential-logistic function. In this case, a stable limit cycle of period 32 arises.

Figure 4: A Cycle with Period 32 for the Exponential-Logistic Equation

Notice the exponential-logistic function is generally not symmetric around any value of x; one tail is heavier than the other. Furthermore, it only has a zero at the origin; nothing corresponds to the zero at x = 1 in the logistic function. So, in some sense, it has a quite different form from the logistic function. Yet, as shown in Figure 5, the structural dynamics for iterative processes for the exponential-logistic function are qualitatively similar to the structural dynamics arising from the logistic function. We see the same shapes in Figures 3 and 5, albeit distorted in some sense.

Figure 5: Structural Dynamics for the Exponential-Logistic Equation
4.0 A Feigenbaum Constant

I now report on some quantitative numerical experiments. Table 1, in the second column, shows the smallest value of the parameter a for which I was able to find a limit cycle of the given period for the logistic equation. Cycles of an infinite number of periods - that is, for all positive integer powers of two (2, 4, 8, 16, ...) - exist in the period-doubling region labeled in Figure 3. As suggested by Table 2, the distance between values of a at which period-doubling occurs gets smaller and smaller. In fact all these limit cycles arise before a = 3.5700..., the point of accumulation of a at which chaos sets in. (I do not fully understand the literature on how to calculate the period of limiting cycles for a. I therefore do not report values of a for larger periods than shown in the table, since I do not fully trust my implementation of certain numeric methods.)

Table 1: Period Doubling in the Logistic Equation
PeriodaDifferenceRatio
22.9999078
43.44945770.4495504.7510
83.54407890.0946214.6556
163.56440290.0203244.6669
323.56875790.0043554.6665
643.56969110.0009334.6666
1283.56981090.000200

Table 1, above shows, in the third column, the difference between values of a at which period-doubling occurs. The fourth column shows the ratio of successive difference. Theoretically, this ratio converges to δ = 4.669201609... My numeric exploration has found this constant to at least two significant figures.

The convergence of this ratio, over limit cycles for periods of powers of two, is not limited to the logistic equation. Table 2 reports the result of a numeric experiment with the exponential-logistic equation. Here too, the constant δ has been found to two significant figures. Interestingly, the ratio would theoretically converge to the same constant if the two tables were infinitely extended. In fact, δ is a universal mathematical constant, like π or e.

Table 2: Period Doubling in the Exponential-Logistic Equation
PeriodaDifferenceRatio
22.7180077
44.60167401.8836662.9506
85.24007860.6384054.2473
165.39038560.1503074.5651
325.42331070.0329254.6456
645.43039820.007087
5.0 Conclusion

The above analysis can be generalized to many other functions, albeit I do not fully understand how to characterize the class of such functions. Feigenbaum states that the period doubling route to chaos is not limited to one-dimensional processes. I believe it also arises in continuous time systems, as defined by certain non-linear differential equations. Do you find it surprising that a universal constant with wide applicably to physical processes (like year-by-year changes in the population demographics of certain species), has been discovered in the lifetime of many now alive?

References
  • Keith Briggs (1991). A Precise Calculation of the Feigenbaum Constants, Mathematics of Computation, V. 57, No. 195 (July): pp. 435-439.
  • Mitchell J. Feigenbaum (1980). Universal Behavior in Nonlinear Systems, Los Alamos Science (Summer): pp. 4-27.
  • Tien-Yien Li & James A. Yorke (1975). Period Three Implies Chaos, American Mathematical Monthly, V. 82, No. 10 (Dec.): pp. 985-992.
  • Robert M. May (1976). Simple Mathematical Models with Very Complicated Dynamics, Nature, 261: pp. 459-467.

Tuesday, December 17, 2013

Purge At Amsterdam?

I have noticed that the recent history of economics has been impacted by various purges in various prominent economics departments. I think of, for example, Harvard, Rutgers, and Notre Dame. I had not noticed this one when it was going on:

"For most of my time over ten years at the University of Amsterdam my research and that of my colleagues was strongly supported. (I taught three courses every second fall term, and took leave from Marquette.) Unfortunately over the last two years people in leadership positions there at the faculty of economics decided that the history and methodology of economics (HME) was not important, and in conditions of a financial emergency associated with chronic budget shortfalls closed down the HME group. That included sacking my very accomplished and, in our field, well-respected colleagues Marcel Boumans and Harro Maas, who had been associate professors there for many years, and ending the chair position in HME, which I held, which had been at the faculty for decades. We had six courses in the history and methodology of economics; engaged and enthusiastic students; a research group of up to a dozen people; a master degree in HME; PhD students; and a required methodology course for bachelor students. I do not think there was a better program in the world in our field. We also had great interaction with the London School of Economics, the history of economics people at Duke University, history of economics people in Paris, and the Erasmus Institute for Philosophy and Economics. The HME group was internationally recognized, and attracted students from across the world. Our financial footprint, in fact, was quite small compared to other groups, and by a number of measures of output per person we were more productive than many other research groups at Amsterdam.

Since I fully believe the faculty financial emergency could have been addressed without eliminating the group, I can only put what happened down to prejudice against our field, plus the usual on-going territorial aggrandizing that has been a key factor in the elimination of history of economics from most American universities. It is interesting to me also, that with a few exceptions, members of the economics faculty at Amsterdam made no effort on the HME group’s behalf to resist what happened or even personally expressed regret or concern to those who lost their jobs. I find this reprehensible.

The loss of this program was a blow to our field. There are now few places in the world training PhD students in history and/or methodology of economics. So in the final analysis the situation for economics and philosophy is mixed: considerable achievement with an uncertain future. Great weight, in my view, should be placed on restoring PhD training in the field, something that is being done, for instance, through generous grants from the Institute for New Economic Thinking at Duke University under Bruce Caldwell." -- John B. Davis (2012). Identity Problems: An interview with John B. Davis, Erasmus Journal for Philosophy and Economics, V. 5, Iss. 2 (Autumn): pp. 81-103.

Wednesday, December 11, 2013

Reminder: Reductionism Is Bad Science

I want to remind myself to try to download the following in a couple of weeks:

Abstract: Causal interactions within complex systems can be analyzed at multiple spatial and temporal scales. For example, the brain can be analyzed at the level of neurons, neuronal groups, and areas, over tens, hundreds, or thousands of milliseconds. It is widely assumed that, once a micro level is fixed, macro levels are fixed too, a relation called supervenience. It is also assumed that, although macro descriptions may be convenient, only the micro level is causally complete, because it includes every detail, thus leaving no room for causation at the macro level. However, this assumption can only be evaluated under a proper measure of causation. Here, we use a measure [effective information (EI)] that depends on both the effectiveness of a system’s mechanisms and the size of its state space: EI is higher the more the mechanisms constrain the system’s possible past and future states. By measuring EI at micro and macro levels in simple systems whose micro mechanisms are fixed, we show that for certain causal architectures EI can peak at a macro level in space and/or time. This happens when coarse-grained macro mechanisms are more effective (more deterministic and/or less degenerate) than the underlying micro mechanisms, to an extent that overcomes the smaller state space. Thus, although the macro level supervenes upon the micro, it can supersede it causally, leading to genuine causal emergence—the gain in EI when moving from a micro to a macro level of analysis. -- Erik P. Hoel, Larissa Albantakis, and Giulio Tononi (2013). Quantifying Causal Emergence Shows that Macro Can Beat Micro, Proceedings of the National Academy of Sciences, V. 110, no. 49.

As far as I can tell, the above article is not specifically about economics. I do not understand download policy for the Proceedings of the National Academy of Sciences. I gather that you must be registered to download articles from the current issues, but can download back issues with no such restriction.

Hat Tip: Philip Ball

Monday, December 09, 2013

Honest Textbooks

1.0 Introduction

Every teacher, I guess, of an introductory or intermediate course has a struggle with how to teach material that requires more advanced concepts, outside the scope of the course, to fully understand. I think it would be nice for the textbooks not to foreclose the possibility of pointing out this requirement. I here provide a couple of examples from some mathematics textbooks that I happen to have.

2.0 Probability

Hogg and Craig (1974) is a book on probability and statistics. I have found many of their examples and theorems of use in a wide variety of areas. They usually do not explain how many of these ideas can be expanded to an entire applied course.

2.1 Borel Sets and Non-Measurable Sets

An axiomatic definition of probability is a fundamental concept for this book. Hogg and Craig recognize that they do not give a completely rigorous and general definition:

Let C denote the set of every possible outcome of a random experiment; that is, C is the sample space. It is our purpose to define a set function P(C) such that if C is a subset of C, then P(C) is the probability that the outcome of the random experiment is an element of C...

Definition 7: If P(C) is defined for a type of subset of the space C, and if,

  • P(C) ≥ 0,
  • Let C be the union of C1, C2, C3, ... Then P(C) = P(C1) + P(C2) + P(C3) + ..., where the sets Ci, i = 1, 2, 3, ..., are such that no two have a point in common...
  • P(C) = 1,

then P(C) is called the probability set function of the outcome of the random experiment. For each subset C of C, the number P(C) is called the probability that the outcome of the random experiment is an element of the set C, or the probability of the event C, or the probability measure of the set C.

Remark. In the definition, the phrase 'a type of subset of the space C' would be explained more fully in a more advanced course. Nevertheless, a few observations can be made about the collection of subsets that are of the type... -- Hogg and Craig (1974): pp. 12-13 (Notation changed from original).

2.2 Moment Generating and Characteristic Functions

Hogg and Craig work with moment generating functions throughout their book. In the chapter in which they introduce them, they state:

Remark: In a more advanced course, we would not work with the moment-generating function because so many distributions do not have moment-generating functions. Instead, we would let i denote the imaginary unit, t an arbitrary real, and we would define φ(t) = E(eitX). This expectation exists for every distribution and it is called the characteristic function of the distribution...

Every distribution has a unique characteristic function; and to each characteristic function there corresponds a unique distribution of probability... Readers who are familiar with complex-valued functions may write φ(t) = M(i t) and, throughout this book, may prove certain theorems in complete generality.

Those who have studied Laplace and Fourier transforms will note a similarity between these transforms and [the moment generating function] M(t) and φ(t); it is the uniqueness of these transforms that allows us to assert the uniqueness of each of the moment-generating and characteristic functions. -- Hogg and Craig (1978): pp. 54-55.

3.0 Fourier Series

Lin and Segel (1974) provides a case study approach to applied mathematics. They introduce certain techniques and concepts in the course of specific problems. Fourier analysis is introduced in the context of the heat equation. They then look at more generals aspects of Fourier series and transforms. They state:

Suppose that we now pose the following problem, which can be regarded as the converse to Parseval's theorem. Given a set of real numbers a0, am, bm, m = 1, 2, ..., such that the series

(1/2) a02 + {[a12 + b12] + [a22 + b22] + ...}

is convergent, is there a function f(x) such that the series

(1/2) a0 + {[a1cos(x) + b1sin(x)]
+ [a2cos(2x) + b2sin(2x)] + ...}

is its Fourier series?

An affirmative answer to this question depends on the introduction of the concepts of Lebesque measure and Lebesque integration. With these notions introduced, we have the Riesz-Fisher theorem, which states that (i) the [above] series ... is indeed the Fourier series of a function f, which is square integrable, and that (ii) the partial sums of the series converge in the mean to f.

The problem we posed is a very natural one from a mathematical point of view. It appears that it might have a simple solution, but it is here that new mathematical concepts and theories emerge. On the other hand, for physical applications, such a mathematical question does not arise naturally. -- C. C. Lin & L. A. Segel (1974): p. 147 (Notation changed from original).

4.0 Discussion

Here is a challenge: point out such candid remarks in textbooks in your field. I suspect many can find such comments in many textbooks. I will not be surprised if some can find some in mainstream intermediate textbooks in economics. Teaching undergraduates in economics, however, presents some challenges and tensions. I think of the acknowledged gap between undergraduate and graduate education. Furthermore, I think some tensions and inconsistencies in microeconomics cannot be and are never resolved in more advanced treatments. Off the top of my head, here are two examples.

  • The theory of the firm requires the absence of transactions costs for perfect competition to prevail. But under the conditions of perfect competition, firms would not exist. Rather workers would be independent contractors, forming temporary collectives when convenient.
  • Under the theory of perfect competition, as taught to undergraduates, firms are not atomistic. Thus, when taking prices as given, the managers are consistently wrong about the response of the market to changes they may each make to the quantity supplied. On the other hand, when firms are atomistic and of measure zero, they do not produce at the strictly positive finite amount required by the theory of a U-shaped average cost curve.

My archives provide many other examples of such tensions, to phrase it nicely.

References
  • Robert V. Hogg & Allen T. Craig (1978). Introduction to Mathematical Statistics, 4th edition, MacMillan.
  • C. C. Lin & L. A. Segel (1974). Mathematics Applied to Deterministic Problems in the Natural Sciencs, Macmillan

Monday, December 02, 2013

A New Order In Economics From Manchester

Some links:

By the way, Ian Steedman, a leading Sraffian economist, was at the University of Manchester not too long ago. And, I believe, he did supervise a number of doctorate theses from students at Manchester. So the closing of the doors to form the current monoculture happened only over the last decade, I guess.

Update: Originally posted on 6 November 2013. Updated on 12 November 2013 to include more links.

Update: Updated on 2 December 2013 to include more links.

Friday, November 29, 2013

Who Wants To Be A Millionaire?

Current Dollars For Millionaires Of Various Eras

How much would you need to have today to live like a millionaire in 1920? Over 10 million dollars, according to the chart1. On the other hand, a millionaire in 1980 would only need a bit less than 1.2 million dollars today to live in comparable luxury.

Footnote
  1. Obviously, any comparison of income over such a long period can be only a rough and ready guide, not an exact function yielding precise results suitable for the application of the the differential calculus. The Bureau of Labor Statistics (BLS) provides information on how the Consumer Price Index (CPI) is computed and how the components in a typical consumption basket changes over time.

Saturday, November 23, 2013

On "Neoclassical Economics"

I just want to document here some usages of the phrase, "Neoclassical economics". I restrict myself to literature in which the use is in the nature of aside, including within critiques of mainstream and neoclassical economics. Such documentation can be multiplied indefinitely. I do not quote from literature (e.g., Colander, Holt, and Rosser 2004; Davis 2008; Lawson 2013; Lee and Lavoie 2012; Varoufakis 2012) which focuses on the meaning of the word and on the sociology of economics. I find it quite silly to attempt to refute a critique of mainstream economics by complaining, without any other argument, that the word 'neoclassical' appears in that critique.

"In the last dozen years what before was simply known as economics in the nonsocialist world has come to be called neo-classical economics. Sometimes, in tribute to John Maynard Keynes's design for government intervention to sustain purchasing power and employment, the reference is to Keynesian or neo-Keynesian economics. From being a general and accepted theory of economic behavior, this has become a special and debatable inter-pretation of such behavior. For a new and notably articulate generation of economists a reference to neoclassical economics has become markedly pejorative. In the world at large the reputation of economists of more mature years is in decline." -- John Kenneth Galbraith (1973).
"One further matter merits consideration before we get down to business. I often refer to neoclassical theory and I had better make clear what I do and do not mean by this designation. For present purposes I shall call a theory neoclassical if (a) an economy is fully described by the preferences and endowments of agents and by the production sets of firms; (b) all agents treat prices parametrically (perfect competition); and (c) all agents are rational and given prices will take that action (or set of actions) from amongst those available to them which is best for them given their preferences. (Firms prefer more profit to less.)" -- Frank Hahn (1984).

"Let us attempt to identify the key characteristics of neoclassical economics; the type of economices that has dominated the twentieth century. One of its exponents, Gary Becker (1967a, p. 5) identified its essence when he described 'the combined assumptions of maximizing behavior, market equilibrium, and stable preferences, used relentlessly and unflinchingly.' Accordingly, neoclassical economics may be conveniently defined as an approach which:

  1. Assumes rational, maximizing behaviour by agents with given and stable preference functions,
  2. Focuses on attained, or movements towards, equilibrium states, and
  3. Is marked by an absence of chronic information problems.

Point (3) requires some brief elaboration. In neoclassical economics, even if information is imperfect, information problems are typically overcome by using the concept of probabilistic risk. Excluded are phenomena such as severe ignorance and divergent perceptions by different individuals of a given reality. It is typically assumed that all individuals will interpret the same information in the same way, ignoring possible variations in the cognitive frameworks that are necessary to make sense of all data. Also excluded is uncertainty, of the radical type explored by Frank Knight (1921) and John Maynard Keynes (1936).

Notably, these three attributes are interconnected. For instance, the attainment of a stable optimum under (1) suggests an equilibrium (2); and rationality under (1) connotes the absence of severe information problems alluded to in (3). It can be freely admitted that some recent developments in modern economic theory - such as in game theory - reach to or even lie outside the boundaries of this definitions. Their precise placement will depend on inspection and refinement of the boundary conditions in the above clauses. But that does not undermine the usefulness of this rough and ready definition.

Although neoclassical economics has dominated the twentieth century, it has changed radically in tone and presentation, as well as in content. Until the 1930s, much neoclassical analysis was in Marshallian, partial equilibrium mode. The following years saw the revival of Walrasian general equilibrium analysis, an approach originally developed in the 1870s. Another transformation during this century has been the increasing use of mathematics, as noted in the preceding chapter. Neoclassical assumptions have proved attractive because of their apparent tractability. To the mathematically inclined economist the assumption that agents are maximizing an exogeneously given and well defined preference function seems preferable to any alternative or more complex model of human behaviour. In its reductionist assumptions, neoclassical economics has contained within itself from its inception an overly formalistic potential, even if this took some time to become fully realized and dominant. Gradually, less and less reliance has been placed on the empirical or other grounding of basic assumptions, and more on the process of deduction from premises that are there simply because they are assumed.

Nevertheless, characteristics (1) to (3) above have remained prominent in mainstream economics from the 1870s to the 1980s. They define an approach that still remains ubiquitous in the economics textbooks and is taught to economics undergraduates throughout the world." -- Geoffrey M. Hodgson (1988).

"The creators of the neoclassical model, the reigning economic paradigm of the twentieth century, ignored the warnings of nineteenth-century and still earlier masters about how information concerns might alter their analyses- perhaps because they could not see how to embrace them in their seemingly precise models, perhaps because doing so would have led to uncomfortable conclusions about the efficiency of markets. For instance, Smith, in anticipating later discussions of adverse selection, wrote that as firms raise interest rates, the best borrowers drop out of the market. If lenders knew perfectly the risks associated with each borrower, this would matter little; each borrower would be charged an appropriate risk premium. It is because lenders do not know the default probabilities of borrowers perfectly that this process of adverse selection has such important consequences." -- Joseph E. Stiglitz (2002).
References
  • David Colander, Richard P. F. Holt, and J. Barkley Rosser, Jr. (2004). The changing face of mainstream economics, Review of Political Economy, V. 16, No. 4: pp. 485-499.
  • John B. Davis (2008). The turn in recent economics and return of orthodoxy, Cambridge Journal of Economics, V. 32: pp. 349-366.
  • John Kenneth Galbraith (1973). Power and the Useful Economist, American Economic Review, Presidential address at the 85th annual meeting of the American Economic Association in Toronto, Canada in December 1972.
  • Frank Hahn (1982). The neo-Ricardians, Cambridge Journal of Economics, V. 6: pp. 353-374.
  • Tony Lawson (2013). What is this 'school' called neoclassical economics?, Cambridge Journal of Economics, 2013.
  • Fred Lee and Marc Lavoie (editors) (2012). In Defense of Post-Keynesian and Heterodox Economics: Responses to their Critics, Routledge. [I've not read the book, but have read some chapters published seperately.]
  • Geoffrey M. Hodgson (1999). False Antagonisms and Doomed Reconcilations, Chapter 2 in Evolution and Institutions: On Evolutionary Economics and the Evolution of Economics, Edward Elgar.
  • Joseph E. Stiglitz (2002). Information and the Change in the Paradigm in Economics, American Economic Review, V. 92, N. 3 (June): pp. 460-501.
  • Yanis Varoufakis (2012). A Most Peculiar Failure: On the dynamic mechanism by which the inescapable theoretical failures of neoclassical economics reinforce its dominance.

Tuesday, November 19, 2013

Thoughts On Davis' Individuals and Identity in Economics

I have previously gone on about multiple selves, also known as Faustian agents. I had not considered how an individual manages these selves in making plans and decisions. My point was to apply Arrow's impossibility theorem at the level of the single agent, thereby demonstrating the necessity of some argument for characterizing an individual by a utility function.

Consider many individuals interacting in a market, each being composed of multiple selves. What, in the analysis, groups together sets of these multiple selves to identify individuals? This problem, and similar problems with many other decision theory analyses, is the theme of John D. Davis' 2011 book, Individuals and Identity in Economics.

By the way, an interesting issue arises with multiple selves interacting through time. One might justify hyperbolic discounting by thinking of an individual as composed of a different self at each moment in time. Why should these selves make consistent plans? Might one self start an action based on a plan for future actions, only to have a future self revise or reject that plan? This is the third or fourth time I have started reading Davis' book. Anyways, on pages 41 and 42, Davis writes:

"...[Herbert] Simon's recommendation to abandon the standard utility function framework was not influential in economics, but Lichtenstein and Slovic's demonstration of preference reversals was. Most economists initially dismissed it on a priori grounds, but David Grether and Charles Plott believed that they could go farther and demonstrate that preference reversals could not possibly exist. They identified thirteen potential errors in psychologists' preference reversal experimental methodology and accordingly set out to show that preference reversals were only an artifact of experimental design. Nonetheless, they ended up confirming their existence as well as Simon's judgement of utility functions:

'Taken at face value the data are simply inconsistent with preference theory and have broad implications about research priorities in economics. The inconsistency is deeper than mere lack of transitivity or even stochastic transitivity. It suggests that no optimization principles of any sort lie behind the simplest of human choices and that uniformities in human choice behavior which lie behind market behavior may result from principles which are of a completely different sort from those generally accepted.'(Grether and Plott 1979, 623; emphasis added)

Published in the American Economic Review, this was a momentous admission for economists. However, for many psychologists the debate was already long over, and research had moved on to which theories best explained preference construction. James Bettman published what is regarded as the first theory of preference construction in the same year Grether and Plott's paper appeared (Bettman 1979), a major review of preference construction theories appeared in 1992 (Payne and Bettman 1992), and Lichtenstein and Slovic's retrospective volume appeared in 2006 (Lichtenstein and Slovic 2006). As Slovic put it in 1995, 'It is now generally recognized among psychologists that utility maximization provides only limited insight into the processes by which decisions are made' (Slovic 1995, 365). Grether and Plott, interestingly, extended their own critique of standard rationality to Kahneman and Tverky's proposed prospect theory replacement, implicitly highlighting the difference between the two currents in Edwards' B[ehavioral] D[ecision] R[esearch] program.

'We need to emphasize that the phenomenon causes problems for preference theory in general, and not for just the expected utility theory. Prospect theory as a special type of preference theory cannot account for the results.' (Grether and Plott 1982; 575)

So, given the data and what economists have said years ago about it in the most prominent and most prestigious economics journal in America, one can expect mainstream economists today to have rejected utility theory, revealed preference theory, prospect theory, and the usual old textbook derivation of market demand curves and factor supply curves. Right?

Saturday, November 09, 2013

Mainstream And Non-Mainstream Economics: Research Areas Transgressing The Boundaries

1.0 Introduction

Mainstream and non-mainstream economics can be read as sociological categories, defined by what conferences economists attend, in which journals they publish, and through patterns of referencing. One might expect the intellectual content of the theories put forth by mainstream and non-mainstream economists to cluster, too. In some sense, non-mainstream economists are also automatically heterodox, where heterodoxy refers to the content of theories. For example, heterodox economists tend to prefer theories in which agents are socially embedded and constituted, in some sense, by society (instead of being pre-existing, asocial monads).

The point of this post, though, is to illustrate that the boundary between mainstream and non-mainstream economists is not hard and fast, at least as far as ideas go. I point out two-and-a-half areas where both categories of economists are developing similar ideas.

2.0 Complex Economic Dynamics

Economic models are available which exhibit complex, non-linear dynamics, including chaos. Richard Goodwin, Steve Keen, and Paul Ormerod are some self-consciously non-mainstream heterodox economists who have developed such models. Jess Benhabib and John Geanakoplos are some authoritative mainstream economists on certain models of this type. I also want to mention some researchers who I do not feel comfortable putting in either category. As I understand it, J. Barkley Rosser, Jr. makes an effort to talk to both mainstream and non-mainstream economists. I do not know enough about, for example, Anna Agliari to say what she would say about these categories. And Donald G. Saari is a mathematician interested in social science; so I am not sure how these categories would apply to him, if at all.

3.0 Multiple Selves

I have previously commented on theories of multiple selves, also known as Faustian agents. I particularly like the conclusion, from the Arrow impossibility theorem, that an agent's preferences cannot necessarily be characterized by a utility function, given a theory of modular minds.

I do not think I know enough about these theories to talk authoritatively on this subject. Specifically, I have some dim awareness that a large literature exists here about time (in)consistency of decisions. But I am aware that this is a topic of research among both non-mainstream and mainstream economists. I cite John B. Davis, Ian Steedman, and Ulrich Krause as non-mainstream, heterodox economists with literature in this area. And I cite E. Glen Weyl as a mainstream economist also with literature here.

4.0 Choice Under Uncertainty

Keynes distinguished between risk and uncertainty. Post Keynesian economists have famously developed this theme. Works seen as part of mainstream economics in their time also distinguish between risk and uncertainty, for example:

"...Let us suppose that a choice must be made between two actions. We shall say that we are in the realm of decision making under:

  • Certainty if each action is known to lead invariably to a specific outcome...
  • Risk if each action leads to one of a set of possible outcomes, each outcome occurring with a known probability. The probabilities are assumed known to the decision maker...
  • Uncertainty if either action or both has as its consequences a set of possible specific outcomes, but where the probabilities of these outcomes are completely unknown or are not even meaningful."

-- R. Duncan Luce and Howard Raiffa, Games and Decisions, Harvard University (1957): p. 13.

I only feel entitled to count this as half an example. I find that other literature on the foundations of decision theory is also clear on assumptions about known outcomes and probabilities necessary to characterize a situation of risk. But I do not know of contemporary mainstream economists researching choice under uncertainty (as opposed to risk). I think elements of Chapter 13 of Luce and Raiffa, on decision making under uncertainty, has entered the teaching of business schools targeted towards, for example, corporate managers.

5.0 Reflections

I do not think that this post has demonstrated an openness in mainstream economics. Further work would need to show an awareness among mainstream researchers of parallel work by non-mainstream economists, a willingness to critically engage that work, and a willingness to cite it in mainstream literature. Furthermore, one would like to show that the implications of such work is transitioning into the teaching of economists at all levels. I have seen some economists verbally affirm that economies are complex dynamic systems and then ignore the implications of such a claim. Some economists - for example, Yanis Varoufakis - have expressed skepticism that cutting edge mainstream economics research, in which unique deterministic outcomes do not obtain, can be successfully transitioned. Nevertheless, I find the parallel research noted above to be intriguing.

Tuesday, October 29, 2013

Immanuel Kant, Crank

One intellectually bankrupt technique to rationalize non-engagement with an argument is to complain that the ones putting forth the argument compare themselves to Nicolaus Copernicus or Galileo:

"So the central laws of the movements of the heavenly bodies established the truth of that which Copernicus, first, assumed only as a hypothesis, and, at the same time, brought to light that invisible force (Newtonian attraction) which holds the universe together. The latter would have remained forever undiscovered, if Copernicus had not ventured on the experiment-contrary to the senses but still just-of looking for the observed movements not in the heavenly bodies, but in the spectator. In this Preface I treat the new metaphysical method as a hypothesis with the view of rendering apparent the first attempts at such a change of method, which are always hypothetical...

This attempt to introduce a complete revolution in the procedure of metaphysics, after the example of the geometricians and natural philosophers, constitutes the aim of the Critique of Pure Speculative Reason." -- Immanuel Kant, Critique of Pure Reason, Preface to Second Edition.

As I understand it, Kant's revolution was not to look for what must humans be to understand an external, given, phenomenal reality. Rather, he asked, what must the phenomena be such that we could observe it, given the properties of our understanding. But I have never read far in this particular book. Nor, given non-Euclidean geometry and other readings in philosophy, do I expect to agree with it.

Saturday, October 26, 2013

An Alternative Economics

Who are the greatest economists of, say, the last fifty years? I suggest the shortlist for many academic economists would include Kenneth Arrow, Milton Friedman, and Paul A. Samuelson. Imagine1 a world in which the typical academic economist would be inclined to name the following as exemplars:

  • Fernand Braudel
  • John Kenneth Galbraith
  • Albert Hirschman
  • Gunnar Myrdal
  • Karl Polanyi.

These economists did not insist on using closed, formal, mathematical models2 everywhere3. They tended to present their findings in detailed historical and qualitative accounts4.

Footnotes
  1. Maybe the history of political economy is overdetermined, in some sense. So I am not sure what this counterfactual would entail.
  2. They did use or, at least, comment on small mathematical models, where appropriate. For example, both Hirshman and Myrdal had cautions about the Harrod-Domar model.
  3. I have been reading a little of Tony Lawson.
  4. As far as I am concerned, these are accounts of empirical research.

Friday, October 25, 2013

Raj Chetty Needs Your Help

I have argued before that whether or not economics is a science is an uninteresting question. Rather, one should be concerned with the quality of arguments economists put forth, how they engage one another, and how they respond to empirical evidence. As regard to the quality of arguments, Raj Chetty's performance contradicts his thesis. (I realize I am talking about an article meant for the general public, not a journal article meant for other members of his profession). Chetty writes:

"...At first blush, Mr. Shiller's thinking about the role of 'irrational exuberance' in stock markets and housing markets appears to contradict Mr. Fama's work showing that such markets efficiently incorporate news into prices.

What kind of science, people wondered, bestows its most distinguished honor on scholars with opposing ideas? 'They should make these politically balanced awards in physics, chemistry and medicine, too,' the Duke sociologist Kieran Healy wrote sardonically on Twitter.

But the headline-grabbing differences between the findings of these Nobel laureates are less significant than the profound agreement in their scientific approach to economic questions, which is characterized by formulating and testing precise hypotheses. I'm troubled by the sense among skeptics that disagreements about the answers to certain questions suggest that economics is a confused discipline, a fake science whose findings cannot be a useful basis for making policy decisions.

That view is unfair and uninformed. It makes demands on economics that are not made of other empirical disciplines, like medicine, and it ignores an emerging body of work, building on the scientific approach of last week's winners, that is transforming economics into a field firmly grounded in fact.

It is true that the answers to many 'big picture' macroeconomic questions — like the causes of recessions or the determinants of growth — remain elusive. But in this respect, the challenges faced by economists are no different from those encountered in medicine and public health. Health researchers have worked for more than a century to understand the 'big picture' questions of how diet and lifestyle affect health and aging, yet they still do not have a full scientific understanding of these connections. Some studies tell us to consume more coffee, wine and chocolate; others recommend the opposite. But few people would argue that medicine should not be approached as a science or that doctors should not make decisions based on the best available evidence..." -- Raj Chetty, Yes, Economics Is a Science, New York Times, 21 October 2013: p. A21.

Chetty's argument is chalk and cheese, apples and oranges. Chetty brings up the existence of open questions in medicine. (I find it easy to decide which studies about health to believe; I pick the ones that tell me to do what I want to do anyways.) But the complaint about this year's Nobel is not about the existence of open questions in economics. It is rather wonder at the simultaneous award of the top prize to two, namely Eugene Fama and Robert Shiller, for apparently making opposite statements about how markets work. (I am ignoring Lars Peter Hansen in this post.) As I understand it, you do not usually get a Nobel in Physics, Chemistry, or Medicine without empirical evidence for your claims. Later empirical findings might overturn Nobel work, but, if so, the prize would not be shared in the same year for both establishing and overturning a theory. So Chetty's argument is a blatant non-sequitur.

What is needed is an argument why the specific work of the researchers sharing the award is complementary, not contradictory. You can find some such arguments on the Internet, but Chetty skates right by this need at too high a level of abstraction to be useful. (Readers should feel free to offer other links in the comments for such arguments.) I do not think this analogy works in the details, but imagine a prize at the relevant time simultaneously honoring Tycho Brahe and Johannes Kepler. Tycho had a mistaken halfway, sort of heliocentric model of solar system, and Kepler had a correct, geocentric model with the planets having elliptic orbits. But Tycho would not be being honored for his mistaken model. Rather, as I understand it, he collected more useful and precise observations of the planets than had been done by anybody before him. And Kepler used these observations to develop and verify his theory. If Tycho's work was not available, Kepler would not have succeeded.

(Note that work by economists with Randomized Control Trials (RCTs), natural experiments, and Instrumental Variables (IVs) is irrelevant to any points I have made above.)

Off topic aside: I want to note this post from Unlearning Economics, the Post-Crash Economics Society at the University of Manchester, an article about this society in The Guardian, and an organization called Rethinking Economics.

Saturday, October 19, 2013

Some Aspects Of A Theory Of Finance

I consider this post to be a complement to some recent comments by Lars Syll on this year's Nobel prize. I here outline an (unoriginal) theory to contrast with the Efficient Market Hypothesis.

Keynes described the prices of bonds, shares, and other financial instruments as, at any time, reflecting a balance of Bulls and Bears. As G. L. S. Shackle points out, if this is an equilibrium, it is an "inherently restless" equilibrium. If a price continues unchanged, Bulls, who expect the price to rise, will eventually be disappointed. Likewise, a symmetrical condition is true for Bears. Furthermore, news from outside the stock market will be changing market expectations, causing some Bulls to become Bears and vice versa.

For this account to make sense, it seems to me, the market must be populated with decision makers who have vastly different ontological and epistemological beliefs. It is not a case of all agents having one model, with different parameter estimates being updated in light of experience. Shackle, in modeling such decision makers, introduces the concept of focal points. If profits exceed the upper focus point at some time, the decision maker will realize that more profit can be gained than is accounted for in his theory. And so he will adopt some other theory, in some sense. Likewise, losses or too little profit will, outside of some range, lead to another change of mind, with consequent changes in plans and actions.

Stock prices are numbers. One might initially think they could be described by probability distributions. One could think of the price of a stock at a given time as a random variable. And, in this special case, perhaps Shackle's model of decision-making reduces to the theory of sequential statistical hypothesis testing.

A random variable is a function from a sample space to the real numbers, where a sample space is a set of all possible outcomes of a random experiment. An event is a (measurable) subset of the sample space. So a stock price is not an event in the sample space. For example, the recent brouhaha in the United States over the debt limit is presumably an aspect of an event that has had some impact on stock prices. Can one assume that all possible events are known beforehand? That all agents who might bet on stock prices know all events? If not, how does it make sense to model stock prices as coming from a probability distribution? Paul Davidson suggests extending the concept of non-ergodicity to cover these cases where the complete sample space is not known, and the mere possibility of some future events are capable of surprising someone:

"In expected utility theory, according to Sugden..., 'A prospect is defined as a list of consequences with an associated list of probabilities, one for each consequence, such that these probabilities sum to unity. Consequences are to be understood to be mutually exclusive possibilities: thus a prospect comprises an exhaustive list of the possible consequences of a particular course of action... An individual's preferences are defined over the set of all conceivable prospects.' Using these definitions, an environment of true uncertainty (that is, one which is nonergodic) occurs whenever an individual cannot specify and/or order a complete set of prospects regarding the future, because the decision maker either cannot conceive of a complete list of consequences that will occur in the future; or cannot assign probabilities to all consequences because 'the evidence is insufficient to establish a probability' so that possible consequences 'are not even orderable' (Hicks)... Hicks associates a violation of the ordering axiom of expected utility theory with 'Keynesian liquidity' ..., since, for Hicks..., 'liquidity is freedom' to delay action that commits claims on real resources whenever the decision maker is ignorant regarding future consequences." -- Paul Davidson (1991, p. 134)

In the above quote, Davidson is contrasting what he calls the True Uncertainty Environment with the Subjective Probability Environment. Davidson argues that my favorite definition of ergodicity, as applying to stochastic processes in which time averages converges to ensemble averages, characterizes the Objective Probability Environment.

One might argue that a more mainstream approach to finance is more empirically applicable than the Post Keynesian approach outlined above. It seems to me that if one wants to argue this, one needs to establish some reason why a diversity of agent views could not persist. I am unaware of any such argument, and I think the literature on noise trading should lean economists to think otherwise.

Traders in stock and bonds constitute one audience for theories of finance. I suppose they prefer mathematical theories that they can use to price financial instruments. I am not sure that the above theory can be cast into mathematics further than Davidson and Shackle have already accomplished. I will note that when Dutch Shell found themselves wrong-footed in the 1970s by the formation of OPEC and the oil crisis, they set up a group to do scenario planning. These are the sort of events whose existence in the sample space might not have been long foreseen. As I understand it - I forget why, some members of that group explicitly took inspiration from Shackle's work.

References
  • Davidson, Paul (1991). Is Probability Theory Relevant for Uncertainty? A Post Keynesian Perspective, Journal of Economic Perspectives, V. 5, No. 1: pp. 129-143.
  • DeLong, J. Bradford, Andrei Shleifer, Lawrence H. Summers, and Robert Waldmann (1990). Noise Trader Risk in Financial Markets, Journal of Political Economy, V. 98, Iss. 4: pp. 703-738.
  • Hogg, Robert V. and Allen T. Craig (1978). Introduction to Mathematical Statistics, Fourth edition, Macmillan.
  • Shackle, G. L. S. (1940). The Nature of the Inducement to Invest, Review of Economic Studies, V. 8, N. 1: pp. 44-48.
  • Shackle, G. L. S. (1988). Treatise, Theory, and Time, in Business, Time and Thought: Selected Papers of G. L. S. Shackle, New York University Press.
  • Shackle, G. L. S. (1988). On the Nature of Profit, in Business, Time and Thought: Selected Papers of G. L. S. Shackle, New York University Press.

Friday, October 04, 2013

Thou Shall Listen To Economists: The Number One Principle Advocated By Economists

Alex Tabarrok, as I understand it, is an economist in the economics department at George Mason University. Brad DeLong has been famously feuding with David Graeber, an anthropologist and anarchist (also known as - to those who speak english - a libertarian), but not an economist. Would it be more or less to DeLong's credit for that background to explain this agreement with Tabarrok's incompetent stupidity? I am not going to link to any of my multiple demonstrations that there is no such thing as a marginal productivity theory of distribution. Not only do I have mathematical proofs, I am in agreement with what good neoclassical economists, from Leon Walras to Christopher Bliss, have explicitly stated.

I am not sure what I think of this article, by Peter Frase, in Jacobin. I am in agreement with the conclusion. I am also in agreement that, in practice, actually existing capitalism contains lots of markets that are not competitive1, in some sense. But I wish more would note that even if markets were competitive, the conclusions of the market fundamentalists would not follow2. It is a major error of contemporary mainstream economics to think that everything would be great if only markets were competitive3.

Footnote
  1. At least two theories of competitive markets exist: (1) The classical theory, in which competitive markets exhibit no barriers to entry, and a tendency exists, across and within industries, towards a uniform rate of profits. (2) The neoclassical theory, in which firms take market prices as given and, perhaps, do not suffer from any problems connected with limits to information, computation, etc.
  2. For example, Piero Sraffa can be read as putting forth a model, in his 1960 book, in which markets are competitive and in which exogenous political forces set the functional distribution of income.
  3. Different schools argue over how important, in practice, deviations from competition are and whether governments can correct these deviations.

Saturday, September 28, 2013

What's Bayesianism Got To Do With It? A Software Reliability Example

Figure 1: The Musa-Okumoto Model Fit To Failure Data
1.0 Introduction

I was partly inspired for this post by a dispute between Matheus Grasselli and Philip Pilkington over the applicability of Bayesian statistics to non-ergodic processes. I happen to know of certain parametric software reliability models that describe software failures as non-stationary stochastic processes. Bev Littlewood is a prominent proponent of Bayesian models in software reliability engineering. And he has also recommended a certain approach to accessing the goodness of fit for software reliability models, including ones not originally put forth as Bayesian models.

Here I provide an example of an application of this approach for assessing goodness of fit. I do not see what Bayesianism has to do with this approach. It is not expounded in terms of prior and posterior distributions. I do see that this approach considers the model(s) being assessed as part of a larger system for making predictions, with the parameter estimates on which those predictions are based being updated with each new observation. Is this approach, therefore, Bayesian?

2.0 Failure Data

The data used for illustration here consists of 367 failures observed over 73 weeks, and they were provided by the Jet Propulsion Laboratory (JPL). The system under test was:

"A facility for tracking and acquiring data from earth resource satellites in high-inclination orbits... About 103,000 uncommented Source Lines Of Code in a mixture of C, Fortran, EQUEL, and OSL... Failure data were obtained from the development organization's anomaly reporting system during software integration and test." -- Nikora and Lyu (1996).

The blue line in Figure 1 above represents the data. This is a step function, where each step occurs after an interval of one week. The height of each step represents the number of failures observed in the week ending at the start of the step. Figure 1 also shows a point estimate of the mean value function, and 90% confidence intervals, based on fitting the Musa-Okumoto software reliability model to the data.

3.0 The Musa-Okumoto Model

As in other software reliability models, the Musa-Okumoto model relies on a distinction between failures and faults. A failure is behavior, including unexpected termination, observed while the software system under test is running. A failure occurs when the system is observed to behave other than specified. A fault, also known as a bug, is a static property of the system. A failure occurs when the inputs to the system cause altered behavior from the software entering a state in which a bug is tripped.

The Musa-Okuomoto model is an example of a software reliability model based on the assumption that the system under test contains an infinite number of faults. No matter how many faults have been found and removed, more faults always exist. Faults found later, however, tend to have an exponentially decreasing impact on the failure rate. More formally, the assumptions of the model are:

  1. The numbers of failures observed in non-overlapping time intervals are stochastically independent.
  2. The probability of one failure occurring in a small enough time interval is roughly proportional to the length of the time interval.
  3. The probability of more than one failure occurring in a small enough time interval is negligible.
  4. Detected faults are immediately corrected.
  5. No new faults are introduced during the fault-removal process.
  6. Aside from such repair of faults, the software is unchanged during test. No new functionality is introduced.
  7. Testing is representative of operational usage.
  8. The failure rate decreases exponentially with the number of removed faults.

The first three assumptions imply that failures occur as a Poisson Process, where a Poisson Process describes events happening randomly in time, in some sense. The sixth and seventh assumptions are unlikely to be met for the data being analyzed here. For best use of these kinds of software reliability models, system test should be designed to randomly select inputs from a specified operational profile. Nevertheless, these sorts of models have often been applied to test data observed naturally. The remaining assumptions imply that failures occur as a Non-Homogeneous Poisson Process (NHPP), where the failure rate is the following function of the mean value function:

λ(t) = λ0 em(t)

where:

  • λ(t) is the failure rate, in units of failures per week. (The distinction between the failure rate and the failure intensity is moot for Poisson Processes.)
  • m(t) is the mean value function, that is, the number of failures expected by week t.
  • λ0 is the initial failure rate.
  • θ is the failure rate decay parameter.

Figure 2 plots the above function for the data for the system under test, with parameter estimates being based on all of the data. The mean value function is plotted with the data along the abscissa for each week by taking the cumulative number of failures observed by the end of that week. The corresponding ordinate value for the data is the quotient of the number of failures observed in each time interval (namely, a week) and the length of the time interval. The graph shows quite a bit of variation on this scale, clustered around the model estimate. A slight improvement in software reliability is noticeable, but the convexity in the estimate from the model is hard to see.

Figure 2: The Failure Rate As A Function Of The Mean Value Functon

4.0 The U-Test: How Am I Doing

An attempt was made to estimate the model parameters after each week. For this model, maximum likelihood estimates are found by an iterative algorithm. The iterative algorithm did not converge at the end of week 49, leaving the twenty four parameter estimates shown in the table in Figure 3. Figure 3 also shows percent variability of the parameter estimates. (Percent variability is the ratio of the absolute difference of the estimates for two successive weeks to the estimate for the earlier of the two weeks.) The model estimates seem to have settled down towards the end of test period. They are unlikely to vary much over succeeding weeks if testing is to continue.

Figure 3: Percent Variability In Parameter Estimates

In the Musa-Okumoto model, the number of failures observed in a given week is a realization of a random variable from a Poisson distribution. The mean of this distribution, for given model parameters, varies with the week. The expected number of failures declines in later weeks. In the approach illustrated here, the model parameters are re-estimated in each successive week. Specifically, define ui; i = 1, 2, ..., 23; as the value of the CDF for the number of failures in week 50 + i, where:

  • The CDF is evaluated for the number of failures actually observed in the given week.
  • The CDF is calculated with estimates of the model parameters based on the number of failures observed in all weeks up to, but not including the given week.

For example, u1 is an estimate of the CDF for the number of failures found in week 51, based on parameter estimates found with the number of failures observed in weeks 1 through 50. This procedure yields a sequence of (temporally ordered) numbers:

u1, u2, ..., u23

Each ui, being a probability, is between zero and one. These numbers, in effect, provide a measure of how well the model, with its estimation procedure, is predicting failures of the system under test. Under the null hypothesis that the model describes the test process generating the failure data, the sequence of uis is the realization of a random sample from a probability distribution uniformly distributed on the interval [0, 1].

The Kolmogorov-Smirnov statistic provides a formal test that the uis are realizations of identical and independently distributed random variables from a uniform distribution. Figure 4 compares and contrasts the CDF from a uniform distribution with the empirical probability distribution found from the data. The theoretical distribution for the uniform distribution is merely a 45 degree line segment sloping upward from the origin to the point (1, 1). That is, the probability that a realization of a random variable uniformly distributed on [0, 1] will be less than some given value u is merely that given value.

Figure 4: The U-Test For The Example

The empirical CDF found from the Musa-Okumoto model applied to the data is a step function. The probability that a random variable from this distribution will be less than a given value u is the proportion of uis in the data less than that given value. The exact location of the step function is found as follows. First, sort the sample values in increasing order:

u(1)u(2) ≤ ... ≤ u(23)

The sorted values are known as order statistics. (Note the temporal order is lost by this sorting; this test does not reject the model if, for example, overestimations of failures in earlier weeks average out underestimations of failures in later weeks.) For this data, u(1) is approximately 0.007271, and u(23) is approximately 0.939. Plot these order statistics along the abscissa. A step occurs at each plotted order statistic, and each step is, in this case, of size 1/(23 + 1) (approximately 0.042) up.

The Kolmogorov-Smirnov statistic is the maximum vertical distance between the theoretical CDF, specified by the null hypothesis, and the empirical CDF, as calculated from the data. In the figure, this maximum vertical distance is 0.18869, found for the second step up, at u(2) (which is approximately 0.23). The p-value of 34.2% is the probability that the empirical CDF, for a random sample of 23 points drawn from a uniform sample, will be at least this far above or below the empirical CDF. Since the p-value exceeds the pre-selected conventional level of statistical significance of 10%, the data does not allow one to conclude that the model does not fit the data. (One should set a fairly low level of statistical significance so as to increase the power of the test.) Thus, one accepts that the Musa-Okumoto model is successful in predicting failures in weeks one step ahead.

5.0 Conclusions

The above analysis suggests that, despite the violation of model assumptions in the test process, JPL could have used the Musa-Okumoto model as an aid in managing this test phase of the tracking and data acquisition system considered here.

In what sense, if any, is the method illustrated here for analyzing the model's goodness-of-fit Bayesian?

References
  • Sarah Brocklehurst and Bev Littlewood (1996). Techniques for Prediction Analysis and Recalibration, in Handbook of Software Reliability (edited by M. R. Lyu), McGraw-Hill.
  • Grasselli, Matheus (2013). Keynes, Bayes, and the law.
  • Nikora, Allen P. and Michael R. Lyu (1996). Software Reliability Measurement Experience, in Handbook of Software Reliability (edited by M. R. Lyu), McGraw-Hill.
  • Pilkington, Philip (2013). Holey Models: A Final Response to Matheus Grasselli.

Wednesday, September 25, 2013

S. Abu Turab Rizvi, Not Steve Keen

Notice the point about the need in neoclassical theory to assume that preferences satisfy Gorman's assumptions of identical and homothetic (non-varying with income) preferences:

"Extensions of the basic arbitrariness results to configurations of preferences and endowments which are in no way 'pathological', and are in fact more and more restrictive, indicate the robustness of S[onnenschein-]M[antel-]D[ebreu] theory.

For instance, an assumption which is often made to improve the chances of meaningful aggregation is that of homothetic preferences, which yields linear Engel curves and so no complicated income effects at the level of the individual. However, with only a slight strengthening of the requirements of Debreu's theorem, Mantel (1976) showed that the assumption of homotheticity does not remove the arbitrariness of A[ggregate] E[xcess] D[emand functions].

Moreover, the possibility that consumers have to be very different, or that unusual endowment effects need to take place, in order for SMD-type results to hold was refuted by Kirman and Koch (1986): even supposing consumers to have identical preferences and collinear endowments does not remove the arbitrariness of the AEDs. Of course, if preferences are simultaneously identical and homothetic, AED is a proportional magnification of individual excess demand (Gorman, 1953; Nataf, 1953) and the whole economy behaves as if it were an individual (obeying the Weak Axiom of Revealed Preference in the aggregate), but this is an extremely special situation. The Mantel and Kirman-Koch theorems effectively countered the criticism of SMD theory raised by Deaton by showing that the primitives can be arrayed in ways which on the face of it are very congenial towards generating well-behaved results, yet the arbitrariness property of AEDs remains." -- S. Abu Turab Rizvi, "The Microfoundations Project in General Equilibrium Theory", Cambridge Journal of Economics, V. 18 (1994): p. 362.

In short, neoclassical economists have proved (by contradiction, in some sense) that neoclassical microeconomics is not even wrong and that methodological individualism has failed.

Monday, September 23, 2013

On The Ideological Function Of Certain Ideas Of Friedman And Barro

A Simple Keynes-Like Model
1.0 Introduction

I want to comment on an ideology that would lead to an acceptance of:

  • Milton Friedman's Permanent Income Hypothesis (PIH)
  • Robert Barro's so-called theory of Ricardian Equivalence

My claim is that Friedman and Barro were each responding, in their own way, to the (policy implications suggested by) Keynes' General Theory. So first, I outline, very superficially, some ideas related to the General Theory. I then briefly describe how Friedman and Barro each tried to downplay these ideas, before finally concluding.

I have a number of inspirations for this post, including Robert Waldmann's assertion that the denial of the PIH is consistent with the data; Brad DeLong noting Simon Wren-Lewis and Chris Dillow commenting on the incompetence of, say, Robert Lucas; and Josh Mason pointing out the nonsense that is taught in graduate macroenonomics about the government budget constraint and interest rates.

2.0 Governments Can End Depressions

The figure above illustrates some basic elements of Keynes' theory. This specification of a discrete-time, dynamic system includes an accounting identity for a closed economy, namely, that national income in any time period is the sum of consumption spending, investment, and government spending. And it includes a behavioral relation, namely, a dynamic formulation of a consumption function. In this system, consumption is the sum of autonomous consumption and a term proportional to national income in the previous period. One should assume that the parameter b lies between zero and one.

A policy consequence follows: government can lift the economy out of a depression by spending more. Government spending increases national income immediately. Through the consumption function, it has a positive feedback on next period's income, as well.

3.0 The Permanent Income Hypothesis

Suppose you are hostile to this policy conclusion and, like the current Republican party in the USA, dislike your fellow countrymen. How might you suggest a theoretical revision to the system structure to mitigate the influence of current government spending? One possibility is to suggest more terms enter the consumption function. With the proper manipulation, current government spending will have a smaller impact, since current income will have a smaller impact on consumption.

So, suppose the consumption function does not contain a term multiplying b by income lagged one period. Instead, assume b multiplies an unobserved and (directly) unobservable state variable which, in turn, is an aggregate of current income lagged multiple periods (Yt - 1, Yt - 2, ..., Yt - n). Call this state variable "permanent income", and assume the aggregation is a matter of forming expectations about this variable based on a number of past values of income.

This accomplishes the goal. Current government spending can directly affect current income. But to have the same size impact as before on future income, it would have to be maintained through many lags. The policy impact of increased government spending is attenuated in this model, as compared to the dynamic system illustrated in the figure.

4.0 "Ricardian" Equivalence

One can go further with unobserved state variables. Suppose that households consume based less on recent income, but, once again, on expected values of future income. And suppose that consumers operate under the mainstream economist's mistaken theory of a government budget constraint. So consumers expect increased income today, if it results from increased government spending, to be accompanied by some combination of future decreased government spending and increased taxes. So the same current upward shock to the system causes an expectation of a future downward shock.

This is the theory of Ricardian equivalence. And, like the PIH, it suggests that Keynesian effects are not as dependable as otherwise would be the case.

5.0 Conclusion

The above story portrays economics as driven by results favorable to the biases and perceived self-interests of the extremely affluent. One would hope that academic economics is not entirely like this.