Modeling the Price Mechanism: Simulation and The Problem of Time

Today’s New York Times article on rapid online repricing by holiday retailers depicts a retail world starting to approach the “flash-trading” status of financial markets:

Amazon dropped its price on the game, Dance Central 3, to $24.99 on Thanksgiving Day, matching Best Buy’s “doorbuster” special, and went to $15 once Walmart stores offered the game at that lower price. Amazon then brought the price up, down, down again, up and up again — in all, seven price changes in seven days.


The parrying could be seen with a Nintendo game, Mario Kart DS.

A week before Thanksgiving, the retailers’ prices varied, with Amazon selling it at $29.17, Walmart at $40.88, and Target at $33.99, according to Dynamite Data. Through Thanksgiving, as Target kept the price stable, Walmart changed prices six times, and Amazon five. On Thanksgiving itself, Walmart marked down the price to its advertised $29.96, which Amazon matched.

This made me think about a Greg Hannsgen post from last May on the Levy Economics Institute’s Multiplier Effect blog, a post I’ve been meaning to write about. It looks at the pricing mechanism based on how fast prices change/adjust.

The especially interesting thing about this post: It uses a dynamic simulation model to display the effects of slower and faster price adjustments, and lets you run the model yourself, right on the page, by moving a slider to change the speed of price adjustment and watch the results. (You need to install a browser plug-in from Wolfram, which only worked for me in Firefox under Mac OS X 10.6.8; it failed in Chrome.)

The gist:

You wonder what will happen when markets finally start working. How about, for example, a market that changes prices and wages quickly in response to fluctuations in demand? …

The pathway shown in the figure just below is followed by public production, capacity utilization, and the markup.

As you move the lever to the right, you are increasing a parameter that controls the speed at which the markup changes in response to high or low levels of customer demand.

What happens as the speed parameter is increased is that the economy’s pathway gradually changes until there is a relatively sudden vertical jump in the middle and much higher markup levels at the end—which means a bigger total rise in capital’s share.


The next pathway is the one followed by a second group of three variables during the same simulation. This second group includes: money, the government deficit (surpluses are negative numbers in this figure), and the employment rate (total work hours divided by total hours supplied).

when the lever is all the way to the right, the pathway begins with an outward spiral, leading to a new inward spiral, and finally an employment “crash” of sorts that occurs as the center of the second spiral is reached. This occurs after the markup has reached very high levels, as seen in the earlier diagram.

That’s a pretty amazing set of conclusions that I don’t think could ever have emerged from mainstream economic modeling techniques. You’ll notice that equilibrium, in particular, is a decidedly problematic concept here. It only seems to emerge when things have gone off the rails and hit the edge of the known world — not in those comfortable middle grounds so fondly envisioned by mainstream models.

I am making no claims of validity for this model’s assumptions, techniques, or conclusions (though I find the conclusions fascinating and plausible) — that’s beyond me. What I want to talk about is the validity of this class or type of dynamic simulation model compared to those commonly employed in “textbook” economic analysis.

In particular it makes me think about some recent posts and comments by Nick Rowe that are (bless him) teaching posts at least partially in response to my constantly demonstrated inability to properly grasp those standard approaches. (No: I’m not being ironic.)

Nick laid out the textbook understanding of the pricing mechanism based on marginal cost of production with wonderful clarity and broad insight here. I want to suggest that the explanation’s key feature is its use of “short-term” and “long-term.” It’s all about time.

But looking at the figures above — which model time fluidly and continuously (realistically?) as opposed to depicting it in two vaguely defined “chunks” — I want to ask whether Nick’s explanation, and the models employed in that explanation, are sufficient (or even proper) to grasp the processes playing out in the marketplace. Could they predict the effects that we see above? Is that type of depiction and prediction even within their inherent realm of capability?

Yes, Hannsgen’s model employs some textbook constructs, but it deploys them in a model that is structurally, qualitatively different from “comparative static” textbook models. Maybe the modeling technique is the message. Or at least, the proper modeling technique is a necessary condition for imparting an accurate and useful message.

Reading Nick’s posts and those of other smart econobloggers and -commenters, I am constantly astounded at his (their) apparent ability to intuitively comprehend and mentally manipulate multidimensional (and multiconceptual) interplays that leave me flummoxed. But I still wonder: is that ability sufficient for Nick to representatively model, in his head — to intuitively understand — the complex interplay of factors at work? His frequent comments about holding one factor constant, and the importance and difficulty of simultanaeity in our thinking, suggest that the answer might be no.

When you add a minimum wage to a free-economy model, is he able to simulate, in his head, all the possibly resultant pathways through the multidimensional space of prices, labor inputs, capital inputs, and output quantities (not to mention redistribution feedback effects, or utility-related factors), all over time — a space where no factors are held constant?

Yes, I’m questioning Nick’s quantitative ability given the models employed to do this kind of simulation in his head. (Suggesting: it’s not just me!, though there is certainly a matter of degree.) But as a result, and also, I’m questioning the qualitative ability of those models as employed to enable such understanding in our limited minds (notably, mine).

As I said recently, science is about really understanding how things work — telling a convincing causative story — not (just) predicting what will (might) happen. (At the extreme, you could say that prediction is only useful for scientists as a test of understanding.) The textbook models seem to provide some predictive power, but you gotta wonder how much of that is false positives. And given that question, you have to ask how well they really “explain” how economies work — how much true understanding they provide.

In other words — back to my apparently congenital inability to really internalize and understand the textbook models — it’s not my fault! (Yes: now I am being ironic.)

All this raises one big question for me: why aren’t mainstream economists all over these kinds of dynamic simulation models? Why do we only see them at the fringes, in work by “heterodox” outsiders like Hannsgen and Keen, and in the ever-about-to-emerge work promised by the guys at the Santa Fe Institute? Why aren’t big, and (within limits) out-of-the-box thinkers like Nick, Scott Sumner, Tyler Cowen, etc. — people who show every indication of really wanting to understand — fascinated by the possibilities of this type of modeling? In the weather and climate business, textbook economic modeling techniques would be laughed out of the room. Are not economies of a similar complex, dynamic, emergent type with weather and climate systems — arguably even more so, and more complex, because economies include the game-theory grist of conscious intentions and expectations (about other people’s intentions and expectations)?

I recently corresponded with a econ Phd candidate who really wanted to work with and build such models for his thesis. He reported that everything about the institutional and intellectual structure of academic economics militated against his doing so. “Just grab a data set, build a model and pull some regressions, call it good and head for the tenure track.”

You don’t have to read Kuhn or Marx to wonder whether this isn’t the result of 1. the irresistible intellectual gravitational pull of institutionally sanctioned models however obviously flawed (miasma, phlogiston, epicycles, equilibrium), and 2. the undeniable (inherent?) effectiveness of existing textbook economic models and modeling techniques in perpetuating and amplifying the established power structures and (increasingly) unequal distribution of income and wealth — wealth that actively seeks to perpetuate and expand itself via institutional structures like universities. (No names, just initials: Mercatus Center.)

I’m not imputing moral corruption here (except perhaps institutional). I both prefer and tend to believe that most of us try to do right, “as God gives us to see the right.” Rather, I tend toward the quite plausible institutional explanation laid out so nicely by Chomsky in Manufacturing Consent. In my words: institutions that are dependent on, are part and parcel of, those larger structures of power and wealth, only hire and promote people who already — perhaps by their very nature — see “right” right. (Or right “right.”)

And who knows? They might be right. Maybe my personal incentive is just to show (myself?) how smart I am relative to the mainstream institutions. But based on the Aha! moments of intuitive understanding that I experience when I see and explore models like Hannsgen’s, I tend to doubt whether that’s the only thing at play.

Cross-posted at Asymptosis.