The Economist on homeopathy in Germany:
IT MAY not be as ancient as acupuncture, but homeopathy is the closest thing Germany has to a native alternative-medicine tradition.
John Oliver’s Last Week Tonight on: “Journalism”
- German statistical office (in German): “Promovierende in Deutschland“:
- They held a survey with 20,000 professors and 20,000 PhD students in Germany.
- There are currrently 33,154 professors in Germany that can graduate a PhD student (“Promotionsrecht”).
- The immense number of PhD students (“Promovierende”) is 196,200 of which 111,400 are enrolled at an university. And 99% of those finishing their PhDs are those that are enrolled at an university. (These numbers are obviously inflated by German peculiarities such as counting medical doctorates as PhDs.)
- 11% of professors have no PhD students, 50% have 1-5 students and 3 percent have more than 20 PhD students. The average is 6 students per professor and that ratio is highest for the engineering subjects.
- 44% () of PhD students are women.
- The modal age is 29.
- 15% are non-German.
- 23% of students are in structured programs, 23% of students are doing a cumulative dissertation (economics PhD-style).
John D. Cook: “One of my favorite proofs: Lagrange multipliers”
- Free online book: “Dynamic Discrete Choice Models: Methods, Matlab Code, and Exercises”, by Jaap Abbring and Tobias Klein (through Jason Blevins)
Samir Okasha in “Philosophy of Science: A Very Short Introduction” gives a good overview of the concept of science.
Okasha explains the difference between deductive and inductive reasoning. A deductive argument follows from its assumptions. An inductive argument is one where you have to reason about new unseen things.
At the root of Hume’s problem is the fact that the premisses of an inductive inference do not guarantee the truth of its conclusion.
Philosophers have responded to Hume’s problem in literally dozens of different ways; this is still an active area of research today.
For inductive reasoning to help us make predictions about the future, we need a new assumption. We have to take as given that along some lines things will remain the same.
This assumption may seem obvious, but as philosophers we want to question it. Why assume that future repetitions of the experiment will yield the same result? How do we know this is true?
A good model is one that’s not too crude about what it accepts about nature’s constancy. If you assume that business cycles just mechanically happen every seven or so years, then that’s fairly crude.
Karl Popper thought that scientists should only argue deductively. We all know Karl Popper and we cite him when we say that theories have to be falsifiable. But philosophy of science didn’t stop with Popper. In particular, Popper’s theory of progress in science doesn’t capture what actually happens:
In general, scientists do not just abandon their theories whenever they conflict with the observational data. […] Obviously if a theory persistently conflicts with more and more data, and no plausible ways of explaining away the conflict are found, it will eventually have to be rejected. But little progress would be made if scientists simply abandoned their theories at the first sign of trouble.
Most philosophers think it’s obvious that science relies heavily on inductive reasoning, indeed so obvious that it hardly needs arguing for. But, remarkably, this was denied by the philosopher Karl Popper, […]. Popper claimed that scientists only need to use deductive inferences.
The weakness of Popper’s argument is obvious. For scientists are not only interested in showing that certain theories are false.
In contrast, Thomas Kuhn speaks of paradigm changes:
In short, a paradigm is an entire scientific outlook – a constellation of shared assumptions, beliefs, and values that unite a scientific community and allow normal science to take place.
But over time anomalies are discovered – phenomena that simply cannot be reconciled with the theoretical assumptions of the paradigm, however hard normal scientists try. When anomalies are few in number they tend to just get ignored. But as more and more anomalies accumulate, a burgeoning sense of crisis envelops the scientific community. Confidence in the existing paradigm breaks down, and the process of normal science temporarily grinds to a halt.
In Kuhn’s words, ‘each paradigm will be shown to satisfy the criteria that it dictates for itself and to fall short of a few of those dictated by its opponent’.
Karl Popper is normative, “How should science be done?”, while Thomas Kuhn is descriptive, “How is science done?”
In rebutting the charge that he had portrayed paradigm shifts as non-rational, Kuhn made the famous claim that there is ‘no algorithm’ for theory choice in science. […] Kuhn’s insistence that there is no algorithm for theory choice in science is almost certainly correct.
The moral of his story is not that paradigm shifts are irrational, but rather that a more relaxed, non-algorithmic concept of rationality is required to make sense of them.
Kuhn’s idea of “theory-ladenness” of data is interesting. Kuhn says that this makes comparisons between theories difficult or impossible. That’s probably exaggerated, but in economics, many of the things we measure (like GDP) are abstract concepts and theory guides how we measure it.
Noah Smith argues that inflation has low costs and central banks should therefore sometimes trade-off higher inflation against better GDP performance. And Olivier Blanchard has made the case to raise the inflation target above the current 2%, to increase the distance to the zero lower bound.
Yet in the mind of many people there’s no place for inflation. People have “money illusion”, so they fail to adjust nominal values for overall price changes and feel richer or poorer when really they ‘re not. Inflation is seen as a bad thing and George Akerlof and Robert Shiller write:
Inflation itself, particularly when it is increasing, can ultimately create a negative effect on the atmosphere of an economy, akin to the effect of broken windows and graffiti on a city. These lead to a breakdown in the sense of civil society, in the sense that all is right with the world. (p65, “Animal Spirits“)
For my bachelor thesis, I read Barry Eichengreen’s “Globalizing Capital”. He explains how modern economies changed after World War I. Larger firm conglomerates and unionization made wages of workers less flexible. And this downward wage rigidity was a problem during the Great Depression.
Nominal rigidities are the reason that monetary policy works at all. If prices and wages were flexible, then when the central bank doubles the money in circulation, all prices would also double immediately. So, I thought, the solution is to index all prices. If inflation from this year to the next is 2 percent, then your wage, your rent and every other price should automatically rise by 2 percent. And if for some reason the aggregate price level falls, then all these prices would also adjust downwards.
But indexing all prices is not workable and people wouldn’t accept it. And because money isn’t neutral, there is a role for intentional monetary policy. One of the most important effects of higher inflation is that, if wages adapt slowly, real wages fall for a while. So it’ll be cheaper for firms to hire people and they’ll be more willing to do so.
Economists are quick to prescribe other people economics lessons, but understanding inflation and the difference between nominal and real values is a basic skill that I wish more people would have.
Chicago Booth Review: “Zip-Code Economics: What loca data reveal about the economy at large”
A non-fiction reading list by J.P. Morgan
Very good visualizations of 2 million chess games from the last two hundred years.
Let Go - Justin Jay, Benny Bridges and Josh Taylor:
In “The Berlin Stock Exchange in Imperial Germany - a Market for New Technology?”, (pdf) Sibylle Lehmann-Hasemeyer and Jochen Streb look at how well the financial market assessed firm innovativeness in pre-1913 Germany. They show that the stock market guessed well which companies would continue to innovate after they went public.
Between 1892 and 1919, 474 companies started trading their shares on the Berlin stock exchange. The authors take the change in the price of the stock on its first day of trading as an measure of “underpricing” which indicates how much asymmetric information there is in the markets.
Underpricing is bad for a firm, as it receives fewer funds than if it had sold its shares at a higher price. For example, Google went to great lengths to determine a good price. And with more capital the firm can invest more into research and so be granted more patents later. So one has to argue that this effect can’t be strong enough to lead to reverse causality.
Lehmann-Hasemeyer and Streb control for what investors knew at the time of the initial public offering (IPO) about how innovative firms already were. For this, they count the number of patents a firm had been granted before. So patents are a proxies for the innovativeness of a firm. This is an example of using patents as “inputs” to the technological process, in Zvi Griliches’ wording.
Research is a risky activity, so there might be more asymmetric information in the price for stocks of research-intensive companies. But that’s not what they find as there was little underpricing in the stocks of firms that continued to be innovative after the IPO. This might be due to the screening of banks:
Overall, German universal banks seemed to be well informed about the market value of firms that planned to go public. The comparatively low underpricing that occurred at the Berlin stock exchange during Germany’s high industrialization might therefore indicate that investors’ uncertainty was rather small because they knew that banks brought only those firms to the market that met certain minimum quality requirements.
They conclude that investors must have had more information than patent counts:
[Investors] were capable of distinguishing between permanently innovative firms and firms with sharply declining innovativeness (Buddenbrooks), even though both types of firms looked very similar at the date of the IPO with respect to their patent history. This observation implies that pure patent counts that are often used in cliometric studies of innovation might not be a good proxy for the knowledge that was available at the date of an IPO.
The paper is forthcoming in the American Economic Review.
There’s a species of books in which a commonly-held view by established researchers is criticized by someone from outside the profession and supposedly shown to be wrong.
There’s nothing wrong with people who are not scientists writing about science. But two lines of argument in those books aren’t convincing:
- The outsider does not have to hold to establish views to advance and has therefore figured out something that people within the profession have missed or aren’t allowed to say.
- The respective branch of science is not an experimental science and can therefore not establish causality.
The first is rarely the case. Science isn’t a closed environment where you’re not allowed to speak your mind. Dani Rodrik writes in his “Ten Commandments for Noneconomists”:
[Nine.] If you think all economists think alike, attend one of their seminars.
[Ten.] If you think economists are especially rude to noneconomists, attend of one of their seminars.
And the second argument is misleading. How did we figure out that the Earth orbits the Sun or that smoking causes cancer? Not through experiments. Also, if scientists can’t claim to identify causality, why should the outsider?
In the “The Nurture Assumption” (which is else an interesting book), Judith Rich Harris writes (added emphasis):
[Socialization research] is a science because it uses some of the methods of science, but it is not, by and large, an experimental science. To do an experiment it is necessary to vary one thing and observe the effects on something else. Since socialization researchers do not, as a rule, have any control over the way parents rear their children, they generally cannot do experiments. Instead, they take advantage of existing variations in parental behavior. They let things vary naturally and, by systematically collecting data, try to to find out what things vary together. In other words, they do correlational studies.
Nina Teicholz uses both arguments in “The Big Fat Surprise”.
Tom Wolfe’s book (which I haven’t read) sounds a lot like that as well:
Evolution, [Tom Wolfe] argues, isn’t a “scientific hypothesis” because nobody’s seen it happen, there’s no observation that could falsify it, it yields no predictions and it doesn’t “illuminate hitherto unknown or baffling areas of science.” Wrong - four times over.
I like Jerry Coyne’s final take-down:
Somewhere on his mission to tear down the famous, elevate the neglected outsider and hit the exclamation-point key as often as possible, Wolfe has forgotten how to think.
John Ellenby […] studied economics and geography at University College London and spent a year in the early 1960s studying at the London School of Economics, where he encountered mainframe computers.
Humans crave patterns. […]
So it is that the Millennial Whoop evokes a kind of primordial sense that everything will be alright. […] In the age of climate change and economic injustice and racial violence, you can take a few moments to forget everything and shout with exuberance at the top of your lungs. Just dance and feel how awesome it is to be alive right now. Wa-oh-wa-oh.
Erdoğan vs. Erdogan (Acemoğlu?):
A fairly quick inspection of web pages suggests that both the New York Times and the Financial Times operate essentially the same policy – diacritics for languages like French, German and Spanish, basic 7-bit ASCII (no diacritics at all) for the rest. […]
For mainstream papers, the Guardian and the Süddeutsche are decidedly to the left of the spectrum, decidedly internationalist/Europeanist, and so on, and you would expect them to resist any suggestion that some languages are more important (or more normal) than others.
Shane Caldwell: “Landing a tenure-track position, 1950’s vs. 2010’s”
Philip Ball in The Atlantic on Occam’s razor:
Much more often, theories are distinguished not by making fewer assumptions but different ones. It’s then not obvious how to weigh them up.
The Economist: “Why investors want alternative data”:
The providers are themselves a disparate group, pumping out databases ranging from satellite imagery to social-media posts. […]
Recent advancements in machine-learning have made it possible for companies to efficiently parse through millions of satellite images a day.
Conducting research with alternative data does not always come easily; it often arrives in messy formats and can be difficult to handle for analysts who lack sophisticated IT operations.
The BBC on the 100 greatest films of the 21st century. (So far, I guess.)