The European Tribune is a forum for thoughtful dialogue of European and international issues. You are invited to post comments and your own articles.
Please REGISTER to post.
But if the paper is a direct calculation in the framework of a standard model, and that the calculation is correct, then whoever made it should not change the view that it is, indeed, correct.
And I don't dispute that neo-classical and Austrians judge the author rather than the argument. But since we complain when they do, it seems that we indeed find it an inappropriate course of action.
That being said, Mankiw was my first example of clearly paid hack during the discussion with the previously mentioned LSE speaker at the following drinks. Earth provides enough to satisfy every man's need, but not every man's greed. Gandhi
As for R&R (which was not a paper), it should be ground for dismissing any such paper until data and calculations are made available. They were hardly trade secrets (which should not be a valid excuse anyway): they were national statistics... Earth provides enough to satisfy every man's need, but not every man's greed. Gandhi
"Proper" peer review would require replicating such calculations.
Peer review should verify that the methodology used is not insane, that the paper properly references its data, that the author has performed adequate robustness and specification tests, and that the data is available to other investigators who wish to replicate the analysis.
It is possible to imagine cases where the analysis is based on data that cannot be made available to the general public for ethical reasons, or because doing so would be an unreasonable commercial loss for the source of said data. However, in those cases I would argue that journals should demand full independent replication rather than the much more cursory process of peer review.
The above is already a higher standard than current academic peer review observes, and I don't think going beyond this is realistic - or necessarily a desirable use of the reviewers' time.
Now, there's a whole issue of replication not receiving the recognition it ought to. But that is a slightly different matter, and one I think can be solved with standard governance methods, like formalized KPIs for researchers requiring them to publish two replications for each original result.
- Jake Friends come and go. Enemies accumulate.
Peer review presumes that the paper is written in good faith. It is the hiring board's job to prevent the infiltration of pseudoscientists into academia; doing it at the paper level is simply not feasible. So if the presumption of good faith is visibly inapplicable, then "review" is not the correct tool for evaluating the paper. "Forensic reconstruction" is, and that is well beyond the scope of what can be expected of an unpaid reviewer.
Now, for general equilibrium models that doesn't really matter, because general equilibrium papers should be rejected out of hand as the pseudoscientific nonsense they are. And since the defining feature of most economic pseudoscientists is the fact that they are incapable or unwilling to operate outside a single, proven false, modeling framework, the distinction between rejecting the man and rejecting the model is in practice not very great.
by Frank Schnittger - Sep 24 2 comments
by Oui - Sep 19 18 comments
by Oui - Sep 13 35 comments
by Frank Schnittger - Sep 11 5 comments
by Cat - Sep 13 9 comments
by Frank Schnittger - Sep 2 2 comments
by Oui - Sep 27
by Oui - Sep 265 comments
by Frank Schnittger - Sep 242 comments
by Oui - Sep 1918 comments
by gmoke - Sep 173 comments
by Oui - Sep 153 comments
by Oui - Sep 15
by Oui - Sep 1411 comments
by Oui - Sep 1335 comments
by Cat - Sep 139 comments
by Oui - Sep 126 comments
by Frank Schnittger - Sep 115 comments
by Oui - Sep 929 comments
by Oui - Sep 713 comments
by Oui - Sep 61 comment
by Frank Schnittger - Sep 22 comments
by gmoke - Sep 2
by Oui - Sep 1181 comments
by Oui - Aug 315 comments
by gmoke - Aug 302 comments