tag:blogger.com,1999:blog-44559727330119454412017-05-27T04:40:35.350-07:00No HesitationsFrancis X. Diebold's BlogFrancis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.comBlogger290125tag:blogger.com,1999:blog-4455972733011945441.post-59091021033144804212017-05-27T04:40:00.001-07:002017-05-27T04:40:35.367-07:00SoFiE 2017 New York<span style="font-size: large;">If you haven't yet been to the <a href="http://sofie.stern.nyu.edu/">Society for Financial Econometrics</a> (SoFiE) annual meeting, now's the time. They're pulling out all the stops for the <a href="http://www.stern.nyu.edu/experience-stern/about/departments-centers-initiatives/centers-of-research/volatility-institute/events/sofie-conference-2017">10th anniversary at NYU Stern, June 21-23, 2017</a>. There will be a good mix of financial econometrics and empirical finance (invited speakers <a href="http://www.stern.nyu.edu/experience-stern/about/departments-centers-initiatives/centers-of-research/volatility-institute/events/sofie-conference-2017/about-conference/invited-speakers">here</a>; full program <a href="http://www.stern.nyu.edu/experience-stern/about/departments-centers-initiatives/centers-of-research/volatility-institute/events/sofie-conference-2017/conference-program">here</a>). The "<a href="http://www.stern.nyu.edu/experience-stern/about/departments-centers-initiatives/centers-of-research/volatility-institute/events/sofie-conference-2017/conference-program">pre-conference</a>" will also continue, this year June 20, with presentations by junior scholars (new/recent Ph.D.'s) and discussions by senior scholars. Lots of information <a href="http://www.stern.nyu.edu/experience-stern/about/departments-centers-initiatives/centers-of-research/volatility-institute/events/sofie-conference-2017">here</a>. See you there!</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-89494572240280596682017-05-22T08:18:00.002-07:002017-05-25T04:00:00.424-07:00Big Data in Econometric Modeling<span style="font-size: large;">Here's a speakers' photo from last week's Penn conference, <a href="http://www.ssc.upenn.edu/~fdiebold/Warren2017/Main.html">Big Data in Dynamic Predictive Econometric Modeling</a>. Click through to find the program, copies of papers and slides, a participant list, and a few more photos. A good and productive time was had by all!</span><br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="https://2.bp.blogspot.com/-Zj8UXu7uZ6w/WSMAw_ZEFfI/AAAAAAAAAwg/RAuYhR81kvMyQbRa3YlVI7PYZq6c_5BlwCLcB/s1600/Speakers.jpg" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="266" src="https://2.bp.blogspot.com/-Zj8UXu7uZ6w/WSMAw_ZEFfI/AAAAAAAAAwg/RAuYhR81kvMyQbRa3YlVI7PYZq6c_5BlwCLcB/s400/Speakers.jpg" width="400" /></a></div><div><br /></div>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-559293240275695682017-05-15T04:41:00.001-07:002017-05-15T04:41:48.697-07:00Statistics in the Computer Age<span style="font-size: large;">Efron and Tibshirani's <i><a href="https://www.amazon.com/s/ref=nb_sb_ss_i_1_13?url=search-alias%3Dstripbooks&field-keywords=computer+age+statistical+inference&sprefix=computer+age+%2Cstripbooks%2C150&crid=2UUGZEZGTDQDM">Computer Age Statistical Inference</a></i> (CASI) is about as good as it gets. Just read it. (Yes, I generally gush about most work in the Efron, Hastie, Tibshirani, Brieman, Friedman, et al. tradition. But there's good reason for that.) As with the earlier Hastie-Tibshirani Springer-published blockbusters (e.g., <a href="https://statweb.stanford.edu/~tibs/ElemStatLearn/printings/ESLII_print10.pdf">here</a>), the CASI publisher (Cambridge) has allowed ungated posting of the pdf (<a href="https://web.stanford.edu/~hastie/CASI_files/PDF/casi.pdf">here</a>). Hats off to Efron, Tibshirani, Springer, and Cambridge.</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-52518382139963304722017-05-08T04:54:00.003-07:002017-05-25T04:10:48.243-07:00Replicating Anomalies<span style="font-size: large;">I <a href="http://fxdiebold.blogspot.com/2017/04/the-latest-on-file-drawer-problem.html">blogged a few weeks ago on "the file drawer problem"</a>. In that vein, check out the interesting new paper below. I like their term "p-hacking". </span><br /><div><span style="font-size: large;"><br /></span></div><div><span style="font-size: large;">Random thought 1: </span></div><div><span style="font-size: large;">Note that <i>reverse </i>p-hacking can also occur, when an author wants <i>low</i> p-values. In the study below, for example, the deck could be stacked with all sorts of dubious/spurious "anomaly variables" that no one ever took seriously. Then of course a very large number would wind up with low p-values. I am <i>not </i>suggesting that the study below is guilty of this; rather, I simply had never thought about reverse p-hacking before, and this paper led me to think of the possibility, so I'm relaying the thought.</span><br /><div><span style="font-size: large;"><br />Related random thought 2: </span></div><div><span style="font-size: large;">It would be interesting to compare anomalies published in "top journals" and "non-top journals" to see whether the top journals are more guilty or less guilty of p-hacking. I can think of competing factors that could tip it either way!</span></div><div><br /></div><span style="font-size: large;"> Replicating Anomalies<br />by Kewei Hou, Chen Xue, Lu Zhang - NBER Working Paper #23394<br />Abstract:<br />The anomalies literature is infested with widespread p-hacking. We replicate the entire anomalies literature in finance and accounting by compiling a largest-to-date data library that contains 447 anomaly variables. With microcaps alleviated via New York Stock Exchange breakpoints and value-weighted returns, 286 anomalies (64%) including 95 out of 102 liquidity variables (93%) are insignificant at the conventional 5% level. Imposing the cutoff t-value of three raises the number of insignificance to 380 (85%). Even for the 161 significant anomalies, their magnitudes are often much lower than originally reported. Out of the 161, the q-factor model leaves 115 alphas insignificant (150 with t < 3). In all, capital markets are more efficient than previously recognized. </span></div><div><span style="font-size: large;"><br /></span></div><div><span style="font-size: large;">NBER version at <a href="http://papers.nber.org/papers/w23394?utm_campaign=ntw&utm_medium=email&utm_source=ntw">http://papers.nber.org/papers/w23394?utm_campaign=ntw&utm_medium=email&utm_source=ntw</a><br />Ungated version at <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2961979">https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2961979</a></span></div><div><br /></div>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-1386382088394805612017-05-04T07:06:00.001-07:002017-05-25T03:53:31.910-07:00Network Tools for Understanding High-Dimensional Dynamic Models<span style="font-size: large;">The slides from my "overview" <a href="http://fxdiebold.blogspot.com/2017/04/econometrics-predictive-modeling-and.html">IMF talk</a> two weeks ago proved popular, so <a href="http://www.ssc.upenn.edu/~fdiebold/NoHesitations/DieboldSlidesNYU.pdf">here are some different overview slides on a different topic</a> ("Estimating and Understanding High-Dimensional Dynamic Stochastic Econometric Models"), from my talk at last week's <a href="http://www.stern.nyu.edu/experience-stern/about/departments-centers-initiatives/centers-of-research/volatility-institute/events/upcoming-conferences">NYU Stern Conference on Volatility and Derivatives</a>.</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-42075829562811291102017-04-30T11:20:00.003-07:002017-05-02T04:35:01.766-07:00One Millionth Birthday...<div class="separator" style="clear: both; text-align: center;"></div><div style="margin-left: 1em; margin-right: 1em;"><div style="clear: right; float: right; margin-bottom: 1em; margin-left: 1em;"><a href="https://www.blogger.com/blogger.g?blogID=4455972733011945441" imageanchor="1" style="clear: right; float: right; margin-bottom: 1em; margin-left: 1em;"></a><img alt="Image result for 1 year birthday cake" src="data:image/jpeg;base64,/9j/4AAQSkZJRgABAQAAAQABAAD/2wCEAAkGBxISEBUSEhIVFRUVFRUVFRUWFRcVFhUVFRUWFxUVFRUYHSggGBolHRUVITEhJSkrLi4uFx8zODMtNygtLisBCgoKDg0OGhAQFysfHR0vLS0tLSstLS0tLystLSstLS0tLS0tLS0tLS0tKy0tLS0tLS0tLS0tLS0tLS0tLS0tLf/AABEIAMQBAQMBEQACEQEDEQH/xAAbAAACAgMBAAAAAAAAAAAAAAAAAQMEAgUGB//EAEIQAAEDAQUEBgYHBwQDAAAAAAEAAhEDBAUSITFBUWFxBhMigZGhMkJiscHRFCNScoKy4QckJTPC8PEWkqLDFUNz/8QAGgEBAQADAQEAAAAAAAAAAAAAAAECAwQFBv/EADURAQACAgEDAgQEBAUFAQAAAAABAgMRBBIhMUFRBRMicTNhwfAjMjSBQpGhsdEUQ1Jy4RX/2gAMAwEAAhEDEQA/AOHVQICEAgEAinCBoBA1AIAIMgipGhSWUJqbFhMttat30duY2iuymNCczuaMyubNlmI1XzPaPu6KUiI6reIdx0t6L0mWVxpsALQDIG5edX52DPWMltxLZTNGas108uq04Xs1lyWqruatjVMIyFWDEohIEUCVAgxKBIBAiiEgSAQCCZA0AgEAgaKEAgaAUAgaDIIySsCxlnC7ZWZrTeXRSHpX7N7EMb6m5sD8R/RcuH68/wD6xv8ARt5X04Yj3l19/UcVmqj2He4rZ8Qx7xxb/wAZiXJxbayx+bw630YJWeK24dWWupayoF0Q5LIXBZtbAoxJAkCVCKBIEgSISAQJAKghQTwgEBCAhAQgcIBFEIGoCEAgYQZBFSsWMtlWwsZzWi8OnHL1b9nf8p54t+K0cSP4t/tH6sudP0U/v+jprx/k1PuO/KV08v8AAv8AZx8f8Wv3h4hevpFc+Dw9HkeWlqLshw2QlZNUoyFUJEKECQEKjEhAoQEIhICEChAoQCCeEBCAhA4QEIohA4QNAQgFAQgcIMgEVm1SWcLNF8LXMN1bPTv2ZWsdXWBOmA/mXJbJGC02n1hsz7yVrEe8uqvO1j6LWcNlN/5StdOV87Has+f/AK1Y8c0y137vELbWklddKabsl9y171vhy2lEVkwYkIxYoFCAhAlQiEChAQiFCAhAoQKFQoUFiEBCBwgIQCKaAhAIHCAhQCAQZBFZBF2zaVjMMolvbgvZ1FtRrNagYBzBPzWjLx65NTPo6MeXpn/Zv7bfdSlRrUKohzqcDOZkwT5FcHEwfMmMseku3l3pXtHlwz3L1Yh5trIismuZYFVixIRChAoQKEBCBQqCEChAQiFCBEIEgIQTQgaAQCKaAhA4QEIBA1AQgEAEDCKyCLDf9CbN1ltpNOgl2zPCCQM+MLVl/kltxfzbn0bj9pzSLQwBv/rkmNSXOynu81q48RWJjwzyTNoj8nEldLRMsSqxYogQKECQKECVBCBQgEChAQgSIUIBBPCAhAQgIRThABA4QCAhAQoCEBCBwgEG1uK4q1rfhpNyBAc85NYDtJ2ngM1ja0QzrXbvrPbbquw4P5lYek8NxvnnozkFzWyV39T0cXw/Plrusaj81oftHsBMHrBxLMvesJyY59Gz/wDJ5EeLR/m03Tm7bNWs/wBMs+GRBJZo8EgGR9oStmK2rajxLjzYbViYvGph52upxkgEAgUKhEIFCAhAoQKEBCBICEChECCeEBCAhFEIhwinCAhAQgIQEKAhAQgIQb/o70ZfafrHnqqDc3VTlIGoZOvPQeSxtbTZjxTedRDd2m+C4fQrrYW0x6dXQmci4uOY09LU7Fy3vMzqHvYOHjwV683n0j9+ZbK5rio0G+iHvPpPcAST7IOgWMYvdqz82157TqFm22CjUaWupMIPsgHmCMwVfkw105d6z2lwV2Vixlts4JLGgxO9riJ78vBXBH1THs2/ErdWKtvdqIXY8EkAgSoIQIhAigUICEChAQgUICEQkCQWIRRCIEDhFEIHCBQgcICEBCAhQEINrcFGhidVtR+qpwcO2o8zhYBt0JPJY3tFY3Lo43Htmv01b51SveIBceosrTDKbdXAZcjz0GwLm+q/2e3vFw41Xvb3b+w2VlFmCm0Nb5k73HaVtrTTzM3ItkndpT4lnpzzZg9yuiLPO7NlaLcPZqfn/VaMMfXZ6XO78en79GthdLxShAKhIBAkAgUIAoEgIQKECQEIBETIBA0AgIRTQEIGAgUIHCAhAQoMX+pIloeC4cMp8pWjPHaHrfC7xHXXfeXpt22mjUaOqqU3ZZNBAIGwYdi2Vms+GrNhy1ndolbIWbkkkYsHKrDgaY/e7aPZf+YLmx/iWepy/wClo1cLoeMUIFCoECQEIEgEAgUIEgECQEIFCInhAIpwiCEU4RBCAhFEICEAgIQNQMIGTvzO/b46rGaVn0dWPm5sfi249p7um6HXs9zn0HuLsIDmEmThyBb5jzWGO07ms+jt5da3x1zVjW/Lq1ueYRCEODiLwtXGm73Uyuan4svW5Hfh1/fu1ELoeKUKghAoQEIFCAhAkBCBICEChAoQCARE0IBFOEQQgcIGgIQJFOEBCAAQOEChAwoNl0REW13GkffTWiv4kvYt34dP37vQWhdDywQiw4Z7P4naBvpn8lMrmr+LL1c39FH792khdLxSQIhAIBAkAQgSBIBAkAgRCBIBETwgcIFCBoGinCIIRRCAhAQgIQEIHCBAINn0UP77zpH+j5Lnif4svYiN8Kv3/WXoTQt7zJMhUhxFRn8Vq8af/Wz5Lmj8WXrZP6KP36ufhdLxChAIFCAhAiEAQgSBIghFKEBCIUIFCAhUWIUAAiiEAUDAQNAIBAIBAIBA0Agv9Fx+/N4sd+X9Fzf917VO/Bj9+r0dgXQ8qTLUIcZaGfxY8af/AF/ouePxXrT34P8Af9XNuGa6XiFCBQgUIBAkChAiEBCBQgEQkCQAQCosQooAQEIABA4QOEBCAIQEICEBCAhAIM6dMuIa0Ek6ACSeQCDaXLZH07fRa9jmOLXGHCDBbUgweS57fiPawf0M/f8AWHobG5Lc8uTLVUhyNppTe7Bvpf0vHwXPP4j1o78Kfv8ArDnrfYsEOxA4i7uIcRBK3xLy8mCaUrf0s3tHoszB23uxb2xhB5EZrxL/ABa3V9NY1/q9Cvw2vT3mdpW9AazqeJlWmSdGkEZbydh4L2MWWMlIvHq83JimlprPo1N69FrVZwXOp4mjVzDiA5jUeC2NemlhVCIQKECQBCBQiFCKMKIUICEChUWFFAQNECBoBAwgmslkfVeGU2lzjoB7+A4lFdldfQMOaHV62Z9WlBA5vOvcFNs5pMTqYS27oCzI0qrhvDwDlwIjPmp1LWlZnvOnF22xPpVDTe0hwOhGu4jerEsJjUlTsbjmcucieSwvkirTfLFJ1Lf9HuiX0gY3VWhoMFrc38Nch5q1vFo3DPHaLxuHeXTcNGzN+qbnGbjm88z8Ak7bI05XpFlfNHLWiPMVgtE/iQ9njxvhW+//AA6Njcl0PKlkWokOStgi+KH/AM/hWXPb8R7GP+it9/8Ahv7H0aph2J4Dxic9rTm0FxJk79dNFs9duLJmrOOMdY7ee/u2v0FgM4Ns7f8AC554uHq6ujux/wCpya11LTYjJdEdo7OeZ3PcYirEsZhxvSrogHA1rO2HZl1MaO2ktGw8NqziWMw4AtWTEiEChAoQKEChAIhFAIEgnQOEUQiHCAhFTWayPqOw02Oc7c0EnyQdVdHQSq/Ou7q2/ZbDnn4DzU2sQ7e7bpo2dhZSYBI7R1c77x28lJlYVrHQfSqYQfqzJ19E7lhuXoZMlMuPc/zQ2QnRTe3GitlMNGLa1ru6dPcrPZtwz1TFJ8TMOdvG5PpNNlQPioAWknPEATE8RvXLmtETM+zHm8St8s9PZf6MXQbO1xJxOeRyAbMe8phtvu58fH+VvvtualcCZ13Bdcd2enC3rU669qOEZ02Q/PIQHkeTh4rVaP4kPYwRNOHabevj/T/h1TWrc8mZBCI5S9W4b1s7zoWETvMVRH/IeK0Wj+JD18E74d4j9+HYWa2NdkeyfetsxLyVgFwOshYxMxP5LMRohtMRvKkaJZYpGX+VY8dmMsmhx2ws4mWMuXd0Cs0yalU5yc258PRWe2Ol7/SNigDqfRM+k7Pg4zmFNrpXq9ErE8uinmZ0e4BuWwStGPlYslprW0TMM5xTEbmHN2zoO5hH14c3aerMg5QIxZzn4cVvmzLFh+ZbW9ILb0Pc2m51MucWwSHNwy0z6I36JWdssnHiuorbbSXldVSg9tN47bmB+EZluIu7J4w3NZObSiQgxIVQKBqiaFAQqHCCayWV9V4ZTaXOOgGv6DioO3ujoI3CHWhxLtSxhgDg52p7oU2yiHV2KxU6LcFKmGN4beJJzPesdrpjb7Y+mA4NBb62/PQjhK2YqddukyXpjxWtO9x4Vad9NcNMJ27VlnwWxeO+2PByRybTXxMeiZtppwTiDo3aSuT/AHd84MnjWkNW8nNeGubDfM96lbbnU9m6nFrakzE90lrtOPsj1+yDw0+KW3aezHDh6Pqn/Cms13Oa0AuHgvN5Pw3NmyTaMmon00xycuszvpO8K7KFIve+ANvwA2ldfG4kYKateZaqdee/TSrQi12i0NmlFCmdHv7VV07Q3QDvldXVa3js6vk4ME/xPrt7R4TXVdFOhLgS57vSe7U5z3CUrXTXyOTfL21qI9GyxBZuPQxIdLWX7dLbQwZ4XsOJjx6p+WQ8Asb126+LyJwz43E+YU7JeT2ObRtTA17jhZUbmyofgf7yUi/pZuy8Sl6zkwz29Y9YdVZmnAEmvs86Z7pIO6D5JqU3Awnd7k1PsdkjWlWIljuGLhkrKwHBT0FJ1iOKWmBrxXiX+FW+f8zHbpjz+f2dkcmOjVo2s9UA4u3gAjYI3eK9uI05pvM1ivsxoxmSNPhosmM7219qsdNzKlSIqVW9Xj9ZrXZQ0nTIuVhhMd3ld92IUq72AENB7MmTh2Ge4rJioQiCECQWQFQ4QACDv/2fXS5gfWe2MYaKc7WnMkcD2fBSVh2Vd2EaTmB5SuXlZ5w1i0Vm3fXZsx16p1vTWWy8IdgYC9+4erMa7tQtkd+8unFgiY6rzqGNSpUfTe11KCG555ExOS2UtNJ6o9DJx8NtVm+4t/s19huo1e0HNaAdMyQY0I711X5HXjiPX1cmPj/9Hypvr6fReZcWGIqZDUYeInbz8VyRTT0p+IxMTuvf7ts6i06tB5gFXph58ZbR4nRtECAAANyumM2mZ3MmjFz/AEvsZrUm0w4NdjDpcHYYggzAJGq0ZbR2q9H4fl+VebzG40dhsNUMa0ua4jIkMeAANCJyPksqwyy5cczNta/vDaU7vG0lbOlxTm9oTGxs3e9XTD5lkD7vzyMeanS2RmjXeEdWyuAyI75A8c1JZVvWfLnbwuy0V6jA4MbTpvD8TXB5JbsnKPBabRNnpY8+HDSZrMzaY1406am4NgSJMytsTDyJiVgOKyYsg5BkHIKt4MqFv1b2t34vhksZh08e2OJ+usz9kVgs1Rs43h06RKx6bTLPkZsdv5a6W6mLZhPAyD4rb0w56zT/ABbRUqpdIc0tIiZiDO47Vr6dM7UiNTWd7Q17QxrSMYk7FemYhl8q899NfbKhwAA6Gdylbdmq1XCdKKZLxUOp7BG4jP4nwW1ploiEQoQGFBNCoYCDJozQevXVRbRotoYpwtEO3xtHfsUlYS1q28+JhSVSU6LdYzIz0nx2hYzDLrnWmFrtgYA2HOccwADvjVJmNN2HBN++4iGdlqmASzCdo5ZZlSLJmrEW1FtwsEwsupo0ayYmgIQNFNERtqNMwQY1jYm2VqWr5jyeJE0JVClQYGmDsHgp0wy6pJ1EQcIAMGDG3YsbVnpnp7StZjcdXhQu+z1Q4l5IjZikO48l5/Ewcit5tltP235dvJy4ZrEY4j/LwuF06eJXobcWlK1W9tMYieGQ1K3YsVsk6hpz5Yw16rJbPaWVmh2vkQdxCl6WxzqWzj8ndeqk+VwcFpmZZAq7TSOo4RmstkTMT2V6bGj0QBnJ7MnXSVjMQ22zXt/NKC9Rjw9WBJIxEgHLgN6lo9lxWrET1e3Z550woOp2lzMRc3svbPEfOVsc8tHCqEoBFTKoYQTWSniqNb9pzR4lB6HdN5ue4McMmt1A0A3rKYIbwNa5ocBOLRYSsJqLYgaf3osRI4RIOkyprbKJYuG5YzX0WJZFwAGX6LOIiPRjM7U74vAWemahEjING9x0H97km2m3Bh+bbpWbK/E0OMdoAxuJ2JEsMlYraYj0SsViWEwyhVEVpaS0gEiREjUTtUtPZnTtaJVbJZhTGUmfSJ1J3lYxOm3NlnJ5XMMHgVltoYkwsZvEdpZRG2sslSs0nrXMIz0BBAViXXeMNu1ImJY2e8nVGVMJAdihkjIRH6pMrfDTHesWjt6qbbztVIh1YMdTGpYCXdwA96nVPq6J4/GyV1jnU/m2NmvHrGvcab6bWicT8pG2Bqm3Jk4/RaKxaJmfZr7NaalqoVHN+rbiik4ek4NkEkbpy7ipv2btVwZYie/u1NttdenSLazA5sj6xokg7J2wtmHL8u29MuRxOPyq6iemZT9GK04nAy0kAGcjHxXTyM1csRp42Lh341rVtO/s6TFB1hcunRtm2vG3yU7gNpGgz4nIeanc0qXhW7GFggnUjdtzU3qPY13U7K7qgTm47IWUEuM6YHE6m46kOHcII95WTGXPFVCUAqJkDAQWbv8A5rPvN8jqrHkdVSqRUxjL7Y1g78th3rOUdJYbRIxAgjgZCxmFXqNXGCHa8Nv6rCYVEHjTrDyzTuqVjxsJKgTg4HFI79IUGFppB7QC0PaDOE54SNoKTG/LPHltSd1nRMtLG5Q4JEJMzMg3mzZi8FdMWbbwnQIaThxcM1JFW3Vi2ADkfPcvD+M5Mta16J1E+dOvi1rMzvyjs9sIMHMbvkvM+H/EM1MsUtMzWe3f9G/NgrNdx5XutadQV9ZNYvHd5+5r4UL8rhtFzoOcCY3n+x3rKI06OLG8kb9Ed00gymBmcszsB1OasQx5GTrvMo7VedIOjFPcCFl0tEW0htBpVGFhcIdrhkKdLox8iaTuGdao8UAyk4NDQA2GnQaNWExMNmG+OcnVeN7c5et9vqNpsgt7YbVGHIg7zvUnTujj0iLTXv27L9jb1Iy0O3YttZ9Hk37r4t0xh7Q2jdySWpYBfEhruE+7goI3VasxhY2N5lUP6UGiCcTjrH95BTXuKlqvARG37I+exUcd0kqYnMy0ads7R8kSWmIVQoQEIJkDQWbsH1zPveexWvklvsUQ6XACQHt9Kn7DxtC2sVoWt7Yfkd1Vn9bdoWOl2v2LpFTnty13DNp4tKxlW2F90DnLfxEA+anSNXfl7hzAKb2tz9UzPCRomtCexXtiYG1A05Qc4P8AlTpXbF1ps9I4hUcw+Pdkmg/9XWbR0njh+CnYZN6Q2I7x+EpqDunZf9iHreRTUHdOy/LO/stqsaDrJgnxU0q3Vo06jIDhwIIK15MNclem0bha3ms7hUpXWWmQ+eYXPj+HYKT1RWNtk57zGtrmBw1eAO5dkV01bai0X3QLsHaeZjPRZaN6ai+b6wiMc7AAIHPkkzEJHdqrLbLOM6lUk/Za0/FYx+csv7Jh0gAIFJkAbSJ7yruE7p6t+wJwVKn2jGQHBu1Jluw1ibfVOmvp0qtd2InqmcTE92gU1MvQvy6Y41SHUG30G0wwua6ABlwyWWoeVa0zO1azW+jTJLA4zsAPvKdvRGNe+y7R5p+yBiPuQVXWhztA93Ew0d6gwNY6THss+JQD3E5ERwHpHmdgRWlv5naZyI8CiTE+WpLEYjAqDq1BkAqMsKC3dLfrm95I4YTIHFWvklvyO0O0GvI7L/Uqt2B3FbWLB4DTmDRfvGdN3wUFG22V2pbr6zO0w8S3UKTCtfgOzP7rgfIqaXaajWqNOpHOnKmhsaF5OAjrmgburVRLNJ+bqtMn7sfFRVd9npzkMfItHvMpoDKYGtF8cCCU0MhZ6bvUe3iQpqBWq3e0HU9zSfNTphdgXbUjsFw55Sr0m0eCs3V7gOZhOk2VV7o9Mu700bVy128Dv+CmpEllY2e2SeQJViDa+WUPsu/2q6hE1OtTboKg/APiE7BuquIgNqnvDR5IqMUDtawcXukqaNsi+NarRwa1AmgHQVH88ggnDXAaMYOOZUClp9ZzzuGQRUj3FgkltNv/AC7pzJUZVrNp1HdEy0gsLxIEmSfScZ0Cm3RHHt19Hq5e/LWXuL2mBTgYQTk07SN8rC3deVx7YsXXSZ3HlJYjjphx2/A6rOJ7OGkzMblN1SrI+rQQwqgCC9c+VUu2tY5w5gK18jfloFXqoBY8SWnMA+zuW1ihr1DSqCmDiYfVf2o5KKltljaxuNhLTuB7PgURq6DxUMPa08Yg+IQWqlhDc2ueOTvmgo1bY9pjFPMA/BTas6N5OJzZTPNoRG2stqxZGnT/ANqK2dOmwCQxo7kC7L5BY2O/5qKKlhYW6EciUGttFhbEhzwRud80RqKtd7dHuPOD7wiqxtzicw082hA/phGjWD8KgYvKoNIHcE2M/wDydU+t4AJs0kZaqjtXu7o+SbVbpWXFq95/EiJ3WGm31Z5klXRtTrWotMNa0cgpIks9V9TV7u4wptVp1ma0YoxH2jKaFenXc4xMD2clNsjt93U3Ui5wJIzGZ1HFTTZjvNZ3Di7wtdQ1GjGQBDQAcgJWubTt2YLzE/crFLq7cTieyZnaBmAeGZSO8uvPH8O2/ZvQFm8I0AqP/9k=" /></div><span style="font-size: large;"> ...in event time. It's true, yesterday <i>No Hesitations</i> passed 1,000,000 page views. Totally humbling. I am grateful for your interest and support.</span><br /><br /></div>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-32980409070881536782017-04-29T06:16:00.000-07:002017-05-02T04:37:37.553-07:00Predictive Modeling and Causal Estimation<div><br /></div><span style="font-size: large;"></span><br /><div class="separator" style="clear: both; text-align: center;"><span style="font-size: large;"><a href="https://3.bp.blogspot.com/-v8zIRRhsih0/WQSRDYTMJAI/AAAAAAAAAv0/62dhPiN9Q2gNcuZ7jaW6d9uFPgsHfM2hQCEw/s1600/Forecasting_group.jpg" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="271" src="https://3.bp.blogspot.com/-v8zIRRhsih0/WQSRDYTMJAI/AAAAAAAAAv0/62dhPiN9Q2gNcuZ7jaW6d9uFPgsHfM2hQCEw/s400/Forecasting_group.jpg" width="400" /></a></span></div><span style="font-size: large;"></span><br /><div><span style="font-size: large;"><br /></span><span style="font-size: large;">Here are the slides, <a href="http://www.ssc.upenn.edu/~fdiebold/NoHesitations/DieboldSlides.pdf">"Econometrics, Predictive Modeling, and Causal Estimation"</a>, from my talk at <a href="http://unassumingeconomist.com/2017/04/workshop-on-forecasting-issues-in-developing-economies/">the recent conference</a> at the IMF in honor of <a href="http://www.albany.edu/~klahiri/">Kajal Lahiri</a>'s 70th (group photo above). They build on <a href="http://fxdiebold.blogspot.com/2017/02/econometrics-angrist-and-pischke-are-at.html">an earlier <i>No Hesitations</i> post</a>.</span></div><span style="font-size: large;"></span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-26044962768984750092017-04-20T09:01:00.000-07:002017-05-02T04:38:39.734-07:00Automated Time-Series Forecasting at Google<span style="font-size: large;">Check out this piece on <a href="http://www.unofficialgoogledatascience.com/2017/04/our-quest-for-robust-time-series.html">automated time-series forecasting at Google</a>. It's a fun and quick read. Several aspects are noteworthy. </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">On the upside:</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- Forecast combination features prominently -- they combine forecasts from an ensemble of models. </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- Uncertainty is acknowledged -- they produce interval forecasts, not just point forecasts.</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">On the downside:</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- There's little to their approach that wasn't well known and widely used in econometrics a quarter century ago (or more). Might not something like <a href="http://www.autobox.com/cms/index.php/home">Autobox</a>, which has been around and evolving since the 1970's, do as well or better?</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-41626511481985921442017-04-14T04:15:00.001-07:002017-05-02T04:40:17.386-07:00On Pseudo Out-of-Sample Model Selection<span style="font-size: large;">Great to see that Hirano and Wright (HW), "Forecasting with Model Uncertainty", finally <a href="https://www.econometricsociety.org/publications/econometrica/2017/03/01/forecasting-model-uncertainty-representations-and-risk">came out in <i>Econometrica</i></a>. (Ungated working paper version <a href="http://www.econ2.jhu.edu/People/Wright/hw.pdf">here</a>.) <br /><br />HW make two key contributions. First, they characterize rigorously the source of the inefficiency in forecast model selection by pseudo out-of-sample methods (expanding-sample, split-sample, ...), adding invaluable precision to more intuitive discussions like <a href="http://amstat.tandfonline.com/doi/abs/10.1080/07350015.2014.983236?src=recsys&journalCode=ubes20">Diebold (2015)</a>. (Ungated working paper version <a href="http://www.ssc.upenn.edu/~fdiebold/papers/paper113/Diebold_DM%20Test.pdf">here</a>.) Second, and very constructively, they show that certain simulation-based estimators (including bagging) can considerably reduce, if not completely eliminate, the inefficiency.</span><br /><div><span style="font-size: large;"><br /></span>Abstract: We consider forecasting with uncertainty about the choice of predictor variables. The researcher wants to select a model, estimate the parameters, and use the parameter estimates for forecasting. We investigate the distributional properties of a number of different schemes for model choice and parameter estimation, including: in‐sample model selection using the Akaike information criterion; out‐of‐sample model selection; and splitting the data into subsamples for model selection and parameter estimation. Using a weak‐predictor local asymptotic scheme, we provide a representation result that facilitates comparison of the distributional properties of the procedures and their associated forecast risks. This representation isolates the source of inefficiency in some of these procedures. We develop a simulation procedure that improves the accuracy of the out‐of‐sample and split‐sample methods uniformly over the local parameter space. We also examine how bootstrap aggregation (bagging) affects the local asymptotic risk of the estimators and their associated forecasts. Numerically, we find that for many values of the local parameter, the out‐of‐sample and split‐sample schemes perform poorly if implemented in the conventional way. But they perform well, if implemented in conjunction with our risk‐reduction method or bagging.</div>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-59308231395766664682017-04-10T08:05:00.001-07:002017-05-02T04:40:40.994-07:00BIg Data, Machine Learning, and the Macroeconomy<span style="font-size: large;">Coming soon at Bank of Norway:</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">CALL FOR PAPERS </span><br /><span style="font-size: large;">Big data, machine learning and the macroeconomy </span><br /><span style="font-size: large;">Norges Bank, Oslo, 2-3 October 2017 </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Data, in both structured and unstructured form, are becoming easily available on an ever increasing scale. To find patterns and make predictions using such big data, machine learning techniques have proven to be extremely valuable in a wide variety of fields. This conference aims to gather researchers using machine learning and big data to answer challenges relevant for central banking. </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Examples of questions, and topics, of interest are: </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Forecasting applications and methods</span><br /><span style="font-size: large;">-Can better predictive performance of key economic aggregates (GDP, inflation, etc.) be achieved by using alternative data sources? </span><br /><span style="font-size: large;">- Does the machine learning tool-kit add value to already well-established forecasting frameworks used at central banks? </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;"> Causal effects</span><br /><span style="font-size: large;">- How can new sources of data and methods be used learn about the causal mechanism underlying economic fluctuations? </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Text as data</span><br /><span style="font-size: large;">- Communication is at the heart of modern central banking. How does this affect markets? </span><br /><span style="font-size: large;">- How can textual data be linked to economic concepts like uncertainty, news, and sentiment? </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Confirmed keynote speakers are: </span><br /><span style="font-size: large;">- Victor Chernozhukov (MIT) </span><br /><span style="font-size: large;">- Matt Taddy (Microsoft, Chicago Booth) </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">The conference will feature 10-12 papers. If you would like to present a paper, please send a draft or an extended abstract to mlconference@norges-bank.no by 31 July 2017. Authors of accepted papers will be notified by 15 August. For other questions regarding this conference, please send an e-mail to mlconference@norges-bank.no. Conference organizers are Vegard H. Larsen and Leif Anders Thorsrud.</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-2825055157643021912017-04-10T04:22:00.003-07:002017-05-02T04:41:08.592-07:0013th Annual Real-Time Conference<span style="font-size: large;">Great news: The Bank of Spain will sponsor the 13th annual conference on real-time data analysis, methods, and applications in macroeconomics and finance, next October 19th and 20th , 2017, in its central headquarters in Madrid, c/ Alcalá, 48. </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">The real-time conference has always been unique and valuable. I'm very happy to see the Bank of Spain confirming and promoting its continued vitality.<br /><br />More information and call for papers <a href="http://www.bde.es/bde/en/secciones/sobreelbanco/Conferencias/13th_Annual_Con_0b762814f0e7a51.html">here</a>. <br /><br />Topics include:<br /><br />• Nowcasting, forecasting and real-time monitoring of macroeconomic and financial conditions. <br />• The use of real-time data in policy formulation and analysis.<br />• New real-time macroeconomic and financial databases. <br />• Real-time modeling and forecasting aspects of high-frequency financial data. <br />• Survey data, and its use in macro model analysis and evaluation. <br />• Evaluation of data revision and real-time forecasts, including point forecasts, probability forecasts, density forecasts, risk assessments and decompositions</span>.Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-23190914645899706752017-04-03T04:47:00.000-07:002017-05-02T04:42:46.457-07:00The Latest on the "File Drawer Problem"<span style="font-size: large;">The term "file drawer problem" was coined long ago. It refers to the bias in published empirical studies toward "large", or "significant", or "good" estimates. That is, "small"/"insignificant"/"bad" estimates remain unpublished, in file drawers (or, in modern times, on hard drives). Correcting the bias is a tough nut to crack, since little is known about the nature or number of unpublished studies. For the latest, together with references to the relevant earlier literature, see the interesting new NBER working paper, <a href="http://papers.nber.org/tmp/96352-w23298.pdf">IDENTIFICATION OF AND CORRECTION FOR PUBLICATION BIAS, by Isaiah AndrewsMaximilian Kasy</a>. There's an ungated version and appendix <a href="http://scholar.harvard.edu/kasy/publications/identification-and-correction-publication-bias">here</a>, and a nice set of slides <a href="http://scholar.harvard.edu/files/kasy/files/publicationbiasslides.pdf">here</a>.<br /><br />Abstract: Some empirical results are more likely to be published than others. Such selective publication leads to biased estimators and distorted inference. This paper proposes two approaches for identifying the conditional probability of publication as a function of a study's results, the first based on systematic replication studies and the second based on meta-studies. For known conditional publication probabilities, we propose median-unbiased estimators and associated confidence sets that correct for selective publication. We apply our methods to recent large-scale replication studies in experimental economics and psychology, and to meta-studies of the effects of minimum wages and de-worming programs. </span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-24766820732394785742017-03-28T11:37:00.002-07:002017-05-02T04:43:48.302-07:00Text as Data<span style="font-size: large;">"Text as data" is a vibrant and by now well-established field. (Just Google "text as data".)<br /><br /> For an informative overview geared toward econometricians, see the new paper, <a href="http://papers.nber.org/papers/w23276?utm_campaign=ntw&utm_medium=email&utm_source=ntw">"Text as Data" </a><a href="http://papers.nber.org/papers/w23276?utm_campaign=ntw&utm_medium=email&utm_source=ntw">by Matthew Gentzkow, Bryan T. Kelly, and Matt Taddy</a> (GKT). (Ungated version <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2934001">here</a>.)<br /><br /> "Text as data" has wide applications in economics. As GKT note:</span><br /><blockquote class="tr_bq"><span style="font-size: large;">... in finance, text from financial news, social media, and company filings is used to predict asset price movements and study the causal impact of new information. In macroeconomics, text is used to forecast variation in inflation and unemployment, and estimate the effects of policy uncertainty. In media economics, text from news and social media is used to study the drivers and effects of political slant. In industrial organization and marketing, text from advertisements and product reviews is used to study the drivers of consumer decision making. In political economy, text from politicians’ speeches is used to study the dynamics of political agendas and debate.</span></blockquote><span style="font-size: large;"><br />There are three key steps:<br /><br /> 1. Represent the raw text D as a numerical array x<br /><br /> 2. Map x into predicted values yhat of outcomes y<br /><br /> 3. Use yhat in subsequent descriptive or causal analysis.<br /><br />GKT emphasize the ultra-high dimensionality inherent in statistical text analyses, with connections to machine learning, etc.</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-72942558904658369912017-03-21T07:00:00.000-07:002017-05-02T04:45:37.378-07:00Forecasting and "As-If" Discounting<span style="font-size: large;">Check out the fascinating and creative new paper, "<a href="http://www.nber.org/papers/w23254.pdf">Myopia and Discounting</a>", by Xavier Gabaix and David Laibson.</span><br /><span style="font-size: large;"><br /></span> <span style="font-size: large;">From their abstract (slightly edited):</span><br /><blockquote class="tr_bq"><span style="font-size: large;">We assume that perfectly patient agents estimate the value of future events by generating noisy, unbiased simulations and combining those signals with priors to form posteriors. These posterior expectations exhibit as-if discounting: agents make choices as if they were maximizing a stream of known utils weighted by a discount function. This as-if discount function reflects the fact that estimated utils are a combination of signals and priors, so average expectations are optimally shaded toward the mean of the prior distribution, generating behavior that partially mimics the properties of classical time preferences. When the simulation noise has variance that is linear in the event's horizon, the as-if discount function is hyperbolic.</span></blockquote><span style="font-size: large;">Among other things, then, they provide a rational foundation for the "myopia" associated with hyperbolic discounting.</span><br /><span style="font-size: large;"><br /></span> <span style="font-size: large;">Note that in the Gabaix-Laibson environment everything depends on how forecast error variance behaves as a function of forecast horizon \(h\). But we know a lot about that. For example, in linear covariance-stationary \(I(0)\) environments, optimal forecast error variance grows with \(h\) at a decreasing rate, approaching the unconditional </span><span style="font-size: large;">variance from below. Hence it cannot grow linearly with \(h\), which is what produces hyperbolic as-if discounting. In contrast, in non-stationary \(I(1)\) environments, optimal forecast error variance <i>does</i> eventually grow linearly with \(h\). In a random walk, for example, \(h\)-step-ahead optimal forecast error variance is just \(h \sigma^2\), where \( \sigma^2\) is the innovation variance. </span><span style="font-size: large;">It would be fascinating to put people in \(I(1)\) vs. \(I(0)\) laboratory environments and see if hyperbolic as-if discounting arises in \(I(1)\) cases but not in \(I(0)\) cases.</span><br /><script src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML" type="text/javascript"></script><br />Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-71093442258277265362017-03-19T15:08:00.000-07:002017-03-20T10:35:24.377-07:00ML and Metrics VIII: The New Predictive Econometric Modeling<span style="font-size: large;">[Click on "Machine Learning" at right for earlier "Machine Learning and Econometrics" posts.]</span><span style="font-size: large;"><br style="background-color: white; color: #222222; font-family: Arial, Tahoma, Helvetica, FreeSans, sans-serif;" /><br /> We econometricians need -- and have always had -- cross section and time series ("micro econometrics" and "macro/financial econometrics"), causal estimation and predictive modeling, structural and non-structural. And all continue to thrive.<br /><br />But there's a new twist, happening now, making this an unusually exciting time in econometrics. P</span><span style="font-size: large;">redictive</span><span style="font-size: large;"> e</span><span style="font-size: large;">conometric modeling is not only alive and well, but also blossoming anew, this time at the interface of micro-econometrics and machine learning. A fine example is the new Kleinberg, Lakkaraju, Leskovic, Ludwig and <a href="http://scholar.harvard.edu/sendhil">Mullainathan</a> paper, “Human Decisions and Machine Predictions”, <a href="http://scholar.harvard.edu/files/sendhil/files/w23180.pdf">NBER Working Paper 23180</a> (February 2017)</span><span style="font-size: large;">.</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Good predictions promote good decisions, and </span><span style="font-size: large;">e</span><span style="font-size: large;">conometrics is ultimately about helping people to make good decisions. Hence the new developments, driven by advances in machine learning, are most welcome contributions to a long and distinguished </span><span style="font-size: large;">predictive econometric modeling</span><span style="font-size: large;"> </span><span style="font-size: large;">tradition.</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-53320733275930700342017-03-13T05:17:00.000-07:002017-03-13T05:17:15.747-07:00ML and Metrics VII: Cross-Section Non-Linearities<span style="font-size: large;">[Click on "Machine Learning" at right for earlier "Machine Learning and Econometrics" posts.]</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Th</span><span style="font-size: large;">e predictive modeling perspective needs not only to be </span><span style="font-size: large;">respected and </span><span style="font-size: large;">embraced in econometrics (as <a href="http://fxdiebold.blogspot.com/2017/02/econometrics-angrist-and-pischke-are-at.html">it routinely <i>is</i></a>, notwithstanding the <a href="http://www.nber.org/papers/w23144?utm_campaign=ntw&utm_medium=email&utm_source=ntw">Angrist-Pischke revisionist agenda</a>), but also to be <i>enhanced </i>by incorporating elements of statistical machine learning (ML). This is particularly true for cross-section econometrics insofar as time-series econometrics is already well ahead in that regard. </span><span style="font-size: large;"> For example, </span><span style="font-size: large;">although <a href="http://fxdiebold.blogspot.com/2017/03/machine-learning-and-econometrics-vi.html">flexible non-parametric ML approaches to estimating conditional-mean functions don't add much to time-series econometrics</a>, they may add lots to cross-section econometric regression and classification analyses, where conditional mean functions may be highly nonlinear for a variety of reasons. Of course econometricians are well aware of traditional non-parametric issues/approaches, especially kernel and series methods, and they have made many contributions, but there's still much more to be learned from ML.</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-10242011694030504662017-03-06T12:39:00.002-08:002017-03-10T07:27:01.146-08:00ML and Metrics VI: A Key Difference Between ML and TS Econometrics<span style="font-size: large;">[Click on "Machine Learning" at right for earlier "Machine Learning and Econometrics" posts.]<br /><br /> Continuing:<br /><br /> So then, statistical machine learning (ML) and </span><span style="font-size: large;">time series econometrics (TS) </span><span style="font-size: large;">have lots in common. But there's also an interesting difference: ML's emphasis on flexible nonparametric modeling of conditional-mean nonlinearity doesn't play a big role in TS. </span><br /><span style="font-size: large;"><br /> Of course there are the traditional TS conditional-mean nonlinearities: smooth non-linear trends, seasonal shifts, and so on. But there's very little evidence of important conditional-mean nonlinearity in the covariance-stationary (de-trended, de-seasonalized) dynamics of most economic time series. Not that people haven't tried hard -- really hard -- to find it, with nearest neighbors, neural nets, random forests, and lots more. </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">So it's no accident that things like linear autoregressions remain overwhelmingly dominant in TS. Indeed I can think of only one type of conditional-mean nonlinearity that has emerged as repeatedly important for (at least some) economic time series: <a href="https://www.ssc.wisc.edu/~bhansen/718/Hamilton1989.pdf">Hamilton-style Markov-switching dynamics</a>.<br /><br /> [Of course there's a non-linear elephant in the room: <a href="http://www.econ.uiuc.edu/~econ536/Papers/engle82.pdf">E</a><a href="http://www.econ.uiuc.edu/~econ536/Papers/engle82.pdf">ngle-style GARCH-type dynamics</a>. They're tremendously important in financial econometrics, and sometimes also in macro-econometrics, but they're about conditional variances, not conditional means.]<br /><br /> So there are basically only two important non-linear models in TS, and only one of them speaks to conditional-mean dynamics. And crucially, they're both very tightly parametric, closely tailored to specialized features of economic and financial data.<br /><br /> Now let's step back and assemble things:<br /><br /> ML emphasizes approximating non-linear conditional-mean functions in highly-flexible non-parametric fashion. That turns out to be doubly unnecessary in TS: There's just not much conditional-mean non-linearity to worry about, and when there occasionally is, it's typically of a highly-specialized nature best approximated in highly-specialized (tightly-parametric) fashion.</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-90919520025890276852017-02-26T10:01:00.001-08:002017-03-03T04:46:11.278-08:00Machine Learning and Econometrics V: Similarities to Time Series<span style="font-size: large;">[Notice that I changed the title from "Machine Learning vs. Econometrics" to "Machine Learning <i>and</i> Econometrics", as the two are complements, not competitors, as this post will begin to emphasize. But I've kept the numbering, so this is number five. For others click on Machine Learning at right.]</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Thanks for the overwhelming response to <a href="http://fxdiebold.blogspot.com/2017/02/econometrics-angrist-and-pischke-are-at.html">my last post</a>, on Angrist-Pischke (AP). I'll have more to say on AP a few posts from now, but first I need to set the stage.</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">A key observation is that statistical machine learning (ML) and time-series econometrics/statistics (TS) are largely about modeling, and they largely have the same foundational perspective. Some of the key ingredients are:</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- George Box got it right: "All models are false; some are useful", so search for good approximating models, not "truth".</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- Be explicit about the loss function, that is, about what defines a "good approximating model" (e.g., 1-step-ahead out-of-sample mean-squared forecast error)</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- Respect and optimize that loss function in model selection (e.g., BIC)</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- Respect and optimize that loss function in estimation (e.g., least squares)</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- Respect and optimize that loss function in forecast construction (e.g., Wiener-Kolmogorov-Kalman)</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- Respect and optimize that loss function in forecast evaluation, comparison, and combination (e.g., Mincer-Zarnowitz evaluations, Diebold-Mariano comparisons, Granger-Ramanathan combinations).</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">So time-series econometrics should <i>embrace</i> ML -- and it <i>is</i>. Just look at <a href="http://www.ssc.upenn.edu/~fdiebold/Warren2017/Program/Program.pdf">recent work like this</a>.</span><br /><br />Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-88730565590693653232017-02-19T16:16:00.001-08:002017-04-28T13:10:58.077-07:00Econometrics: Angrist and Pischke are at it Again<span style="font-size: large;">Check out the new Angrist-Pischke (AP), "<a href="http://www.nber.org/papers/w23144?utm_campaign=ntw&utm_medium=email&utm_source=ntw">Undergraduate Econometrics Instruction: Through Our Classes, Darkly</a>". <br /><br />I guess I have no choice but to weigh in. The issues are important, and my earlier AP post, "<a href="http://fxdiebold.blogspot.com/2015/01/mostly-harmless-econometrics.html">Mostly Harmless Econometrics?</a>", is my all-time most popular.<br /><br />Basically AP want all econometrics texts to look a lot more like theirs. But their books and their new essay unfortunately miss (read: dismiss) half of econometrics. <br /><br />Here's what AP get right:<br /><br />(Goal G1) One of the major goals in econometrics is predicting the effects of exogenous "treatments" or "interventions" or "policies". Phrased in the language of estimation, the question is "If I intervene and give someone a certain treatment \({\partial x}, x \in X\), what is my minimum-MSE estimate of her \(\ \partial y\)?" So we are estimating the partial derivative \({\partial y / \partial x}\).<br /><br />AP argue the virtues and trumpet the successes of a "design-based" approach to G1. In my view they make many good points as regards G1: discontinuity designs, dif-in-dif designs, and other clever modern approaches for approximating random experiments indeed take us far beyond "Stones'-age" approaches to G1. </span><span style="font-size: large;">(AP sure turn a great phrase...)</span><span style="font-size: large;">. And the econometric simplicity of the design-based approach is intoxicating: it's mostly just linear regression of \(y\) on \(x\) and a few cleverly-chosen control variables -- you don't need a full model -- with White-washed standard errors. Nice work if you can get it. And yes, moving forward, any good text should feature a solid chapter on those methods.</span><br /><span style="font-size: large;"><br />Here's what AP miss/dismiss:<br /><br />(Goal G2) The other major goal in econometrics is predicting \(y\). In the language of estimation, the question is "If a new person \(i\) arrives with covariates \(X_i\), what is my minimum-MSE estimate of her \(y_i\)? So we are estimating a conditional mean \(E(y | X) \), which in general is very different from estimating a partial derivative \({\partial y / \partial x}\).<br /><br />The problem with the AP paradigm is that it doesn't work for goal G2. Modeling nonlinear functional form is important, as the conditional mean function \(E(y | X) \) may be highly nonlinear in \(X\); systematic model selection is important, as it's not clear a priori what subset of \(X\) (i.e., what model) might be most important for approximating \(E(y | X) \); detecting and modeling heteroskedasticity is important (in both cross sections and time series), as it's the key to accurate interval and density prediction; detecting and modeling serial correlation is crucially important in time-series contexts, as "the past" is the key conditioning information for predicting "the future"; etc., etc, ... </span><br /><div><span style="font-size: large;"><br /></span></div><div><span style="font-size: large;">(Notice how often "model" and "modeling" appear in the above paragraph. That's precisely what AP dismiss, even in their abstract, which very precisely, and incorrectly, declares that "Applied econometrics ...[now prioritizes]... the estimation of specific causal effects and empirical policy analysis over general models of outcome determination".)<br /><br />The AP approach to goal G2 is to ignore it, in a thinly-veiled attempt to equate econometrics exclusively with G1, which nicely feathers the AP nest. Sorry guys, but no one's buying it. That's why the textbooks continue to feature G2 tools and techniques so prominently, as well they should.</span></div><script src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML" type="text/javascript"></script><br /><br /><br />Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-20463757957031506512017-02-13T06:23:00.001-08:002017-02-15T05:11:59.570-08:00Predictive Loss vs. Predictive Regret<span style="font-size: large;">It's interesting to contrast two prediction paradigms.</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">A. The universal statistical/econometric approach to prediction: </span><br /><span style="font-size: large;">Take a stand on a loss function and find/use a predictor that minimizes conditionally expected loss. Note that this is an <i>absolute </i>standard. We minimize loss, not some sort of relative loss.</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">B. An alternative approach to prediction, common in certain communities/literatures:</span><br /><span style="font-size: large;">Take a stand on a loss function and find/use a predictor that minimizes regret. Note that this is a <i>relative </i>standard. Regret minimization is relative loss minimization, i.e., striving to do no worse than others.</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Approach A strikes me as natural and appropriate, whereas B strikes me as as quirky and "behavioral". That is, it seems to me that we generally want tools that perform well, not tools that merely perform no worse than others.</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">There's also another issue, the <i>ex ante</i> nature of A (standing in the present, conditioning on available information, looking forward) vs. the <i>ex post</i> nature of B (standing in the future, looking backward). Approach A again seems more natural and appropriate.</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-74933111159306944322017-02-05T10:49:00.000-08:002017-02-15T05:08:58.294-08:00Data for the People<span style="font-size: large;"><i><a href="http://ourdata.squarespace.com/">Data for the People</a></i>, by Andreas Weigend, is coming out this week, or maybe it came out last week. Andreas is a leading technologist (at least that's the most accurate one-word description I can think of), and I have valued his insights ever since we were colleagues at NYU almost twenty years ago. Since then he's moved on to many other things; see <a href="http://www.weigend.com/">http://www.weigend.com</a>. </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">Andreas challenges prevailing views about data creation and "data privacy". Rather than perpetuating a romanticized view of data privacy, he argues that we need increased data transparency, combined with increased data literacy, so that people can take command of their own data. Drawing on his work with numerous firms, he proposes six "data rights":</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">-- The right to access data</span><br /><span style="font-size: large;">-- The right to amend data</span><br /><span style="font-size: large;">-- The right to blur data</span><br /><span style="font-size: large;">-- The right to port data</span><br /><span style="font-size: large;">-- The right to inspect data refineries</span><br /><span style="font-size: large;">-- The right to experiment with data refineries</span><br /><div><br /></div><span style="font-size: large;">Check out <i>Data for the People</i> at <a href="http://ourdata.com/">http://ourdata.com</a>.</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;"><br /></span><span style="font-size: large;">[Acknowledgment: Parts of this post were adapted from the book's web site.]</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-52084362822939192472017-01-30T05:34:00.000-08:002017-02-15T05:08:12.273-08:00Randomization Tests for Regime Switching<span style="font-size: large;">I have always been fascinated by distribution-free non-parametric tests, or randomization tests, or Monte Carlo tests -- whatever you want to call them. (For example, I used some in ancient work like <a href="http://www.ssc.upenn.edu/~fdiebold/papers2/Diebold-Rudebusch%20(1992).pdf">Diebold-Rudebusch 1992</a>.) They seem almost too good to be true: exact finite-sample tests without distributional assumptions! They also still seem curiously underutilized in econometrics, notwithstanding, for example, the path-breaking and well-known contributions over many decades by <a href="https://dl.dropboxusercontent.com/u/11900540/Web_Site_JMDufour/dufour.html">Jean-Marie Dufour</a>, <a href="http://ecares.ulb.ac.be/index.php?option=com_comprofiler&task=userProfile&user=114&Itemid=263">Marc Hallin</a>, and others.</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">For the latest, see the <a href="http://www.cireqmontreal.com/wp-content/uploads/cahiers/15-2016-cah.pdf">fascinating new contribution by Jean-Marie Dufour and Richard Luger</a>. They show how to use randomization to perform simple tests of the null of linearity against the alternative of Markov switching in dynamic environments. That's a very hard problem (nuisance parameters not identified under the null, singular information matrix under the null), and several top researchers have wrestled with it (e.g., <a href="http://econpapers.repec.org/article/ieriecrev/v_3a39_3ay_3a1998_3ai_3a3_3ap_3a763-88.htm">Garcia</a>, <a href="http://www.ssc.wisc.edu/~bhansen/papers/jae_92.pdf">Hansen</a>, <a href="http://onlinelibrary.wiley.com/doi/10.3982/ECTA8609/abstract">Carasco-Hu-Ploberger</a>). Randomization delivers tests that are exact, distribution-free, and <i>simple</i>. And power looks pretty good too. </span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-50893845230278469442017-01-23T04:43:00.000-08:002017-02-15T04:58:21.149-08:00Bayes Stifling Creativity?<span style="font-size: large;">Some twenty years ago, a leading Bayesian econometrician startled me during an office visit at Penn. We were discussing Bayesian vs. frequentist approaches to a few things, when all of a sudden he declared that "There must be something about Bayesian analysis that stifles creativity. It seems that frequentists invent all the great stuff, and Bayesians just trail behind, telling them how to do it right".</span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">His </span><span style="font-size: large;">characterization rings true in certain significant respects, which is why it's so funny. But the intellectually interesting thing is that it doesn't have to be that way. As Chris Sims notes in a recent communication: </span><br /><blockquote class="tr_bq"><span style="font-size: large;">...</span><span style="font-size: large;"> frequentists are in the habit of inventing easily computed, intuitively appealing estimators and then deriving their properties without insisting that the method whose properties they derive is optimal. ... Bayesians are more likely to go from model to optimal inference, [but] they don't have to, and [they] ought to work more on Bayesian analysis of methods based on conveniently calculated statistics.</span></blockquote><span style="font-size: large;"><br /></span><span style="font-size: large;">See Chris' thought-provoking unpublished paper draft, "<a href="http://sims.princeton.edu/yftp/UndrstndgNnBsns/GewekeBookChpter.pdf">Understanding Non-Bayesians</a>". </span><br /><span style="font-size: large;"><br /></span><span style="font-size: large;">[As noted on <a href="http://www.princeton.edu/~sims/">Chris' web site</a>, he wrote that paper for </span><span style="font-size: large;">the Oxford University Press <i>Handbook of Bayesian Econometrics</i>, but he "withheld [it] from publication there because of the Draconian copyright agreement that OUP insisted on --- forbidding posting even a late draft like this one on a personal web site."] </span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-86513355031212118632017-01-16T08:39:00.000-08:002017-02-15T05:04:45.635-08:00Impulse Responses From Smooth Local Projections<span style="font-size: large;">Check out <a href="https://papers.ssrn.com/sol3/papers2.cfm?abstract_id=2892508">Barnichon-Brownlees (2017)</a> (BB). As proposed and developed in <a href="https://www.aeaweb.org/articles?id=10.1257/0002828053828518">Jorda (2005)</a>, they estimate impulse-response functions (IRF's) directly by projecting outcomes on estimates of structural shocks at various horizons, as opposed to inverting a fitted autoregression. The BB enhancement relative to Jorda is the effective incorporation of a smoothness prior in IRF estimation. (Notice that the traditional approach of inverting a low-ordered autoregression automatically promotes IRF smoothness.) In my view, smoothness is a natural IRF shrinkage direction, and BB convincingly show that it's likely to enhance estimation efficiency relative to Jorda's original approach. I always liked the idea of attempting to go after IRF's directly, and Jorda/BB seems appealing.</span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0tag:blogger.com,1999:blog-4455972733011945441.post-69423797027523341472017-01-13T05:22:00.003-08:002017-01-13T05:26:34.113-08:00Math Rendering Problem Fixed<span style="font-size: large;">The problem with math rendering in the recent post, <a href="http://fxdiebold.blogspot.com/2017/01/all-of-machine-learning-in-one.html">"All of Machine Learning in One Expression"</a>, is now fixed (I hope). That is, the math should now look like math, not LaTeX code, on all devices. </span>Francis Dieboldhttps://plus.google.com/104011662239494052073noreply@blogger.com0