Last week my research group discussed Hal Varian’s interesting new paper on “Big data: new tricks for econometrics”, Journal of Economic Perspectives, 28(2): 3–28. It’s a nice introduction to trees, bagging and forests, plus a very brief entrée to the LASSO and the elastic net, and to slab and spike regression. Not enough to be able to use them, but ok if you’ve no idea what they are.
Posts Tagged ‘research team’:
Last week, my research group discussed Galit Shmueli’s paper “To explain or to predict?”, Statistical Science, 25(3), 289–310. (See her website for further materials.) This is a paper everyone doing statistics and econometrics should read as it helps to clarify a distinction that is often blurred. In the discussion, the following issues were covered amongst other things. The AIC is better suited to model selection for prediction as it is asymptotically equivalent to leave-one-out cross-validation in regression, or one-step-cross-validation in time series. On the other hand, it might be argued that the BIC is better suited to model selection for explanation, as it is consistent. P-values are associated with explanation, not prediction. It makes little sense to use p-values to determine the variables in a model that is being used for prediction. (There are problems in using p-values for variable selection in any context, but that is a different issue.) Multicollinearity has a very different impact if your goal is prediction from when your goal is estimation. When predicting, multicollinearity is not really a problem provided the values of your predictors lie within the hyper-region of the predictors used when estimating the model. An ARIMA model has no explanatory use, but is great at short-term prediction. How to
My research group meets every two weeks. It is always fun to talk about general research issues and new tools and tips we have discovered. We also use some of the time to discuss a paper that I choose for them. Today we discussed Breiman’s classic (2001) two cultures paper — something every statistician should read, including the discussion. I select papers that I want every member of research team to be familiar with. Usually they are classics in forecasting, or they are recent survey papers. In the last couple of months we have also read the following papers: Timmermann (2008) Elusive return predictability Diebold (2013) Comparing predictive accuracy, twenty years later: A personal perspective on the use and abuse of Diebold-Mariano tests Gneiting and Katsfuss (2014) Probabilistic forecasting Makridakis and Hibon (1978) Accuracy of forecasting: an empirical investigation
We are looking for a new post-doctoral research fellow to work on the project “Macroeconomic Forecasting in a Big Data World”. Details are given at the link below jobs.monash.edu.au/jobDetails.asp?sJobIDs=519824 This is a two year position, funded by the Australian Research Council, and working with me, George Athanasopoulos, Farshid Vahid and Anastasios Panagiotelis. We are looking for someone with a PhD in econometrics, statistics or machine learning, who is well-trained in computationally intensive methods, and who has a background in at least one of time series analysis, macroeconomic modelling, or Bayesian econometrics.
If you find this blog helpful (or even if you don’t but you’re interested in blogs on research issues and tools), there are a few other blogs about doing research that you might find useful. Here are a few that I read. Patter — Pat Thomson. The Thesis Whisperer — Inger Mewburn. The Research Whisperer – several RMIT researchers. the (research) supervisor’s friend — Geof Hill. My Research Rants – Jordi Cabot. The Three Month Thesis – James Hayton. profserious – Anthony Finkelstein. Academic Life — Marialuisa Aliotta. Help for New Professors — Faye Hicks. The Art of Scientific Writing – Faye Hicks. Explorations of style– Rachael Cayley. sharmanedit — Anna Sharman. GradHacker – writers from several universities. PhD Life – Warwick Uni students. PhD Comics — essential reading for every PhD student, and good therapy. I’ve created a bundle so you can subscribe to all of these in one go. Of course, there are lots of statistics blogs as well, and blogs about other research disciplines. The ones above are those that concentrate on generic research issues.
Journal Clubs are a great way to learn new research ideas and to keep up with the literature. The idea is that a group of people get together every week or so to discuss a paper of joint interest. This can happen within your own research group or department, or virtually online. There is now a virtual journal club operating in conjunction with CrossValidated.com. The first paper discussed was on text data mining. It appears that the next paper may be on collaborative filtering. The emphasis is on Open Access papers, preferably with associated software that is freely available. Some of the discussion tends to centre on how to implement the ideas in R. For those of us in Australia, the timing is tricky. The first discussion took place at 3am local time! If you can’t make the CrossValidated Journal Club chats, why not start your own local club?
Today I gave a workshop for supervisors of postgraduate students. Mostly I talked about creating a team environment for postgraduate students rather than the traditional model (at least in statistics and econometrics) of each student working in isolation. The slides are available here in presentation form or in handout form. Actually, these are an edited version of the slides as I accidentally left out a couple of the phởtographs in the workshop, and I’ve omitted slides that I didn’t end up covering in the workshop. An important part of my research group is this blog. So if you haven’t been here before, please take a look around. For those people who attended, feel free to add comments below if you would like to provide feedback on the workshop.
Muhammad Akram was my PhD student a few years ago, and has remained a good friend since he moved on. Here is an interview he recently gave about moving to Australia. Thanks, Akram, for the kind words about me!