But then I found a pointer to the TensorFlow project, and decided to give it another try. Reading through their tutorials, I started to have a suspicion that I might actually understand what was going on. And then I tracked down this article, https://medium.com/safegraph/a-non-
When they call it machine learning, it's easy enough to assume that the model continues learning as you use it, giving it more data. No, by the time you begin using the model, all the learning is done, at least with the systems I've looked at. The learning part happens when you let the code tweak the equation to better fit what you want to see as a result, and that's where the dangerous part is. There are a number of decisions the model binder has to make and they all influence the outcome of the model, and that's not even touching the quality of the data set or whether the assumed correlation is real enough to be useful. And even if you don't make a wrong step in making and training and tuning the model, it's still dependent upon past data. If the nature of the correlation changes, it will invalidate the model.
All of which is a fancy way of saying that if someone comes to you with a fancy machine learning model, don't treat it like a magical black box, because it isn't. Also, brush up on your statistics if r-squared is still greek to you. Really, it's not that bad, and it can come in handy when reviewing scientific studies or listening to economists.